@smake/eigen 1.1.0 → 1.1.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (431) hide show
  1. package/README.md +1 -1
  2. package/eigen/Eigen/AccelerateSupport +52 -0
  3. package/eigen/Eigen/Cholesky +18 -20
  4. package/eigen/Eigen/CholmodSupport +28 -28
  5. package/eigen/Eigen/Core +187 -120
  6. package/eigen/Eigen/Eigenvalues +16 -13
  7. package/eigen/Eigen/Geometry +18 -18
  8. package/eigen/Eigen/Householder +9 -7
  9. package/eigen/Eigen/IterativeLinearSolvers +8 -4
  10. package/eigen/Eigen/Jacobi +14 -13
  11. package/eigen/Eigen/KLUSupport +23 -21
  12. package/eigen/Eigen/LU +15 -16
  13. package/eigen/Eigen/MetisSupport +12 -12
  14. package/eigen/Eigen/OrderingMethods +54 -51
  15. package/eigen/Eigen/PaStiXSupport +23 -21
  16. package/eigen/Eigen/PardisoSupport +17 -14
  17. package/eigen/Eigen/QR +18 -20
  18. package/eigen/Eigen/QtAlignedMalloc +5 -12
  19. package/eigen/Eigen/SPQRSupport +21 -14
  20. package/eigen/Eigen/SVD +23 -17
  21. package/eigen/Eigen/Sparse +1 -2
  22. package/eigen/Eigen/SparseCholesky +18 -15
  23. package/eigen/Eigen/SparseCore +18 -17
  24. package/eigen/Eigen/SparseLU +9 -9
  25. package/eigen/Eigen/SparseQR +16 -14
  26. package/eigen/Eigen/StdDeque +5 -2
  27. package/eigen/Eigen/StdList +5 -2
  28. package/eigen/Eigen/StdVector +5 -2
  29. package/eigen/Eigen/SuperLUSupport +30 -24
  30. package/eigen/Eigen/ThreadPool +80 -0
  31. package/eigen/Eigen/UmfPackSupport +19 -17
  32. package/eigen/Eigen/Version +14 -0
  33. package/eigen/Eigen/src/AccelerateSupport/AccelerateSupport.h +423 -0
  34. package/eigen/Eigen/src/AccelerateSupport/InternalHeaderCheck.h +3 -0
  35. package/eigen/Eigen/src/Cholesky/InternalHeaderCheck.h +3 -0
  36. package/eigen/Eigen/src/Cholesky/LDLT.h +366 -405
  37. package/eigen/Eigen/src/Cholesky/LLT.h +323 -367
  38. package/eigen/Eigen/src/Cholesky/LLT_LAPACKE.h +81 -56
  39. package/eigen/Eigen/src/CholmodSupport/CholmodSupport.h +585 -529
  40. package/eigen/Eigen/src/CholmodSupport/InternalHeaderCheck.h +3 -0
  41. package/eigen/Eigen/src/Core/ArithmeticSequence.h +143 -317
  42. package/eigen/Eigen/src/Core/Array.h +329 -370
  43. package/eigen/Eigen/src/Core/ArrayBase.h +190 -203
  44. package/eigen/Eigen/src/Core/ArrayWrapper.h +126 -170
  45. package/eigen/Eigen/src/Core/Assign.h +30 -40
  46. package/eigen/Eigen/src/Core/AssignEvaluator.h +651 -604
  47. package/eigen/Eigen/src/Core/Assign_MKL.h +125 -120
  48. package/eigen/Eigen/src/Core/BandMatrix.h +267 -282
  49. package/eigen/Eigen/src/Core/Block.h +371 -390
  50. package/eigen/Eigen/src/Core/CommaInitializer.h +85 -100
  51. package/eigen/Eigen/src/Core/ConditionEstimator.h +51 -53
  52. package/eigen/Eigen/src/Core/CoreEvaluators.h +1214 -937
  53. package/eigen/Eigen/src/Core/CoreIterators.h +72 -63
  54. package/eigen/Eigen/src/Core/CwiseBinaryOp.h +112 -129
  55. package/eigen/Eigen/src/Core/CwiseNullaryOp.h +676 -702
  56. package/eigen/Eigen/src/Core/CwiseTernaryOp.h +77 -103
  57. package/eigen/Eigen/src/Core/CwiseUnaryOp.h +55 -67
  58. package/eigen/Eigen/src/Core/CwiseUnaryView.h +127 -92
  59. package/eigen/Eigen/src/Core/DenseBase.h +630 -658
  60. package/eigen/Eigen/src/Core/DenseCoeffsBase.h +511 -628
  61. package/eigen/Eigen/src/Core/DenseStorage.h +511 -590
  62. package/eigen/Eigen/src/Core/DeviceWrapper.h +153 -0
  63. package/eigen/Eigen/src/Core/Diagonal.h +168 -207
  64. package/eigen/Eigen/src/Core/DiagonalMatrix.h +346 -317
  65. package/eigen/Eigen/src/Core/DiagonalProduct.h +12 -10
  66. package/eigen/Eigen/src/Core/Dot.h +167 -217
  67. package/eigen/Eigen/src/Core/EigenBase.h +74 -85
  68. package/eigen/Eigen/src/Core/Fill.h +138 -0
  69. package/eigen/Eigen/src/Core/FindCoeff.h +464 -0
  70. package/eigen/Eigen/src/Core/ForceAlignedAccess.h +90 -113
  71. package/eigen/Eigen/src/Core/Fuzzy.h +82 -105
  72. package/eigen/Eigen/src/Core/GeneralProduct.h +315 -261
  73. package/eigen/Eigen/src/Core/GenericPacketMath.h +1182 -520
  74. package/eigen/Eigen/src/Core/GlobalFunctions.h +193 -157
  75. package/eigen/Eigen/src/Core/IO.h +131 -156
  76. package/eigen/Eigen/src/Core/IndexedView.h +209 -125
  77. package/eigen/Eigen/src/Core/InnerProduct.h +260 -0
  78. package/eigen/Eigen/src/Core/InternalHeaderCheck.h +3 -0
  79. package/eigen/Eigen/src/Core/Inverse.h +50 -59
  80. package/eigen/Eigen/src/Core/Map.h +123 -141
  81. package/eigen/Eigen/src/Core/MapBase.h +255 -282
  82. package/eigen/Eigen/src/Core/MathFunctions.h +1247 -1201
  83. package/eigen/Eigen/src/Core/MathFunctionsImpl.h +162 -99
  84. package/eigen/Eigen/src/Core/Matrix.h +463 -494
  85. package/eigen/Eigen/src/Core/MatrixBase.h +468 -470
  86. package/eigen/Eigen/src/Core/NestByValue.h +58 -52
  87. package/eigen/Eigen/src/Core/NoAlias.h +79 -86
  88. package/eigen/Eigen/src/Core/NumTraits.h +206 -206
  89. package/eigen/Eigen/src/Core/PartialReduxEvaluator.h +163 -142
  90. package/eigen/Eigen/src/Core/PermutationMatrix.h +461 -511
  91. package/eigen/Eigen/src/Core/PlainObjectBase.h +858 -972
  92. package/eigen/Eigen/src/Core/Product.h +246 -130
  93. package/eigen/Eigen/src/Core/ProductEvaluators.h +779 -671
  94. package/eigen/Eigen/src/Core/Random.h +153 -164
  95. package/eigen/Eigen/src/Core/RandomImpl.h +262 -0
  96. package/eigen/Eigen/src/Core/RealView.h +250 -0
  97. package/eigen/Eigen/src/Core/Redux.h +334 -314
  98. package/eigen/Eigen/src/Core/Ref.h +259 -257
  99. package/eigen/Eigen/src/Core/Replicate.h +92 -104
  100. package/eigen/Eigen/src/Core/Reshaped.h +215 -271
  101. package/eigen/Eigen/src/Core/ReturnByValue.h +47 -55
  102. package/eigen/Eigen/src/Core/Reverse.h +133 -148
  103. package/eigen/Eigen/src/Core/Select.h +68 -140
  104. package/eigen/Eigen/src/Core/SelfAdjointView.h +254 -290
  105. package/eigen/Eigen/src/Core/SelfCwiseBinaryOp.h +23 -20
  106. package/eigen/Eigen/src/Core/SkewSymmetricMatrix3.h +382 -0
  107. package/eigen/Eigen/src/Core/Solve.h +88 -102
  108. package/eigen/Eigen/src/Core/SolveTriangular.h +126 -124
  109. package/eigen/Eigen/src/Core/SolverBase.h +132 -133
  110. package/eigen/Eigen/src/Core/StableNorm.h +113 -147
  111. package/eigen/Eigen/src/Core/StlIterators.h +404 -248
  112. package/eigen/Eigen/src/Core/Stride.h +90 -92
  113. package/eigen/Eigen/src/Core/Swap.h +70 -39
  114. package/eigen/Eigen/src/Core/Transpose.h +258 -295
  115. package/eigen/Eigen/src/Core/Transpositions.h +270 -333
  116. package/eigen/Eigen/src/Core/TriangularMatrix.h +642 -743
  117. package/eigen/Eigen/src/Core/VectorBlock.h +59 -72
  118. package/eigen/Eigen/src/Core/VectorwiseOp.h +653 -704
  119. package/eigen/Eigen/src/Core/Visitor.h +464 -308
  120. package/eigen/Eigen/src/Core/arch/AVX/Complex.h +380 -187
  121. package/eigen/Eigen/src/Core/arch/AVX/MathFunctions.h +65 -163
  122. package/eigen/Eigen/src/Core/arch/AVX/PacketMath.h +2145 -638
  123. package/eigen/Eigen/src/Core/arch/AVX/Reductions.h +353 -0
  124. package/eigen/Eigen/src/Core/arch/AVX/TypeCasting.h +253 -60
  125. package/eigen/Eigen/src/Core/arch/AVX512/Complex.h +278 -228
  126. package/eigen/Eigen/src/Core/arch/AVX512/GemmKernel.h +1245 -0
  127. package/eigen/Eigen/src/Core/arch/AVX512/MathFunctions.h +48 -269
  128. package/eigen/Eigen/src/Core/arch/AVX512/MathFunctionsFP16.h +75 -0
  129. package/eigen/Eigen/src/Core/arch/AVX512/PacketMath.h +1597 -754
  130. package/eigen/Eigen/src/Core/arch/AVX512/PacketMathFP16.h +1413 -0
  131. package/eigen/Eigen/src/Core/arch/AVX512/Reductions.h +297 -0
  132. package/eigen/Eigen/src/Core/arch/AVX512/TrsmKernel.h +1167 -0
  133. package/eigen/Eigen/src/Core/arch/AVX512/TrsmUnrolls.inc +1219 -0
  134. package/eigen/Eigen/src/Core/arch/AVX512/TypeCasting.h +229 -41
  135. package/eigen/Eigen/src/Core/arch/AVX512/TypeCastingFP16.h +130 -0
  136. package/eigen/Eigen/src/Core/arch/AltiVec/Complex.h +420 -184
  137. package/eigen/Eigen/src/Core/arch/AltiVec/MathFunctions.h +40 -49
  138. package/eigen/Eigen/src/Core/arch/AltiVec/MatrixProduct.h +2962 -2213
  139. package/eigen/Eigen/src/Core/arch/AltiVec/MatrixProductCommon.h +196 -212
  140. package/eigen/Eigen/src/Core/arch/AltiVec/MatrixProductMMA.h +713 -441
  141. package/eigen/Eigen/src/Core/arch/AltiVec/MatrixProductMMAbfloat16.h +742 -0
  142. package/eigen/Eigen/src/Core/arch/AltiVec/MatrixVectorProduct.inc +2818 -0
  143. package/eigen/Eigen/src/Core/arch/AltiVec/PacketMath.h +2380 -1362
  144. package/eigen/Eigen/src/Core/arch/AltiVec/TypeCasting.h +153 -0
  145. package/eigen/Eigen/src/Core/arch/Default/BFloat16.h +390 -224
  146. package/eigen/Eigen/src/Core/arch/Default/ConjHelper.h +78 -67
  147. package/eigen/Eigen/src/Core/arch/Default/GenericPacketMathFunctions.h +1784 -799
  148. package/eigen/Eigen/src/Core/arch/Default/GenericPacketMathFunctionsFwd.h +167 -50
  149. package/eigen/Eigen/src/Core/arch/Default/Half.h +528 -379
  150. package/eigen/Eigen/src/Core/arch/Default/Settings.h +10 -12
  151. package/eigen/Eigen/src/Core/arch/GPU/Complex.h +244 -0
  152. package/eigen/Eigen/src/Core/arch/GPU/MathFunctions.h +41 -40
  153. package/eigen/Eigen/src/Core/arch/GPU/PacketMath.h +550 -523
  154. package/eigen/Eigen/src/Core/arch/GPU/Tuple.h +268 -0
  155. package/eigen/Eigen/src/Core/arch/GPU/TypeCasting.h +27 -30
  156. package/eigen/Eigen/src/Core/arch/HIP/hcc/math_constants.h +8 -8
  157. package/eigen/Eigen/src/Core/arch/HVX/PacketMath.h +1088 -0
  158. package/eigen/Eigen/src/Core/arch/LSX/Complex.h +520 -0
  159. package/eigen/Eigen/src/Core/arch/LSX/GeneralBlockPanelKernel.h +23 -0
  160. package/eigen/Eigen/src/Core/arch/LSX/MathFunctions.h +43 -0
  161. package/eigen/Eigen/src/Core/arch/LSX/PacketMath.h +2866 -0
  162. package/eigen/Eigen/src/Core/arch/LSX/TypeCasting.h +526 -0
  163. package/eigen/Eigen/src/Core/arch/MSA/Complex.h +54 -82
  164. package/eigen/Eigen/src/Core/arch/MSA/MathFunctions.h +84 -92
  165. package/eigen/Eigen/src/Core/arch/MSA/PacketMath.h +51 -47
  166. package/eigen/Eigen/src/Core/arch/NEON/Complex.h +454 -306
  167. package/eigen/Eigen/src/Core/arch/NEON/GeneralBlockPanelKernel.h +175 -115
  168. package/eigen/Eigen/src/Core/arch/NEON/MathFunctions.h +23 -30
  169. package/eigen/Eigen/src/Core/arch/NEON/PacketMath.h +4366 -2857
  170. package/eigen/Eigen/src/Core/arch/NEON/TypeCasting.h +616 -393
  171. package/eigen/Eigen/src/Core/arch/NEON/UnaryFunctors.h +57 -0
  172. package/eigen/Eigen/src/Core/arch/SSE/Complex.h +350 -198
  173. package/eigen/Eigen/src/Core/arch/SSE/MathFunctions.h +38 -149
  174. package/eigen/Eigen/src/Core/arch/SSE/PacketMath.h +1791 -912
  175. package/eigen/Eigen/src/Core/arch/SSE/Reductions.h +324 -0
  176. package/eigen/Eigen/src/Core/arch/SSE/TypeCasting.h +128 -40
  177. package/eigen/Eigen/src/Core/arch/SVE/MathFunctions.h +10 -6
  178. package/eigen/Eigen/src/Core/arch/SVE/PacketMath.h +156 -234
  179. package/eigen/Eigen/src/Core/arch/SVE/TypeCasting.h +6 -3
  180. package/eigen/Eigen/src/Core/arch/SYCL/InteropHeaders.h +27 -32
  181. package/eigen/Eigen/src/Core/arch/SYCL/MathFunctions.h +119 -117
  182. package/eigen/Eigen/src/Core/arch/SYCL/PacketMath.h +325 -419
  183. package/eigen/Eigen/src/Core/arch/SYCL/TypeCasting.h +15 -17
  184. package/eigen/Eigen/src/Core/arch/ZVector/Complex.h +325 -181
  185. package/eigen/Eigen/src/Core/arch/ZVector/MathFunctions.h +94 -83
  186. package/eigen/Eigen/src/Core/arch/ZVector/PacketMath.h +811 -458
  187. package/eigen/Eigen/src/Core/functors/AssignmentFunctors.h +121 -124
  188. package/eigen/Eigen/src/Core/functors/BinaryFunctors.h +576 -370
  189. package/eigen/Eigen/src/Core/functors/NullaryFunctors.h +194 -109
  190. package/eigen/Eigen/src/Core/functors/StlFunctors.h +95 -112
  191. package/eigen/Eigen/src/Core/functors/TernaryFunctors.h +34 -7
  192. package/eigen/Eigen/src/Core/functors/UnaryFunctors.h +1038 -749
  193. package/eigen/Eigen/src/Core/products/GeneralBlockPanelKernel.h +1883 -1375
  194. package/eigen/Eigen/src/Core/products/GeneralMatrixMatrix.h +312 -370
  195. package/eigen/Eigen/src/Core/products/GeneralMatrixMatrixTriangular.h +189 -176
  196. package/eigen/Eigen/src/Core/products/GeneralMatrixMatrixTriangular_BLAS.h +84 -81
  197. package/eigen/Eigen/src/Core/products/GeneralMatrixMatrix_BLAS.h +154 -73
  198. package/eigen/Eigen/src/Core/products/GeneralMatrixVector.h +292 -337
  199. package/eigen/Eigen/src/Core/products/GeneralMatrixVector_BLAS.h +80 -77
  200. package/eigen/Eigen/src/Core/products/Parallelizer.h +207 -105
  201. package/eigen/Eigen/src/Core/products/SelfadjointMatrixMatrix.h +327 -388
  202. package/eigen/Eigen/src/Core/products/SelfadjointMatrixMatrix_BLAS.h +206 -224
  203. package/eigen/Eigen/src/Core/products/SelfadjointMatrixVector.h +138 -147
  204. package/eigen/Eigen/src/Core/products/SelfadjointMatrixVector_BLAS.h +58 -61
  205. package/eigen/Eigen/src/Core/products/SelfadjointProduct.h +71 -71
  206. package/eigen/Eigen/src/Core/products/SelfadjointRank2Update.h +48 -47
  207. package/eigen/Eigen/src/Core/products/TriangularMatrixMatrix.h +294 -369
  208. package/eigen/Eigen/src/Core/products/TriangularMatrixMatrix_BLAS.h +246 -238
  209. package/eigen/Eigen/src/Core/products/TriangularMatrixVector.h +244 -247
  210. package/eigen/Eigen/src/Core/products/TriangularMatrixVector_BLAS.h +212 -192
  211. package/eigen/Eigen/src/Core/products/TriangularSolverMatrix.h +328 -277
  212. package/eigen/Eigen/src/Core/products/TriangularSolverMatrix_BLAS.h +108 -109
  213. package/eigen/Eigen/src/Core/products/TriangularSolverVector.h +68 -94
  214. package/eigen/Eigen/src/Core/util/Assert.h +158 -0
  215. package/eigen/Eigen/src/Core/util/BlasUtil.h +342 -303
  216. package/eigen/Eigen/src/Core/util/ConfigureVectorization.h +348 -317
  217. package/eigen/Eigen/src/Core/util/Constants.h +297 -262
  218. package/eigen/Eigen/src/Core/util/DisableStupidWarnings.h +130 -90
  219. package/eigen/Eigen/src/Core/util/EmulateArray.h +270 -0
  220. package/eigen/Eigen/src/Core/util/ForwardDeclarations.h +449 -247
  221. package/eigen/Eigen/src/Core/util/GpuHipCudaDefines.inc +101 -0
  222. package/eigen/Eigen/src/Core/util/GpuHipCudaUndefines.inc +45 -0
  223. package/eigen/Eigen/src/Core/util/IndexedViewHelper.h +417 -116
  224. package/eigen/Eigen/src/Core/util/IntegralConstant.h +211 -204
  225. package/eigen/Eigen/src/Core/util/MKL_support.h +39 -37
  226. package/eigen/Eigen/src/Core/util/Macros.h +655 -773
  227. package/eigen/Eigen/src/Core/util/MaxSizeVector.h +139 -0
  228. package/eigen/Eigen/src/Core/util/Memory.h +970 -748
  229. package/eigen/Eigen/src/Core/util/Meta.h +581 -633
  230. package/eigen/Eigen/src/Core/util/MoreMeta.h +638 -0
  231. package/eigen/Eigen/src/Core/util/ReenableStupidWarnings.h +32 -19
  232. package/eigen/Eigen/src/Core/util/ReshapedHelper.h +17 -17
  233. package/eigen/Eigen/src/Core/util/Serializer.h +209 -0
  234. package/eigen/Eigen/src/Core/util/StaticAssert.h +50 -166
  235. package/eigen/Eigen/src/Core/util/SymbolicIndex.h +377 -225
  236. package/eigen/Eigen/src/Core/util/XprHelper.h +784 -547
  237. package/eigen/Eigen/src/Eigenvalues/ComplexEigenSolver.h +246 -277
  238. package/eigen/Eigen/src/Eigenvalues/ComplexSchur.h +299 -319
  239. package/eigen/Eigen/src/Eigenvalues/ComplexSchur_LAPACKE.h +52 -48
  240. package/eigen/Eigen/src/Eigenvalues/EigenSolver.h +413 -456
  241. package/eigen/Eigen/src/Eigenvalues/GeneralizedEigenSolver.h +309 -325
  242. package/eigen/Eigen/src/Eigenvalues/GeneralizedSelfAdjointEigenSolver.h +157 -171
  243. package/eigen/Eigen/src/Eigenvalues/HessenbergDecomposition.h +292 -310
  244. package/eigen/Eigen/src/Eigenvalues/InternalHeaderCheck.h +3 -0
  245. package/eigen/Eigen/src/Eigenvalues/MatrixBaseEigenvalues.h +89 -105
  246. package/eigen/Eigen/src/Eigenvalues/RealQZ.h +537 -607
  247. package/eigen/Eigen/src/Eigenvalues/RealSchur.h +342 -381
  248. package/eigen/Eigen/src/Eigenvalues/RealSchur_LAPACKE.h +41 -35
  249. package/eigen/Eigen/src/Eigenvalues/SelfAdjointEigenSolver.h +541 -595
  250. package/eigen/Eigen/src/Eigenvalues/SelfAdjointEigenSolver_LAPACKE.h +47 -44
  251. package/eigen/Eigen/src/Eigenvalues/Tridiagonalization.h +430 -462
  252. package/eigen/Eigen/src/Geometry/AlignedBox.h +226 -227
  253. package/eigen/Eigen/src/Geometry/AngleAxis.h +131 -133
  254. package/eigen/Eigen/src/Geometry/EulerAngles.h +163 -74
  255. package/eigen/Eigen/src/Geometry/Homogeneous.h +285 -333
  256. package/eigen/Eigen/src/Geometry/Hyperplane.h +151 -160
  257. package/eigen/Eigen/src/Geometry/InternalHeaderCheck.h +3 -0
  258. package/eigen/Eigen/src/Geometry/OrthoMethods.h +168 -146
  259. package/eigen/Eigen/src/Geometry/ParametrizedLine.h +127 -127
  260. package/eigen/Eigen/src/Geometry/Quaternion.h +566 -506
  261. package/eigen/Eigen/src/Geometry/Rotation2D.h +107 -105
  262. package/eigen/Eigen/src/Geometry/RotationBase.h +148 -145
  263. package/eigen/Eigen/src/Geometry/Scaling.h +113 -106
  264. package/eigen/Eigen/src/Geometry/Transform.h +858 -936
  265. package/eigen/Eigen/src/Geometry/Translation.h +94 -92
  266. package/eigen/Eigen/src/Geometry/Umeyama.h +79 -84
  267. package/eigen/Eigen/src/Geometry/arch/Geometry_SIMD.h +90 -104
  268. package/eigen/Eigen/src/Householder/BlockHouseholder.h +51 -46
  269. package/eigen/Eigen/src/Householder/Householder.h +102 -124
  270. package/eigen/Eigen/src/Householder/HouseholderSequence.h +412 -453
  271. package/eigen/Eigen/src/Householder/InternalHeaderCheck.h +3 -0
  272. package/eigen/Eigen/src/IterativeLinearSolvers/BasicPreconditioners.h +149 -162
  273. package/eigen/Eigen/src/IterativeLinearSolvers/BiCGSTAB.h +124 -119
  274. package/eigen/Eigen/src/IterativeLinearSolvers/ConjugateGradient.h +92 -104
  275. package/eigen/Eigen/src/IterativeLinearSolvers/IncompleteCholesky.h +251 -243
  276. package/eigen/Eigen/src/IterativeLinearSolvers/IncompleteLUT.h +224 -228
  277. package/eigen/Eigen/src/IterativeLinearSolvers/InternalHeaderCheck.h +3 -0
  278. package/eigen/Eigen/src/IterativeLinearSolvers/IterativeSolverBase.h +178 -227
  279. package/eigen/Eigen/src/IterativeLinearSolvers/LeastSquareConjugateGradient.h +79 -84
  280. package/eigen/Eigen/src/IterativeLinearSolvers/SolveWithGuess.h +54 -60
  281. package/eigen/Eigen/src/Jacobi/InternalHeaderCheck.h +3 -0
  282. package/eigen/Eigen/src/Jacobi/Jacobi.h +252 -308
  283. package/eigen/Eigen/src/KLUSupport/InternalHeaderCheck.h +3 -0
  284. package/eigen/Eigen/src/KLUSupport/KLUSupport.h +208 -227
  285. package/eigen/Eigen/src/LU/Determinant.h +50 -69
  286. package/eigen/Eigen/src/LU/FullPivLU.h +545 -596
  287. package/eigen/Eigen/src/LU/InternalHeaderCheck.h +3 -0
  288. package/eigen/Eigen/src/LU/InverseImpl.h +206 -285
  289. package/eigen/Eigen/src/LU/PartialPivLU.h +390 -428
  290. package/eigen/Eigen/src/LU/PartialPivLU_LAPACKE.h +54 -40
  291. package/eigen/Eigen/src/LU/arch/InverseSize4.h +72 -70
  292. package/eigen/Eigen/src/MetisSupport/InternalHeaderCheck.h +3 -0
  293. package/eigen/Eigen/src/MetisSupport/MetisSupport.h +81 -93
  294. package/eigen/Eigen/src/OrderingMethods/Amd.h +243 -265
  295. package/eigen/Eigen/src/OrderingMethods/Eigen_Colamd.h +831 -1004
  296. package/eigen/Eigen/src/OrderingMethods/InternalHeaderCheck.h +3 -0
  297. package/eigen/Eigen/src/OrderingMethods/Ordering.h +112 -119
  298. package/eigen/Eigen/src/PaStiXSupport/InternalHeaderCheck.h +3 -0
  299. package/eigen/Eigen/src/PaStiXSupport/PaStiXSupport.h +524 -570
  300. package/eigen/Eigen/src/PardisoSupport/InternalHeaderCheck.h +3 -0
  301. package/eigen/Eigen/src/PardisoSupport/PardisoSupport.h +385 -430
  302. package/eigen/Eigen/src/QR/ColPivHouseholderQR.h +479 -479
  303. package/eigen/Eigen/src/QR/ColPivHouseholderQR_LAPACKE.h +120 -56
  304. package/eigen/Eigen/src/QR/CompleteOrthogonalDecomposition.h +166 -153
  305. package/eigen/Eigen/src/QR/FullPivHouseholderQR.h +495 -475
  306. package/eigen/Eigen/src/QR/HouseholderQR.h +394 -285
  307. package/eigen/Eigen/src/QR/HouseholderQR_LAPACKE.h +32 -23
  308. package/eigen/Eigen/src/QR/InternalHeaderCheck.h +3 -0
  309. package/eigen/Eigen/src/SPQRSupport/InternalHeaderCheck.h +3 -0
  310. package/eigen/Eigen/src/SPQRSupport/SuiteSparseQRSupport.h +244 -264
  311. package/eigen/Eigen/src/SVD/BDCSVD.h +817 -713
  312. package/eigen/Eigen/src/SVD/BDCSVD_LAPACKE.h +174 -0
  313. package/eigen/Eigen/src/SVD/InternalHeaderCheck.h +3 -0
  314. package/eigen/Eigen/src/SVD/JacobiSVD.h +577 -543
  315. package/eigen/Eigen/src/SVD/JacobiSVD_LAPACKE.h +85 -49
  316. package/eigen/Eigen/src/SVD/SVDBase.h +242 -182
  317. package/eigen/Eigen/src/SVD/UpperBidiagonalization.h +200 -235
  318. package/eigen/Eigen/src/SparseCholesky/InternalHeaderCheck.h +3 -0
  319. package/eigen/Eigen/src/SparseCholesky/SimplicialCholesky.h +765 -594
  320. package/eigen/Eigen/src/SparseCholesky/SimplicialCholesky_impl.h +308 -94
  321. package/eigen/Eigen/src/SparseCore/AmbiVector.h +202 -251
  322. package/eigen/Eigen/src/SparseCore/CompressedStorage.h +184 -252
  323. package/eigen/Eigen/src/SparseCore/ConservativeSparseSparseProduct.h +134 -178
  324. package/eigen/Eigen/src/SparseCore/InternalHeaderCheck.h +3 -0
  325. package/eigen/Eigen/src/SparseCore/SparseAssign.h +149 -140
  326. package/eigen/Eigen/src/SparseCore/SparseBlock.h +403 -440
  327. package/eigen/Eigen/src/SparseCore/SparseColEtree.h +100 -112
  328. package/eigen/Eigen/src/SparseCore/SparseCompressedBase.h +525 -303
  329. package/eigen/Eigen/src/SparseCore/SparseCwiseBinaryOp.h +555 -339
  330. package/eigen/Eigen/src/SparseCore/SparseCwiseUnaryOp.h +100 -108
  331. package/eigen/Eigen/src/SparseCore/SparseDenseProduct.h +169 -197
  332. package/eigen/Eigen/src/SparseCore/SparseDiagonalProduct.h +71 -71
  333. package/eigen/Eigen/src/SparseCore/SparseDot.h +49 -47
  334. package/eigen/Eigen/src/SparseCore/SparseFuzzy.h +13 -11
  335. package/eigen/Eigen/src/SparseCore/SparseMap.h +243 -253
  336. package/eigen/Eigen/src/SparseCore/SparseMatrix.h +1603 -1245
  337. package/eigen/Eigen/src/SparseCore/SparseMatrixBase.h +403 -350
  338. package/eigen/Eigen/src/SparseCore/SparsePermutation.h +186 -115
  339. package/eigen/Eigen/src/SparseCore/SparseProduct.h +94 -97
  340. package/eigen/Eigen/src/SparseCore/SparseRedux.h +22 -24
  341. package/eigen/Eigen/src/SparseCore/SparseRef.h +268 -295
  342. package/eigen/Eigen/src/SparseCore/SparseSelfAdjointView.h +370 -416
  343. package/eigen/Eigen/src/SparseCore/SparseSolverBase.h +78 -87
  344. package/eigen/Eigen/src/SparseCore/SparseSparseProductWithPruning.h +81 -95
  345. package/eigen/Eigen/src/SparseCore/SparseTranspose.h +62 -71
  346. package/eigen/Eigen/src/SparseCore/SparseTriangularView.h +132 -144
  347. package/eigen/Eigen/src/SparseCore/SparseUtil.h +138 -115
  348. package/eigen/Eigen/src/SparseCore/SparseVector.h +426 -372
  349. package/eigen/Eigen/src/SparseCore/SparseView.h +164 -193
  350. package/eigen/Eigen/src/SparseCore/TriangularSolver.h +129 -170
  351. package/eigen/Eigen/src/SparseLU/InternalHeaderCheck.h +3 -0
  352. package/eigen/Eigen/src/SparseLU/SparseLU.h +756 -710
  353. package/eigen/Eigen/src/SparseLU/SparseLUImpl.h +61 -48
  354. package/eigen/Eigen/src/SparseLU/SparseLU_Memory.h +102 -118
  355. package/eigen/Eigen/src/SparseLU/SparseLU_Structs.h +38 -35
  356. package/eigen/Eigen/src/SparseLU/SparseLU_SupernodalMatrix.h +245 -301
  357. package/eigen/Eigen/src/SparseLU/SparseLU_Utils.h +44 -49
  358. package/eigen/Eigen/src/SparseLU/SparseLU_column_bmod.h +104 -108
  359. package/eigen/Eigen/src/SparseLU/SparseLU_column_dfs.h +89 -100
  360. package/eigen/Eigen/src/SparseLU/SparseLU_copy_to_ucol.h +57 -58
  361. package/eigen/Eigen/src/SparseLU/SparseLU_heap_relax_snode.h +43 -55
  362. package/eigen/Eigen/src/SparseLU/SparseLU_kernel_bmod.h +74 -71
  363. package/eigen/Eigen/src/SparseLU/SparseLU_panel_bmod.h +124 -132
  364. package/eigen/Eigen/src/SparseLU/SparseLU_panel_dfs.h +136 -159
  365. package/eigen/Eigen/src/SparseLU/SparseLU_pivotL.h +51 -52
  366. package/eigen/Eigen/src/SparseLU/SparseLU_pruneL.h +67 -73
  367. package/eigen/Eigen/src/SparseLU/SparseLU_relax_snode.h +24 -26
  368. package/eigen/Eigen/src/SparseQR/InternalHeaderCheck.h +3 -0
  369. package/eigen/Eigen/src/SparseQR/SparseQR.h +450 -502
  370. package/eigen/Eigen/src/StlSupport/StdDeque.h +28 -93
  371. package/eigen/Eigen/src/StlSupport/StdList.h +28 -84
  372. package/eigen/Eigen/src/StlSupport/StdVector.h +28 -108
  373. package/eigen/Eigen/src/StlSupport/details.h +48 -50
  374. package/eigen/Eigen/src/SuperLUSupport/InternalHeaderCheck.h +3 -0
  375. package/eigen/Eigen/src/SuperLUSupport/SuperLUSupport.h +634 -730
  376. package/eigen/Eigen/src/ThreadPool/Barrier.h +70 -0
  377. package/eigen/Eigen/src/ThreadPool/CoreThreadPoolDevice.h +336 -0
  378. package/eigen/Eigen/src/ThreadPool/EventCount.h +241 -0
  379. package/eigen/Eigen/src/ThreadPool/ForkJoin.h +140 -0
  380. package/eigen/Eigen/src/ThreadPool/InternalHeaderCheck.h +4 -0
  381. package/eigen/Eigen/src/ThreadPool/NonBlockingThreadPool.h +587 -0
  382. package/eigen/Eigen/src/ThreadPool/RunQueue.h +230 -0
  383. package/eigen/Eigen/src/ThreadPool/ThreadCancel.h +21 -0
  384. package/eigen/Eigen/src/ThreadPool/ThreadEnvironment.h +43 -0
  385. package/eigen/Eigen/src/ThreadPool/ThreadLocal.h +289 -0
  386. package/eigen/Eigen/src/ThreadPool/ThreadPoolInterface.h +50 -0
  387. package/eigen/Eigen/src/ThreadPool/ThreadYield.h +16 -0
  388. package/eigen/Eigen/src/UmfPackSupport/InternalHeaderCheck.h +3 -0
  389. package/eigen/Eigen/src/UmfPackSupport/UmfPackSupport.h +428 -464
  390. package/eigen/Eigen/src/misc/Image.h +41 -43
  391. package/eigen/Eigen/src/misc/InternalHeaderCheck.h +3 -0
  392. package/eigen/Eigen/src/misc/Kernel.h +39 -41
  393. package/eigen/Eigen/src/misc/RealSvd2x2.h +19 -21
  394. package/eigen/Eigen/src/misc/blas.h +83 -426
  395. package/eigen/Eigen/src/misc/lapacke.h +9972 -16179
  396. package/eigen/Eigen/src/misc/lapacke_helpers.h +163 -0
  397. package/eigen/Eigen/src/misc/lapacke_mangling.h +4 -5
  398. package/eigen/Eigen/src/plugins/ArrayCwiseBinaryOps.inc +344 -0
  399. package/eigen/Eigen/src/plugins/ArrayCwiseUnaryOps.inc +544 -0
  400. package/eigen/Eigen/src/plugins/{BlockMethods.h → BlockMethods.inc} +434 -506
  401. package/eigen/Eigen/src/plugins/CommonCwiseBinaryOps.inc +116 -0
  402. package/eigen/Eigen/src/plugins/{CommonCwiseUnaryOps.h → CommonCwiseUnaryOps.inc} +58 -68
  403. package/eigen/Eigen/src/plugins/IndexedViewMethods.inc +192 -0
  404. package/eigen/Eigen/src/plugins/InternalHeaderCheck.inc +3 -0
  405. package/eigen/Eigen/src/plugins/MatrixCwiseBinaryOps.inc +331 -0
  406. package/eigen/Eigen/src/plugins/MatrixCwiseUnaryOps.inc +118 -0
  407. package/eigen/Eigen/src/plugins/ReshapedMethods.inc +133 -0
  408. package/package.json +1 -1
  409. package/eigen/COPYING.APACHE +0 -203
  410. package/eigen/COPYING.BSD +0 -26
  411. package/eigen/COPYING.GPL +0 -674
  412. package/eigen/COPYING.LGPL +0 -502
  413. package/eigen/COPYING.MINPACK +0 -51
  414. package/eigen/COPYING.MPL2 +0 -373
  415. package/eigen/COPYING.README +0 -18
  416. package/eigen/Eigen/src/Core/BooleanRedux.h +0 -162
  417. package/eigen/Eigen/src/Core/arch/CUDA/Complex.h +0 -258
  418. package/eigen/Eigen/src/Core/arch/Default/TypeCasting.h +0 -120
  419. package/eigen/Eigen/src/Core/arch/SYCL/SyclMemoryModel.h +0 -694
  420. package/eigen/Eigen/src/Core/util/NonMPL2.h +0 -3
  421. package/eigen/Eigen/src/SparseCore/MappedSparseMatrix.h +0 -67
  422. package/eigen/Eigen/src/SparseLU/SparseLU_gemm_kernel.h +0 -280
  423. package/eigen/Eigen/src/misc/lapack.h +0 -152
  424. package/eigen/Eigen/src/plugins/ArrayCwiseBinaryOps.h +0 -358
  425. package/eigen/Eigen/src/plugins/ArrayCwiseUnaryOps.h +0 -696
  426. package/eigen/Eigen/src/plugins/CommonCwiseBinaryOps.h +0 -115
  427. package/eigen/Eigen/src/plugins/IndexedViewMethods.h +0 -262
  428. package/eigen/Eigen/src/plugins/MatrixCwiseBinaryOps.h +0 -152
  429. package/eigen/Eigen/src/plugins/MatrixCwiseUnaryOps.h +0 -95
  430. package/eigen/Eigen/src/plugins/ReshapedMethods.h +0 -149
  431. package/eigen/README.md +0 -5
@@ -10,6 +10,10 @@
10
10
  #ifndef EIGEN_PACKET_MATH_SSE_H
11
11
  #define EIGEN_PACKET_MATH_SSE_H
12
12
 
13
+ #include <cstdint>
14
+ // IWYU pragma: private
15
+ #include "../../InternalHeaderCheck.h"
16
+
13
17
  namespace Eigen {
14
18
 
15
19
  namespace internal {
@@ -21,7 +25,7 @@ namespace internal {
21
25
  #if !defined(EIGEN_VECTORIZE_AVX) && !defined(EIGEN_ARCH_DEFAULT_NUMBER_OF_REGISTERS)
22
26
  // 32 bits => 8 registers
23
27
  // 64 bits => 16 registers
24
- #define EIGEN_ARCH_DEFAULT_NUMBER_OF_REGISTERS (2*sizeof(void*))
28
+ #define EIGEN_ARCH_DEFAULT_NUMBER_OF_REGISTERS (2 * sizeof(void*))
25
29
  #endif
26
30
 
27
31
  #ifdef EIGEN_VECTORIZE_FMA
@@ -30,94 +34,137 @@ namespace internal {
30
34
  #endif
31
35
  #endif
32
36
 
33
- #if ((defined EIGEN_VECTORIZE_AVX) && (EIGEN_COMP_GNUC_STRICT || EIGEN_COMP_MINGW) && (__GXX_ABI_VERSION < 1004)) || EIGEN_OS_QNX
37
+ #if ((defined EIGEN_VECTORIZE_AVX) && (EIGEN_COMP_GNUC_STRICT || EIGEN_COMP_MINGW || EIGEN_COMP_LCC) && \
38
+ (__GXX_ABI_VERSION < 1004)) || \
39
+ EIGEN_OS_QNX
34
40
  // With GCC's default ABI version, a __m128 or __m256 are the same types and therefore we cannot
35
41
  // have overloads for both types without linking error.
36
42
  // One solution is to increase ABI version using -fabi-version=4 (or greater).
37
43
  // Otherwise, we workaround this inconvenience by wrapping 128bit types into the following helper
38
44
  // structure:
39
- typedef eigen_packet_wrapper<__m128> Packet4f;
45
+ typedef eigen_packet_wrapper<__m128> Packet4f;
40
46
  typedef eigen_packet_wrapper<__m128d> Packet2d;
41
47
  #else
42
- typedef __m128 Packet4f;
48
+ typedef __m128 Packet4f;
43
49
  typedef __m128d Packet2d;
44
50
  #endif
45
51
 
46
52
  typedef eigen_packet_wrapper<__m128i, 0> Packet4i;
47
53
  typedef eigen_packet_wrapper<__m128i, 1> Packet16b;
54
+ typedef eigen_packet_wrapper<__m128i, 4> Packet4ui;
55
+ typedef eigen_packet_wrapper<__m128i, 5> Packet2l;
48
56
 
49
- template<> struct is_arithmetic<__m128> { enum { value = true }; };
50
- template<> struct is_arithmetic<__m128i> { enum { value = true }; };
51
- template<> struct is_arithmetic<__m128d> { enum { value = true }; };
52
- template<> struct is_arithmetic<Packet4i> { enum { value = true }; };
53
- template<> struct is_arithmetic<Packet16b> { enum { value = true }; };
57
+ template <>
58
+ struct is_arithmetic<__m128> {
59
+ enum { value = true };
60
+ };
61
+ template <>
62
+ struct is_arithmetic<__m128i> {
63
+ enum { value = true };
64
+ };
65
+ template <>
66
+ struct is_arithmetic<__m128d> {
67
+ enum { value = true };
68
+ };
69
+ template <>
70
+ struct is_arithmetic<Packet4i> {
71
+ enum { value = true };
72
+ };
73
+ template <>
74
+ struct is_arithmetic<Packet2l> {
75
+ enum { value = true };
76
+ };
77
+ // Note that `Packet4ui` uses the underlying type `__m128i`, which is
78
+ // interpreted as a vector of _signed_ `int32`s, which breaks some arithmetic
79
+ // operations used in `GenericPacketMath.h`.
80
+ template <>
81
+ struct is_arithmetic<Packet4ui> {
82
+ enum { value = false };
83
+ };
84
+ template <>
85
+ struct is_arithmetic<Packet16b> {
86
+ enum { value = true };
87
+ };
54
88
 
55
- template<int p, int q, int r, int s>
56
- struct shuffle_mask{
57
- enum { mask = (s)<<6|(r)<<4|(q)<<2|(p) };
89
+ template <int p, int q, int r, int s>
90
+ struct shuffle_mask {
91
+ enum { mask = (s) << 6 | (r) << 4 | (q) << 2 | (p) };
58
92
  };
59
93
 
60
94
  // TODO: change the implementation of all swizzle* ops from macro to template,
61
- #define vec4f_swizzle1(v,p,q,r,s) \
62
- Packet4f(_mm_castsi128_ps(_mm_shuffle_epi32( _mm_castps_si128(v), (shuffle_mask<p,q,r,s>::mask))))
95
+ #define vec4f_swizzle1(v, p, q, r, s) \
96
+ Packet4f(_mm_castsi128_ps(_mm_shuffle_epi32(_mm_castps_si128(v), (shuffle_mask<p, q, r, s>::mask))))
63
97
 
64
- #define vec4i_swizzle1(v,p,q,r,s) \
65
- Packet4i(_mm_shuffle_epi32( v, (shuffle_mask<p,q,r,s>::mask)))
98
+ #define vec4i_swizzle1(v, p, q, r, s) Packet4i(_mm_shuffle_epi32(v, (shuffle_mask<p, q, r, s>::mask)))
66
99
 
67
- #define vec2d_swizzle1(v,p,q) \
68
- Packet2d(_mm_castsi128_pd(_mm_shuffle_epi32( _mm_castpd_si128(v), (shuffle_mask<2*p,2*p+1,2*q,2*q+1>::mask))))
100
+ #define vec4ui_swizzle1(v, p, q, r, s) Packet4ui(vec4i_swizzle1(v, p, q, r, s))
69
101
 
70
- #define vec4f_swizzle2(a,b,p,q,r,s) \
71
- Packet4f(_mm_shuffle_ps( (a), (b), (shuffle_mask<p,q,r,s>::mask)))
102
+ #define vec2d_swizzle1(v, p, q) \
103
+ Packet2d(_mm_castsi128_pd( \
104
+ _mm_shuffle_epi32(_mm_castpd_si128(v), (shuffle_mask<2 * p, 2 * p + 1, 2 * q, 2 * q + 1>::mask))))
72
105
 
73
- #define vec4i_swizzle2(a,b,p,q,r,s) \
74
- Packet4i(_mm_castps_si128( (_mm_shuffle_ps( _mm_castsi128_ps(a), _mm_castsi128_ps(b), (shuffle_mask<p,q,r,s>::mask)))))
106
+ #define vec4f_swizzle2(a, b, p, q, r, s) Packet4f(_mm_shuffle_ps((a), (b), (shuffle_mask<p, q, r, s>::mask)))
75
107
 
76
- EIGEN_STRONG_INLINE Packet4f vec4f_movelh(const Packet4f& a, const Packet4f& b)
77
- {
78
- return Packet4f(_mm_movelh_ps(a,b));
108
+ #define vec4i_swizzle2(a, b, p, q, r, s) \
109
+ Packet4i( \
110
+ _mm_castps_si128((_mm_shuffle_ps(_mm_castsi128_ps(a), _mm_castsi128_ps(b), (shuffle_mask<p, q, r, s>::mask)))))
111
+
112
+ #define vec4ui_swizzle2(a, b, p, q, r, s) Packet4i(vec4i_swizzle2(a, b, p, q, r, s))
113
+
114
+ EIGEN_STRONG_INLINE Packet4f vec4f_movelh(const Packet4f& a, const Packet4f& b) {
115
+ return Packet4f(_mm_movelh_ps(a, b));
79
116
  }
80
- EIGEN_STRONG_INLINE Packet4f vec4f_movehl(const Packet4f& a, const Packet4f& b)
81
- {
82
- return Packet4f(_mm_movehl_ps(a,b));
117
+ EIGEN_STRONG_INLINE Packet4f vec4f_movehl(const Packet4f& a, const Packet4f& b) {
118
+ return Packet4f(_mm_movehl_ps(a, b));
83
119
  }
84
- EIGEN_STRONG_INLINE Packet4f vec4f_unpacklo(const Packet4f& a, const Packet4f& b)
85
- {
86
- return Packet4f(_mm_unpacklo_ps(a,b));
120
+ EIGEN_STRONG_INLINE Packet4f vec4f_unpacklo(const Packet4f& a, const Packet4f& b) {
121
+ return Packet4f(_mm_unpacklo_ps(a, b));
87
122
  }
88
- EIGEN_STRONG_INLINE Packet4f vec4f_unpackhi(const Packet4f& a, const Packet4f& b)
89
- {
90
- return Packet4f(_mm_unpackhi_ps(a,b));
123
+ EIGEN_STRONG_INLINE Packet4f vec4f_unpackhi(const Packet4f& a, const Packet4f& b) {
124
+ return Packet4f(_mm_unpackhi_ps(a, b));
91
125
  }
92
- #define vec4f_duplane(a,p) \
93
- vec4f_swizzle2(a,a,p,p,p,p)
126
+ #define vec4f_duplane(a, p) vec4f_swizzle2(a, a, p, p, p, p)
94
127
 
95
- #define vec2d_swizzle2(a,b,mask) \
96
- Packet2d(_mm_shuffle_pd(a,b,mask))
128
+ #define vec2d_swizzle2(a, b, mask) Packet2d(_mm_shuffle_pd(a, b, mask))
97
129
 
98
- EIGEN_STRONG_INLINE Packet2d vec2d_unpacklo(const Packet2d& a, const Packet2d& b)
99
- {
100
- return Packet2d(_mm_unpacklo_pd(a,b));
130
+ EIGEN_STRONG_INLINE Packet2d vec2d_unpacklo(const Packet2d& a, const Packet2d& b) {
131
+ return Packet2d(_mm_unpacklo_pd(a, b));
101
132
  }
102
- EIGEN_STRONG_INLINE Packet2d vec2d_unpackhi(const Packet2d& a, const Packet2d& b)
103
- {
104
- return Packet2d(_mm_unpackhi_pd(a,b));
133
+ EIGEN_STRONG_INLINE Packet2d vec2d_unpackhi(const Packet2d& a, const Packet2d& b) {
134
+ return Packet2d(_mm_unpackhi_pd(a, b));
105
135
  }
106
- #define vec2d_duplane(a,p) \
107
- vec2d_swizzle2(a,a,(p<<1)|p)
136
+ #define vec2d_duplane(a, p) vec2d_swizzle2(a, a, (p << 1) | p)
137
+
138
+ #define EIGEN_DECLARE_CONST_Packet4f(NAME, X) const Packet4f p4f_##NAME = pset1<Packet4f>(X)
108
139
 
109
- #define _EIGEN_DECLARE_CONST_Packet4f(NAME,X) \
110
- const Packet4f p4f_##NAME = pset1<Packet4f>(X)
140
+ #define EIGEN_DECLARE_CONST_Packet2d(NAME, X) const Packet2d p2d_##NAME = pset1<Packet2d>(X)
111
141
 
112
- #define _EIGEN_DECLARE_CONST_Packet2d(NAME,X) \
113
- const Packet2d p2d_##NAME = pset1<Packet2d>(X)
142
+ #define EIGEN_DECLARE_CONST_Packet4f_FROM_INT(NAME, X) const Packet4f p4f_##NAME = pset1frombits<Packet4f>(X)
114
143
 
115
- #define _EIGEN_DECLARE_CONST_Packet4f_FROM_INT(NAME,X) \
116
- const Packet4f p4f_##NAME = pset1frombits<Packet4f>(X)
144
+ #define EIGEN_DECLARE_CONST_Packet4i(NAME, X) const Packet4i p4i_##NAME = pset1<Packet4i>(X)
117
145
 
118
- #define _EIGEN_DECLARE_CONST_Packet4i(NAME,X) \
119
- const Packet4i p4i_##NAME = pset1<Packet4i>(X)
146
+ #define EIGEN_DECLARE_CONST_Packet4ui(NAME, X) const Packet4ui p4ui_##NAME = pset1<Packet4ui>(X)
120
147
 
148
+ // Work around lack of extract/cvt for epi64 when compiling for 32-bit.
149
+ #if EIGEN_ARCH_x86_64
150
+ EIGEN_ALWAYS_INLINE int64_t _mm_extract_epi64_0(const __m128i& a) { return _mm_cvtsi128_si64(a); }
151
+ #ifdef EIGEN_VECTORIZE_SSE4_1
152
+ EIGEN_ALWAYS_INLINE int64_t _mm_extract_epi64_1(const __m128i& a) { return _mm_extract_epi64(a, 1); }
153
+ #else
154
+ EIGEN_ALWAYS_INLINE int64_t _mm_extract_epi64_1(const __m128i& a) {
155
+ return _mm_cvtsi128_si64(_mm_castpd_si128(_mm_shuffle_pd(_mm_castsi128_pd(a), _mm_castsi128_pd(a), 0x1)));
156
+ }
157
+ #endif
158
+ #else
159
+ // epi64 instructions are not available. The following seems to generate the same instructions
160
+ // with -O2 in GCC/Clang.
161
+ EIGEN_ALWAYS_INLINE int64_t _mm_extract_epi64_0(const __m128i& a) {
162
+ return numext::bit_cast<int64_t>(_mm_cvtsd_f64(_mm_castsi128_pd(a)));
163
+ }
164
+ EIGEN_ALWAYS_INLINE int64_t _mm_extract_epi64_1(const __m128i& a) {
165
+ return numext::bit_cast<int64_t>(_mm_cvtsd_f64(_mm_shuffle_pd(_mm_castsi128_pd(a), _mm_castsi128_pd(a), 0x1)));
166
+ }
167
+ #endif
121
168
 
122
169
  // Use the packet_traits defined in AVX/PacketMath.h instead if we're going
123
170
  // to leverage AVX instructions.
@@ -130,29 +177,31 @@ struct packet_traits<float> : default_packet_traits {
130
177
  Vectorizable = 1,
131
178
  AlignedOnScalar = 1,
132
179
  size = 4,
133
- HasHalfPacket = 0,
134
180
 
135
- HasCmp = 1,
181
+ HasCmp = 1,
136
182
  HasDiv = 1,
183
+ HasReciprocal = EIGEN_FAST_MATH,
137
184
  HasSin = EIGEN_FAST_MATH,
138
185
  HasCos = EIGEN_FAST_MATH,
186
+ HasACos = 1,
187
+ HasASin = 1,
188
+ HasATan = 1,
189
+ HasATanh = 1,
139
190
  HasLog = 1,
140
191
  HasLog1p = 1,
141
192
  HasExpm1 = 1,
142
193
  HasNdtri = 1,
143
194
  HasExp = 1,
195
+ HasPow = 1,
144
196
  HasBessel = 1,
145
197
  HasSqrt = 1,
146
198
  HasRsqrt = 1,
199
+ HasCbrt = 1,
147
200
  HasTanh = EIGEN_FAST_MATH,
148
201
  HasErf = EIGEN_FAST_MATH,
202
+ HasErfc = EIGEN_FAST_MATH,
149
203
  HasBlend = 1,
150
- HasCeil = 1,
151
- HasFloor = 1,
152
- #ifdef EIGEN_VECTORIZE_SSE4_1
153
- HasRound = 1,
154
- #endif
155
- HasRint = 1
204
+ HasSign = 0 // The manually vectorized version is slightly slower for SSE.
156
205
  };
157
206
  };
158
207
  template <>
@@ -162,114 +211,256 @@ struct packet_traits<double> : default_packet_traits {
162
211
  enum {
163
212
  Vectorizable = 1,
164
213
  AlignedOnScalar = 1,
165
- size=2,
166
- HasHalfPacket = 0,
214
+ size = 2,
167
215
 
168
- HasCmp = 1,
169
- HasDiv = 1,
170
- HasLog = 1,
171
- HasExp = 1,
216
+ HasCmp = 1,
217
+ HasDiv = 1,
218
+ HasSin = EIGEN_FAST_MATH,
219
+ HasCos = EIGEN_FAST_MATH,
220
+ HasTanh = EIGEN_FAST_MATH,
221
+ HasLog = 1,
222
+ HasErf = EIGEN_FAST_MATH,
223
+ HasErfc = EIGEN_FAST_MATH,
224
+ HasExp = 1,
225
+ HasPow = 1,
172
226
  HasSqrt = 1,
173
227
  HasRsqrt = 1,
174
- HasBlend = 1,
175
- HasFloor = 1,
176
- HasCeil = 1,
177
- #ifdef EIGEN_VECTORIZE_SSE4_1
178
- HasRound = 1,
179
- #endif
180
- HasRint = 1
228
+ HasCbrt = 1,
229
+ HasATan = 1,
230
+ HasATanh = 1,
231
+ HasBlend = 1
181
232
  };
182
233
  };
183
- #endif
184
- template<> struct packet_traits<int> : default_packet_traits
185
- {
234
+ template <>
235
+ struct packet_traits<int> : default_packet_traits {
186
236
  typedef Packet4i type;
187
237
  typedef Packet4i half;
188
238
  enum {
189
239
  Vectorizable = 1,
190
240
  AlignedOnScalar = 1,
191
- size=4,
241
+ size = 4,
192
242
 
243
+ HasCmp = 1,
244
+ HasDiv = 1,
193
245
  HasShift = 1,
194
246
  HasBlend = 1
195
247
  };
196
248
  };
249
+ template <>
250
+ struct packet_traits<uint32_t> : default_packet_traits {
251
+ typedef Packet4ui type;
252
+ typedef Packet4ui half;
253
+ enum {
254
+ Vectorizable = 1,
255
+ AlignedOnScalar = 1,
256
+ size = 4,
197
257
 
198
- template<> struct packet_traits<bool> : default_packet_traits
199
- {
258
+ HasDiv = 0,
259
+ HasNegate = 0,
260
+ HasCmp = 1,
261
+ HasShift = 1,
262
+ HasBlend = 1
263
+ };
264
+ };
265
+ template <>
266
+ struct packet_traits<int64_t> : default_packet_traits {
267
+ typedef Packet2l type;
268
+ typedef Packet2l half;
269
+ enum {
270
+ Vectorizable = 1,
271
+ AlignedOnScalar = 1,
272
+ size = 2,
273
+
274
+ HasDiv = 0,
275
+ HasCmp = 1,
276
+ HasShift = 1,
277
+ HasBlend = 1
278
+ };
279
+ };
280
+ #endif
281
+ template <>
282
+ struct packet_traits<bool> : default_packet_traits {
200
283
  typedef Packet16b type;
201
284
  typedef Packet16b half;
202
285
  enum {
203
286
  Vectorizable = 1,
204
287
  AlignedOnScalar = 1,
205
- HasHalfPacket = 0,
206
- size=16,
207
-
208
- HasAdd = 1,
209
- HasSub = 1,
210
- HasShift = 0,
211
- HasMul = 1,
212
- HasNegate = 1,
213
- HasAbs = 0,
214
- HasAbs2 = 0,
215
- HasMin = 0,
216
- HasMax = 0,
217
- HasConj = 0,
218
- HasSqrt = 1
288
+ size = 16,
289
+
290
+ HasCmp = 1,
291
+ HasShift = 0,
292
+ HasAbs = 0,
293
+ HasAbs2 = 0,
294
+ HasMin = 0,
295
+ HasMax = 0,
296
+ HasConj = 0,
297
+ HasSqrt = 1,
298
+ HasNegate = 0,
299
+ HasSign = 0 // Don't try to vectorize psign<bool> = identity.
219
300
  };
220
301
  };
221
302
 
222
- template<> struct unpacket_traits<Packet4f> {
223
- typedef float type;
224
- typedef Packet4f half;
225
- typedef Packet4i integer_packet;
226
- enum {size=4, alignment=Aligned16, vectorizable=true, masked_load_available=false, masked_store_available=false};
303
+ template <>
304
+ struct unpacket_traits<Packet4f> {
305
+ typedef float type;
306
+ typedef Packet4f half;
307
+ typedef Packet4i integer_packet;
308
+ enum {
309
+ size = 4,
310
+ alignment = Aligned16,
311
+ vectorizable = true,
312
+ masked_load_available = false,
313
+ masked_store_available = false
314
+ };
315
+ };
316
+ template <>
317
+ struct unpacket_traits<Packet2d> {
318
+ typedef double type;
319
+ typedef Packet2d half;
320
+ typedef Packet2l integer_packet;
321
+ enum {
322
+ size = 2,
323
+ alignment = Aligned16,
324
+ vectorizable = true,
325
+ masked_load_available = false,
326
+ masked_store_available = false
327
+ };
328
+ };
329
+ template <>
330
+ struct unpacket_traits<Packet2l> {
331
+ typedef int64_t type;
332
+ typedef Packet2l half;
333
+ enum {
334
+ size = 2,
335
+ alignment = Aligned16,
336
+ vectorizable = true,
337
+ masked_load_available = false,
338
+ masked_store_available = false
339
+ };
227
340
  };
228
- template<> struct unpacket_traits<Packet2d> {
229
- typedef double type;
230
- typedef Packet2d half;
231
- enum {size=2, alignment=Aligned16, vectorizable=true, masked_load_available=false, masked_store_available=false};
341
+ template <>
342
+ struct unpacket_traits<Packet4i> {
343
+ typedef int type;
344
+ typedef Packet4i half;
345
+ enum {
346
+ size = 4,
347
+ alignment = Aligned16,
348
+ vectorizable = true,
349
+ masked_load_available = false,
350
+ masked_store_available = false
351
+ };
232
352
  };
233
- template<> struct unpacket_traits<Packet4i> {
234
- typedef int type;
235
- typedef Packet4i half;
236
- enum {size=4, alignment=Aligned16, vectorizable=false, masked_load_available=false, masked_store_available=false};
353
+ template <>
354
+ struct unpacket_traits<Packet4ui> {
355
+ typedef uint32_t type;
356
+ typedef Packet4ui half;
357
+ enum {
358
+ size = 4,
359
+ alignment = Aligned16,
360
+ vectorizable = true,
361
+ masked_load_available = false,
362
+ masked_store_available = false
363
+ };
237
364
  };
238
- template<> struct unpacket_traits<Packet16b> {
239
- typedef bool type;
240
- typedef Packet16b half;
241
- enum {size=16, alignment=Aligned16, vectorizable=true, masked_load_available=false, masked_store_available=false};
365
+ template <>
366
+ struct unpacket_traits<Packet16b> {
367
+ typedef bool type;
368
+ typedef Packet16b half;
369
+ enum {
370
+ size = 16,
371
+ alignment = Aligned16,
372
+ vectorizable = true,
373
+ masked_load_available = false,
374
+ masked_store_available = false
375
+ };
242
376
  };
243
377
 
244
378
  #ifndef EIGEN_VECTORIZE_AVX
245
- template<> struct scalar_div_cost<float,true> { enum { value = 7 }; };
246
- template<> struct scalar_div_cost<double,true> { enum { value = 8 }; };
379
+ template <>
380
+ struct scalar_div_cost<float, true> {
381
+ enum { value = 7 };
382
+ };
383
+ template <>
384
+ struct scalar_div_cost<double, true> {
385
+ enum { value = 8 };
386
+ };
247
387
  #endif
248
388
 
249
- #if EIGEN_COMP_MSVC==1500
250
- // Workaround MSVC 9 internal compiler error.
251
- // TODO: It has been detected with win64 builds (amd64), so let's check whether it also happens in 32bits+SSE mode
252
- // TODO: let's check whether there does not exist a better fix, like adding a pset0() function. (it crashed on pset1(0)).
253
- template<> EIGEN_STRONG_INLINE Packet4f pset1<Packet4f>(const float& from) { return _mm_set_ps(from,from,from,from); }
254
- template<> EIGEN_STRONG_INLINE Packet2d pset1<Packet2d>(const double& from) { return _mm_set_pd(from,from); }
255
- template<> EIGEN_STRONG_INLINE Packet4i pset1<Packet4i>(const int& from) { return _mm_set_epi32(from,from,from,from); }
256
- #else
257
- template<> EIGEN_STRONG_INLINE Packet4f pset1<Packet4f>(const float& from) { return _mm_set_ps1(from); }
258
- template<> EIGEN_STRONG_INLINE Packet2d pset1<Packet2d>(const double& from) { return _mm_set1_pd(from); }
259
- template<> EIGEN_STRONG_INLINE Packet4i pset1<Packet4i>(const int& from) { return _mm_set1_epi32(from); }
260
- #endif
261
- template<> EIGEN_STRONG_INLINE Packet16b pset1<Packet16b>(const bool& from) { return _mm_set1_epi8(static_cast<char>(from)); }
389
+ template <>
390
+ EIGEN_STRONG_INLINE Packet4f pset1<Packet4f>(const float& from) {
391
+ return _mm_set_ps1(from);
392
+ }
393
+ template <>
394
+ EIGEN_STRONG_INLINE Packet2d pset1<Packet2d>(const double& from) {
395
+ return _mm_set1_pd(from);
396
+ }
397
+ template <>
398
+ EIGEN_STRONG_INLINE Packet2l pset1<Packet2l>(const int64_t& from) {
399
+ return _mm_set1_epi64x(from);
400
+ }
401
+ template <>
402
+ EIGEN_STRONG_INLINE Packet4i pset1<Packet4i>(const int& from) {
403
+ return _mm_set1_epi32(from);
404
+ }
405
+ template <>
406
+ EIGEN_STRONG_INLINE Packet4ui pset1<Packet4ui>(const uint32_t& from) {
407
+ return _mm_set1_epi32(numext::bit_cast<int32_t>(from));
408
+ }
409
+ template <>
410
+ EIGEN_STRONG_INLINE Packet16b pset1<Packet16b>(const bool& from) {
411
+ return _mm_set1_epi8(static_cast<char>(from));
412
+ }
262
413
 
263
- template<> EIGEN_STRONG_INLINE Packet4f pset1frombits<Packet4f>(unsigned int from) { return _mm_castsi128_ps(pset1<Packet4i>(from)); }
264
- template<> EIGEN_STRONG_INLINE Packet2d pset1frombits<Packet2d>(uint64_t from) { return _mm_castsi128_pd(_mm_set1_epi64x(from)); }
414
+ template <>
415
+ EIGEN_STRONG_INLINE Packet4f pset1frombits<Packet4f>(unsigned int from) {
416
+ return _mm_castsi128_ps(pset1<Packet4i>(from));
417
+ }
418
+ template <>
419
+ EIGEN_STRONG_INLINE Packet2d pset1frombits<Packet2d>(uint64_t from) {
420
+ return _mm_castsi128_pd(_mm_set1_epi64x(from));
421
+ }
265
422
 
266
- template<> EIGEN_STRONG_INLINE Packet4f peven_mask(const Packet4f& /*a*/) { return _mm_castsi128_ps(_mm_set_epi32(0, -1, 0, -1)); }
267
- template<> EIGEN_STRONG_INLINE Packet4i peven_mask(const Packet4i& /*a*/) { return _mm_set_epi32(0, -1, 0, -1); }
268
- template<> EIGEN_STRONG_INLINE Packet2d peven_mask(const Packet2d& /*a*/) { return _mm_castsi128_pd(_mm_set_epi32(0, 0, -1, -1)); }
423
+ template <>
424
+ EIGEN_STRONG_INLINE Packet4f peven_mask(const Packet4f& /*a*/) {
425
+ return _mm_castsi128_ps(_mm_set_epi32(0, -1, 0, -1));
426
+ }
427
+ template <>
428
+ EIGEN_STRONG_INLINE Packet2l peven_mask(const Packet2l& /*a*/) {
429
+ return _mm_set_epi32(0, 0, -1, -1);
430
+ }
431
+ template <>
432
+ EIGEN_STRONG_INLINE Packet4i peven_mask(const Packet4i& /*a*/) {
433
+ return _mm_set_epi32(0, -1, 0, -1);
434
+ }
435
+ template <>
436
+ EIGEN_STRONG_INLINE Packet4ui peven_mask(const Packet4ui& /*a*/) {
437
+ return _mm_set_epi32(0, -1, 0, -1);
438
+ }
439
+ template <>
440
+ EIGEN_STRONG_INLINE Packet2d peven_mask(const Packet2d& /*a*/) {
441
+ return _mm_castsi128_pd(_mm_set_epi32(0, 0, -1, -1));
442
+ }
269
443
 
270
- template<> EIGEN_STRONG_INLINE Packet4f pzero(const Packet4f& /*a*/) { return _mm_setzero_ps(); }
271
- template<> EIGEN_STRONG_INLINE Packet2d pzero(const Packet2d& /*a*/) { return _mm_setzero_pd(); }
272
- template<> EIGEN_STRONG_INLINE Packet4i pzero(const Packet4i& /*a*/) { return _mm_setzero_si128(); }
444
+ template <>
445
+ EIGEN_STRONG_INLINE Packet4f pzero(const Packet4f& /*a*/) {
446
+ return _mm_setzero_ps();
447
+ }
448
+ template <>
449
+ EIGEN_STRONG_INLINE Packet2d pzero(const Packet2d& /*a*/) {
450
+ return _mm_setzero_pd();
451
+ }
452
+ template <>
453
+ EIGEN_STRONG_INLINE Packet2l pzero(const Packet2l& /*a*/) {
454
+ return _mm_setzero_si128();
455
+ }
456
+ template <>
457
+ EIGEN_STRONG_INLINE Packet4i pzero(const Packet4i& /*a*/) {
458
+ return _mm_setzero_si128();
459
+ }
460
+ template <>
461
+ EIGEN_STRONG_INLINE Packet4ui pzero(const Packet4ui& /*a*/) {
462
+ return _mm_setzero_si128();
463
+ }
273
464
 
274
465
  // GCC generates a shufps instruction for _mm_set1_ps/_mm_load1_ps instead of the more efficient pshufd instruction.
275
466
  // However, using inrinsics for pset1 makes gcc to generate crappy code in some cases (see bug 203)
@@ -277,268 +468,664 @@ template<> EIGEN_STRONG_INLINE Packet4i pzero(const Packet4i& /*a*/) { return _m
277
468
  // Therefore, we introduced the pload1 functions to be used in product kernels for which bug 203 does not apply.
278
469
  // Also note that with AVX, we want it to generate a vbroadcastss.
279
470
  #if EIGEN_COMP_GNUC_STRICT && (!defined __AVX__)
280
- template<> EIGEN_STRONG_INLINE Packet4f pload1<Packet4f>(const float *from) {
281
- return vec4f_swizzle1(_mm_load_ss(from),0,0,0,0);
471
+ template <>
472
+ EIGEN_STRONG_INLINE Packet4f pload1<Packet4f>(const float* from) {
473
+ return vec4f_swizzle1(_mm_load_ss(from), 0, 0, 0, 0);
282
474
  }
283
475
  #endif
284
476
 
285
- template<> EIGEN_STRONG_INLINE Packet4f plset<Packet4f>(const float& a) { return _mm_add_ps(pset1<Packet4f>(a), _mm_set_ps(3,2,1,0)); }
286
- template<> EIGEN_STRONG_INLINE Packet2d plset<Packet2d>(const double& a) { return _mm_add_pd(pset1<Packet2d>(a),_mm_set_pd(1,0)); }
287
- template<> EIGEN_STRONG_INLINE Packet4i plset<Packet4i>(const int& a) { return _mm_add_epi32(pset1<Packet4i>(a),_mm_set_epi32(3,2,1,0)); }
477
+ template <>
478
+ EIGEN_STRONG_INLINE Packet4f plset<Packet4f>(const float& a) {
479
+ return _mm_add_ps(pset1<Packet4f>(a), _mm_set_ps(3, 2, 1, 0));
480
+ }
481
+ template <>
482
+ EIGEN_STRONG_INLINE Packet2d plset<Packet2d>(const double& a) {
483
+ return _mm_add_pd(pset1<Packet2d>(a), _mm_set_pd(1, 0));
484
+ }
485
+ template <>
486
+ EIGEN_STRONG_INLINE Packet2l plset<Packet2l>(const int64_t& a) {
487
+ return _mm_add_epi32(pset1<Packet2l>(a), _mm_set_epi64x(1, 0));
488
+ }
489
+ template <>
490
+ EIGEN_STRONG_INLINE Packet4i plset<Packet4i>(const int& a) {
491
+ return _mm_add_epi32(pset1<Packet4i>(a), _mm_set_epi32(3, 2, 1, 0));
492
+ }
493
+ template <>
494
+ EIGEN_STRONG_INLINE Packet4ui plset<Packet4ui>(const uint32_t& a) {
495
+ return _mm_add_epi32(pset1<Packet4ui>(a), _mm_set_epi32(3, 2, 1, 0));
496
+ }
497
+
498
+ template <>
499
+ EIGEN_STRONG_INLINE Packet4f padd<Packet4f>(const Packet4f& a, const Packet4f& b) {
500
+ return _mm_add_ps(a, b);
501
+ }
502
+ template <>
503
+ EIGEN_STRONG_INLINE Packet2d padd<Packet2d>(const Packet2d& a, const Packet2d& b) {
504
+ return _mm_add_pd(a, b);
505
+ }
506
+ template <>
507
+ EIGEN_STRONG_INLINE Packet2l padd<Packet2l>(const Packet2l& a, const Packet2l& b) {
508
+ return _mm_add_epi64(a, b);
509
+ }
510
+ template <>
511
+ EIGEN_STRONG_INLINE Packet4i padd<Packet4i>(const Packet4i& a, const Packet4i& b) {
512
+ return _mm_add_epi32(a, b);
513
+ }
514
+ template <>
515
+ EIGEN_STRONG_INLINE Packet4ui padd<Packet4ui>(const Packet4ui& a, const Packet4ui& b) {
516
+ return _mm_add_epi32(a, b);
517
+ }
288
518
 
289
- template<> EIGEN_STRONG_INLINE Packet4f padd<Packet4f>(const Packet4f& a, const Packet4f& b) { return _mm_add_ps(a,b); }
290
- template<> EIGEN_STRONG_INLINE Packet2d padd<Packet2d>(const Packet2d& a, const Packet2d& b) { return _mm_add_pd(a,b); }
291
- template<> EIGEN_STRONG_INLINE Packet4i padd<Packet4i>(const Packet4i& a, const Packet4i& b) { return _mm_add_epi32(a,b); }
519
+ template <>
520
+ EIGEN_STRONG_INLINE Packet16b padd<Packet16b>(const Packet16b& a, const Packet16b& b) {
521
+ return _mm_or_si128(a, b);
522
+ }
292
523
 
293
- template<> EIGEN_STRONG_INLINE Packet16b padd<Packet16b>(const Packet16b& a, const Packet16b& b) { return _mm_or_si128(a,b); }
524
+ template <typename Packet>
525
+ EIGEN_STRONG_INLINE Packet padds(const Packet& a, const Packet& b);
526
+ template <>
527
+ EIGEN_STRONG_INLINE Packet4f padds<Packet4f>(const Packet4f& a, const Packet4f& b) {
528
+ return _mm_add_ss(a, b);
529
+ }
530
+ template <>
531
+ EIGEN_STRONG_INLINE Packet2d padds<Packet2d>(const Packet2d& a, const Packet2d& b) {
532
+ return _mm_add_sd(a, b);
533
+ }
294
534
 
295
- template<> EIGEN_STRONG_INLINE Packet4f psub<Packet4f>(const Packet4f& a, const Packet4f& b) { return _mm_sub_ps(a,b); }
296
- template<> EIGEN_STRONG_INLINE Packet2d psub<Packet2d>(const Packet2d& a, const Packet2d& b) { return _mm_sub_pd(a,b); }
297
- template<> EIGEN_STRONG_INLINE Packet4i psub<Packet4i>(const Packet4i& a, const Packet4i& b) { return _mm_sub_epi32(a,b); }
298
- template<> EIGEN_STRONG_INLINE Packet16b psub<Packet16b>(const Packet16b& a, const Packet16b& b) { return _mm_xor_si128(a,b); }
535
+ template <>
536
+ EIGEN_STRONG_INLINE Packet4f psub<Packet4f>(const Packet4f& a, const Packet4f& b) {
537
+ return _mm_sub_ps(a, b);
538
+ }
539
+ template <>
540
+ EIGEN_STRONG_INLINE Packet2d psub<Packet2d>(const Packet2d& a, const Packet2d& b) {
541
+ return _mm_sub_pd(a, b);
542
+ }
543
+ template <>
544
+ EIGEN_STRONG_INLINE Packet2l psub<Packet2l>(const Packet2l& a, const Packet2l& b) {
545
+ return _mm_sub_epi64(a, b);
546
+ }
547
+ template <>
548
+ EIGEN_STRONG_INLINE Packet4i psub<Packet4i>(const Packet4i& a, const Packet4i& b) {
549
+ return _mm_sub_epi32(a, b);
550
+ }
551
+ template <>
552
+ EIGEN_STRONG_INLINE Packet4ui psub<Packet4ui>(const Packet4ui& a, const Packet4ui& b) {
553
+ return _mm_sub_epi32(a, b);
554
+ }
555
+ template <>
556
+ EIGEN_STRONG_INLINE Packet16b psub<Packet16b>(const Packet16b& a, const Packet16b& b) {
557
+ return _mm_xor_si128(a, b);
558
+ }
299
559
 
300
- template<> EIGEN_STRONG_INLINE Packet4f pxor<Packet4f>(const Packet4f& a, const Packet4f& b);
301
- template<> EIGEN_STRONG_INLINE Packet4f paddsub<Packet4f>(const Packet4f& a, const Packet4f& b)
302
- {
560
+ template <>
561
+ EIGEN_STRONG_INLINE Packet4f pxor<Packet4f>(const Packet4f& a, const Packet4f& b);
562
+ template <>
563
+ EIGEN_STRONG_INLINE Packet4f paddsub<Packet4f>(const Packet4f& a, const Packet4f& b) {
303
564
  #ifdef EIGEN_VECTORIZE_SSE3
304
- return _mm_addsub_ps(a,b);
565
+ return _mm_addsub_ps(a, b);
305
566
  #else
306
- const Packet4f mask = _mm_castsi128_ps(_mm_setr_epi32(0x80000000,0x0,0x80000000,0x0));
567
+ const Packet4f mask = _mm_castsi128_ps(_mm_setr_epi32(0x80000000, 0x0, 0x80000000, 0x0));
307
568
  return padd(a, pxor(mask, b));
308
569
  #endif
309
570
  }
310
571
 
311
- template<> EIGEN_STRONG_INLINE Packet2d pxor<Packet2d>(const Packet2d& , const Packet2d& );
312
- template<> EIGEN_STRONG_INLINE Packet2d paddsub<Packet2d>(const Packet2d& a, const Packet2d& b)
313
- {
314
- #ifdef EIGEN_VECTORIZE_SSE3
315
- return _mm_addsub_pd(a,b);
572
+ template <>
573
+ EIGEN_STRONG_INLINE Packet2d pxor<Packet2d>(const Packet2d&, const Packet2d&);
574
+ template <>
575
+ EIGEN_STRONG_INLINE Packet2d paddsub<Packet2d>(const Packet2d& a, const Packet2d& b) {
576
+ #ifdef EIGEN_VECTORIZE_SSE3
577
+ return _mm_addsub_pd(a, b);
316
578
  #else
317
- const Packet2d mask = _mm_castsi128_pd(_mm_setr_epi32(0x0,0x80000000,0x0,0x0));
579
+ const Packet2d mask = _mm_castsi128_pd(_mm_setr_epi32(0x0, 0x80000000, 0x0, 0x0));
318
580
  return padd(a, pxor(mask, b));
319
581
  #endif
320
582
  }
321
583
 
322
- template<> EIGEN_STRONG_INLINE Packet4f pnegate(const Packet4f& a)
323
- {
324
- const Packet4f mask = _mm_castsi128_ps(_mm_setr_epi32(0x80000000,0x80000000,0x80000000,0x80000000));
325
- return _mm_xor_ps(a,mask);
326
- }
327
- template<> EIGEN_STRONG_INLINE Packet2d pnegate(const Packet2d& a)
328
- {
329
- const Packet2d mask = _mm_castsi128_pd(_mm_setr_epi32(0x0,0x80000000,0x0,0x80000000));
330
- return _mm_xor_pd(a,mask);
584
+ template <>
585
+ EIGEN_STRONG_INLINE Packet4f pnegate(const Packet4f& a) {
586
+ const Packet4f mask = _mm_castsi128_ps(_mm_setr_epi32(0x80000000, 0x80000000, 0x80000000, 0x80000000));
587
+ return _mm_xor_ps(a, mask);
331
588
  }
332
- template<> EIGEN_STRONG_INLINE Packet4i pnegate(const Packet4i& a)
333
- {
334
- return psub(Packet4i(_mm_setr_epi32(0,0,0,0)), a);
589
+ template <>
590
+ EIGEN_STRONG_INLINE Packet2d pnegate(const Packet2d& a) {
591
+ const Packet2d mask = _mm_castsi128_pd(_mm_setr_epi32(0x0, 0x80000000, 0x0, 0x80000000));
592
+ return _mm_xor_pd(a, mask);
335
593
  }
336
-
337
- template<> EIGEN_STRONG_INLINE Packet16b pnegate(const Packet16b& a)
338
- {
339
- return psub(pset1<Packet16b>(false), a);
594
+ template <>
595
+ EIGEN_STRONG_INLINE Packet2l pnegate(const Packet2l& a) {
596
+ return psub(pzero(a), a);
340
597
  }
341
598
 
342
- template<> EIGEN_STRONG_INLINE Packet4f pconj(const Packet4f& a) { return a; }
343
- template<> EIGEN_STRONG_INLINE Packet2d pconj(const Packet2d& a) { return a; }
344
- template<> EIGEN_STRONG_INLINE Packet4i pconj(const Packet4i& a) { return a; }
345
-
346
- template<> EIGEN_STRONG_INLINE Packet4f pmul<Packet4f>(const Packet4f& a, const Packet4f& b) { return _mm_mul_ps(a,b); }
347
- template<> EIGEN_STRONG_INLINE Packet2d pmul<Packet2d>(const Packet2d& a, const Packet2d& b) { return _mm_mul_pd(a,b); }
348
- template<> EIGEN_STRONG_INLINE Packet4i pmul<Packet4i>(const Packet4i& a, const Packet4i& b)
349
- {
350
- #ifdef EIGEN_VECTORIZE_SSE4_1
351
- return _mm_mullo_epi32(a,b);
352
- #else
353
- // this version is slightly faster than 4 scalar products
354
- return vec4i_swizzle1(
355
- vec4i_swizzle2(
356
- _mm_mul_epu32(a,b),
357
- _mm_mul_epu32(vec4i_swizzle1(a,1,0,3,2),
358
- vec4i_swizzle1(b,1,0,3,2)),
359
- 0,2,0,2),
360
- 0,2,1,3);
361
- #endif
599
+ template <>
600
+ EIGEN_STRONG_INLINE Packet4i pnegate(const Packet4i& a) {
601
+ return psub(pzero(a), a);
362
602
  }
363
603
 
364
- template<> EIGEN_STRONG_INLINE Packet16b pmul<Packet16b>(const Packet16b& a, const Packet16b& b) { return _mm_and_si128(a,b); }
365
-
366
- template<> EIGEN_STRONG_INLINE Packet4f pdiv<Packet4f>(const Packet4f& a, const Packet4f& b) { return _mm_div_ps(a,b); }
367
- template<> EIGEN_STRONG_INLINE Packet2d pdiv<Packet2d>(const Packet2d& a, const Packet2d& b) { return _mm_div_pd(a,b); }
368
-
369
- // for some weird raisons, it has to be overloaded for packet of integers
370
- template<> EIGEN_STRONG_INLINE Packet4i pmadd(const Packet4i& a, const Packet4i& b, const Packet4i& c) { return padd(pmul(a,b), c); }
371
- #ifdef EIGEN_VECTORIZE_FMA
372
- template<> EIGEN_STRONG_INLINE Packet4f pmadd(const Packet4f& a, const Packet4f& b, const Packet4f& c) { return _mm_fmadd_ps(a,b,c); }
373
- template<> EIGEN_STRONG_INLINE Packet2d pmadd(const Packet2d& a, const Packet2d& b, const Packet2d& c) { return _mm_fmadd_pd(a,b,c); }
374
- #endif
375
-
376
- #ifdef EIGEN_VECTORIZE_SSE4_1
377
- template<> EIGEN_DEVICE_FUNC inline Packet4f pselect(const Packet4f& mask, const Packet4f& a, const Packet4f& b) {
378
- return _mm_blendv_ps(b,a,mask);
604
+ template <>
605
+ EIGEN_STRONG_INLINE Packet4f pconj(const Packet4f& a) {
606
+ return a;
379
607
  }
380
-
381
- template<> EIGEN_DEVICE_FUNC inline Packet4i pselect(const Packet4i& mask, const Packet4i& a, const Packet4i& b) {
382
- return _mm_castps_si128(_mm_blendv_ps(_mm_castsi128_ps(b),_mm_castsi128_ps(a),_mm_castsi128_ps(mask)));
608
+ template <>
609
+ EIGEN_STRONG_INLINE Packet2d pconj(const Packet2d& a) {
610
+ return a;
383
611
  }
384
-
385
- template<> EIGEN_DEVICE_FUNC inline Packet2d pselect(const Packet2d& mask, const Packet2d& a, const Packet2d& b) { return _mm_blendv_pd(b,a,mask); }
386
-
387
- template<> EIGEN_DEVICE_FUNC inline Packet16b pselect(const Packet16b& mask, const Packet16b& a, const Packet16b& b) {
388
- return _mm_blendv_epi8(b,a,mask);
612
+ template <>
613
+ EIGEN_STRONG_INLINE Packet2l pconj(const Packet2l& a) {
614
+ return a;
389
615
  }
390
- #else
391
- template<> EIGEN_DEVICE_FUNC inline Packet16b pselect(const Packet16b& mask, const Packet16b& a, const Packet16b& b) {
392
- Packet16b a_part = _mm_and_si128(mask, a);
393
- Packet16b b_part = _mm_andnot_si128(mask, b);
394
- return _mm_or_si128(a_part, b_part);
616
+ template <>
617
+ EIGEN_STRONG_INLINE Packet4i pconj(const Packet4i& a) {
618
+ return a;
395
619
  }
396
- #endif
397
620
 
398
- template<> EIGEN_STRONG_INLINE Packet4i ptrue<Packet4i>(const Packet4i& a) { return _mm_cmpeq_epi32(a, a); }
399
- template<> EIGEN_STRONG_INLINE Packet16b ptrue<Packet16b>(const Packet16b& a) { return _mm_cmpeq_epi8(a, a); }
400
- template<> EIGEN_STRONG_INLINE Packet4f
401
- ptrue<Packet4f>(const Packet4f& a) {
402
- Packet4i b = _mm_castps_si128(a);
403
- return _mm_castsi128_ps(_mm_cmpeq_epi32(b, b));
621
+ template <>
622
+ EIGEN_STRONG_INLINE Packet4f pmul<Packet4f>(const Packet4f& a, const Packet4f& b) {
623
+ return _mm_mul_ps(a, b);
404
624
  }
405
- template<> EIGEN_STRONG_INLINE Packet2d
406
- ptrue<Packet2d>(const Packet2d& a) {
407
- Packet4i b = _mm_castpd_si128(a);
408
- return _mm_castsi128_pd(_mm_cmpeq_epi32(b, b));
625
+ template <>
626
+ EIGEN_STRONG_INLINE Packet2d pmul<Packet2d>(const Packet2d& a, const Packet2d& b) {
627
+ return _mm_mul_pd(a, b);
409
628
  }
629
+ template <>
630
+ EIGEN_STRONG_INLINE Packet2l pmul<Packet2l>(const Packet2l& a, const Packet2l& b) {
631
+ // 64-bit mul requires avx512, so do this with 32-bit multiplication
632
+ __m128i upper32_a = _mm_srli_epi64(a, 32);
633
+ __m128i upper32_b = _mm_srli_epi64(b, 32);
410
634
 
635
+ // upper * lower
636
+ __m128i mul1 = _mm_mul_epu32(upper32_a, b);
637
+ __m128i mul2 = _mm_mul_epu32(upper32_b, a);
638
+ // Gives us both upper*upper and lower*lower
639
+ __m128i mul3 = _mm_mul_epu32(a, b);
411
640
 
412
- template<> EIGEN_STRONG_INLINE Packet4f pand<Packet4f>(const Packet4f& a, const Packet4f& b) { return _mm_and_ps(a,b); }
413
- template<> EIGEN_STRONG_INLINE Packet2d pand<Packet2d>(const Packet2d& a, const Packet2d& b) { return _mm_and_pd(a,b); }
414
- template<> EIGEN_STRONG_INLINE Packet4i pand<Packet4i>(const Packet4i& a, const Packet4i& b) { return _mm_and_si128(a,b); }
415
- template<> EIGEN_STRONG_INLINE Packet16b pand<Packet16b>(const Packet16b& a, const Packet16b& b) { return _mm_and_si128(a,b); }
416
-
417
- template<> EIGEN_STRONG_INLINE Packet4f por<Packet4f>(const Packet4f& a, const Packet4f& b) { return _mm_or_ps(a,b); }
418
- template<> EIGEN_STRONG_INLINE Packet2d por<Packet2d>(const Packet2d& a, const Packet2d& b) { return _mm_or_pd(a,b); }
419
- template<> EIGEN_STRONG_INLINE Packet4i por<Packet4i>(const Packet4i& a, const Packet4i& b) { return _mm_or_si128(a,b); }
420
- template<> EIGEN_STRONG_INLINE Packet16b por<Packet16b>(const Packet16b& a, const Packet16b& b) { return _mm_or_si128(a,b); }
421
-
422
- template<> EIGEN_STRONG_INLINE Packet4f pxor<Packet4f>(const Packet4f& a, const Packet4f& b) { return _mm_xor_ps(a,b); }
423
- template<> EIGEN_STRONG_INLINE Packet2d pxor<Packet2d>(const Packet2d& a, const Packet2d& b) { return _mm_xor_pd(a,b); }
424
- template<> EIGEN_STRONG_INLINE Packet4i pxor<Packet4i>(const Packet4i& a, const Packet4i& b) { return _mm_xor_si128(a,b); }
425
- template<> EIGEN_STRONG_INLINE Packet16b pxor<Packet16b>(const Packet16b& a, const Packet16b& b) { return _mm_xor_si128(a,b); }
426
-
427
- template<> EIGEN_STRONG_INLINE Packet4f pandnot<Packet4f>(const Packet4f& a, const Packet4f& b) { return _mm_andnot_ps(b,a); }
428
- template<> EIGEN_STRONG_INLINE Packet2d pandnot<Packet2d>(const Packet2d& a, const Packet2d& b) { return _mm_andnot_pd(b,a); }
429
- template<> EIGEN_STRONG_INLINE Packet4i pandnot<Packet4i>(const Packet4i& a, const Packet4i& b) { return _mm_andnot_si128(b,a); }
430
-
431
- template<> EIGEN_STRONG_INLINE Packet4f pcmp_le(const Packet4f& a, const Packet4f& b) { return _mm_cmple_ps(a,b); }
432
- template<> EIGEN_STRONG_INLINE Packet4f pcmp_lt(const Packet4f& a, const Packet4f& b) { return _mm_cmplt_ps(a,b); }
433
- template<> EIGEN_STRONG_INLINE Packet4f pcmp_lt_or_nan(const Packet4f& a, const Packet4f& b) { return _mm_cmpnge_ps(a,b); }
434
- template<> EIGEN_STRONG_INLINE Packet4f pcmp_eq(const Packet4f& a, const Packet4f& b) { return _mm_cmpeq_ps(a,b); }
435
-
436
- template<> EIGEN_STRONG_INLINE Packet2d pcmp_le(const Packet2d& a, const Packet2d& b) { return _mm_cmple_pd(a,b); }
437
- template<> EIGEN_STRONG_INLINE Packet2d pcmp_lt(const Packet2d& a, const Packet2d& b) { return _mm_cmplt_pd(a,b); }
438
- template<> EIGEN_STRONG_INLINE Packet2d pcmp_lt_or_nan(const Packet2d& a, const Packet2d& b) { return _mm_cmpnge_pd(a,b); }
439
- template<> EIGEN_STRONG_INLINE Packet2d pcmp_eq(const Packet2d& a, const Packet2d& b) { return _mm_cmpeq_pd(a,b); }
440
-
441
- template<> EIGEN_STRONG_INLINE Packet4i pcmp_lt(const Packet4i& a, const Packet4i& b) { return _mm_cmplt_epi32(a,b); }
442
- template<> EIGEN_STRONG_INLINE Packet4i pcmp_eq(const Packet4i& a, const Packet4i& b) { return _mm_cmpeq_epi32(a,b); }
443
- template<> EIGEN_STRONG_INLINE Packet16b pcmp_eq(const Packet16b& a, const Packet16b& b) { return _mm_cmpeq_epi8(a,b); }
444
- template<> EIGEN_STRONG_INLINE Packet4i pcmp_le(const Packet4i& a, const Packet4i& b) { return por(pcmp_lt(a,b), pcmp_eq(a,b)); }
445
-
446
- template<> EIGEN_STRONG_INLINE Packet4f pmin<Packet4f>(const Packet4f& a, const Packet4f& b) {
447
- #if EIGEN_COMP_GNUC && EIGEN_COMP_GNUC < 63
448
- // There appears to be a bug in GCC, by which the optimizer may
449
- // flip the argument order in calls to _mm_min_ps, so we have to
450
- // resort to inline ASM here. This is supposed to be fixed in gcc6.3,
451
- // see also: https://gcc.gnu.org/bugzilla/show_bug.cgi?id=72867
452
- #ifdef EIGEN_VECTORIZE_AVX
453
- Packet4f res;
454
- asm("vminps %[a], %[b], %[res]" : [res] "=x" (res) : [a] "x" (a), [b] "x" (b));
455
- #else
456
- Packet4f res = b;
457
- asm("minps %[a], %[res]" : [res] "+x" (res) : [a] "x" (a));
458
- #endif
459
- return res;
460
- #else
461
- // Arguments are reversed to match NaN propagation behavior of std::min.
462
- return _mm_min_ps(b, a);
463
- #endif
641
+ __m128i high = _mm_slli_epi64(_mm_add_epi64(mul1, mul2), 32);
642
+ return _mm_add_epi64(high, mul3);
464
643
  }
465
- template<> EIGEN_STRONG_INLINE Packet2d pmin<Packet2d>(const Packet2d& a, const Packet2d& b) {
466
- #if EIGEN_COMP_GNUC && EIGEN_COMP_GNUC < 63
467
- // There appears to be a bug in GCC, by which the optimizer may
468
- // flip the argument order in calls to _mm_min_pd, so we have to
469
- // resort to inline ASM here. This is supposed to be fixed in gcc6.3,
470
- // see also: https://gcc.gnu.org/bugzilla/show_bug.cgi?id=72867
471
- #ifdef EIGEN_VECTORIZE_AVX
472
- Packet2d res;
473
- asm("vminpd %[a], %[b], %[res]" : [res] "=x" (res) : [a] "x" (a), [b] "x" (b));
474
- #else
475
- Packet2d res = b;
476
- asm("minpd %[a], %[res]" : [res] "+x" (res) : [a] "x" (a));
477
- #endif
478
- return res;
644
+ template <>
645
+ EIGEN_STRONG_INLINE Packet4i pmul<Packet4i>(const Packet4i& a, const Packet4i& b) {
646
+ #ifdef EIGEN_VECTORIZE_SSE4_1
647
+ return _mm_mullo_epi32(a, b);
479
648
  #else
480
- // Arguments are reversed to match NaN propagation behavior of std::min.
481
- return _mm_min_pd(b, a);
649
+ // this version is slightly faster than 4 scalar products
650
+ return vec4i_swizzle1(
651
+ vec4i_swizzle2(_mm_mul_epu32(a, b), _mm_mul_epu32(vec4i_swizzle1(a, 1, 0, 3, 2), vec4i_swizzle1(b, 1, 0, 3, 2)),
652
+ 0, 2, 0, 2),
653
+ 0, 2, 1, 3);
482
654
  #endif
483
655
  }
484
- template<> EIGEN_STRONG_INLINE Packet4i pmin<Packet4i>(const Packet4i& a, const Packet4i& b)
485
- {
656
+ template <>
657
+ EIGEN_STRONG_INLINE Packet4ui pmul<Packet4ui>(const Packet4ui& a, const Packet4ui& b) {
486
658
  #ifdef EIGEN_VECTORIZE_SSE4_1
487
- return _mm_min_epi32(a,b);
659
+ return _mm_mullo_epi32(a, b);
488
660
  #else
489
- // after some bench, this version *is* faster than a scalar implementation
490
- Packet4i mask = _mm_cmplt_epi32(a,b);
491
- return _mm_or_si128(_mm_and_si128(mask,a),_mm_andnot_si128(mask,b));
661
+ // this version is slightly faster than 4 scalar products
662
+ return vec4ui_swizzle1(
663
+ vec4ui_swizzle2(_mm_mul_epu32(a, b),
664
+ _mm_mul_epu32(vec4ui_swizzle1(a, 1, 0, 3, 2), vec4ui_swizzle1(b, 1, 0, 3, 2)), 0, 2, 0, 2),
665
+ 0, 2, 1, 3);
492
666
  #endif
493
667
  }
494
668
 
669
+ template <>
670
+ EIGEN_STRONG_INLINE Packet16b pmul<Packet16b>(const Packet16b& a, const Packet16b& b) {
671
+ return _mm_and_si128(a, b);
672
+ }
495
673
 
496
- template<> EIGEN_STRONG_INLINE Packet4f pmax<Packet4f>(const Packet4f& a, const Packet4f& b) {
497
- #if EIGEN_COMP_GNUC && EIGEN_COMP_GNUC < 63
498
- // There appears to be a bug in GCC, by which the optimizer may
499
- // flip the argument order in calls to _mm_max_ps, so we have to
500
- // resort to inline ASM here. This is supposed to be fixed in gcc6.3,
501
- // see also: https://gcc.gnu.org/bugzilla/show_bug.cgi?id=72867
502
- #ifdef EIGEN_VECTORIZE_AVX
503
- Packet4f res;
504
- asm("vmaxps %[a], %[b], %[res]" : [res] "=x" (res) : [a] "x" (a), [b] "x" (b));
505
- #else
506
- Packet4f res = b;
507
- asm("maxps %[a], %[res]" : [res] "+x" (res) : [a] "x" (a));
508
- #endif
509
- return res;
510
- #else
511
- // Arguments are reversed to match NaN propagation behavior of std::max.
674
+ template <>
675
+ EIGEN_STRONG_INLINE Packet4f pdiv<Packet4f>(const Packet4f& a, const Packet4f& b) {
676
+ return _mm_div_ps(a, b);
677
+ }
678
+ template <>
679
+ EIGEN_STRONG_INLINE Packet2d pdiv<Packet2d>(const Packet2d& a, const Packet2d& b) {
680
+ return _mm_div_pd(a, b);
681
+ }
682
+
683
+ template <>
684
+ EIGEN_STRONG_INLINE Packet4i pdiv<Packet4i>(const Packet4i& a, const Packet4i& b) {
685
+ #ifdef EIGEN_VECTORIZE_AVX
686
+ return _mm256_cvttpd_epi32(_mm256_div_pd(_mm256_cvtepi32_pd(a), _mm256_cvtepi32_pd(b)));
687
+ #else
688
+ __m128i q_lo = _mm_cvttpd_epi32(_mm_div_pd(_mm_cvtepi32_pd(a), _mm_cvtepi32_pd(b)));
689
+ __m128i q_hi = _mm_cvttpd_epi32(
690
+ _mm_div_pd(_mm_cvtepi32_pd(vec4i_swizzle1(a, 2, 3, 0, 1)), _mm_cvtepi32_pd(vec4i_swizzle1(b, 2, 3, 0, 1))));
691
+ return vec4i_swizzle1(_mm_unpacklo_epi32(q_lo, q_hi), 0, 2, 1, 3);
692
+ #endif
693
+ }
694
+
695
+ #ifdef EIGEN_VECTORIZE_FMA
696
+ template <>
697
+ EIGEN_STRONG_INLINE Packet4f pmadd(const Packet4f& a, const Packet4f& b, const Packet4f& c) {
698
+ return _mm_fmadd_ps(a, b, c);
699
+ }
700
+ template <>
701
+ EIGEN_STRONG_INLINE Packet2d pmadd(const Packet2d& a, const Packet2d& b, const Packet2d& c) {
702
+ return _mm_fmadd_pd(a, b, c);
703
+ }
704
+ template <>
705
+ EIGEN_STRONG_INLINE Packet4f pmsub(const Packet4f& a, const Packet4f& b, const Packet4f& c) {
706
+ return _mm_fmsub_ps(a, b, c);
707
+ }
708
+ template <>
709
+ EIGEN_STRONG_INLINE Packet2d pmsub(const Packet2d& a, const Packet2d& b, const Packet2d& c) {
710
+ return _mm_fmsub_pd(a, b, c);
711
+ }
712
+ template <>
713
+ EIGEN_STRONG_INLINE Packet4f pnmadd(const Packet4f& a, const Packet4f& b, const Packet4f& c) {
714
+ return _mm_fnmadd_ps(a, b, c);
715
+ }
716
+ template <>
717
+ EIGEN_STRONG_INLINE Packet2d pnmadd(const Packet2d& a, const Packet2d& b, const Packet2d& c) {
718
+ return _mm_fnmadd_pd(a, b, c);
719
+ }
720
+ template <>
721
+ EIGEN_STRONG_INLINE Packet4f pnmsub(const Packet4f& a, const Packet4f& b, const Packet4f& c) {
722
+ return _mm_fnmsub_ps(a, b, c);
723
+ }
724
+ template <>
725
+ EIGEN_STRONG_INLINE Packet2d pnmsub(const Packet2d& a, const Packet2d& b, const Packet2d& c) {
726
+ return _mm_fnmsub_pd(a, b, c);
727
+ }
728
+
729
+ template <typename Packet>
730
+ EIGEN_STRONG_INLINE Packet pmadds(const Packet& a, const Packet& b, const Packet& c);
731
+ template <>
732
+ EIGEN_STRONG_INLINE Packet4f pmadds<Packet4f>(const Packet4f& a, const Packet4f& b, const Packet4f& c) {
733
+ return _mm_fmadd_ss(a, b, c);
734
+ }
735
+ template <>
736
+ EIGEN_STRONG_INLINE Packet2d pmadds<Packet2d>(const Packet2d& a, const Packet2d& b, const Packet2d& c) {
737
+ return _mm_fmadd_sd(a, b, c);
738
+ }
739
+ #endif
740
+
741
+ #ifdef EIGEN_VECTORIZE_SSE4_1
742
+ template <>
743
+ EIGEN_STRONG_INLINE Packet4f pselect(const Packet4f& mask, const Packet4f& a, const Packet4f& b) {
744
+ return _mm_blendv_ps(b, a, mask);
745
+ }
746
+
747
+ template <>
748
+ EIGEN_STRONG_INLINE Packet2l pselect(const Packet2l& mask, const Packet2l& a, const Packet2l& b) {
749
+ return _mm_castpd_si128(_mm_blendv_pd(_mm_castsi128_pd(b), _mm_castsi128_pd(a), _mm_castsi128_pd(mask)));
750
+ }
751
+
752
+ template <>
753
+ EIGEN_STRONG_INLINE Packet4i pselect(const Packet4i& mask, const Packet4i& a, const Packet4i& b) {
754
+ return _mm_castps_si128(_mm_blendv_ps(_mm_castsi128_ps(b), _mm_castsi128_ps(a), _mm_castsi128_ps(mask)));
755
+ }
756
+
757
+ template <>
758
+ EIGEN_STRONG_INLINE Packet4ui pselect(const Packet4ui& mask, const Packet4ui& a, const Packet4ui& b) {
759
+ return _mm_castps_si128(_mm_blendv_ps(_mm_castsi128_ps(b), _mm_castsi128_ps(a), _mm_castsi128_ps(mask)));
760
+ }
761
+
762
+ template <>
763
+ EIGEN_STRONG_INLINE Packet2d pselect(const Packet2d& mask, const Packet2d& a, const Packet2d& b) {
764
+ return _mm_blendv_pd(b, a, mask);
765
+ }
766
+ #endif
767
+
768
+ template <>
769
+ EIGEN_STRONG_INLINE Packet2l ptrue<Packet2l>(const Packet2l& a) {
770
+ return _mm_cmpeq_epi32(a, a);
771
+ }
772
+ template <>
773
+ EIGEN_STRONG_INLINE Packet4i ptrue<Packet4i>(const Packet4i& a) {
774
+ return _mm_cmpeq_epi32(a, a);
775
+ }
776
+ template <>
777
+ EIGEN_STRONG_INLINE Packet16b ptrue<Packet16b>(const Packet16b& /*a*/) {
778
+ return pset1<Packet16b>(true);
779
+ }
780
+ template <>
781
+ EIGEN_STRONG_INLINE Packet4f ptrue<Packet4f>(const Packet4f& a) {
782
+ Packet4i b = _mm_castps_si128(a);
783
+ return _mm_castsi128_ps(_mm_cmpeq_epi32(b, b));
784
+ }
785
+ template <>
786
+ EIGEN_STRONG_INLINE Packet2d ptrue<Packet2d>(const Packet2d& a) {
787
+ Packet4i b = _mm_castpd_si128(a);
788
+ return _mm_castsi128_pd(_mm_cmpeq_epi32(b, b));
789
+ }
790
+
791
+ template <>
792
+ EIGEN_STRONG_INLINE Packet4f pand<Packet4f>(const Packet4f& a, const Packet4f& b) {
793
+ return _mm_and_ps(a, b);
794
+ }
795
+ template <>
796
+ EIGEN_STRONG_INLINE Packet2d pand<Packet2d>(const Packet2d& a, const Packet2d& b) {
797
+ return _mm_and_pd(a, b);
798
+ }
799
+ template <>
800
+ EIGEN_STRONG_INLINE Packet2l pand<Packet2l>(const Packet2l& a, const Packet2l& b) {
801
+ return _mm_and_si128(a, b);
802
+ }
803
+ template <>
804
+ EIGEN_STRONG_INLINE Packet4i pand<Packet4i>(const Packet4i& a, const Packet4i& b) {
805
+ return _mm_and_si128(a, b);
806
+ }
807
+ template <>
808
+ EIGEN_STRONG_INLINE Packet4ui pand<Packet4ui>(const Packet4ui& a, const Packet4ui& b) {
809
+ return _mm_and_si128(a, b);
810
+ }
811
+ template <>
812
+ EIGEN_STRONG_INLINE Packet16b pand<Packet16b>(const Packet16b& a, const Packet16b& b) {
813
+ return _mm_and_si128(a, b);
814
+ }
815
+
816
+ template <>
817
+ EIGEN_STRONG_INLINE Packet4f por<Packet4f>(const Packet4f& a, const Packet4f& b) {
818
+ return _mm_or_ps(a, b);
819
+ }
820
+ template <>
821
+ EIGEN_STRONG_INLINE Packet2d por<Packet2d>(const Packet2d& a, const Packet2d& b) {
822
+ return _mm_or_pd(a, b);
823
+ }
824
+ template <>
825
+ EIGEN_STRONG_INLINE Packet2l por<Packet2l>(const Packet2l& a, const Packet2l& b) {
826
+ return _mm_or_si128(a, b);
827
+ }
828
+ template <>
829
+ EIGEN_STRONG_INLINE Packet4i por<Packet4i>(const Packet4i& a, const Packet4i& b) {
830
+ return _mm_or_si128(a, b);
831
+ }
832
+ template <>
833
+ EIGEN_STRONG_INLINE Packet4ui por<Packet4ui>(const Packet4ui& a, const Packet4ui& b) {
834
+ return _mm_or_si128(a, b);
835
+ }
836
+ template <>
837
+ EIGEN_STRONG_INLINE Packet16b por<Packet16b>(const Packet16b& a, const Packet16b& b) {
838
+ return _mm_or_si128(a, b);
839
+ }
840
+
841
+ template <>
842
+ EIGEN_STRONG_INLINE Packet4f pxor<Packet4f>(const Packet4f& a, const Packet4f& b) {
843
+ return _mm_xor_ps(a, b);
844
+ }
845
+ template <>
846
+ EIGEN_STRONG_INLINE Packet2d pxor<Packet2d>(const Packet2d& a, const Packet2d& b) {
847
+ return _mm_xor_pd(a, b);
848
+ }
849
+ template <>
850
+ EIGEN_STRONG_INLINE Packet2l pxor<Packet2l>(const Packet2l& a, const Packet2l& b) {
851
+ return _mm_xor_si128(a, b);
852
+ }
853
+ template <>
854
+ EIGEN_STRONG_INLINE Packet4i pxor<Packet4i>(const Packet4i& a, const Packet4i& b) {
855
+ return _mm_xor_si128(a, b);
856
+ }
857
+ template <>
858
+ EIGEN_STRONG_INLINE Packet4ui pxor<Packet4ui>(const Packet4ui& a, const Packet4ui& b) {
859
+ return _mm_xor_si128(a, b);
860
+ }
861
+ template <>
862
+ EIGEN_STRONG_INLINE Packet16b pxor<Packet16b>(const Packet16b& a, const Packet16b& b) {
863
+ return _mm_xor_si128(a, b);
864
+ }
865
+
866
+ template <>
867
+ EIGEN_STRONG_INLINE Packet4f pandnot<Packet4f>(const Packet4f& a, const Packet4f& b) {
868
+ return _mm_andnot_ps(b, a);
869
+ }
870
+ template <>
871
+ EIGEN_STRONG_INLINE Packet2d pandnot<Packet2d>(const Packet2d& a, const Packet2d& b) {
872
+ return _mm_andnot_pd(b, a);
873
+ }
874
+ template <>
875
+ EIGEN_STRONG_INLINE Packet2l pandnot<Packet2l>(const Packet2l& a, const Packet2l& b) {
876
+ return _mm_andnot_si128(b, a);
877
+ }
878
+ template <>
879
+ EIGEN_STRONG_INLINE Packet4i pandnot<Packet4i>(const Packet4i& a, const Packet4i& b) {
880
+ return _mm_andnot_si128(b, a);
881
+ }
882
+ template <>
883
+ EIGEN_STRONG_INLINE Packet4ui pandnot<Packet4ui>(const Packet4ui& a, const Packet4ui& b) {
884
+ return _mm_andnot_si128(b, a);
885
+ }
886
+ template <>
887
+ EIGEN_STRONG_INLINE Packet16b pandnot<Packet16b>(const Packet16b& a, const Packet16b& b) {
888
+ return _mm_andnot_si128(b, a);
889
+ }
890
+ template <>
891
+ EIGEN_STRONG_INLINE Packet16b pcmp_lt(const Packet16b& a, const Packet16b& b) {
892
+ return _mm_andnot_si128(a, b);
893
+ }
894
+ template <>
895
+ EIGEN_STRONG_INLINE Packet4f pcmp_le(const Packet4f& a, const Packet4f& b) {
896
+ return _mm_cmple_ps(a, b);
897
+ }
898
+ template <>
899
+ EIGEN_STRONG_INLINE Packet4f pcmp_lt(const Packet4f& a, const Packet4f& b) {
900
+ return _mm_cmplt_ps(a, b);
901
+ }
902
+ template <>
903
+ EIGEN_STRONG_INLINE Packet4f pcmp_lt_or_nan(const Packet4f& a, const Packet4f& b) {
904
+ return _mm_cmpnge_ps(a, b);
905
+ }
906
+ template <>
907
+ EIGEN_STRONG_INLINE Packet4f pcmp_eq(const Packet4f& a, const Packet4f& b) {
908
+ return _mm_cmpeq_ps(a, b);
909
+ }
910
+
911
+ template <>
912
+ EIGEN_STRONG_INLINE Packet2d pcmp_le(const Packet2d& a, const Packet2d& b) {
913
+ return _mm_cmple_pd(a, b);
914
+ }
915
+ template <>
916
+ EIGEN_STRONG_INLINE Packet2d pcmp_lt(const Packet2d& a, const Packet2d& b) {
917
+ return _mm_cmplt_pd(a, b);
918
+ }
919
+ template <>
920
+ EIGEN_STRONG_INLINE Packet2d pcmp_lt_or_nan(const Packet2d& a, const Packet2d& b) {
921
+ return _mm_cmpnge_pd(a, b);
922
+ }
923
+ template <>
924
+ EIGEN_STRONG_INLINE Packet2d pcmp_eq(const Packet2d& a, const Packet2d& b) {
925
+ return _mm_cmpeq_pd(a, b);
926
+ }
927
+ template <>
928
+ EIGEN_STRONG_INLINE Packet4i pcmp_lt(const Packet4i& a, const Packet4i& b) {
929
+ return _mm_cmplt_epi32(a, b);
930
+ }
931
+ template <>
932
+ EIGEN_STRONG_INLINE Packet4i pcmp_eq(const Packet4i& a, const Packet4i& b) {
933
+ return _mm_cmpeq_epi32(a, b);
934
+ }
935
+ template <>
936
+ EIGEN_STRONG_INLINE Packet4i pcmp_le(const Packet4i& a, const Packet4i& b) {
937
+ #ifdef EIGEN_VECTORIZE_SSE4_1
938
+ return _mm_cmpeq_epi32(a, _mm_min_epi32(a, b));
939
+ #else
940
+ return por(pcmp_lt(a, b), pcmp_eq(a, b));
941
+ #endif
942
+ }
943
+ template <>
944
+ EIGEN_STRONG_INLINE Packet2l pcmp_lt(const Packet2l& a, const Packet2l& b) {
945
+ #ifdef EIGEN_VECTORIZE_SSE4_2
946
+ return _mm_cmpgt_epi64(b, a);
947
+ #else
948
+ Packet4i eq = pcmp_eq<Packet4i>(Packet4i(a), Packet4i(b));
949
+ Packet2l hi_eq = Packet2l(_mm_shuffle_epi32(eq, (shuffle_mask<1, 1, 3, 3>::mask)));
950
+ Packet4i lt = pcmp_lt<Packet4i>(Packet4i(a), Packet4i(b));
951
+ Packet2l hi_lt = Packet2l(_mm_shuffle_epi32(lt, (shuffle_mask<1, 1, 3, 3>::mask)));
952
+ Packet2l lo_lt = Packet2l(_mm_shuffle_epi32(lt, (shuffle_mask<0, 0, 2, 2>::mask)));
953
+ // return hi(a) < hi(b) || (hi(a) == hi(b) && lo(a) < lo(b))
954
+ return por(hi_lt, pand(hi_eq, lo_lt));
955
+ #endif
956
+ }
957
+ template <>
958
+ EIGEN_STRONG_INLINE Packet2l pcmp_eq(const Packet2l& a, const Packet2l& b) {
959
+ #ifdef EIGEN_VECTORIZE_SSE4_1
960
+ return _mm_cmpeq_epi64(a, b);
961
+ #else
962
+ Packet4i tmp = pcmp_eq<Packet4i>(Packet4i(a), Packet4i(b));
963
+ return Packet2l(pand<Packet4i>(tmp, _mm_shuffle_epi32(tmp, (shuffle_mask<1, 0, 3, 2>::mask))));
964
+ #endif
965
+ }
966
+ template <>
967
+ EIGEN_STRONG_INLINE Packet2l pcmp_le(const Packet2l& a, const Packet2l& b) {
968
+ return por(pcmp_lt(a, b), pcmp_eq(a, b));
969
+ }
970
+ template <>
971
+ EIGEN_STRONG_INLINE Packet16b pcmp_eq(const Packet16b& a, const Packet16b& b) {
972
+ // Mask out invalid bool bits to avoid UB.
973
+ const Packet16b kBoolMask = pset1<Packet16b>(true);
974
+ return _mm_and_si128(_mm_cmpeq_epi8(a, b), kBoolMask);
975
+ }
976
+ template <>
977
+ EIGEN_STRONG_INLINE Packet4ui pcmp_eq(const Packet4ui& a, const Packet4ui& b) {
978
+ return _mm_cmpeq_epi32(a, b);
979
+ }
980
+
981
+ template <>
982
+ EIGEN_STRONG_INLINE Packet4f pmin<Packet4f>(const Packet4f& a, const Packet4f& b) {
983
+ #if EIGEN_GNUC_STRICT_LESS_THAN(6, 3, 0)
984
+ // There appears to be a bug in GCC, by which the optimizer may
985
+ // flip the argument order in calls to _mm_min_ps, so we have to
986
+ // resort to inline ASM here. This is supposed to be fixed in gcc6.3,
987
+ // see also: https://gcc.gnu.org/bugzilla/show_bug.cgi?id=72867
988
+ #ifdef EIGEN_VECTORIZE_AVX
989
+ Packet4f res;
990
+ asm("vminps %[a], %[b], %[res]" : [res] "=x"(res) : [a] "x"(a), [b] "x"(b));
991
+ #else
992
+ Packet4f res = b;
993
+ asm("minps %[a], %[res]" : [res] "+x"(res) : [a] "x"(a));
994
+ #endif
995
+ return res;
996
+ #else
997
+ // Arguments are reversed to match NaN propagation behavior of std::min.
998
+ return _mm_min_ps(b, a);
999
+ #endif
1000
+ }
1001
+ template <>
1002
+ EIGEN_STRONG_INLINE Packet2d pmin<Packet2d>(const Packet2d& a, const Packet2d& b) {
1003
+ #if EIGEN_GNUC_STRICT_LESS_THAN(6, 3, 0)
1004
+ // There appears to be a bug in GCC, by which the optimizer may
1005
+ // flip the argument order in calls to _mm_min_pd, so we have to
1006
+ // resort to inline ASM here. This is supposed to be fixed in gcc6.3,
1007
+ // see also: https://gcc.gnu.org/bugzilla/show_bug.cgi?id=72867
1008
+ #ifdef EIGEN_VECTORIZE_AVX
1009
+ Packet2d res;
1010
+ asm("vminpd %[a], %[b], %[res]" : [res] "=x"(res) : [a] "x"(a), [b] "x"(b));
1011
+ #else
1012
+ Packet2d res = b;
1013
+ asm("minpd %[a], %[res]" : [res] "+x"(res) : [a] "x"(a));
1014
+ #endif
1015
+ return res;
1016
+ #else
1017
+ // Arguments are reversed to match NaN propagation behavior of std::min.
1018
+ return _mm_min_pd(b, a);
1019
+ #endif
1020
+ }
1021
+ template <>
1022
+ EIGEN_STRONG_INLINE Packet2l pmin<Packet2l>(const Packet2l& a, const Packet2l& b) {
1023
+ Packet2l a_lt_mask = pcmp_lt(a, b);
1024
+ return por(pandnot(b, a_lt_mask), pand(a, a_lt_mask));
1025
+ }
1026
+ template <>
1027
+ EIGEN_STRONG_INLINE Packet4i pmin<Packet4i>(const Packet4i& a, const Packet4i& b) {
1028
+ #ifdef EIGEN_VECTORIZE_SSE4_1
1029
+ return _mm_min_epi32(a, b);
1030
+ #else
1031
+ // after some bench, this version *is* faster than a scalar implementation
1032
+ Packet4i mask = _mm_cmplt_epi32(a, b);
1033
+ return _mm_or_si128(_mm_and_si128(mask, a), _mm_andnot_si128(mask, b));
1034
+ #endif
1035
+ }
1036
+ template <>
1037
+ EIGEN_STRONG_INLINE Packet4ui pmin<Packet4ui>(const Packet4ui& a, const Packet4ui& b) {
1038
+ #ifdef EIGEN_VECTORIZE_SSE4_1
1039
+ return _mm_min_epu32(a, b);
1040
+ #else
1041
+ return padd((Packet4ui)pmin((Packet4i)psub(a, pset1<Packet4ui>(0x80000000UL)),
1042
+ (Packet4i)psub(b, pset1<Packet4ui>(0x80000000UL))),
1043
+ pset1<Packet4ui>(0x80000000UL));
1044
+ #endif
1045
+ }
1046
+
1047
+ template <>
1048
+ EIGEN_STRONG_INLINE Packet4f pmax<Packet4f>(const Packet4f& a, const Packet4f& b) {
1049
+ #if EIGEN_GNUC_STRICT_LESS_THAN(6, 3, 0)
1050
+ // There appears to be a bug in GCC, by which the optimizer may
1051
+ // flip the argument order in calls to _mm_max_ps, so we have to
1052
+ // resort to inline ASM here. This is supposed to be fixed in gcc6.3,
1053
+ // see also: https://gcc.gnu.org/bugzilla/show_bug.cgi?id=72867
1054
+ #ifdef EIGEN_VECTORIZE_AVX
1055
+ Packet4f res;
1056
+ asm("vmaxps %[a], %[b], %[res]" : [res] "=x"(res) : [a] "x"(a), [b] "x"(b));
1057
+ #else
1058
+ Packet4f res = b;
1059
+ asm("maxps %[a], %[res]" : [res] "+x"(res) : [a] "x"(a));
1060
+ #endif
1061
+ return res;
1062
+ #else
1063
+ // Arguments are reversed to match NaN propagation behavior of std::max.
512
1064
  return _mm_max_ps(b, a);
513
1065
  #endif
514
1066
  }
515
- template<> EIGEN_STRONG_INLINE Packet2d pmax<Packet2d>(const Packet2d& a, const Packet2d& b) {
516
- #if EIGEN_COMP_GNUC && EIGEN_COMP_GNUC < 63
517
- // There appears to be a bug in GCC, by which the optimizer may
518
- // flip the argument order in calls to _mm_max_pd, so we have to
519
- // resort to inline ASM here. This is supposed to be fixed in gcc6.3,
520
- // see also: https://gcc.gnu.org/bugzilla/show_bug.cgi?id=72867
521
- #ifdef EIGEN_VECTORIZE_AVX
1067
+ template <>
1068
+ EIGEN_STRONG_INLINE Packet2d pmax<Packet2d>(const Packet2d& a, const Packet2d& b) {
1069
+ #if EIGEN_GNUC_STRICT_LESS_THAN(6, 3, 0)
1070
+ // There appears to be a bug in GCC, by which the optimizer may
1071
+ // flip the argument order in calls to _mm_max_pd, so we have to
1072
+ // resort to inline ASM here. This is supposed to be fixed in gcc6.3,
1073
+ // see also: https://gcc.gnu.org/bugzilla/show_bug.cgi?id=72867
1074
+ #ifdef EIGEN_VECTORIZE_AVX
522
1075
  Packet2d res;
523
- asm("vmaxpd %[a], %[b], %[res]" : [res] "=x" (res) : [a] "x" (a), [b] "x" (b));
524
- #else
1076
+ asm("vmaxpd %[a], %[b], %[res]" : [res] "=x"(res) : [a] "x"(a), [b] "x"(b));
1077
+ #else
525
1078
  Packet2d res = b;
526
- asm("maxpd %[a], %[res]" : [res] "+x" (res) : [a] "x" (a));
527
- #endif
1079
+ asm("maxpd %[a], %[res]" : [res] "+x"(res) : [a] "x"(a));
1080
+ #endif
528
1081
  return res;
529
1082
  #else
530
1083
  // Arguments are reversed to match NaN propagation behavior of std::max.
531
1084
  return _mm_max_pd(b, a);
532
1085
  #endif
533
1086
  }
534
- template<> EIGEN_STRONG_INLINE Packet4i pmax<Packet4i>(const Packet4i& a, const Packet4i& b)
535
- {
1087
+ template <>
1088
+ EIGEN_STRONG_INLINE Packet2l pmax<Packet2l>(const Packet2l& a, const Packet2l& b) {
1089
+ Packet2l a_lt_mask = pcmp_lt(a, b);
1090
+ return por(pandnot(a, a_lt_mask), pand(b, a_lt_mask));
1091
+ }
1092
+ template <>
1093
+ EIGEN_STRONG_INLINE Packet4i pmax<Packet4i>(const Packet4i& a, const Packet4i& b) {
536
1094
  #ifdef EIGEN_VECTORIZE_SSE4_1
537
- return _mm_max_epi32(a,b);
1095
+ return _mm_max_epi32(a, b);
538
1096
  #else
539
1097
  // after some bench, this version *is* faster than a scalar implementation
540
- Packet4i mask = _mm_cmpgt_epi32(a,b);
541
- return _mm_or_si128(_mm_and_si128(mask,a),_mm_andnot_si128(mask,b));
1098
+ Packet4i mask = _mm_cmpgt_epi32(a, b);
1099
+ return _mm_or_si128(_mm_and_si128(mask, a), _mm_andnot_si128(mask, b));
1100
+ #endif
1101
+ }
1102
+ template <>
1103
+ EIGEN_STRONG_INLINE Packet4ui pmax<Packet4ui>(const Packet4ui& a, const Packet4ui& b) {
1104
+ #ifdef EIGEN_VECTORIZE_SSE4_1
1105
+ return _mm_max_epu32(a, b);
1106
+ #else
1107
+ return padd((Packet4ui)pmax((Packet4i)psub(a, pset1<Packet4ui>(0x80000000UL)),
1108
+ (Packet4i)psub(b, pset1<Packet4ui>(0x80000000UL))),
1109
+ pset1<Packet4ui>(0x80000000UL));
1110
+ #endif
1111
+ }
1112
+
1113
+ template <>
1114
+ EIGEN_STRONG_INLINE Packet4ui pcmp_lt(const Packet4ui& a, const Packet4ui& b) {
1115
+ #ifdef EIGEN_VECTORIZE_SSE4_1
1116
+ return pxor(pcmp_eq(a, pmax(a, b)), ptrue(a));
1117
+ #else
1118
+ return (Packet4ui)pcmp_lt((Packet4i)psub(a, pset1<Packet4ui>(0x80000000UL)),
1119
+ (Packet4i)psub(b, pset1<Packet4ui>(0x80000000UL)));
1120
+ #endif
1121
+ }
1122
+ template <>
1123
+ EIGEN_STRONG_INLINE Packet4ui pcmp_le(const Packet4ui& a, const Packet4ui& b) {
1124
+ #ifdef EIGEN_VECTORIZE_SSE4_1
1125
+ return pcmp_eq(a, pmin(a, b));
1126
+ #else
1127
+ return (Packet4ui)pcmp_le((Packet4i)psub(a, pset1<Packet4ui>(0x80000000UL)),
1128
+ (Packet4i)psub(b, pset1<Packet4ui>(0x80000000UL)));
542
1129
  #endif
543
1130
  }
544
1131
 
@@ -560,580 +1147,746 @@ EIGEN_STRONG_INLINE Packet pminmax_propagate_nan(const Packet& a, const Packet&
560
1147
  return pselect<Packet>(not_nan_mask_a, m, a);
561
1148
  }
562
1149
 
563
- // Add specializations for min/max with prescribed NaN progation.
564
- template<>
1150
+ // Add specializations for min/max with prescribed NaN propagation.
1151
+ template <>
565
1152
  EIGEN_STRONG_INLINE Packet4f pmin<PropagateNumbers, Packet4f>(const Packet4f& a, const Packet4f& b) {
566
1153
  return pminmax_propagate_numbers(a, b, pmin<Packet4f>);
567
1154
  }
568
- template<>
1155
+ template <>
569
1156
  EIGEN_STRONG_INLINE Packet2d pmin<PropagateNumbers, Packet2d>(const Packet2d& a, const Packet2d& b) {
570
1157
  return pminmax_propagate_numbers(a, b, pmin<Packet2d>);
571
1158
  }
572
- template<>
1159
+ template <>
573
1160
  EIGEN_STRONG_INLINE Packet4f pmax<PropagateNumbers, Packet4f>(const Packet4f& a, const Packet4f& b) {
574
1161
  return pminmax_propagate_numbers(a, b, pmax<Packet4f>);
575
1162
  }
576
- template<>
1163
+ template <>
577
1164
  EIGEN_STRONG_INLINE Packet2d pmax<PropagateNumbers, Packet2d>(const Packet2d& a, const Packet2d& b) {
578
1165
  return pminmax_propagate_numbers(a, b, pmax<Packet2d>);
579
1166
  }
580
- template<>
1167
+ template <>
581
1168
  EIGEN_STRONG_INLINE Packet4f pmin<PropagateNaN, Packet4f>(const Packet4f& a, const Packet4f& b) {
582
1169
  return pminmax_propagate_nan(a, b, pmin<Packet4f>);
583
1170
  }
584
- template<>
1171
+ template <>
585
1172
  EIGEN_STRONG_INLINE Packet2d pmin<PropagateNaN, Packet2d>(const Packet2d& a, const Packet2d& b) {
586
1173
  return pminmax_propagate_nan(a, b, pmin<Packet2d>);
587
1174
  }
588
- template<>
1175
+ template <>
589
1176
  EIGEN_STRONG_INLINE Packet4f pmax<PropagateNaN, Packet4f>(const Packet4f& a, const Packet4f& b) {
590
1177
  return pminmax_propagate_nan(a, b, pmax<Packet4f>);
591
1178
  }
592
- template<>
1179
+ template <>
593
1180
  EIGEN_STRONG_INLINE Packet2d pmax<PropagateNaN, Packet2d>(const Packet2d& a, const Packet2d& b) {
594
1181
  return pminmax_propagate_nan(a, b, pmax<Packet2d>);
595
1182
  }
596
1183
 
597
- template<int N> EIGEN_STRONG_INLINE Packet4i parithmetic_shift_right(const Packet4i& a) { return _mm_srai_epi32(a,N); }
598
- template<int N> EIGEN_STRONG_INLINE Packet4i plogical_shift_right (const Packet4i& a) { return _mm_srli_epi32(a,N); }
599
- template<int N> EIGEN_STRONG_INLINE Packet4i plogical_shift_left (const Packet4i& a) { return _mm_slli_epi32(a,N); }
600
-
601
- template<> EIGEN_STRONG_INLINE Packet4f pabs(const Packet4f& a)
602
- {
603
- const Packet4f mask = _mm_castsi128_ps(_mm_setr_epi32(0x7FFFFFFF,0x7FFFFFFF,0x7FFFFFFF,0x7FFFFFFF));
604
- return _mm_and_ps(a,mask);
1184
+ template <>
1185
+ EIGEN_STRONG_INLINE Packet4f psignbit(const Packet4f& a) {
1186
+ return _mm_castsi128_ps(_mm_srai_epi32(_mm_castps_si128(a), 31));
1187
+ }
1188
+ template <>
1189
+ EIGEN_STRONG_INLINE Packet2d psignbit(const Packet2d& a) {
1190
+ Packet4f tmp = psignbit<Packet4f>(_mm_castpd_ps(a));
1191
+ #ifdef EIGEN_VECTORIZE_AVX
1192
+ return _mm_castps_pd(_mm_permute_ps(tmp, (shuffle_mask<1, 1, 3, 3>::mask)));
1193
+ #else
1194
+ return _mm_castps_pd(_mm_shuffle_ps(tmp, tmp, (shuffle_mask<1, 1, 3, 3>::mask)));
1195
+ #endif // EIGEN_VECTORIZE_AVX
1196
+ }
1197
+ template <>
1198
+ EIGEN_STRONG_INLINE Packet4i psignbit(const Packet4i& a) {
1199
+ return _mm_srai_epi32(a, 31);
1200
+ }
1201
+ template <>
1202
+ EIGEN_STRONG_INLINE Packet4ui psignbit(const Packet4ui& a) {
1203
+ return pzero(a);
1204
+ }
1205
+ template <>
1206
+ EIGEN_STRONG_INLINE Packet2l psignbit(const Packet2l& a) {
1207
+ Packet4i tmp = psignbit<Packet4i>(Packet4i(a));
1208
+ return Packet2l(_mm_shuffle_epi32(tmp, (shuffle_mask<1, 1, 3, 3>::mask)));
1209
+ }
1210
+
1211
+ template <int N>
1212
+ EIGEN_STRONG_INLINE Packet2l parithmetic_shift_right(const Packet2l& a) {
1213
+ Packet2l signbit = psignbit(a);
1214
+ return por(_mm_slli_epi64(signbit, 64 - N), _mm_srli_epi64(a, N));
1215
+ }
1216
+ template <int N>
1217
+ EIGEN_STRONG_INLINE Packet2l plogical_shift_right(const Packet2l& a) {
1218
+ return _mm_srli_epi64(a, N);
1219
+ }
1220
+ template <int N>
1221
+ EIGEN_STRONG_INLINE Packet2l plogical_shift_left(const Packet2l& a) {
1222
+ return _mm_slli_epi64(a, N);
1223
+ }
1224
+ template <int N>
1225
+ EIGEN_STRONG_INLINE Packet4i parithmetic_shift_right(const Packet4i& a) {
1226
+ return _mm_srai_epi32(a, N);
1227
+ }
1228
+ template <int N>
1229
+ EIGEN_STRONG_INLINE Packet4i plogical_shift_right(const Packet4i& a) {
1230
+ return _mm_srli_epi32(a, N);
1231
+ }
1232
+ template <int N>
1233
+ EIGEN_STRONG_INLINE Packet4i plogical_shift_left(const Packet4i& a) {
1234
+ return _mm_slli_epi32(a, N);
1235
+ }
1236
+ template <int N>
1237
+ EIGEN_STRONG_INLINE Packet4ui parithmetic_shift_right(const Packet4ui& a) {
1238
+ return _mm_srli_epi32(a, N);
1239
+ }
1240
+ template <int N>
1241
+ EIGEN_STRONG_INLINE Packet4ui plogical_shift_right(const Packet4ui& a) {
1242
+ return _mm_srli_epi32(a, N);
1243
+ }
1244
+ template <int N>
1245
+ EIGEN_STRONG_INLINE Packet4ui plogical_shift_left(const Packet4ui& a) {
1246
+ return _mm_slli_epi32(a, N);
1247
+ }
1248
+
1249
+ template <>
1250
+ EIGEN_STRONG_INLINE Packet4f pabs(const Packet4f& a) {
1251
+ const __m128i mask = _mm_setr_epi32(0x7FFFFFFF, 0x7FFFFFFF, 0x7FFFFFFF, 0x7FFFFFFF);
1252
+ return _mm_castsi128_ps(_mm_and_si128(mask, _mm_castps_si128(a)));
1253
+ }
1254
+ template <>
1255
+ EIGEN_STRONG_INLINE Packet2d pabs(const Packet2d& a) {
1256
+ const __m128i mask = _mm_setr_epi32(0xFFFFFFFF, 0x7FFFFFFF, 0xFFFFFFFF, 0x7FFFFFFF);
1257
+ return _mm_castsi128_pd(_mm_and_si128(mask, _mm_castpd_si128(a)));
1258
+ }
1259
+ template <>
1260
+ EIGEN_STRONG_INLINE Packet2l pabs(const Packet2l& a) {
1261
+ Packet2l signbit = psignbit(a);
1262
+ return _mm_sub_epi64(_mm_xor_si128(a, signbit), signbit);
1263
+ }
1264
+ template <>
1265
+ EIGEN_STRONG_INLINE Packet4i pabs(const Packet4i& a) {
1266
+ #ifdef EIGEN_VECTORIZE_SSSE3
1267
+ return _mm_abs_epi32(a);
1268
+ #else
1269
+ Packet4i signbit = psignbit(a);
1270
+ return _mm_sub_epi32(_mm_xor_si128(a, signbit), signbit);
1271
+ #endif
1272
+ }
1273
+ template <>
1274
+ EIGEN_STRONG_INLINE Packet4ui pabs(const Packet4ui& a) {
1275
+ return a;
1276
+ }
1277
+
1278
+ #ifdef EIGEN_VECTORIZE_SSE4_1
1279
+ template <>
1280
+ EIGEN_STRONG_INLINE Packet4f pround<Packet4f>(const Packet4f& a) {
1281
+ // Unfortunately _mm_round_ps doesn't have a rounding mode to implement numext::round.
1282
+ const Packet4f mask = pset1frombits<Packet4f>(0x80000000u);
1283
+ const Packet4f prev0dot5 = pset1frombits<Packet4f>(0x3EFFFFFFu);
1284
+ return _mm_round_ps(padd(por(pand(a, mask), prev0dot5), a), _MM_FROUND_TO_ZERO);
1285
+ }
1286
+
1287
+ template <>
1288
+ EIGEN_STRONG_INLINE Packet2d pround<Packet2d>(const Packet2d& a) {
1289
+ const Packet2d mask = _mm_castsi128_pd(_mm_set_epi64x(0x8000000000000000ull, 0x8000000000000000ull));
1290
+ const Packet2d prev0dot5 = _mm_castsi128_pd(_mm_set_epi64x(0x3FDFFFFFFFFFFFFFull, 0x3FDFFFFFFFFFFFFFull));
1291
+ return _mm_round_pd(padd(por(pand(a, mask), prev0dot5), a), _MM_FROUND_TO_ZERO);
1292
+ }
1293
+
1294
+ template <>
1295
+ EIGEN_STRONG_INLINE Packet4f print<Packet4f>(const Packet4f& a) {
1296
+ return _mm_round_ps(a, _MM_FROUND_CUR_DIRECTION);
1297
+ }
1298
+ template <>
1299
+ EIGEN_STRONG_INLINE Packet2d print<Packet2d>(const Packet2d& a) {
1300
+ return _mm_round_pd(a, _MM_FROUND_CUR_DIRECTION);
1301
+ }
1302
+
1303
+ template <>
1304
+ EIGEN_STRONG_INLINE Packet4f pceil<Packet4f>(const Packet4f& a) {
1305
+ return _mm_ceil_ps(a);
1306
+ }
1307
+ template <>
1308
+ EIGEN_STRONG_INLINE Packet2d pceil<Packet2d>(const Packet2d& a) {
1309
+ return _mm_ceil_pd(a);
1310
+ }
1311
+
1312
+ template <>
1313
+ EIGEN_STRONG_INLINE Packet4f pfloor<Packet4f>(const Packet4f& a) {
1314
+ return _mm_floor_ps(a);
1315
+ }
1316
+ template <>
1317
+ EIGEN_STRONG_INLINE Packet2d pfloor<Packet2d>(const Packet2d& a) {
1318
+ return _mm_floor_pd(a);
1319
+ }
1320
+
1321
+ template <>
1322
+ EIGEN_STRONG_INLINE Packet4f ptrunc<Packet4f>(const Packet4f& a) {
1323
+ return _mm_round_ps(a, _MM_FROUND_TRUNC);
1324
+ }
1325
+ template <>
1326
+ EIGEN_STRONG_INLINE Packet2d ptrunc<Packet2d>(const Packet2d& a) {
1327
+ return _mm_round_pd(a, _MM_FROUND_TRUNC);
1328
+ }
1329
+ #endif
1330
+
1331
+ template <>
1332
+ EIGEN_STRONG_INLINE Packet4f pload<Packet4f>(const float* from) {
1333
+ EIGEN_DEBUG_ALIGNED_LOAD return _mm_load_ps(from);
1334
+ }
1335
+ template <>
1336
+ EIGEN_STRONG_INLINE Packet2d pload<Packet2d>(const double* from) {
1337
+ EIGEN_DEBUG_ALIGNED_LOAD return _mm_load_pd(from);
1338
+ }
1339
+ template <>
1340
+ EIGEN_STRONG_INLINE Packet2l pload<Packet2l>(const int64_t* from) {
1341
+ EIGEN_DEBUG_ALIGNED_LOAD return _mm_load_si128(reinterpret_cast<const __m128i*>(from));
1342
+ }
1343
+ template <>
1344
+ EIGEN_STRONG_INLINE Packet4i pload<Packet4i>(const int* from) {
1345
+ EIGEN_DEBUG_ALIGNED_LOAD return _mm_load_si128(reinterpret_cast<const __m128i*>(from));
1346
+ }
1347
+ template <>
1348
+ EIGEN_STRONG_INLINE Packet4ui pload<Packet4ui>(const uint32_t* from) {
1349
+ EIGEN_DEBUG_ALIGNED_LOAD return _mm_load_si128(reinterpret_cast<const __m128i*>(from));
1350
+ }
1351
+ template <>
1352
+ EIGEN_STRONG_INLINE Packet16b pload<Packet16b>(const bool* from) {
1353
+ EIGEN_DEBUG_ALIGNED_LOAD return _mm_load_si128(reinterpret_cast<const __m128i*>(from));
1354
+ }
1355
+
1356
+ #if EIGEN_COMP_MSVC
1357
+ template <>
1358
+ EIGEN_STRONG_INLINE Packet4f ploadu<Packet4f>(const float* from) {
1359
+ EIGEN_DEBUG_UNALIGNED_LOAD
1360
+ return _mm_loadu_ps(from);
1361
+ }
1362
+ #else
1363
+ // NOTE: with the code below, MSVC's compiler crashes!
1364
+
1365
+ template <>
1366
+ EIGEN_STRONG_INLINE Packet4f ploadu<Packet4f>(const float* from) {
1367
+ EIGEN_DEBUG_UNALIGNED_LOAD
1368
+ return _mm_loadu_ps(from);
1369
+ }
1370
+ #endif
1371
+
1372
+ template <>
1373
+ EIGEN_STRONG_INLINE Packet2d ploadu<Packet2d>(const double* from) {
1374
+ EIGEN_DEBUG_UNALIGNED_LOAD
1375
+ return _mm_loadu_pd(from);
1376
+ }
1377
+ template <>
1378
+ EIGEN_STRONG_INLINE Packet2l ploadu<Packet2l>(const int64_t* from) {
1379
+ EIGEN_DEBUG_UNALIGNED_LOAD
1380
+ return _mm_loadu_si128(reinterpret_cast<const __m128i*>(from));
1381
+ }
1382
+ template <>
1383
+ EIGEN_STRONG_INLINE Packet4i ploadu<Packet4i>(const int* from) {
1384
+ EIGEN_DEBUG_UNALIGNED_LOAD
1385
+ return _mm_loadu_si128(reinterpret_cast<const __m128i*>(from));
1386
+ }
1387
+ template <>
1388
+ EIGEN_STRONG_INLINE Packet4ui ploadu<Packet4ui>(const uint32_t* from) {
1389
+ EIGEN_DEBUG_UNALIGNED_LOAD
1390
+ return _mm_loadu_si128(reinterpret_cast<const __m128i*>(from));
1391
+ }
1392
+ template <>
1393
+ EIGEN_STRONG_INLINE Packet16b ploadu<Packet16b>(const bool* from) {
1394
+ EIGEN_DEBUG_UNALIGNED_LOAD
1395
+ return _mm_loadu_si128(reinterpret_cast<const __m128i*>(from));
1396
+ }
1397
+
1398
+ // Load lower part of packet zero extending.
1399
+ template <typename Packet>
1400
+ EIGEN_STRONG_INLINE Packet ploadl(const typename unpacket_traits<Packet>::type* from);
1401
+ template <>
1402
+ EIGEN_STRONG_INLINE Packet4f ploadl<Packet4f>(const float* from) {
1403
+ EIGEN_DEBUG_UNALIGNED_LOAD return _mm_castpd_ps(_mm_load_sd(reinterpret_cast<const double*>(from)));
1404
+ }
1405
+ template <>
1406
+ EIGEN_STRONG_INLINE Packet2d ploadl<Packet2d>(const double* from) {
1407
+ EIGEN_DEBUG_UNALIGNED_LOAD return _mm_load_sd(from);
1408
+ }
1409
+
1410
+ // Load scalar
1411
+ template <typename Packet>
1412
+ EIGEN_STRONG_INLINE Packet ploads(const typename unpacket_traits<Packet>::type* from);
1413
+ template <>
1414
+ EIGEN_STRONG_INLINE Packet4f ploads<Packet4f>(const float* from) {
1415
+ EIGEN_DEBUG_UNALIGNED_LOAD return _mm_load_ss(from);
1416
+ }
1417
+ template <>
1418
+ EIGEN_STRONG_INLINE Packet2d ploads<Packet2d>(const double* from) {
1419
+ EIGEN_DEBUG_UNALIGNED_LOAD return _mm_load_sd(from);
1420
+ }
1421
+
1422
+ template <>
1423
+ EIGEN_STRONG_INLINE Packet4f ploaddup<Packet4f>(const float* from) {
1424
+ return vec4f_swizzle1(_mm_castpd_ps(_mm_load_sd(reinterpret_cast<const double*>(from))), 0, 0, 1, 1);
1425
+ }
1426
+ template <>
1427
+ EIGEN_STRONG_INLINE Packet2d ploaddup<Packet2d>(const double* from) {
1428
+ return pset1<Packet2d>(from[0]);
1429
+ }
1430
+ template <>
1431
+ EIGEN_STRONG_INLINE Packet2l ploaddup<Packet2l>(const int64_t* from) {
1432
+ return pset1<Packet2l>(from[0]);
1433
+ }
1434
+ template <>
1435
+ EIGEN_STRONG_INLINE Packet4i ploaddup<Packet4i>(const int* from) {
1436
+ Packet4i tmp;
1437
+ tmp = _mm_loadl_epi64(reinterpret_cast<const __m128i*>(from));
1438
+ return vec4i_swizzle1(tmp, 0, 0, 1, 1);
1439
+ }
1440
+ template <>
1441
+ EIGEN_STRONG_INLINE Packet4ui ploaddup<Packet4ui>(const uint32_t* from) {
1442
+ Packet4ui tmp;
1443
+ tmp = _mm_loadl_epi64(reinterpret_cast<const __m128i*>(from));
1444
+ return vec4ui_swizzle1(tmp, 0, 0, 1, 1);
1445
+ }
1446
+
1447
+ // Loads 8 bools from memory and returns the packet
1448
+ // {b0, b0, b1, b1, b2, b2, b3, b3, b4, b4, b5, b5, b6, b6, b7, b7}
1449
+ template <>
1450
+ EIGEN_STRONG_INLINE Packet16b ploaddup<Packet16b>(const bool* from) {
1451
+ __m128i tmp = _mm_castpd_si128(pload1<Packet2d>(reinterpret_cast<const double*>(from)));
1452
+ return _mm_unpacklo_epi8(tmp, tmp);
1453
+ }
1454
+
1455
+ // Loads 4 bools from memory and returns the packet
1456
+ // {b0, b0 b0, b0, b1, b1, b1, b1, b2, b2, b2, b2, b3, b3, b3, b3}
1457
+ template <>
1458
+ EIGEN_STRONG_INLINE Packet16b ploadquad<Packet16b>(const bool* from) {
1459
+ __m128i tmp = _mm_castps_si128(pload1<Packet4f>(reinterpret_cast<const float*>(from)));
1460
+ tmp = _mm_unpacklo_epi8(tmp, tmp);
1461
+ return _mm_unpacklo_epi16(tmp, tmp);
1462
+ }
1463
+
1464
+ template <>
1465
+ EIGEN_STRONG_INLINE void pstore<float>(float* to, const Packet4f& from) {
1466
+ EIGEN_DEBUG_ALIGNED_STORE _mm_store_ps(to, from);
1467
+ }
1468
+ template <>
1469
+ EIGEN_STRONG_INLINE void pstore<double>(double* to, const Packet2d& from) {
1470
+ EIGEN_DEBUG_ALIGNED_STORE _mm_store_pd(to, from);
1471
+ }
1472
+ template <>
1473
+ EIGEN_STRONG_INLINE void pstore<int64_t>(int64_t* to, const Packet2l& from) {
1474
+ EIGEN_DEBUG_ALIGNED_STORE _mm_store_si128(reinterpret_cast<__m128i*>(to), from);
1475
+ }
1476
+ template <>
1477
+ EIGEN_STRONG_INLINE void pstore<int>(int* to, const Packet4i& from) {
1478
+ EIGEN_DEBUG_ALIGNED_STORE _mm_store_si128(reinterpret_cast<__m128i*>(to), from);
1479
+ }
1480
+ template <>
1481
+ EIGEN_STRONG_INLINE void pstore<uint32_t>(uint32_t* to, const Packet4ui& from) {
1482
+ EIGEN_DEBUG_ALIGNED_STORE _mm_store_si128(reinterpret_cast<__m128i*>(to), from);
1483
+ }
1484
+ template <>
1485
+ EIGEN_STRONG_INLINE void pstore<bool>(bool* to, const Packet16b& from) {
1486
+ EIGEN_DEBUG_ALIGNED_STORE _mm_store_si128(reinterpret_cast<__m128i*>(to), from);
1487
+ }
1488
+
1489
+ template <>
1490
+ EIGEN_STRONG_INLINE void pstoreu<double>(double* to, const Packet2d& from) {
1491
+ EIGEN_DEBUG_UNALIGNED_STORE _mm_storeu_pd(to, from);
1492
+ }
1493
+ template <>
1494
+ EIGEN_STRONG_INLINE void pstoreu<float>(float* to, const Packet4f& from) {
1495
+ EIGEN_DEBUG_UNALIGNED_STORE _mm_storeu_ps(to, from);
1496
+ }
1497
+ template <>
1498
+ EIGEN_STRONG_INLINE void pstoreu<int64_t>(int64_t* to, const Packet2l& from) {
1499
+ EIGEN_DEBUG_UNALIGNED_STORE _mm_storeu_si128(reinterpret_cast<__m128i*>(to), from);
1500
+ }
1501
+ template <>
1502
+ EIGEN_STRONG_INLINE void pstoreu<int>(int* to, const Packet4i& from) {
1503
+ EIGEN_DEBUG_UNALIGNED_STORE _mm_storeu_si128(reinterpret_cast<__m128i*>(to), from);
1504
+ }
1505
+ template <>
1506
+ EIGEN_STRONG_INLINE void pstoreu<uint32_t>(uint32_t* to, const Packet4ui& from) {
1507
+ EIGEN_DEBUG_UNALIGNED_STORE _mm_storeu_si128(reinterpret_cast<__m128i*>(to), from);
1508
+ }
1509
+ template <>
1510
+ EIGEN_STRONG_INLINE void pstoreu<bool>(bool* to, const Packet16b& from) {
1511
+ EIGEN_DEBUG_UNALIGNED_STORE _mm_storeu_si128(reinterpret_cast<__m128i*>(to), from);
1512
+ }
1513
+
1514
+ template <typename Scalar, typename Packet>
1515
+ EIGEN_STRONG_INLINE void pstorel(Scalar* to, const Packet& from);
1516
+ template <>
1517
+ EIGEN_STRONG_INLINE void pstorel(float* to, const Packet4f& from) {
1518
+ EIGEN_DEBUG_UNALIGNED_STORE _mm_storel_pi(reinterpret_cast<__m64*>(to), from);
1519
+ }
1520
+ template <>
1521
+ EIGEN_STRONG_INLINE void pstorel(double* to, const Packet2d& from) {
1522
+ EIGEN_DEBUG_UNALIGNED_STORE _mm_storel_pd(to, from);
1523
+ }
1524
+
1525
+ template <typename Scalar, typename Packet>
1526
+ EIGEN_STRONG_INLINE void pstores(Scalar* to, const Packet& from);
1527
+ template <>
1528
+ EIGEN_STRONG_INLINE void pstores(float* to, const Packet4f& from) {
1529
+ EIGEN_DEBUG_UNALIGNED_STORE _mm_store_ss(to, from);
1530
+ }
1531
+ template <>
1532
+ EIGEN_STRONG_INLINE void pstores(double* to, const Packet2d& from) {
1533
+ EIGEN_DEBUG_UNALIGNED_STORE _mm_store_sd(to, from);
1534
+ }
1535
+
1536
+ template <>
1537
+ EIGEN_STRONG_INLINE Packet4f preverse(const Packet4f& a) {
1538
+ return _mm_shuffle_ps(a, a, 0x1B);
1539
+ }
1540
+ template <>
1541
+ EIGEN_STRONG_INLINE Packet2d preverse(const Packet2d& a) {
1542
+ return _mm_shuffle_pd(a, a, 0x1);
1543
+ }
1544
+ template <>
1545
+ EIGEN_STRONG_INLINE Packet2l preverse(const Packet2l& a) {
1546
+ return _mm_castpd_si128(preverse(_mm_castsi128_pd(a)));
1547
+ }
1548
+ template <>
1549
+ EIGEN_STRONG_INLINE Packet4i preverse(const Packet4i& a) {
1550
+ return _mm_shuffle_epi32(a, 0x1B);
1551
+ }
1552
+ template <>
1553
+ EIGEN_STRONG_INLINE Packet4ui preverse(const Packet4ui& a) {
1554
+ return _mm_shuffle_epi32(a, 0x1B);
1555
+ }
1556
+ template <>
1557
+ EIGEN_STRONG_INLINE Packet16b preverse(const Packet16b& a) {
1558
+ #ifdef EIGEN_VECTORIZE_SSSE3
1559
+ __m128i mask = _mm_set_epi8(0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
1560
+ return _mm_shuffle_epi8(a, mask);
1561
+ #else
1562
+ Packet16b tmp = _mm_shuffle_epi32(a, _MM_SHUFFLE(0, 1, 2, 3));
1563
+ tmp = _mm_shufflehi_epi16(_mm_shufflelo_epi16(tmp, _MM_SHUFFLE(2, 3, 0, 1)), _MM_SHUFFLE(2, 3, 0, 1));
1564
+ return _mm_or_si128(_mm_slli_epi16(tmp, 8), _mm_srli_epi16(tmp, 8));
1565
+ #endif
1566
+ }
1567
+
1568
+ #if EIGEN_COMP_MSVC_STRICT && EIGEN_OS_WIN64
1569
+ // The temporary variable fixes an internal compilation error in vs <= 2008 and a wrong-result bug in vs 2010
1570
+ // Direct of the struct members fixed bug #62.
1571
+ template <>
1572
+ EIGEN_STRONG_INLINE float pfirst<Packet4f>(const Packet4f& a) {
1573
+ return a.m128_f32[0];
605
1574
  }
606
- template<> EIGEN_STRONG_INLINE Packet2d pabs(const Packet2d& a)
607
- {
608
- const Packet2d mask = _mm_castsi128_pd(_mm_setr_epi32(0xFFFFFFFF,0x7FFFFFFF,0xFFFFFFFF,0x7FFFFFFF));
609
- return _mm_and_pd(a,mask);
1575
+ template <>
1576
+ EIGEN_STRONG_INLINE double pfirst<Packet2d>(const Packet2d& a) {
1577
+ return a.m128d_f64[0];
610
1578
  }
611
- template<> EIGEN_STRONG_INLINE Packet4i pabs(const Packet4i& a)
612
- {
613
- #ifdef EIGEN_VECTORIZE_SSSE3
614
- return _mm_abs_epi32(a);
615
- #else
616
- Packet4i aux = _mm_srai_epi32(a,31);
617
- return _mm_sub_epi32(_mm_xor_si128(a,aux),aux);
618
- #endif
1579
+ template <>
1580
+ EIGEN_STRONG_INLINE int64_t pfirst<Packet2l>(const Packet2l& a) {
1581
+ int64_t x = _mm_extract_epi64_0(a);
1582
+ return x;
619
1583
  }
620
-
621
- #ifdef EIGEN_VECTORIZE_SSE4_1
622
- template<> EIGEN_STRONG_INLINE Packet4f pround<Packet4f>(const Packet4f& a)
623
- {
624
- // Unfortunatly _mm_round_ps doesn't have a rounding mode to implement numext::round.
625
- const Packet4f mask = pset1frombits<Packet4f>(0x80000000u);
626
- const Packet4f prev0dot5 = pset1frombits<Packet4f>(0x3EFFFFFFu);
627
- return _mm_round_ps(padd(por(pand(a, mask), prev0dot5), a), _MM_FROUND_TO_ZERO);
1584
+ template <>
1585
+ EIGEN_STRONG_INLINE int pfirst<Packet4i>(const Packet4i& a) {
1586
+ int x = _mm_cvtsi128_si32(a);
1587
+ return x;
628
1588
  }
629
-
630
- template<> EIGEN_STRONG_INLINE Packet2d pround<Packet2d>(const Packet2d& a)
631
- {
632
- const Packet2d mask = _mm_castsi128_pd(_mm_set_epi64x(0x8000000000000000ull, 0x8000000000000000ull));
633
- const Packet2d prev0dot5 = _mm_castsi128_pd(_mm_set_epi64x(0x3FDFFFFFFFFFFFFFull, 0x3FDFFFFFFFFFFFFFull));
634
- return _mm_round_pd(padd(por(pand(a, mask), prev0dot5), a), _MM_FROUND_TO_ZERO);
1589
+ template <>
1590
+ EIGEN_STRONG_INLINE uint32_t pfirst<Packet4ui>(const Packet4ui& a) {
1591
+ uint32_t x = numext::bit_cast<uint32_t>(_mm_cvtsi128_si32(a));
1592
+ return x;
635
1593
  }
636
-
637
- template<> EIGEN_STRONG_INLINE Packet4f print<Packet4f>(const Packet4f& a) { return _mm_round_ps(a, _MM_FROUND_CUR_DIRECTION); }
638
- template<> EIGEN_STRONG_INLINE Packet2d print<Packet2d>(const Packet2d& a) { return _mm_round_pd(a, _MM_FROUND_CUR_DIRECTION); }
639
-
640
- template<> EIGEN_STRONG_INLINE Packet4f pceil<Packet4f>(const Packet4f& a) { return _mm_ceil_ps(a); }
641
- template<> EIGEN_STRONG_INLINE Packet2d pceil<Packet2d>(const Packet2d& a) { return _mm_ceil_pd(a); }
642
-
643
- template<> EIGEN_STRONG_INLINE Packet4f pfloor<Packet4f>(const Packet4f& a) { return _mm_floor_ps(a); }
644
- template<> EIGEN_STRONG_INLINE Packet2d pfloor<Packet2d>(const Packet2d& a) { return _mm_floor_pd(a); }
645
- #else
646
- template<> EIGEN_STRONG_INLINE Packet4f print(const Packet4f& a) {
647
- // Adds and subtracts signum(a) * 2^23 to force rounding.
648
- const Packet4f limit = pset1<Packet4f>(static_cast<float>(1<<23));
649
- const Packet4f abs_a = pabs(a);
650
- Packet4f r = padd(abs_a, limit);
651
- // Don't compile-away addition and subtraction.
652
- EIGEN_OPTIMIZATION_BARRIER(r);
653
- r = psub(r, limit);
654
- // If greater than limit, simply return a. Otherwise, account for sign.
655
- r = pselect(pcmp_lt(abs_a, limit),
656
- pselect(pcmp_lt(a, pzero(a)), pnegate(r), r), a);
657
- return r;
658
- }
659
-
660
- template<> EIGEN_STRONG_INLINE Packet2d print(const Packet2d& a) {
661
- // Adds and subtracts signum(a) * 2^52 to force rounding.
662
- const Packet2d limit = pset1<Packet2d>(static_cast<double>(1ull<<52));
663
- const Packet2d abs_a = pabs(a);
664
- Packet2d r = padd(abs_a, limit);
665
- // Don't compile-away addition and subtraction.
666
- EIGEN_OPTIMIZATION_BARRIER(r);
667
- r = psub(r, limit);
668
- // If greater than limit, simply return a. Otherwise, account for sign.
669
- r = pselect(pcmp_lt(abs_a, limit),
670
- pselect(pcmp_lt(a, pzero(a)), pnegate(r), r), a);
671
- return r;
672
- }
673
-
674
- template<> EIGEN_STRONG_INLINE Packet4f pfloor<Packet4f>(const Packet4f& a)
675
- {
676
- const Packet4f cst_1 = pset1<Packet4f>(1.0f);
677
- Packet4f tmp = print<Packet4f>(a);
678
- // If greater, subtract one.
679
- Packet4f mask = _mm_cmpgt_ps(tmp, a);
680
- mask = pand(mask, cst_1);
681
- return psub(tmp, mask);
1594
+ #elif EIGEN_COMP_MSVC_STRICT
1595
+ // The temporary variable fixes an internal compilation error in vs <= 2008 and a wrong-result bug in vs 2010
1596
+ template <>
1597
+ EIGEN_STRONG_INLINE float pfirst<Packet4f>(const Packet4f& a) {
1598
+ float x = _mm_cvtss_f32(a);
1599
+ return x;
682
1600
  }
683
-
684
- template<> EIGEN_STRONG_INLINE Packet2d pfloor<Packet2d>(const Packet2d& a)
685
- {
686
- const Packet2d cst_1 = pset1<Packet2d>(1.0);
687
- Packet2d tmp = print<Packet2d>(a);
688
- // If greater, subtract one.
689
- Packet2d mask = _mm_cmpgt_pd(tmp, a);
690
- mask = pand(mask, cst_1);
691
- return psub(tmp, mask);
1601
+ template <>
1602
+ EIGEN_STRONG_INLINE double pfirst<Packet2d>(const Packet2d& a) {
1603
+ double x = _mm_cvtsd_f64(a);
1604
+ return x;
692
1605
  }
693
-
694
- template<> EIGEN_STRONG_INLINE Packet4f pceil<Packet4f>(const Packet4f& a)
695
- {
696
- const Packet4f cst_1 = pset1<Packet4f>(1.0f);
697
- Packet4f tmp = print<Packet4f>(a);
698
- // If smaller, add one.
699
- Packet4f mask = _mm_cmplt_ps(tmp, a);
700
- mask = pand(mask, cst_1);
701
- return padd(tmp, mask);
1606
+ template <>
1607
+ EIGEN_STRONG_INLINE int64_t pfirst<Packet2l>(const Packet2l& a) {
1608
+ int64_t x = _mm_extract_epi64_0(a);
1609
+ return x;
702
1610
  }
703
-
704
- template<> EIGEN_STRONG_INLINE Packet2d pceil<Packet2d>(const Packet2d& a)
705
- {
706
- const Packet2d cst_1 = pset1<Packet2d>(1.0);
707
- Packet2d tmp = print<Packet2d>(a);
708
- // If smaller, add one.
709
- Packet2d mask = _mm_cmplt_pd(tmp, a);
710
- mask = pand(mask, cst_1);
711
- return padd(tmp, mask);
1611
+ template <>
1612
+ EIGEN_STRONG_INLINE int pfirst<Packet4i>(const Packet4i& a) {
1613
+ int x = _mm_cvtsi128_si32(a);
1614
+ return x;
1615
+ }
1616
+ template <>
1617
+ EIGEN_STRONG_INLINE uint32_t pfirst<Packet4ui>(const Packet4ui& a) {
1618
+ uint32_t x = numext::bit_cast<uint32_t>(_mm_cvtsi128_si32(a));
1619
+ return x;
712
1620
  }
713
- #endif
714
-
715
- template<> EIGEN_STRONG_INLINE Packet4f pload<Packet4f>(const float* from) { EIGEN_DEBUG_ALIGNED_LOAD return _mm_load_ps(from); }
716
- template<> EIGEN_STRONG_INLINE Packet2d pload<Packet2d>(const double* from) { EIGEN_DEBUG_ALIGNED_LOAD return _mm_load_pd(from); }
717
- template<> EIGEN_STRONG_INLINE Packet4i pload<Packet4i>(const int* from) { EIGEN_DEBUG_ALIGNED_LOAD return _mm_load_si128(reinterpret_cast<const __m128i*>(from)); }
718
- template<> EIGEN_STRONG_INLINE Packet16b pload<Packet16b>(const bool* from) { EIGEN_DEBUG_ALIGNED_LOAD return _mm_load_si128(reinterpret_cast<const __m128i*>(from)); }
719
-
720
- #if EIGEN_COMP_MSVC
721
- template<> EIGEN_STRONG_INLINE Packet4f ploadu<Packet4f>(const float* from) {
722
- EIGEN_DEBUG_UNALIGNED_LOAD
723
- #if (EIGEN_COMP_MSVC==1600)
724
- // NOTE Some version of MSVC10 generates bad code when using _mm_loadu_ps
725
- // (i.e., it does not generate an unaligned load!!
726
- __m128 res = _mm_loadl_pi(_mm_set1_ps(0.0f), (const __m64*)(from));
727
- res = _mm_loadh_pi(res, (const __m64*)(from+2));
728
- return res;
729
- #else
730
- return _mm_loadu_ps(from);
731
- #endif
732
- }
733
1621
  #else
734
- // NOTE: with the code below, MSVC's compiler crashes!
735
-
736
- template<> EIGEN_STRONG_INLINE Packet4f ploadu<Packet4f>(const float* from)
737
- {
738
- EIGEN_DEBUG_UNALIGNED_LOAD
739
- return _mm_loadu_ps(from);
1622
+ template <>
1623
+ EIGEN_STRONG_INLINE float pfirst<Packet4f>(const Packet4f& a) {
1624
+ return _mm_cvtss_f32(a);
740
1625
  }
741
- #endif
742
-
743
- template<> EIGEN_STRONG_INLINE Packet2d ploadu<Packet2d>(const double* from)
744
- {
745
- EIGEN_DEBUG_UNALIGNED_LOAD
746
- return _mm_loadu_pd(from);
1626
+ template <>
1627
+ EIGEN_STRONG_INLINE double pfirst<Packet2d>(const Packet2d& a) {
1628
+ return _mm_cvtsd_f64(a);
747
1629
  }
748
- template<> EIGEN_STRONG_INLINE Packet4i ploadu<Packet4i>(const int* from)
749
- {
750
- EIGEN_DEBUG_UNALIGNED_LOAD
751
- return _mm_loadu_si128(reinterpret_cast<const __m128i*>(from));
1630
+ template <>
1631
+ EIGEN_STRONG_INLINE int64_t pfirst<Packet2l>(const Packet2l& a) {
1632
+ return _mm_extract_epi64_0(a);
752
1633
  }
753
- template<> EIGEN_STRONG_INLINE Packet16b ploadu<Packet16b>(const bool* from) {
754
- EIGEN_DEBUG_UNALIGNED_LOAD
755
- return _mm_loadu_si128(reinterpret_cast<const __m128i*>(from));
1634
+ template <>
1635
+ EIGEN_STRONG_INLINE int pfirst<Packet4i>(const Packet4i& a) {
1636
+ return _mm_cvtsi128_si32(a);
756
1637
  }
757
-
758
-
759
- template<> EIGEN_STRONG_INLINE Packet4f ploaddup<Packet4f>(const float* from)
760
- {
761
- return vec4f_swizzle1(_mm_castpd_ps(_mm_load_sd(reinterpret_cast<const double*>(from))), 0, 0, 1, 1);
1638
+ template <>
1639
+ EIGEN_STRONG_INLINE uint32_t pfirst<Packet4ui>(const Packet4ui& a) {
1640
+ return numext::bit_cast<uint32_t>(_mm_cvtsi128_si32(a));
762
1641
  }
763
- template<> EIGEN_STRONG_INLINE Packet2d ploaddup<Packet2d>(const double* from)
764
- { return pset1<Packet2d>(from[0]); }
765
- template<> EIGEN_STRONG_INLINE Packet4i ploaddup<Packet4i>(const int* from)
766
- {
767
- Packet4i tmp;
768
- tmp = _mm_loadl_epi64(reinterpret_cast<const __m128i*>(from));
769
- return vec4i_swizzle1(tmp, 0, 0, 1, 1);
1642
+ #endif
1643
+ template <>
1644
+ EIGEN_STRONG_INLINE bool pfirst<Packet16b>(const Packet16b& a) {
1645
+ int x = _mm_cvtsi128_si32(a);
1646
+ return static_cast<bool>(x & 1);
770
1647
  }
771
1648
 
772
- // Loads 8 bools from memory and returns the packet
773
- // {b0, b0, b1, b1, b2, b2, b3, b3, b4, b4, b5, b5, b6, b6, b7, b7}
774
- template<> EIGEN_STRONG_INLINE Packet16b ploaddup<Packet16b>(const bool* from)
775
- {
776
- __m128i tmp = _mm_castpd_si128(pload1<Packet2d>(reinterpret_cast<const double*>(from)));
777
- return _mm_unpacklo_epi8(tmp, tmp);
1649
+ template <>
1650
+ EIGEN_STRONG_INLINE Packet4f pgather<float, Packet4f>(const float* from, Index stride) {
1651
+ return _mm_set_ps(from[3 * stride], from[2 * stride], from[1 * stride], from[0 * stride]);
778
1652
  }
779
-
780
- // Loads 4 bools from memory and returns the packet
781
- // {b0, b0 b0, b0, b1, b1, b1, b1, b2, b2, b2, b2, b3, b3, b3, b3}
782
- template<> EIGEN_STRONG_INLINE Packet16b
783
- ploadquad<Packet16b>(const bool* from) {
784
- __m128i tmp = _mm_castps_si128(pload1<Packet4f>(reinterpret_cast<const float*>(from)));
785
- tmp = _mm_unpacklo_epi8(tmp, tmp);
786
- return _mm_unpacklo_epi16(tmp, tmp);
1653
+ template <>
1654
+ EIGEN_STRONG_INLINE Packet2d pgather<double, Packet2d>(const double* from, Index stride) {
1655
+ return _mm_set_pd(from[1 * stride], from[0 * stride]);
787
1656
  }
788
-
789
- template<> EIGEN_STRONG_INLINE void pstore<float>(float* to, const Packet4f& from) { EIGEN_DEBUG_ALIGNED_STORE _mm_store_ps(to, from); }
790
- template<> EIGEN_STRONG_INLINE void pstore<double>(double* to, const Packet2d& from) { EIGEN_DEBUG_ALIGNED_STORE _mm_store_pd(to, from); }
791
- template<> EIGEN_STRONG_INLINE void pstore<int>(int* to, const Packet4i& from) { EIGEN_DEBUG_ALIGNED_STORE _mm_store_si128(reinterpret_cast<__m128i*>(to), from); }
792
- template<> EIGEN_STRONG_INLINE void pstore<bool>(bool* to, const Packet16b& from) { EIGEN_DEBUG_ALIGNED_STORE _mm_store_si128(reinterpret_cast<__m128i*>(to), from); }
793
-
794
- template<> EIGEN_STRONG_INLINE void pstoreu<double>(double* to, const Packet2d& from) { EIGEN_DEBUG_UNALIGNED_STORE _mm_storeu_pd(to, from); }
795
- template<> EIGEN_STRONG_INLINE void pstoreu<float>(float* to, const Packet4f& from) { EIGEN_DEBUG_UNALIGNED_STORE _mm_storeu_ps(to, from); }
796
- template<> EIGEN_STRONG_INLINE void pstoreu<int>(int* to, const Packet4i& from) { EIGEN_DEBUG_UNALIGNED_STORE _mm_storeu_si128(reinterpret_cast<__m128i*>(to), from); }
797
- template<> EIGEN_STRONG_INLINE void pstoreu<bool>(bool* to, const Packet16b& from) { EIGEN_DEBUG_ALIGNED_STORE _mm_storeu_si128(reinterpret_cast<__m128i*>(to), from); }
798
-
799
- template<> EIGEN_DEVICE_FUNC inline Packet4f pgather<float, Packet4f>(const float* from, Index stride)
800
- {
801
- return _mm_set_ps(from[3*stride], from[2*stride], from[1*stride], from[0*stride]);
1657
+ template <>
1658
+ EIGEN_STRONG_INLINE Packet2l pgather<int64_t, Packet2l>(const int64_t* from, Index stride) {
1659
+ return _mm_set_epi64x(from[1 * stride], from[0 * stride]);
802
1660
  }
803
- template<> EIGEN_DEVICE_FUNC inline Packet2d pgather<double, Packet2d>(const double* from, Index stride)
804
- {
805
- return _mm_set_pd(from[1*stride], from[0*stride]);
1661
+ template <>
1662
+ EIGEN_STRONG_INLINE Packet4i pgather<int, Packet4i>(const int* from, Index stride) {
1663
+ return _mm_set_epi32(from[3 * stride], from[2 * stride], from[1 * stride], from[0 * stride]);
806
1664
  }
807
- template<> EIGEN_DEVICE_FUNC inline Packet4i pgather<int, Packet4i>(const int* from, Index stride)
808
- {
809
- return _mm_set_epi32(from[3*stride], from[2*stride], from[1*stride], from[0*stride]);
1665
+ template <>
1666
+ EIGEN_STRONG_INLINE Packet4ui pgather<uint32_t, Packet4ui>(const uint32_t* from, Index stride) {
1667
+ return _mm_set_epi32(numext::bit_cast<int32_t>(from[3 * stride]), numext::bit_cast<int32_t>(from[2 * stride]),
1668
+ numext::bit_cast<int32_t>(from[1 * stride]), numext::bit_cast<int32_t>(from[0 * stride]));
810
1669
  }
811
1670
 
812
- template<> EIGEN_DEVICE_FUNC inline Packet16b pgather<bool, Packet16b>(const bool* from, Index stride)
813
- {
814
- return _mm_set_epi8(from[15*stride], from[14*stride], from[13*stride], from[12*stride],
815
- from[11*stride], from[10*stride], from[9*stride], from[8*stride],
816
- from[7*stride], from[6*stride], from[5*stride], from[4*stride],
817
- from[3*stride], from[2*stride], from[1*stride], from[0*stride]);
1671
+ template <>
1672
+ EIGEN_STRONG_INLINE Packet16b pgather<bool, Packet16b>(const bool* from, Index stride) {
1673
+ return _mm_set_epi8(from[15 * stride], from[14 * stride], from[13 * stride], from[12 * stride], from[11 * stride],
1674
+ from[10 * stride], from[9 * stride], from[8 * stride], from[7 * stride], from[6 * stride],
1675
+ from[5 * stride], from[4 * stride], from[3 * stride], from[2 * stride], from[1 * stride],
1676
+ from[0 * stride]);
818
1677
  }
819
1678
 
820
- template<> EIGEN_DEVICE_FUNC inline void pscatter<float, Packet4f>(float* to, const Packet4f& from, Index stride)
821
- {
822
- to[stride*0] = _mm_cvtss_f32(from);
823
- to[stride*1] = _mm_cvtss_f32(_mm_shuffle_ps(from, from, 1));
824
- to[stride*2] = _mm_cvtss_f32(_mm_shuffle_ps(from, from, 2));
825
- to[stride*3] = _mm_cvtss_f32(_mm_shuffle_ps(from, from, 3));
1679
+ template <>
1680
+ EIGEN_STRONG_INLINE void pscatter<float, Packet4f>(float* to, const Packet4f& from, Index stride) {
1681
+ to[stride * 0] = pfirst(from);
1682
+ to[stride * 1] = pfirst(Packet4f(_mm_shuffle_ps(from, from, 1)));
1683
+ to[stride * 2] = pfirst(Packet4f(_mm_shuffle_ps(from, from, 2)));
1684
+ to[stride * 3] = pfirst(Packet4f(_mm_shuffle_ps(from, from, 3)));
826
1685
  }
827
- template<> EIGEN_DEVICE_FUNC inline void pscatter<double, Packet2d>(double* to, const Packet2d& from, Index stride)
828
- {
829
- to[stride*0] = _mm_cvtsd_f64(from);
830
- to[stride*1] = _mm_cvtsd_f64(_mm_shuffle_pd(from, from, 1));
1686
+ template <>
1687
+ EIGEN_STRONG_INLINE void pscatter<double, Packet2d>(double* to, const Packet2d& from, Index stride) {
1688
+ to[stride * 0] = pfirst(from);
1689
+ to[stride * 1] = pfirst(preverse(from));
831
1690
  }
832
- template<> EIGEN_DEVICE_FUNC inline void pscatter<int, Packet4i>(int* to, const Packet4i& from, Index stride)
833
- {
834
- to[stride*0] = _mm_cvtsi128_si32(from);
835
- to[stride*1] = _mm_cvtsi128_si32(_mm_shuffle_epi32(from, 1));
836
- to[stride*2] = _mm_cvtsi128_si32(_mm_shuffle_epi32(from, 2));
837
- to[stride*3] = _mm_cvtsi128_si32(_mm_shuffle_epi32(from, 3));
1691
+ template <>
1692
+ EIGEN_STRONG_INLINE void pscatter<int64_t, Packet2l>(int64_t* to, const Packet2l& from, Index stride) {
1693
+ to[stride * 0] = pfirst(from);
1694
+ to[stride * 1] = pfirst(preverse(from));
838
1695
  }
839
- template<> EIGEN_DEVICE_FUNC inline void pscatter<bool, Packet16b>(bool* to, const Packet16b& from, Index stride)
840
- {
841
- to[4*stride*0] = _mm_cvtsi128_si32(from);
842
- to[4*stride*1] = _mm_cvtsi128_si32(_mm_shuffle_epi32(from, 1));
843
- to[4*stride*2] = _mm_cvtsi128_si32(_mm_shuffle_epi32(from, 2));
844
- to[4*stride*3] = _mm_cvtsi128_si32(_mm_shuffle_epi32(from, 3));
1696
+ template <>
1697
+ EIGEN_STRONG_INLINE void pscatter<int, Packet4i>(int* to, const Packet4i& from, Index stride) {
1698
+ to[stride * 0] = _mm_cvtsi128_si32(from);
1699
+ to[stride * 1] = _mm_cvtsi128_si32(_mm_shuffle_epi32(from, 1));
1700
+ to[stride * 2] = _mm_cvtsi128_si32(_mm_shuffle_epi32(from, 2));
1701
+ to[stride * 3] = _mm_cvtsi128_si32(_mm_shuffle_epi32(from, 3));
1702
+ }
1703
+ template <>
1704
+ EIGEN_STRONG_INLINE void pscatter<uint32_t, Packet4ui>(uint32_t* to, const Packet4ui& from, Index stride) {
1705
+ to[stride * 0] = numext::bit_cast<uint32_t>(_mm_cvtsi128_si32(from));
1706
+ to[stride * 1] = numext::bit_cast<uint32_t>(_mm_cvtsi128_si32(_mm_shuffle_epi32(from, 1)));
1707
+ to[stride * 2] = numext::bit_cast<uint32_t>(_mm_cvtsi128_si32(_mm_shuffle_epi32(from, 2)));
1708
+ to[stride * 3] = numext::bit_cast<uint32_t>(_mm_cvtsi128_si32(_mm_shuffle_epi32(from, 3)));
1709
+ }
1710
+ template <>
1711
+ EIGEN_STRONG_INLINE void pscatter<bool, Packet16b>(bool* to, const Packet16b& from, Index stride) {
1712
+ EIGEN_ALIGN16 bool tmp[16];
1713
+ pstore(tmp, from);
1714
+ to[stride * 0] = tmp[0];
1715
+ to[stride * 1] = tmp[1];
1716
+ to[stride * 2] = tmp[2];
1717
+ to[stride * 3] = tmp[3];
1718
+ to[stride * 4] = tmp[4];
1719
+ to[stride * 5] = tmp[5];
1720
+ to[stride * 6] = tmp[6];
1721
+ to[stride * 7] = tmp[7];
1722
+ to[stride * 8] = tmp[8];
1723
+ to[stride * 9] = tmp[9];
1724
+ to[stride * 10] = tmp[10];
1725
+ to[stride * 11] = tmp[11];
1726
+ to[stride * 12] = tmp[12];
1727
+ to[stride * 13] = tmp[13];
1728
+ to[stride * 14] = tmp[14];
1729
+ to[stride * 15] = tmp[15];
845
1730
  }
846
-
847
1731
 
848
1732
  // some compilers might be tempted to perform multiple moves instead of using a vector path.
849
- template<> EIGEN_STRONG_INLINE void pstore1<Packet4f>(float* to, const float& a)
850
- {
1733
+ template <>
1734
+ EIGEN_STRONG_INLINE void pstore1<Packet4f>(float* to, const float& a) {
851
1735
  Packet4f pa = _mm_set_ss(a);
852
- pstore(to, Packet4f(vec4f_swizzle1(pa,0,0,0,0)));
1736
+ pstore(to, Packet4f(vec4f_swizzle1(pa, 0, 0, 0, 0)));
853
1737
  }
854
1738
  // some compilers might be tempted to perform multiple moves instead of using a vector path.
855
- template<> EIGEN_STRONG_INLINE void pstore1<Packet2d>(double* to, const double& a)
856
- {
1739
+ template <>
1740
+ EIGEN_STRONG_INLINE void pstore1<Packet2d>(double* to, const double& a) {
857
1741
  Packet2d pa = _mm_set_sd(a);
858
- pstore(to, Packet2d(vec2d_swizzle1(pa,0,0)));
1742
+ pstore(to, Packet2d(vec2d_swizzle1(pa, 0, 0)));
859
1743
  }
860
1744
 
861
1745
  #if EIGEN_COMP_PGI && EIGEN_COMP_PGI < 1900
862
- typedef const void * SsePrefetchPtrType;
1746
+ typedef const void* SsePrefetchPtrType;
863
1747
  #else
864
- typedef const char * SsePrefetchPtrType;
1748
+ typedef const char* SsePrefetchPtrType;
865
1749
  #endif
866
1750
 
867
1751
  #ifndef EIGEN_VECTORIZE_AVX
868
- template<> EIGEN_STRONG_INLINE void prefetch<float>(const float* addr) { _mm_prefetch((SsePrefetchPtrType)(addr), _MM_HINT_T0); }
869
- template<> EIGEN_STRONG_INLINE void prefetch<double>(const double* addr) { _mm_prefetch((SsePrefetchPtrType)(addr), _MM_HINT_T0); }
870
- template<> EIGEN_STRONG_INLINE void prefetch<int>(const int* addr) { _mm_prefetch((SsePrefetchPtrType)(addr), _MM_HINT_T0); }
871
- #endif
872
-
873
- #if EIGEN_COMP_MSVC_STRICT && EIGEN_OS_WIN64
874
- // The temporary variable fixes an internal compilation error in vs <= 2008 and a wrong-result bug in vs 2010
875
- // Direct of the struct members fixed bug #62.
876
- template<> EIGEN_STRONG_INLINE float pfirst<Packet4f>(const Packet4f& a) { return a.m128_f32[0]; }
877
- template<> EIGEN_STRONG_INLINE double pfirst<Packet2d>(const Packet2d& a) { return a.m128d_f64[0]; }
878
- template<> EIGEN_STRONG_INLINE int pfirst<Packet4i>(const Packet4i& a) { int x = _mm_cvtsi128_si32(a); return x; }
879
- #elif EIGEN_COMP_MSVC_STRICT
880
- // The temporary variable fixes an internal compilation error in vs <= 2008 and a wrong-result bug in vs 2010
881
- template<> EIGEN_STRONG_INLINE float pfirst<Packet4f>(const Packet4f& a) { float x = _mm_cvtss_f32(a); return x; }
882
- template<> EIGEN_STRONG_INLINE double pfirst<Packet2d>(const Packet2d& a) { double x = _mm_cvtsd_f64(a); return x; }
883
- template<> EIGEN_STRONG_INLINE int pfirst<Packet4i>(const Packet4i& a) { int x = _mm_cvtsi128_si32(a); return x; }
884
- #else
885
- template<> EIGEN_STRONG_INLINE float pfirst<Packet4f>(const Packet4f& a) { return _mm_cvtss_f32(a); }
886
- template<> EIGEN_STRONG_INLINE double pfirst<Packet2d>(const Packet2d& a) { return _mm_cvtsd_f64(a); }
887
- template<> EIGEN_STRONG_INLINE int pfirst<Packet4i>(const Packet4i& a) { return _mm_cvtsi128_si32(a); }
888
- #endif
889
- template<> EIGEN_STRONG_INLINE bool pfirst<Packet16b>(const Packet16b& a) { int x = _mm_cvtsi128_si32(a); return static_cast<bool>(x & 1); }
890
-
891
- template<> EIGEN_STRONG_INLINE Packet4f preverse(const Packet4f& a) { return _mm_shuffle_ps(a,a,0x1B); }
892
- template<> EIGEN_STRONG_INLINE Packet2d preverse(const Packet2d& a) { return _mm_shuffle_pd(a,a,0x1); }
893
- template<> EIGEN_STRONG_INLINE Packet4i preverse(const Packet4i& a) { return _mm_shuffle_epi32(a,0x1B); }
894
- template<> EIGEN_STRONG_INLINE Packet16b preverse(const Packet16b& a) {
895
- #ifdef EIGEN_VECTORIZE_SSSE3
896
- __m128i mask = _mm_set_epi8(0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
897
- return _mm_shuffle_epi8(a, mask);
898
- #else
899
- Packet16b tmp = _mm_shuffle_epi32(a, _MM_SHUFFLE(0, 1, 2, 3));
900
- tmp = _mm_shufflehi_epi16(_mm_shufflelo_epi16(tmp, _MM_SHUFFLE(2, 3, 0, 1)), _MM_SHUFFLE(2, 3, 0, 1));
901
- return _mm_or_si128(_mm_slli_epi16(tmp, 8), _mm_srli_epi16(tmp, 8));
902
- #endif
1752
+ template <>
1753
+ EIGEN_STRONG_INLINE void prefetch<float>(const float* addr) {
1754
+ _mm_prefetch((SsePrefetchPtrType)(addr), _MM_HINT_T0);
1755
+ }
1756
+ template <>
1757
+ EIGEN_STRONG_INLINE void prefetch<double>(const double* addr) {
1758
+ _mm_prefetch((SsePrefetchPtrType)(addr), _MM_HINT_T0);
1759
+ }
1760
+ template <>
1761
+ EIGEN_STRONG_INLINE void prefetch<int>(const int* addr) {
1762
+ _mm_prefetch((SsePrefetchPtrType)(addr), _MM_HINT_T0);
1763
+ }
1764
+ template <>
1765
+ EIGEN_STRONG_INLINE void prefetch<int64_t>(const int64_t* addr) {
1766
+ _mm_prefetch((SsePrefetchPtrType)(addr), _MM_HINT_T0);
1767
+ }
1768
+ template <>
1769
+ EIGEN_STRONG_INLINE void prefetch<uint32_t>(const uint32_t* addr) {
1770
+ _mm_prefetch((SsePrefetchPtrType)(addr), _MM_HINT_T0);
903
1771
  }
1772
+ #endif
904
1773
 
905
- template<> EIGEN_STRONG_INLINE Packet4f pfrexp<Packet4f>(const Packet4f& a, Packet4f& exponent) {
906
- return pfrexp_generic(a,exponent);
1774
+ template <>
1775
+ EIGEN_STRONG_INLINE Packet4f pfrexp<Packet4f>(const Packet4f& a, Packet4f& exponent) {
1776
+ return pfrexp_generic(a, exponent);
907
1777
  }
908
1778
 
909
1779
  // Extract exponent without existence of Packet2l.
910
- template<>
911
- EIGEN_STRONG_INLINE
912
- Packet2d pfrexp_generic_get_biased_exponent(const Packet2d& a) {
913
- const Packet2d cst_exp_mask = pset1frombits<Packet2d>(static_cast<uint64_t>(0x7ff0000000000000ull));
1780
+ template <>
1781
+ EIGEN_STRONG_INLINE Packet2d pfrexp_generic_get_biased_exponent(const Packet2d& a) {
1782
+ const Packet2d cst_exp_mask = pset1frombits<Packet2d>(static_cast<uint64_t>(0x7ff0000000000000ull));
914
1783
  __m128i a_expo = _mm_srli_epi64(_mm_castpd_si128(pand(a, cst_exp_mask)), 52);
915
1784
  return _mm_cvtepi32_pd(vec4i_swizzle1(a_expo, 0, 2, 1, 3));
916
1785
  }
917
1786
 
918
- template<> EIGEN_STRONG_INLINE Packet2d pfrexp<Packet2d>(const Packet2d& a, Packet2d& exponent) {
1787
+ template <>
1788
+ EIGEN_STRONG_INLINE Packet2d pfrexp<Packet2d>(const Packet2d& a, Packet2d& exponent) {
919
1789
  return pfrexp_generic(a, exponent);
920
1790
  }
921
1791
 
922
- template<> EIGEN_STRONG_INLINE Packet4f pldexp<Packet4f>(const Packet4f& a, const Packet4f& exponent) {
923
- return pldexp_generic(a,exponent);
1792
+ template <>
1793
+ EIGEN_STRONG_INLINE Packet4f pldexp<Packet4f>(const Packet4f& a, const Packet4f& exponent) {
1794
+ return pldexp_generic(a, exponent);
924
1795
  }
925
1796
 
926
1797
  // We specialize pldexp here, since the generic implementation uses Packet2l, which is not well
927
1798
  // supported by SSE, and has more range than is needed for exponents.
928
- template<> EIGEN_STRONG_INLINE Packet2d pldexp<Packet2d>(const Packet2d& a, const Packet2d& exponent) {
1799
+ template <>
1800
+ EIGEN_STRONG_INLINE Packet2d pldexp<Packet2d>(const Packet2d& a, const Packet2d& exponent) {
929
1801
  // Clamp exponent to [-2099, 2099]
930
1802
  const Packet2d max_exponent = pset1<Packet2d>(2099.0);
931
1803
  const Packet2d e = pmin(pmax(exponent, pnegate(max_exponent)), max_exponent);
932
-
1804
+
933
1805
  // Convert e to integer and swizzle to low-order bits.
934
1806
  const Packet4i ei = vec4i_swizzle1(_mm_cvtpd_epi32(e), 0, 3, 1, 3);
935
-
1807
+
936
1808
  // Split 2^e into four factors and multiply:
937
1809
  const Packet4i bias = _mm_set_epi32(0, 1023, 0, 1023);
938
- Packet4i b = parithmetic_shift_right<2>(ei); // floor(e/4)
1810
+ Packet4i b = parithmetic_shift_right<2>(ei); // floor(e/4)
939
1811
  Packet2d c = _mm_castsi128_pd(_mm_slli_epi64(padd(b, bias), 52)); // 2^b
940
- Packet2d out = pmul(pmul(pmul(a, c), c), c); // a * 2^(3b)
941
- b = psub(psub(psub(ei, b), b), b); // e - 3b
942
- c = _mm_castsi128_pd(_mm_slli_epi64(padd(b, bias), 52)); // 2^(e - 3b)
943
- out = pmul(out, c); // a * 2^e
1812
+ Packet2d out = pmul(pmul(pmul(a, c), c), c); // a * 2^(3b)
1813
+ b = psub(psub(psub(ei, b), b), b); // e - 3b
1814
+ c = _mm_castsi128_pd(_mm_slli_epi64(padd(b, bias), 52)); // 2^(e - 3b)
1815
+ out = pmul(out, c); // a * 2^e
944
1816
  return out;
945
1817
  }
946
1818
 
1819
+ // We specialize pldexp here, since the generic implementation uses Packet2l, which is not well
1820
+ // supported by SSE, and has more range than is needed for exponents.
1821
+ template <>
1822
+ EIGEN_STRONG_INLINE Packet2d pldexp_fast<Packet2d>(const Packet2d& a, const Packet2d& exponent) {
1823
+ // Clamp exponent to [-1023, 1024]
1824
+ const Packet2d min_exponent = pset1<Packet2d>(-1023.0);
1825
+ const Packet2d max_exponent = pset1<Packet2d>(1024.0);
1826
+ const Packet2d e = pmin(pmax(exponent, min_exponent), max_exponent);
1827
+
1828
+ // Convert e to integer and swizzle to low-order bits.
1829
+ const Packet4i ei = vec4i_swizzle1(_mm_cvtpd_epi32(e), 0, 3, 1, 3);
1830
+
1831
+ // Compute 2^e multiply:
1832
+ const Packet4i bias = _mm_set_epi32(0, 1023, 0, 1023);
1833
+ const Packet2d c = _mm_castsi128_pd(_mm_slli_epi64(padd(ei, bias), 52)); // 2^e
1834
+ return pmul(a, c);
1835
+ }
1836
+
947
1837
  // with AVX, the default implementations based on pload1 are faster
948
1838
  #ifndef __AVX__
949
- template<> EIGEN_STRONG_INLINE void
950
- pbroadcast4<Packet4f>(const float *a,
951
- Packet4f& a0, Packet4f& a1, Packet4f& a2, Packet4f& a3)
952
- {
1839
+ template <>
1840
+ EIGEN_STRONG_INLINE void pbroadcast4<Packet4f>(const float* a, Packet4f& a0, Packet4f& a1, Packet4f& a2, Packet4f& a3) {
953
1841
  a3 = pload<Packet4f>(a);
954
- a0 = vec4f_swizzle1(a3, 0,0,0,0);
955
- a1 = vec4f_swizzle1(a3, 1,1,1,1);
956
- a2 = vec4f_swizzle1(a3, 2,2,2,2);
957
- a3 = vec4f_swizzle1(a3, 3,3,3,3);
958
- }
959
- template<> EIGEN_STRONG_INLINE void
960
- pbroadcast4<Packet2d>(const double *a,
961
- Packet2d& a0, Packet2d& a1, Packet2d& a2, Packet2d& a3)
962
- {
1842
+ a0 = vec4f_swizzle1(a3, 0, 0, 0, 0);
1843
+ a1 = vec4f_swizzle1(a3, 1, 1, 1, 1);
1844
+ a2 = vec4f_swizzle1(a3, 2, 2, 2, 2);
1845
+ a3 = vec4f_swizzle1(a3, 3, 3, 3, 3);
1846
+ }
1847
+ template <>
1848
+ EIGEN_STRONG_INLINE void pbroadcast4<Packet2d>(const double* a, Packet2d& a0, Packet2d& a1, Packet2d& a2,
1849
+ Packet2d& a3) {
963
1850
  #ifdef EIGEN_VECTORIZE_SSE3
964
- a0 = _mm_loaddup_pd(a+0);
965
- a1 = _mm_loaddup_pd(a+1);
966
- a2 = _mm_loaddup_pd(a+2);
967
- a3 = _mm_loaddup_pd(a+3);
1851
+ a0 = _mm_loaddup_pd(a + 0);
1852
+ a1 = _mm_loaddup_pd(a + 1);
1853
+ a2 = _mm_loaddup_pd(a + 2);
1854
+ a3 = _mm_loaddup_pd(a + 3);
968
1855
  #else
969
1856
  a1 = pload<Packet2d>(a);
970
- a0 = vec2d_swizzle1(a1, 0,0);
971
- a1 = vec2d_swizzle1(a1, 1,1);
972
- a3 = pload<Packet2d>(a+2);
973
- a2 = vec2d_swizzle1(a3, 0,0);
974
- a3 = vec2d_swizzle1(a3, 1,1);
1857
+ a0 = vec2d_swizzle1(a1, 0, 0);
1858
+ a1 = vec2d_swizzle1(a1, 1, 1);
1859
+ a3 = pload<Packet2d>(a + 2);
1860
+ a2 = vec2d_swizzle1(a3, 0, 0);
1861
+ a3 = vec2d_swizzle1(a3, 1, 1);
975
1862
  #endif
976
1863
  }
977
1864
  #endif
978
1865
 
979
- EIGEN_STRONG_INLINE void punpackp(Packet4f* vecs)
980
- {
1866
+ EIGEN_STRONG_INLINE void punpackp(Packet4f* vecs) {
981
1867
  vecs[1] = _mm_castsi128_ps(_mm_shuffle_epi32(_mm_castps_si128(vecs[0]), 0x55));
982
1868
  vecs[2] = _mm_castsi128_ps(_mm_shuffle_epi32(_mm_castps_si128(vecs[0]), 0xAA));
983
1869
  vecs[3] = _mm_castsi128_ps(_mm_shuffle_epi32(_mm_castps_si128(vecs[0]), 0xFF));
984
1870
  vecs[0] = _mm_castsi128_ps(_mm_shuffle_epi32(_mm_castps_si128(vecs[0]), 0x00));
985
1871
  }
986
1872
 
987
- template<> EIGEN_STRONG_INLINE float predux<Packet4f>(const Packet4f& a)
988
- {
989
- // Disable SSE3 _mm_hadd_pd that is extremely slow on all existing Intel's architectures
990
- // (from Nehalem to Haswell)
991
- // #ifdef EIGEN_VECTORIZE_SSE3
992
- // Packet4f tmp = _mm_add_ps(a, vec4f_swizzle1(a,2,3,2,3));
993
- // return pfirst<Packet4f>(_mm_hadd_ps(tmp, tmp));
994
- // #else
995
- Packet4f tmp = _mm_add_ps(a, _mm_movehl_ps(a,a));
996
- return pfirst<Packet4f>(_mm_add_ss(tmp, _mm_shuffle_ps(tmp,tmp, 1)));
997
- // #endif
998
- }
999
-
1000
- template<> EIGEN_STRONG_INLINE double predux<Packet2d>(const Packet2d& a)
1001
- {
1002
- // Disable SSE3 _mm_hadd_pd that is extremely slow on all existing Intel's architectures
1003
- // (from Nehalem to Haswell)
1004
- // #ifdef EIGEN_VECTORIZE_SSE3
1005
- // return pfirst<Packet2d>(_mm_hadd_pd(a, a));
1006
- // #else
1007
- return pfirst<Packet2d>(_mm_add_sd(a, _mm_unpackhi_pd(a,a)));
1008
- // #endif
1009
- }
1010
-
1011
- #ifdef EIGEN_VECTORIZE_SSSE3
1012
- template<> EIGEN_STRONG_INLINE int predux<Packet4i>(const Packet4i& a)
1013
- {
1014
- Packet4i tmp0 = _mm_hadd_epi32(a,a);
1015
- return pfirst<Packet4i>(_mm_hadd_epi32(tmp0,tmp0));
1016
- }
1017
-
1018
- #else
1019
- template<> EIGEN_STRONG_INLINE int predux<Packet4i>(const Packet4i& a)
1020
- {
1021
- Packet4i tmp = _mm_add_epi32(a, _mm_unpackhi_epi64(a,a));
1022
- return pfirst(tmp) + pfirst<Packet4i>(_mm_shuffle_epi32(tmp, 1));
1023
- }
1024
- #endif
1025
-
1026
- template<> EIGEN_STRONG_INLINE bool predux<Packet16b>(const Packet16b& a) {
1027
- Packet4i tmp = _mm_or_si128(a, _mm_unpackhi_epi64(a,a));
1028
- return (pfirst(tmp) != 0) || (pfirst<Packet4i>(_mm_shuffle_epi32(tmp, 1)) != 0);
1029
- }
1030
-
1031
- // Other reduction functions:
1032
-
1033
-
1034
- // mul
1035
- template<> EIGEN_STRONG_INLINE float predux_mul<Packet4f>(const Packet4f& a)
1036
- {
1037
- Packet4f tmp = _mm_mul_ps(a, _mm_movehl_ps(a,a));
1038
- return pfirst<Packet4f>(_mm_mul_ss(tmp, _mm_shuffle_ps(tmp,tmp, 1)));
1039
- }
1040
- template<> EIGEN_STRONG_INLINE double predux_mul<Packet2d>(const Packet2d& a)
1041
- {
1042
- return pfirst<Packet2d>(_mm_mul_sd(a, _mm_unpackhi_pd(a,a)));
1043
- }
1044
- template<> EIGEN_STRONG_INLINE int predux_mul<Packet4i>(const Packet4i& a)
1045
- {
1046
- // after some experiments, it is seems this is the fastest way to implement it
1047
- // for GCC (eg., reusing pmul is very slow !)
1048
- // TODO try to call _mm_mul_epu32 directly
1049
- EIGEN_ALIGN16 int aux[4];
1050
- pstore(aux, a);
1051
- return (aux[0] * aux[1]) * (aux[2] * aux[3]);
1052
- }
1053
-
1054
- template<> EIGEN_STRONG_INLINE bool predux_mul<Packet16b>(const Packet16b& a) {
1055
- Packet4i tmp = _mm_and_si128(a, _mm_unpackhi_epi64(a,a));
1056
- return ((pfirst<Packet4i>(tmp) == 0x01010101) &&
1057
- (pfirst<Packet4i>(_mm_shuffle_epi32(tmp, 1)) == 0x01010101));
1058
- }
1059
-
1060
- // min
1061
- template<> EIGEN_STRONG_INLINE float predux_min<Packet4f>(const Packet4f& a)
1062
- {
1063
- Packet4f tmp = _mm_min_ps(a, _mm_movehl_ps(a,a));
1064
- return pfirst<Packet4f>(_mm_min_ss(tmp, _mm_shuffle_ps(tmp,tmp, 1)));
1065
- }
1066
- template<> EIGEN_STRONG_INLINE double predux_min<Packet2d>(const Packet2d& a)
1067
- {
1068
- return pfirst<Packet2d>(_mm_min_sd(a, _mm_unpackhi_pd(a,a)));
1069
- }
1070
- template<> EIGEN_STRONG_INLINE int predux_min<Packet4i>(const Packet4i& a)
1071
- {
1072
- #ifdef EIGEN_VECTORIZE_SSE4_1
1073
- Packet4i tmp = _mm_min_epi32(a, _mm_shuffle_epi32(a, _MM_SHUFFLE(0,0,3,2)));
1074
- return pfirst<Packet4i>(_mm_min_epi32(tmp,_mm_shuffle_epi32(tmp, 1)));
1075
- #else
1076
- // after some experiments, it is seems this is the fastest way to implement it
1077
- // for GCC (eg., it does not like using std::min after the pstore !!)
1078
- EIGEN_ALIGN16 int aux[4];
1079
- pstore(aux, a);
1080
- int aux0 = aux[0]<aux[1] ? aux[0] : aux[1];
1081
- int aux2 = aux[2]<aux[3] ? aux[2] : aux[3];
1082
- return aux0<aux2 ? aux0 : aux2;
1083
- #endif // EIGEN_VECTORIZE_SSE4_1
1084
- }
1085
-
1086
- // max
1087
- template<> EIGEN_STRONG_INLINE float predux_max<Packet4f>(const Packet4f& a)
1088
- {
1089
- Packet4f tmp = _mm_max_ps(a, _mm_movehl_ps(a,a));
1090
- return pfirst<Packet4f>(_mm_max_ss(tmp, _mm_shuffle_ps(tmp,tmp, 1)));
1091
- }
1092
- template<> EIGEN_STRONG_INLINE double predux_max<Packet2d>(const Packet2d& a)
1093
- {
1094
- return pfirst<Packet2d>(_mm_max_sd(a, _mm_unpackhi_pd(a,a)));
1095
- }
1096
- template<> EIGEN_STRONG_INLINE int predux_max<Packet4i>(const Packet4i& a)
1097
- {
1098
- #ifdef EIGEN_VECTORIZE_SSE4_1
1099
- Packet4i tmp = _mm_max_epi32(a, _mm_shuffle_epi32(a, _MM_SHUFFLE(0,0,3,2)));
1100
- return pfirst<Packet4i>(_mm_max_epi32(tmp,_mm_shuffle_epi32(tmp, 1)));
1101
- #else
1102
- // after some experiments, it is seems this is the fastest way to implement it
1103
- // for GCC (eg., it does not like using std::min after the pstore !!)
1104
- EIGEN_ALIGN16 int aux[4];
1105
- pstore(aux, a);
1106
- int aux0 = aux[0]>aux[1] ? aux[0] : aux[1];
1107
- int aux2 = aux[2]>aux[3] ? aux[2] : aux[3];
1108
- return aux0>aux2 ? aux0 : aux2;
1109
- #endif // EIGEN_VECTORIZE_SSE4_1
1110
- }
1111
-
1112
- // not needed yet
1113
- // template<> EIGEN_STRONG_INLINE bool predux_all(const Packet4f& x)
1114
- // {
1115
- // return _mm_movemask_ps(x) == 0xF;
1116
- // }
1117
-
1118
- template<> EIGEN_STRONG_INLINE bool predux_any(const Packet4f& x)
1119
- {
1120
- return _mm_movemask_ps(x) != 0x0;
1121
- }
1122
-
1123
- EIGEN_DEVICE_FUNC inline void
1124
- ptranspose(PacketBlock<Packet4f,4>& kernel) {
1873
+ EIGEN_STRONG_INLINE void ptranspose(PacketBlock<Packet4f, 4>& kernel) {
1125
1874
  _MM_TRANSPOSE4_PS(kernel.packet[0], kernel.packet[1], kernel.packet[2], kernel.packet[3]);
1126
1875
  }
1127
1876
 
1128
- EIGEN_DEVICE_FUNC inline void
1129
- ptranspose(PacketBlock<Packet2d,2>& kernel) {
1877
+ EIGEN_STRONG_INLINE void ptranspose(PacketBlock<Packet2d, 2>& kernel) {
1130
1878
  __m128d tmp = _mm_unpackhi_pd(kernel.packet[0], kernel.packet[1]);
1131
1879
  kernel.packet[0] = _mm_unpacklo_pd(kernel.packet[0], kernel.packet[1]);
1132
1880
  kernel.packet[1] = tmp;
1133
1881
  }
1134
1882
 
1135
- EIGEN_DEVICE_FUNC inline void
1136
- ptranspose(PacketBlock<Packet4i,4>& kernel) {
1883
+ EIGEN_STRONG_INLINE void ptranspose(PacketBlock<Packet2l, 2>& kernel) {
1884
+ __m128i tmp = _mm_unpackhi_epi64(kernel.packet[0], kernel.packet[1]);
1885
+ kernel.packet[0] = _mm_unpacklo_epi64(kernel.packet[0], kernel.packet[1]);
1886
+ kernel.packet[1] = tmp;
1887
+ }
1888
+
1889
+ EIGEN_STRONG_INLINE void ptranspose(PacketBlock<Packet4i, 4>& kernel) {
1137
1890
  __m128i T0 = _mm_unpacklo_epi32(kernel.packet[0], kernel.packet[1]);
1138
1891
  __m128i T1 = _mm_unpacklo_epi32(kernel.packet[2], kernel.packet[3]);
1139
1892
  __m128i T2 = _mm_unpackhi_epi32(kernel.packet[0], kernel.packet[1]);
@@ -1144,21 +1897,22 @@ ptranspose(PacketBlock<Packet4i,4>& kernel) {
1144
1897
  kernel.packet[2] = _mm_unpacklo_epi64(T2, T3);
1145
1898
  kernel.packet[3] = _mm_unpackhi_epi64(T2, T3);
1146
1899
  }
1900
+ EIGEN_STRONG_INLINE void ptranspose(PacketBlock<Packet4ui, 4>& kernel) {
1901
+ ptranspose((PacketBlock<Packet4i, 4>&)kernel);
1902
+ }
1147
1903
 
1148
- EIGEN_DEVICE_FUNC inline void
1149
- ptranspose(PacketBlock<Packet16b,4>& kernel) {
1150
- __m128i T0 = _mm_unpacklo_epi8(kernel.packet[0], kernel.packet[1]);
1151
- __m128i T1 = _mm_unpackhi_epi8(kernel.packet[0], kernel.packet[1]);
1152
- __m128i T2 = _mm_unpacklo_epi8(kernel.packet[2], kernel.packet[3]);
1153
- __m128i T3 = _mm_unpackhi_epi8(kernel.packet[2], kernel.packet[3]);
1904
+ EIGEN_STRONG_INLINE void ptranspose(PacketBlock<Packet16b, 4>& kernel) {
1905
+ __m128i T0 = _mm_unpacklo_epi8(kernel.packet[0], kernel.packet[1]);
1906
+ __m128i T1 = _mm_unpackhi_epi8(kernel.packet[0], kernel.packet[1]);
1907
+ __m128i T2 = _mm_unpacklo_epi8(kernel.packet[2], kernel.packet[3]);
1908
+ __m128i T3 = _mm_unpackhi_epi8(kernel.packet[2], kernel.packet[3]);
1154
1909
  kernel.packet[0] = _mm_unpacklo_epi16(T0, T2);
1155
1910
  kernel.packet[1] = _mm_unpackhi_epi16(T0, T2);
1156
1911
  kernel.packet[2] = _mm_unpacklo_epi16(T1, T3);
1157
1912
  kernel.packet[3] = _mm_unpackhi_epi16(T1, T3);
1158
1913
  }
1159
1914
 
1160
- EIGEN_DEVICE_FUNC inline void
1161
- ptranspose(PacketBlock<Packet16b,16>& kernel) {
1915
+ EIGEN_STRONG_INLINE void ptranspose(PacketBlock<Packet16b, 16>& kernel) {
1162
1916
  // If we number the elements in the input thus:
1163
1917
  // kernel.packet[ 0] = {00, 01, 02, 03, 04, 05, 06, 07, 08, 09, 0a, 0b, 0c, 0d, 0e, 0f}
1164
1918
  // kernel.packet[ 1] = {10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 1a, 1b, 1c, 1d, 1e, 1f}
@@ -1170,67 +1924,72 @@ ptranspose(PacketBlock<Packet16b,16>& kernel) {
1170
1924
  // kernel.packet[ 1] = {01, 11, 21, 31, 41, 51, 61, 71, 81, 91, a1, b1, c1, d1, e1, f1}
1171
1925
  // ...
1172
1926
  // kernel.packet[15] = {0f, 1f, 2f, 3f, 4f, 5f, 6f, 7f, 8f, 9f, af, bf, cf, df, ef, ff},
1173
- __m128i t0 = _mm_unpacklo_epi8(kernel.packet[0], kernel.packet[1]); // 00 10 01 11 02 12 03 13 04 14 05 15 06 16 07 17
1174
- __m128i t1 = _mm_unpackhi_epi8(kernel.packet[0], kernel.packet[1]); // 08 18 09 19 0a 1a 0b 1b 0c 1c 0d 1d 0e 1e 0f 1f
1175
- __m128i t2 = _mm_unpacklo_epi8(kernel.packet[2], kernel.packet[3]); // 20 30 21 31 22 32 ... 27 37
1176
- __m128i t3 = _mm_unpackhi_epi8(kernel.packet[2], kernel.packet[3]); // 28 38 29 39 2a 3a ... 2f 3f
1177
- __m128i t4 = _mm_unpacklo_epi8(kernel.packet[4], kernel.packet[5]); // 40 50 41 51 42 52 47 57
1178
- __m128i t5 = _mm_unpackhi_epi8(kernel.packet[4], kernel.packet[5]); // 48 58 49 59 4a 5a
1179
- __m128i t6 = _mm_unpacklo_epi8(kernel.packet[6], kernel.packet[7]);
1180
- __m128i t7 = _mm_unpackhi_epi8(kernel.packet[6], kernel.packet[7]);
1181
- __m128i t8 = _mm_unpacklo_epi8(kernel.packet[8], kernel.packet[9]);
1182
- __m128i t9 = _mm_unpackhi_epi8(kernel.packet[8], kernel.packet[9]);
1183
- __m128i ta = _mm_unpacklo_epi8(kernel.packet[10], kernel.packet[11]);
1184
- __m128i tb = _mm_unpackhi_epi8(kernel.packet[10], kernel.packet[11]);
1185
- __m128i tc = _mm_unpacklo_epi8(kernel.packet[12], kernel.packet[13]);
1186
- __m128i td = _mm_unpackhi_epi8(kernel.packet[12], kernel.packet[13]);
1187
- __m128i te = _mm_unpacklo_epi8(kernel.packet[14], kernel.packet[15]);
1188
- __m128i tf = _mm_unpackhi_epi8(kernel.packet[14], kernel.packet[15]);
1189
-
1190
- __m128i s0 = _mm_unpacklo_epi16(t0, t2); // 00 10 20 30 01 11 21 31 02 12 22 32 03 13 23 33
1191
- __m128i s1 = _mm_unpackhi_epi16(t0, t2); // 04 14 24 34
1192
- __m128i s2 = _mm_unpacklo_epi16(t1, t3); // 08 18 28 38 ...
1193
- __m128i s3 = _mm_unpackhi_epi16(t1, t3); // 0c 1c 2c 3c ...
1194
- __m128i s4 = _mm_unpacklo_epi16(t4, t6); // 40 50 60 70 41 51 61 71 42 52 62 72 43 53 63 73
1195
- __m128i s5 = _mm_unpackhi_epi16(t4, t6); // 44 54 64 74 ...
1196
- __m128i s6 = _mm_unpacklo_epi16(t5, t7);
1197
- __m128i s7 = _mm_unpackhi_epi16(t5, t7);
1198
- __m128i s8 = _mm_unpacklo_epi16(t8, ta);
1199
- __m128i s9 = _mm_unpackhi_epi16(t8, ta);
1200
- __m128i sa = _mm_unpacklo_epi16(t9, tb);
1201
- __m128i sb = _mm_unpackhi_epi16(t9, tb);
1202
- __m128i sc = _mm_unpacklo_epi16(tc, te);
1203
- __m128i sd = _mm_unpackhi_epi16(tc, te);
1204
- __m128i se = _mm_unpacklo_epi16(td, tf);
1205
- __m128i sf = _mm_unpackhi_epi16(td, tf);
1206
-
1207
- __m128i u0 = _mm_unpacklo_epi32(s0, s4); // 00 10 20 30 40 50 60 70 01 11 21 31 41 51 61 71
1208
- __m128i u1 = _mm_unpackhi_epi32(s0, s4); // 02 12 22 32 42 52 62 72 03 13 23 33 43 53 63 73
1209
- __m128i u2 = _mm_unpacklo_epi32(s1, s5);
1210
- __m128i u3 = _mm_unpackhi_epi32(s1, s5);
1211
- __m128i u4 = _mm_unpacklo_epi32(s2, s6);
1212
- __m128i u5 = _mm_unpackhi_epi32(s2, s6);
1213
- __m128i u6 = _mm_unpacklo_epi32(s3, s7);
1214
- __m128i u7 = _mm_unpackhi_epi32(s3, s7);
1215
- __m128i u8 = _mm_unpacklo_epi32(s8, sc);
1216
- __m128i u9 = _mm_unpackhi_epi32(s8, sc);
1217
- __m128i ua = _mm_unpacklo_epi32(s9, sd);
1218
- __m128i ub = _mm_unpackhi_epi32(s9, sd);
1219
- __m128i uc = _mm_unpacklo_epi32(sa, se);
1220
- __m128i ud = _mm_unpackhi_epi32(sa, se);
1221
- __m128i ue = _mm_unpacklo_epi32(sb, sf);
1222
- __m128i uf = _mm_unpackhi_epi32(sb, sf);
1223
-
1224
- kernel.packet[0] = _mm_unpacklo_epi64(u0, u8);
1225
- kernel.packet[1] = _mm_unpackhi_epi64(u0, u8);
1226
- kernel.packet[2] = _mm_unpacklo_epi64(u1, u9);
1227
- kernel.packet[3] = _mm_unpackhi_epi64(u1, u9);
1228
- kernel.packet[4] = _mm_unpacklo_epi64(u2, ua);
1229
- kernel.packet[5] = _mm_unpackhi_epi64(u2, ua);
1230
- kernel.packet[6] = _mm_unpacklo_epi64(u3, ub);
1231
- kernel.packet[7] = _mm_unpackhi_epi64(u3, ub);
1232
- kernel.packet[8] = _mm_unpacklo_epi64(u4, uc);
1233
- kernel.packet[9] = _mm_unpackhi_epi64(u4, uc);
1927
+ __m128i t0 =
1928
+ _mm_unpacklo_epi8(kernel.packet[0], kernel.packet[1]); // 00 10 01 11 02 12 03 13 04 14 05 15 06 16 07 17
1929
+ __m128i t1 =
1930
+ _mm_unpackhi_epi8(kernel.packet[0], kernel.packet[1]); // 08 18 09 19 0a 1a 0b 1b 0c 1c 0d 1d 0e 1e 0f 1f
1931
+ __m128i t2 =
1932
+ _mm_unpacklo_epi8(kernel.packet[2], kernel.packet[3]); // 20 30 21 31 22 32 ... 27 37
1933
+ __m128i t3 =
1934
+ _mm_unpackhi_epi8(kernel.packet[2], kernel.packet[3]); // 28 38 29 39 2a 3a ... 2f 3f
1935
+ __m128i t4 =
1936
+ _mm_unpacklo_epi8(kernel.packet[4], kernel.packet[5]); // 40 50 41 51 42 52 47 57
1937
+ __m128i t5 = _mm_unpackhi_epi8(kernel.packet[4], kernel.packet[5]); // 48 58 49 59 4a 5a
1938
+ __m128i t6 = _mm_unpacklo_epi8(kernel.packet[6], kernel.packet[7]);
1939
+ __m128i t7 = _mm_unpackhi_epi8(kernel.packet[6], kernel.packet[7]);
1940
+ __m128i t8 = _mm_unpacklo_epi8(kernel.packet[8], kernel.packet[9]);
1941
+ __m128i t9 = _mm_unpackhi_epi8(kernel.packet[8], kernel.packet[9]);
1942
+ __m128i ta = _mm_unpacklo_epi8(kernel.packet[10], kernel.packet[11]);
1943
+ __m128i tb = _mm_unpackhi_epi8(kernel.packet[10], kernel.packet[11]);
1944
+ __m128i tc = _mm_unpacklo_epi8(kernel.packet[12], kernel.packet[13]);
1945
+ __m128i td = _mm_unpackhi_epi8(kernel.packet[12], kernel.packet[13]);
1946
+ __m128i te = _mm_unpacklo_epi8(kernel.packet[14], kernel.packet[15]);
1947
+ __m128i tf = _mm_unpackhi_epi8(kernel.packet[14], kernel.packet[15]);
1948
+
1949
+ __m128i s0 = _mm_unpacklo_epi16(t0, t2); // 00 10 20 30 01 11 21 31 02 12 22 32 03 13 23 33
1950
+ __m128i s1 = _mm_unpackhi_epi16(t0, t2); // 04 14 24 34
1951
+ __m128i s2 = _mm_unpacklo_epi16(t1, t3); // 08 18 28 38 ...
1952
+ __m128i s3 = _mm_unpackhi_epi16(t1, t3); // 0c 1c 2c 3c ...
1953
+ __m128i s4 = _mm_unpacklo_epi16(t4, t6); // 40 50 60 70 41 51 61 71 42 52 62 72 43 53 63 73
1954
+ __m128i s5 = _mm_unpackhi_epi16(t4, t6); // 44 54 64 74 ...
1955
+ __m128i s6 = _mm_unpacklo_epi16(t5, t7);
1956
+ __m128i s7 = _mm_unpackhi_epi16(t5, t7);
1957
+ __m128i s8 = _mm_unpacklo_epi16(t8, ta);
1958
+ __m128i s9 = _mm_unpackhi_epi16(t8, ta);
1959
+ __m128i sa = _mm_unpacklo_epi16(t9, tb);
1960
+ __m128i sb = _mm_unpackhi_epi16(t9, tb);
1961
+ __m128i sc = _mm_unpacklo_epi16(tc, te);
1962
+ __m128i sd = _mm_unpackhi_epi16(tc, te);
1963
+ __m128i se = _mm_unpacklo_epi16(td, tf);
1964
+ __m128i sf = _mm_unpackhi_epi16(td, tf);
1965
+
1966
+ __m128i u0 = _mm_unpacklo_epi32(s0, s4); // 00 10 20 30 40 50 60 70 01 11 21 31 41 51 61 71
1967
+ __m128i u1 = _mm_unpackhi_epi32(s0, s4); // 02 12 22 32 42 52 62 72 03 13 23 33 43 53 63 73
1968
+ __m128i u2 = _mm_unpacklo_epi32(s1, s5);
1969
+ __m128i u3 = _mm_unpackhi_epi32(s1, s5);
1970
+ __m128i u4 = _mm_unpacklo_epi32(s2, s6);
1971
+ __m128i u5 = _mm_unpackhi_epi32(s2, s6);
1972
+ __m128i u6 = _mm_unpacklo_epi32(s3, s7);
1973
+ __m128i u7 = _mm_unpackhi_epi32(s3, s7);
1974
+ __m128i u8 = _mm_unpacklo_epi32(s8, sc);
1975
+ __m128i u9 = _mm_unpackhi_epi32(s8, sc);
1976
+ __m128i ua = _mm_unpacklo_epi32(s9, sd);
1977
+ __m128i ub = _mm_unpackhi_epi32(s9, sd);
1978
+ __m128i uc = _mm_unpacklo_epi32(sa, se);
1979
+ __m128i ud = _mm_unpackhi_epi32(sa, se);
1980
+ __m128i ue = _mm_unpacklo_epi32(sb, sf);
1981
+ __m128i uf = _mm_unpackhi_epi32(sb, sf);
1982
+
1983
+ kernel.packet[0] = _mm_unpacklo_epi64(u0, u8);
1984
+ kernel.packet[1] = _mm_unpackhi_epi64(u0, u8);
1985
+ kernel.packet[2] = _mm_unpacklo_epi64(u1, u9);
1986
+ kernel.packet[3] = _mm_unpackhi_epi64(u1, u9);
1987
+ kernel.packet[4] = _mm_unpacklo_epi64(u2, ua);
1988
+ kernel.packet[5] = _mm_unpackhi_epi64(u2, ua);
1989
+ kernel.packet[6] = _mm_unpacklo_epi64(u3, ub);
1990
+ kernel.packet[7] = _mm_unpackhi_epi64(u3, ub);
1991
+ kernel.packet[8] = _mm_unpacklo_epi64(u4, uc);
1992
+ kernel.packet[9] = _mm_unpackhi_epi64(u4, uc);
1234
1993
  kernel.packet[10] = _mm_unpacklo_epi64(u5, ud);
1235
1994
  kernel.packet[11] = _mm_unpackhi_epi64(u5, ud);
1236
1995
  kernel.packet[12] = _mm_unpacklo_epi64(u6, ue);
@@ -1239,51 +1998,178 @@ ptranspose(PacketBlock<Packet16b,16>& kernel) {
1239
1998
  kernel.packet[15] = _mm_unpackhi_epi64(u7, uf);
1240
1999
  }
1241
2000
 
1242
- template<> EIGEN_STRONG_INLINE Packet4i pblend(const Selector<4>& ifPacket, const Packet4i& thenPacket, const Packet4i& elsePacket) {
1243
- const __m128i zero = _mm_setzero_si128();
1244
- const __m128i select = _mm_set_epi32(ifPacket.select[3], ifPacket.select[2], ifPacket.select[1], ifPacket.select[0]);
1245
- __m128i false_mask = _mm_cmpeq_epi32(select, zero);
1246
- #ifdef EIGEN_VECTORIZE_SSE4_1
1247
- return _mm_blendv_epi8(thenPacket, elsePacket, false_mask);
1248
- #else
1249
- return _mm_or_si128(_mm_andnot_si128(false_mask, thenPacket), _mm_and_si128(false_mask, elsePacket));
1250
- #endif
2001
+ EIGEN_STRONG_INLINE __m128i sse_blend_mask(const Selector<2>& ifPacket) {
2002
+ return _mm_set_epi64x(0 - ifPacket.select[1], 0 - ifPacket.select[0]);
1251
2003
  }
1252
- template<> EIGEN_STRONG_INLINE Packet4f pblend(const Selector<4>& ifPacket, const Packet4f& thenPacket, const Packet4f& elsePacket) {
1253
- const __m128 zero = _mm_setzero_ps();
1254
- const __m128 select = _mm_set_ps(ifPacket.select[3], ifPacket.select[2], ifPacket.select[1], ifPacket.select[0]);
1255
- __m128 false_mask = _mm_cmpeq_ps(select, zero);
1256
- #ifdef EIGEN_VECTORIZE_SSE4_1
1257
- return _mm_blendv_ps(thenPacket, elsePacket, false_mask);
1258
- #else
1259
- return _mm_or_ps(_mm_andnot_ps(false_mask, thenPacket), _mm_and_ps(false_mask, elsePacket));
1260
- #endif
2004
+
2005
+ EIGEN_STRONG_INLINE __m128i sse_blend_mask(const Selector<4>& ifPacket) {
2006
+ return _mm_set_epi32(0 - ifPacket.select[3], 0 - ifPacket.select[2], 0 - ifPacket.select[1], 0 - ifPacket.select[0]);
1261
2007
  }
1262
- template<> EIGEN_STRONG_INLINE Packet2d pblend(const Selector<2>& ifPacket, const Packet2d& thenPacket, const Packet2d& elsePacket) {
1263
- const __m128d zero = _mm_setzero_pd();
1264
- const __m128d select = _mm_set_pd(ifPacket.select[1], ifPacket.select[0]);
1265
- __m128d false_mask = _mm_cmpeq_pd(select, zero);
1266
- #ifdef EIGEN_VECTORIZE_SSE4_1
1267
- return _mm_blendv_pd(thenPacket, elsePacket, false_mask);
1268
- #else
1269
- return _mm_or_pd(_mm_andnot_pd(false_mask, thenPacket), _mm_and_pd(false_mask, elsePacket));
1270
- #endif
2008
+
2009
+ template <>
2010
+ EIGEN_STRONG_INLINE Packet2l pblend(const Selector<2>& ifPacket, const Packet2l& thenPacket,
2011
+ const Packet2l& elsePacket) {
2012
+ const __m128i true_mask = sse_blend_mask(ifPacket);
2013
+ return pselect<Packet2l>(true_mask, thenPacket, elsePacket);
2014
+ }
2015
+ template <>
2016
+ EIGEN_STRONG_INLINE Packet4i pblend(const Selector<4>& ifPacket, const Packet4i& thenPacket,
2017
+ const Packet4i& elsePacket) {
2018
+ const __m128i true_mask = sse_blend_mask(ifPacket);
2019
+ return pselect<Packet4i>(true_mask, thenPacket, elsePacket);
2020
+ }
2021
+ template <>
2022
+ EIGEN_STRONG_INLINE Packet4ui pblend(const Selector<4>& ifPacket, const Packet4ui& thenPacket,
2023
+ const Packet4ui& elsePacket) {
2024
+ return (Packet4ui)pblend(ifPacket, (Packet4i)thenPacket, (Packet4i)elsePacket);
2025
+ }
2026
+ template <>
2027
+ EIGEN_STRONG_INLINE Packet4f pblend(const Selector<4>& ifPacket, const Packet4f& thenPacket,
2028
+ const Packet4f& elsePacket) {
2029
+ const __m128i true_mask = sse_blend_mask(ifPacket);
2030
+ return pselect<Packet4f>(_mm_castsi128_ps(true_mask), thenPacket, elsePacket);
2031
+ }
2032
+ template <>
2033
+ EIGEN_STRONG_INLINE Packet2d pblend(const Selector<2>& ifPacket, const Packet2d& thenPacket,
2034
+ const Packet2d& elsePacket) {
2035
+ const __m128i true_mask = sse_blend_mask(ifPacket);
2036
+ return pselect<Packet2d>(_mm_castsi128_pd(true_mask), thenPacket, elsePacket);
1271
2037
  }
1272
2038
 
1273
2039
  // Scalar path for pmadd with FMA to ensure consistency with vectorized path.
1274
- #ifdef EIGEN_VECTORIZE_FMA
1275
- template<> EIGEN_STRONG_INLINE float pmadd(const float& a, const float& b, const float& c) {
1276
- return ::fmaf(a,b,c);
2040
+ #if defined(EIGEN_VECTORIZE_FMA)
2041
+ template <>
2042
+ EIGEN_STRONG_INLINE float pmadd(const float& a, const float& b, const float& c) {
2043
+ return std::fmaf(a, b, c);
2044
+ }
2045
+ template <>
2046
+ EIGEN_STRONG_INLINE double pmadd(const double& a, const double& b, const double& c) {
2047
+ return std::fma(a, b, c);
2048
+ }
2049
+ template <>
2050
+ EIGEN_STRONG_INLINE float pmsub(const float& a, const float& b, const float& c) {
2051
+ return std::fmaf(a, b, -c);
2052
+ }
2053
+ template <>
2054
+ EIGEN_STRONG_INLINE double pmsub(const double& a, const double& b, const double& c) {
2055
+ return std::fma(a, b, -c);
2056
+ }
2057
+ template <>
2058
+ EIGEN_STRONG_INLINE float pnmadd(const float& a, const float& b, const float& c) {
2059
+ return std::fmaf(-a, b, c);
2060
+ }
2061
+ template <>
2062
+ EIGEN_STRONG_INLINE double pnmadd(const double& a, const double& b, const double& c) {
2063
+ return std::fma(-a, b, c);
2064
+ }
2065
+ template <>
2066
+ EIGEN_STRONG_INLINE float pnmsub(const float& a, const float& b, const float& c) {
2067
+ return std::fmaf(-a, b, -c);
1277
2068
  }
1278
- template<> EIGEN_STRONG_INLINE double pmadd(const double& a, const double& b, const double& c) {
1279
- return ::fma(a,b,c);
2069
+ template <>
2070
+ EIGEN_STRONG_INLINE double pnmsub(const double& a, const double& b, const double& c) {
2071
+ return std::fma(-a, b, -c);
1280
2072
  }
1281
2073
  #endif
1282
2074
 
2075
+ #ifdef EIGEN_VECTORIZE_SSE4_1
2076
+ // Helpers for half->float and float->half conversions.
2077
+ // Currently only used by the AVX code.
2078
+ EIGEN_STRONG_INLINE __m128i half2floatsse(__m128i h) {
2079
+ __m128i input = _mm_cvtepu16_epi32(h);
2080
+
2081
+ // Direct vectorization of half_to_float, C parts in the comments.
2082
+ __m128i shifted_exp = _mm_set1_epi32(0x7c00 << 13);
2083
+ // o.u = (h.x & 0x7fff) << 13; // exponent/mantissa bits
2084
+ __m128i ou = _mm_slli_epi32(_mm_and_si128(input, _mm_set1_epi32(0x7fff)), 13);
2085
+ // exp = shifted_exp & o.u; // just the exponent
2086
+ __m128i exp = _mm_and_si128(ou, shifted_exp);
2087
+ // o.u += (127 - 15) << 23;
2088
+ ou = _mm_add_epi32(ou, _mm_set1_epi32((127 - 15) << 23));
2089
+
2090
+ // Inf/NaN?
2091
+ __m128i naninf_mask = _mm_cmpeq_epi32(exp, shifted_exp);
2092
+ // Inf/NaN adjust
2093
+ __m128i naninf_adj = _mm_and_si128(_mm_set1_epi32((128 - 16) << 23), naninf_mask);
2094
+ // extra exp adjust for Inf/NaN
2095
+ ou = _mm_add_epi32(ou, naninf_adj);
2096
+
2097
+ // Zero/Denormal?
2098
+ __m128i zeroden_mask = _mm_cmpeq_epi32(exp, _mm_setzero_si128());
2099
+ __m128i zeroden_adj = _mm_and_si128(zeroden_mask, _mm_set1_epi32(1 << 23));
2100
+ // o.u += 1 << 23;
2101
+ ou = _mm_add_epi32(ou, zeroden_adj);
2102
+ // magic.u = 113 << 23
2103
+ __m128i magic = _mm_and_si128(zeroden_mask, _mm_set1_epi32(113 << 23));
2104
+ // o.f -= magic.f
2105
+ ou = _mm_castps_si128(_mm_sub_ps(_mm_castsi128_ps(ou), _mm_castsi128_ps(magic)));
2106
+
2107
+ __m128i sign = _mm_slli_epi32(_mm_and_si128(input, _mm_set1_epi32(0x8000)), 16);
2108
+ // o.u |= (h.x & 0x8000) << 16; // sign bit
2109
+ ou = _mm_or_si128(ou, sign);
2110
+ // return o.f;
2111
+ // We are actually returning uint version, to make
2112
+ // _mm256_insertf128_si256 work.
2113
+ return ou;
2114
+ }
2115
+
2116
+ EIGEN_STRONG_INLINE __m128i float2half(__m128 f) {
2117
+ // unsigned int sign_mask = 0x80000000u;
2118
+ __m128i sign = _mm_set1_epi32(0x80000000u);
2119
+ // unsigned int sign = f.u & sign_mask;
2120
+ sign = _mm_and_si128(sign, _mm_castps_si128(f));
2121
+ // f.u ^= sign;
2122
+ f = _mm_xor_ps(f, _mm_castsi128_ps(sign));
2123
+
2124
+ __m128i fu = _mm_castps_si128(f);
2125
+
2126
+ __m128i f16max = _mm_set1_epi32((127 + 16) << 23);
2127
+ __m128i f32infty = _mm_set1_epi32(255 << 23);
2128
+ // if (f.u >= f16max.u) // result is Inf or NaN (all exponent bits set)
2129
+ // there is no _mm_cmpge_epi32, so use lt and swap operands
2130
+ __m128i infnan_mask = _mm_cmplt_epi32(f16max, _mm_castps_si128(f));
2131
+ __m128i inf_mask = _mm_cmpgt_epi32(_mm_castps_si128(f), f32infty);
2132
+ __m128i nan_mask = _mm_andnot_si128(inf_mask, infnan_mask);
2133
+ __m128i inf_value = _mm_and_si128(inf_mask, _mm_set1_epi32(0x7e00));
2134
+ __m128i nan_value = _mm_and_si128(nan_mask, _mm_set1_epi32(0x7c00));
2135
+ // o.x = (f.u > f32infty.u) ? 0x7e00 : 0x7c00; // NaN->qNaN and Inf->Inf
2136
+ __m128i naninf_value = _mm_or_si128(inf_value, nan_value);
2137
+
2138
+ __m128i denorm_magic = _mm_set1_epi32(((127 - 15) + (23 - 10) + 1) << 23);
2139
+ __m128i subnorm_mask = _mm_cmplt_epi32(_mm_castps_si128(f), _mm_set1_epi32(113 << 23));
2140
+ // f.f += denorm_magic.f;
2141
+ f = _mm_add_ps(f, _mm_castsi128_ps(denorm_magic));
2142
+ // f.u - denorm_magic.u
2143
+ __m128i o = _mm_sub_epi32(_mm_castps_si128(f), denorm_magic);
2144
+ o = _mm_and_si128(o, subnorm_mask);
2145
+ // Correct result for inf/nan/zero/subnormal, 0 otherwise
2146
+ o = _mm_or_si128(o, naninf_value);
2147
+
2148
+ __m128i mask = _mm_or_si128(infnan_mask, subnorm_mask);
2149
+ o = _mm_and_si128(o, mask);
2150
+
2151
+ // mant_odd = (f.u >> 13) & 1;
2152
+ __m128i mand_odd = _mm_and_si128(_mm_srli_epi32(fu, 13), _mm_set1_epi32(0x1));
2153
+ // f.u += 0xc8000fffU;
2154
+ fu = _mm_add_epi32(fu, _mm_set1_epi32(0xc8000fffU));
2155
+ // f.u += mant_odd;
2156
+ fu = _mm_add_epi32(fu, mand_odd);
2157
+ fu = _mm_andnot_si128(mask, fu);
2158
+ // f.u >> 13
2159
+ fu = _mm_srli_epi32(fu, 13);
2160
+ o = _mm_or_si128(fu, o);
2161
+
2162
+ // o.x |= static_cast<numext::uint16_t>(sign >> 16);
2163
+ o = _mm_or_si128(o, _mm_srli_epi32(sign, 16));
2164
+
2165
+ // 16 bit values
2166
+ return _mm_and_si128(o, _mm_set1_epi32(0xffff));
2167
+ }
2168
+ #endif
1283
2169
 
1284
2170
  // Packet math for Eigen::half
1285
2171
  // Disable the following code since it's broken on too many platforms / compilers.
1286
- //#elif defined(EIGEN_VECTORIZE_SSE) && (!EIGEN_ARCH_x86_64) && (!EIGEN_COMP_MSVC)
2172
+ // #elif defined(EIGEN_VECTORIZE_SSE) && (!EIGEN_ARCH_x86_64) && (!EIGEN_COMP_MSVC)
1287
2173
  #if 0
1288
2174
 
1289
2175
  typedef struct {
@@ -1302,7 +2188,6 @@ struct packet_traits<Eigen::half> : default_packet_traits {
1302
2188
  Vectorizable = 1,
1303
2189
  AlignedOnScalar = 1,
1304
2190
  size = 4,
1305
- HasHalfPacket = 0,
1306
2191
  HasAdd = 1,
1307
2192
  HasSub = 1,
1308
2193
  HasMul = 1,
@@ -1314,11 +2199,6 @@ struct packet_traits<Eigen::half> : default_packet_traits {
1314
2199
  HasMax = 0,
1315
2200
  HasConj = 0,
1316
2201
  HasSetLinear = 0,
1317
- HasSqrt = 0,
1318
- HasRsqrt = 0,
1319
- HasExp = 0,
1320
- HasLog = 0,
1321
- HasBlend = 0
1322
2202
  };
1323
2203
  };
1324
2204
 
@@ -1487,19 +2367,18 @@ ptranspose(PacketBlock<Packet4h,4>& kernel) {
1487
2367
 
1488
2368
  #endif
1489
2369
 
2370
+ } // end namespace internal
1490
2371
 
1491
- } // end namespace internal
1492
-
1493
- } // end namespace Eigen
2372
+ } // end namespace Eigen
1494
2373
 
1495
2374
  #if EIGEN_COMP_PGI && EIGEN_COMP_PGI < 1900
1496
2375
  // PGI++ does not define the following intrinsics in C++ mode.
1497
- static inline __m128 _mm_castpd_ps (__m128d x) { return reinterpret_cast<__m128&>(x); }
2376
+ static inline __m128 _mm_castpd_ps(__m128d x) { return reinterpret_cast<__m128&>(x); }
1498
2377
  static inline __m128i _mm_castpd_si128(__m128d x) { return reinterpret_cast<__m128i&>(x); }
1499
- static inline __m128d _mm_castps_pd (__m128 x) { return reinterpret_cast<__m128d&>(x); }
1500
- static inline __m128i _mm_castps_si128(__m128 x) { return reinterpret_cast<__m128i&>(x); }
1501
- static inline __m128 _mm_castsi128_ps(__m128i x) { return reinterpret_cast<__m128&>(x); }
2378
+ static inline __m128d _mm_castps_pd(__m128 x) { return reinterpret_cast<__m128d&>(x); }
2379
+ static inline __m128i _mm_castps_si128(__m128 x) { return reinterpret_cast<__m128i&>(x); }
2380
+ static inline __m128 _mm_castsi128_ps(__m128i x) { return reinterpret_cast<__m128&>(x); }
1502
2381
  static inline __m128d _mm_castsi128_pd(__m128i x) { return reinterpret_cast<__m128d&>(x); }
1503
2382
  #endif
1504
2383
 
1505
- #endif // EIGEN_PACKET_MATH_SSE_H
2384
+ #endif // EIGEN_PACKET_MATH_SSE_H