ring-native 0.0.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (261) hide show
  1. checksums.yaml +7 -0
  2. data/.gitignore +9 -0
  3. data/Gemfile +3 -0
  4. data/README.md +22 -0
  5. data/Rakefile +1 -0
  6. data/ext/ring/extconf.rb +29 -0
  7. data/lib/ring/native.rb +8 -0
  8. data/lib/ring/native/version.rb +5 -0
  9. data/ring-native.gemspec +25 -0
  10. data/vendor/ring/BUILDING.md +40 -0
  11. data/vendor/ring/Cargo.toml +43 -0
  12. data/vendor/ring/LICENSE +185 -0
  13. data/vendor/ring/Makefile +35 -0
  14. data/vendor/ring/PORTING.md +163 -0
  15. data/vendor/ring/README.md +113 -0
  16. data/vendor/ring/STYLE.md +197 -0
  17. data/vendor/ring/appveyor.yml +27 -0
  18. data/vendor/ring/build.rs +108 -0
  19. data/vendor/ring/crypto/aes/aes.c +1142 -0
  20. data/vendor/ring/crypto/aes/aes_test.Windows.vcxproj +25 -0
  21. data/vendor/ring/crypto/aes/aes_test.cc +93 -0
  22. data/vendor/ring/crypto/aes/asm/aes-586.pl +2368 -0
  23. data/vendor/ring/crypto/aes/asm/aes-armv4.pl +1249 -0
  24. data/vendor/ring/crypto/aes/asm/aes-x86_64.pl +2246 -0
  25. data/vendor/ring/crypto/aes/asm/aesni-x86.pl +1318 -0
  26. data/vendor/ring/crypto/aes/asm/aesni-x86_64.pl +2084 -0
  27. data/vendor/ring/crypto/aes/asm/aesv8-armx.pl +675 -0
  28. data/vendor/ring/crypto/aes/asm/bsaes-armv7.pl +1364 -0
  29. data/vendor/ring/crypto/aes/asm/bsaes-x86_64.pl +1565 -0
  30. data/vendor/ring/crypto/aes/asm/vpaes-x86.pl +841 -0
  31. data/vendor/ring/crypto/aes/asm/vpaes-x86_64.pl +1116 -0
  32. data/vendor/ring/crypto/aes/internal.h +87 -0
  33. data/vendor/ring/crypto/aes/mode_wrappers.c +61 -0
  34. data/vendor/ring/crypto/bn/add.c +394 -0
  35. data/vendor/ring/crypto/bn/asm/armv4-mont.pl +694 -0
  36. data/vendor/ring/crypto/bn/asm/armv8-mont.pl +1503 -0
  37. data/vendor/ring/crypto/bn/asm/bn-586.pl +774 -0
  38. data/vendor/ring/crypto/bn/asm/co-586.pl +287 -0
  39. data/vendor/ring/crypto/bn/asm/rsaz-avx2.pl +1882 -0
  40. data/vendor/ring/crypto/bn/asm/x86-mont.pl +592 -0
  41. data/vendor/ring/crypto/bn/asm/x86_64-gcc.c +599 -0
  42. data/vendor/ring/crypto/bn/asm/x86_64-mont.pl +1393 -0
  43. data/vendor/ring/crypto/bn/asm/x86_64-mont5.pl +3507 -0
  44. data/vendor/ring/crypto/bn/bn.c +352 -0
  45. data/vendor/ring/crypto/bn/bn_asn1.c +74 -0
  46. data/vendor/ring/crypto/bn/bn_test.Windows.vcxproj +25 -0
  47. data/vendor/ring/crypto/bn/bn_test.cc +1696 -0
  48. data/vendor/ring/crypto/bn/cmp.c +200 -0
  49. data/vendor/ring/crypto/bn/convert.c +433 -0
  50. data/vendor/ring/crypto/bn/ctx.c +311 -0
  51. data/vendor/ring/crypto/bn/div.c +594 -0
  52. data/vendor/ring/crypto/bn/exponentiation.c +1335 -0
  53. data/vendor/ring/crypto/bn/gcd.c +711 -0
  54. data/vendor/ring/crypto/bn/generic.c +1019 -0
  55. data/vendor/ring/crypto/bn/internal.h +316 -0
  56. data/vendor/ring/crypto/bn/montgomery.c +516 -0
  57. data/vendor/ring/crypto/bn/mul.c +888 -0
  58. data/vendor/ring/crypto/bn/prime.c +829 -0
  59. data/vendor/ring/crypto/bn/random.c +334 -0
  60. data/vendor/ring/crypto/bn/rsaz_exp.c +262 -0
  61. data/vendor/ring/crypto/bn/rsaz_exp.h +53 -0
  62. data/vendor/ring/crypto/bn/shift.c +276 -0
  63. data/vendor/ring/crypto/bytestring/bytestring_test.Windows.vcxproj +25 -0
  64. data/vendor/ring/crypto/bytestring/bytestring_test.cc +421 -0
  65. data/vendor/ring/crypto/bytestring/cbb.c +399 -0
  66. data/vendor/ring/crypto/bytestring/cbs.c +227 -0
  67. data/vendor/ring/crypto/bytestring/internal.h +46 -0
  68. data/vendor/ring/crypto/chacha/chacha_generic.c +140 -0
  69. data/vendor/ring/crypto/chacha/chacha_vec.c +323 -0
  70. data/vendor/ring/crypto/chacha/chacha_vec_arm.S +1447 -0
  71. data/vendor/ring/crypto/chacha/chacha_vec_arm_generate.go +153 -0
  72. data/vendor/ring/crypto/cipher/cipher_test.Windows.vcxproj +25 -0
  73. data/vendor/ring/crypto/cipher/e_aes.c +390 -0
  74. data/vendor/ring/crypto/cipher/e_chacha20poly1305.c +208 -0
  75. data/vendor/ring/crypto/cipher/internal.h +173 -0
  76. data/vendor/ring/crypto/cipher/test/aes_128_gcm_tests.txt +543 -0
  77. data/vendor/ring/crypto/cipher/test/aes_128_key_wrap_tests.txt +9 -0
  78. data/vendor/ring/crypto/cipher/test/aes_256_gcm_tests.txt +475 -0
  79. data/vendor/ring/crypto/cipher/test/aes_256_key_wrap_tests.txt +23 -0
  80. data/vendor/ring/crypto/cipher/test/chacha20_poly1305_old_tests.txt +422 -0
  81. data/vendor/ring/crypto/cipher/test/chacha20_poly1305_tests.txt +484 -0
  82. data/vendor/ring/crypto/cipher/test/cipher_test.txt +100 -0
  83. data/vendor/ring/crypto/constant_time_test.Windows.vcxproj +25 -0
  84. data/vendor/ring/crypto/constant_time_test.c +304 -0
  85. data/vendor/ring/crypto/cpu-arm-asm.S +32 -0
  86. data/vendor/ring/crypto/cpu-arm.c +199 -0
  87. data/vendor/ring/crypto/cpu-intel.c +261 -0
  88. data/vendor/ring/crypto/crypto.c +151 -0
  89. data/vendor/ring/crypto/curve25519/asm/x25519-arm.S +2118 -0
  90. data/vendor/ring/crypto/curve25519/curve25519.c +4888 -0
  91. data/vendor/ring/crypto/curve25519/x25519_test.cc +128 -0
  92. data/vendor/ring/crypto/digest/md32_common.h +181 -0
  93. data/vendor/ring/crypto/ec/asm/p256-x86_64-asm.pl +2725 -0
  94. data/vendor/ring/crypto/ec/ec.c +193 -0
  95. data/vendor/ring/crypto/ec/ec_curves.c +61 -0
  96. data/vendor/ring/crypto/ec/ec_key.c +228 -0
  97. data/vendor/ring/crypto/ec/ec_montgomery.c +114 -0
  98. data/vendor/ring/crypto/ec/example_mul.Windows.vcxproj +25 -0
  99. data/vendor/ring/crypto/ec/internal.h +243 -0
  100. data/vendor/ring/crypto/ec/oct.c +253 -0
  101. data/vendor/ring/crypto/ec/p256-64.c +1794 -0
  102. data/vendor/ring/crypto/ec/p256-x86_64-table.h +9548 -0
  103. data/vendor/ring/crypto/ec/p256-x86_64.c +509 -0
  104. data/vendor/ring/crypto/ec/simple.c +1007 -0
  105. data/vendor/ring/crypto/ec/util-64.c +183 -0
  106. data/vendor/ring/crypto/ec/wnaf.c +508 -0
  107. data/vendor/ring/crypto/ecdh/ecdh.c +155 -0
  108. data/vendor/ring/crypto/ecdsa/ecdsa.c +304 -0
  109. data/vendor/ring/crypto/ecdsa/ecdsa_asn1.c +193 -0
  110. data/vendor/ring/crypto/ecdsa/ecdsa_test.Windows.vcxproj +25 -0
  111. data/vendor/ring/crypto/ecdsa/ecdsa_test.cc +327 -0
  112. data/vendor/ring/crypto/header_removed.h +17 -0
  113. data/vendor/ring/crypto/internal.h +495 -0
  114. data/vendor/ring/crypto/libring.Windows.vcxproj +101 -0
  115. data/vendor/ring/crypto/mem.c +98 -0
  116. data/vendor/ring/crypto/modes/asm/aesni-gcm-x86_64.pl +1045 -0
  117. data/vendor/ring/crypto/modes/asm/ghash-armv4.pl +517 -0
  118. data/vendor/ring/crypto/modes/asm/ghash-x86.pl +1393 -0
  119. data/vendor/ring/crypto/modes/asm/ghash-x86_64.pl +1741 -0
  120. data/vendor/ring/crypto/modes/asm/ghashv8-armx.pl +422 -0
  121. data/vendor/ring/crypto/modes/ctr.c +226 -0
  122. data/vendor/ring/crypto/modes/gcm.c +1206 -0
  123. data/vendor/ring/crypto/modes/gcm_test.Windows.vcxproj +25 -0
  124. data/vendor/ring/crypto/modes/gcm_test.c +348 -0
  125. data/vendor/ring/crypto/modes/internal.h +299 -0
  126. data/vendor/ring/crypto/perlasm/arm-xlate.pl +170 -0
  127. data/vendor/ring/crypto/perlasm/readme +100 -0
  128. data/vendor/ring/crypto/perlasm/x86_64-xlate.pl +1164 -0
  129. data/vendor/ring/crypto/perlasm/x86asm.pl +292 -0
  130. data/vendor/ring/crypto/perlasm/x86gas.pl +263 -0
  131. data/vendor/ring/crypto/perlasm/x86masm.pl +200 -0
  132. data/vendor/ring/crypto/perlasm/x86nasm.pl +187 -0
  133. data/vendor/ring/crypto/poly1305/poly1305.c +331 -0
  134. data/vendor/ring/crypto/poly1305/poly1305_arm.c +301 -0
  135. data/vendor/ring/crypto/poly1305/poly1305_arm_asm.S +2015 -0
  136. data/vendor/ring/crypto/poly1305/poly1305_test.Windows.vcxproj +25 -0
  137. data/vendor/ring/crypto/poly1305/poly1305_test.cc +80 -0
  138. data/vendor/ring/crypto/poly1305/poly1305_test.txt +52 -0
  139. data/vendor/ring/crypto/poly1305/poly1305_vec.c +892 -0
  140. data/vendor/ring/crypto/rand/asm/rdrand-x86_64.pl +75 -0
  141. data/vendor/ring/crypto/rand/internal.h +32 -0
  142. data/vendor/ring/crypto/rand/rand.c +189 -0
  143. data/vendor/ring/crypto/rand/urandom.c +219 -0
  144. data/vendor/ring/crypto/rand/windows.c +56 -0
  145. data/vendor/ring/crypto/refcount_c11.c +66 -0
  146. data/vendor/ring/crypto/refcount_lock.c +53 -0
  147. data/vendor/ring/crypto/refcount_test.Windows.vcxproj +25 -0
  148. data/vendor/ring/crypto/refcount_test.c +58 -0
  149. data/vendor/ring/crypto/rsa/blinding.c +462 -0
  150. data/vendor/ring/crypto/rsa/internal.h +108 -0
  151. data/vendor/ring/crypto/rsa/padding.c +300 -0
  152. data/vendor/ring/crypto/rsa/rsa.c +450 -0
  153. data/vendor/ring/crypto/rsa/rsa_asn1.c +261 -0
  154. data/vendor/ring/crypto/rsa/rsa_impl.c +944 -0
  155. data/vendor/ring/crypto/rsa/rsa_test.Windows.vcxproj +25 -0
  156. data/vendor/ring/crypto/rsa/rsa_test.cc +437 -0
  157. data/vendor/ring/crypto/sha/asm/sha-armv8.pl +436 -0
  158. data/vendor/ring/crypto/sha/asm/sha-x86_64.pl +2390 -0
  159. data/vendor/ring/crypto/sha/asm/sha256-586.pl +1275 -0
  160. data/vendor/ring/crypto/sha/asm/sha256-armv4.pl +735 -0
  161. data/vendor/ring/crypto/sha/asm/sha256-armv8.pl +14 -0
  162. data/vendor/ring/crypto/sha/asm/sha256-x86_64.pl +14 -0
  163. data/vendor/ring/crypto/sha/asm/sha512-586.pl +911 -0
  164. data/vendor/ring/crypto/sha/asm/sha512-armv4.pl +666 -0
  165. data/vendor/ring/crypto/sha/asm/sha512-armv8.pl +14 -0
  166. data/vendor/ring/crypto/sha/asm/sha512-x86_64.pl +14 -0
  167. data/vendor/ring/crypto/sha/sha1.c +271 -0
  168. data/vendor/ring/crypto/sha/sha256.c +204 -0
  169. data/vendor/ring/crypto/sha/sha512.c +355 -0
  170. data/vendor/ring/crypto/test/file_test.cc +326 -0
  171. data/vendor/ring/crypto/test/file_test.h +181 -0
  172. data/vendor/ring/crypto/test/malloc.cc +150 -0
  173. data/vendor/ring/crypto/test/scoped_types.h +95 -0
  174. data/vendor/ring/crypto/test/test.Windows.vcxproj +35 -0
  175. data/vendor/ring/crypto/test/test_util.cc +46 -0
  176. data/vendor/ring/crypto/test/test_util.h +41 -0
  177. data/vendor/ring/crypto/thread_none.c +55 -0
  178. data/vendor/ring/crypto/thread_pthread.c +165 -0
  179. data/vendor/ring/crypto/thread_test.Windows.vcxproj +25 -0
  180. data/vendor/ring/crypto/thread_test.c +200 -0
  181. data/vendor/ring/crypto/thread_win.c +282 -0
  182. data/vendor/ring/examples/checkdigest.rs +103 -0
  183. data/vendor/ring/include/openssl/aes.h +121 -0
  184. data/vendor/ring/include/openssl/arm_arch.h +129 -0
  185. data/vendor/ring/include/openssl/base.h +156 -0
  186. data/vendor/ring/include/openssl/bn.h +794 -0
  187. data/vendor/ring/include/openssl/buffer.h +18 -0
  188. data/vendor/ring/include/openssl/bytestring.h +235 -0
  189. data/vendor/ring/include/openssl/chacha.h +37 -0
  190. data/vendor/ring/include/openssl/cmac.h +76 -0
  191. data/vendor/ring/include/openssl/cpu.h +184 -0
  192. data/vendor/ring/include/openssl/crypto.h +43 -0
  193. data/vendor/ring/include/openssl/curve25519.h +88 -0
  194. data/vendor/ring/include/openssl/ec.h +225 -0
  195. data/vendor/ring/include/openssl/ec_key.h +129 -0
  196. data/vendor/ring/include/openssl/ecdh.h +110 -0
  197. data/vendor/ring/include/openssl/ecdsa.h +156 -0
  198. data/vendor/ring/include/openssl/err.h +201 -0
  199. data/vendor/ring/include/openssl/mem.h +101 -0
  200. data/vendor/ring/include/openssl/obj_mac.h +71 -0
  201. data/vendor/ring/include/openssl/opensslfeatures.h +68 -0
  202. data/vendor/ring/include/openssl/opensslv.h +18 -0
  203. data/vendor/ring/include/openssl/ossl_typ.h +18 -0
  204. data/vendor/ring/include/openssl/poly1305.h +51 -0
  205. data/vendor/ring/include/openssl/rand.h +70 -0
  206. data/vendor/ring/include/openssl/rsa.h +399 -0
  207. data/vendor/ring/include/openssl/thread.h +133 -0
  208. data/vendor/ring/include/openssl/type_check.h +71 -0
  209. data/vendor/ring/mk/Common.props +63 -0
  210. data/vendor/ring/mk/Windows.props +42 -0
  211. data/vendor/ring/mk/WindowsTest.props +18 -0
  212. data/vendor/ring/mk/appveyor.bat +62 -0
  213. data/vendor/ring/mk/bottom_of_makefile.mk +54 -0
  214. data/vendor/ring/mk/ring.mk +266 -0
  215. data/vendor/ring/mk/top_of_makefile.mk +214 -0
  216. data/vendor/ring/mk/travis.sh +40 -0
  217. data/vendor/ring/mk/update-travis-yml.py +229 -0
  218. data/vendor/ring/ring.sln +153 -0
  219. data/vendor/ring/src/aead.rs +682 -0
  220. data/vendor/ring/src/agreement.rs +248 -0
  221. data/vendor/ring/src/c.rs +129 -0
  222. data/vendor/ring/src/constant_time.rs +37 -0
  223. data/vendor/ring/src/der.rs +96 -0
  224. data/vendor/ring/src/digest.rs +690 -0
  225. data/vendor/ring/src/digest_tests.txt +57 -0
  226. data/vendor/ring/src/ecc.rs +28 -0
  227. data/vendor/ring/src/ecc_build.rs +279 -0
  228. data/vendor/ring/src/ecc_curves.rs +117 -0
  229. data/vendor/ring/src/ed25519_tests.txt +2579 -0
  230. data/vendor/ring/src/exe_tests.rs +46 -0
  231. data/vendor/ring/src/ffi.rs +29 -0
  232. data/vendor/ring/src/file_test.rs +187 -0
  233. data/vendor/ring/src/hkdf.rs +153 -0
  234. data/vendor/ring/src/hkdf_tests.txt +59 -0
  235. data/vendor/ring/src/hmac.rs +414 -0
  236. data/vendor/ring/src/hmac_tests.txt +97 -0
  237. data/vendor/ring/src/input.rs +312 -0
  238. data/vendor/ring/src/lib.rs +41 -0
  239. data/vendor/ring/src/pbkdf2.rs +265 -0
  240. data/vendor/ring/src/pbkdf2_tests.txt +113 -0
  241. data/vendor/ring/src/polyfill.rs +57 -0
  242. data/vendor/ring/src/rand.rs +28 -0
  243. data/vendor/ring/src/signature.rs +314 -0
  244. data/vendor/ring/third-party/NIST/README.md +9 -0
  245. data/vendor/ring/third-party/NIST/SHAVS/SHA1LongMsg.rsp +263 -0
  246. data/vendor/ring/third-party/NIST/SHAVS/SHA1Monte.rsp +309 -0
  247. data/vendor/ring/third-party/NIST/SHAVS/SHA1ShortMsg.rsp +267 -0
  248. data/vendor/ring/third-party/NIST/SHAVS/SHA224LongMsg.rsp +263 -0
  249. data/vendor/ring/third-party/NIST/SHAVS/SHA224Monte.rsp +309 -0
  250. data/vendor/ring/third-party/NIST/SHAVS/SHA224ShortMsg.rsp +267 -0
  251. data/vendor/ring/third-party/NIST/SHAVS/SHA256LongMsg.rsp +263 -0
  252. data/vendor/ring/third-party/NIST/SHAVS/SHA256Monte.rsp +309 -0
  253. data/vendor/ring/third-party/NIST/SHAVS/SHA256ShortMsg.rsp +267 -0
  254. data/vendor/ring/third-party/NIST/SHAVS/SHA384LongMsg.rsp +519 -0
  255. data/vendor/ring/third-party/NIST/SHAVS/SHA384Monte.rsp +309 -0
  256. data/vendor/ring/third-party/NIST/SHAVS/SHA384ShortMsg.rsp +523 -0
  257. data/vendor/ring/third-party/NIST/SHAVS/SHA512LongMsg.rsp +519 -0
  258. data/vendor/ring/third-party/NIST/SHAVS/SHA512Monte.rsp +309 -0
  259. data/vendor/ring/third-party/NIST/SHAVS/SHA512ShortMsg.rsp +523 -0
  260. data/vendor/ring/third-party/NIST/sha256sums.txt +1 -0
  261. metadata +333 -0
@@ -0,0 +1,422 @@
1
+ #!/usr/bin/env perl
2
+ #
3
+ # ====================================================================
4
+ # Written by Andy Polyakov <appro@openssl.org> for the OpenSSL
5
+ # project. The module is, however, dual licensed under OpenSSL and
6
+ # CRYPTOGAMS licenses depending on where you obtain it. For further
7
+ # details see http://www.openssl.org/~appro/cryptogams/.
8
+ # ====================================================================
9
+ #
10
+ # GHASH for ARMv8 Crypto Extension, 64-bit polynomial multiplication.
11
+ #
12
+ # June 2014
13
+ #
14
+ # Initial version was developed in tight cooperation with Ard
15
+ # Biesheuvel <ard.biesheuvel@linaro.org> from bits-n-pieces from
16
+ # other assembly modules. Just like aesv8-armx.pl this module
17
+ # supports both AArch32 and AArch64 execution modes.
18
+ #
19
+ # July 2014
20
+ #
21
+ # Implement 2x aggregated reduction [see ghash-x86.pl for background
22
+ # information].
23
+ #
24
+ # Current performance in cycles per processed byte:
25
+ #
26
+ # PMULL[2] 32-bit NEON(*)
27
+ # Apple A7 0.92 5.62
28
+ # Cortex-A53 1.01 8.39
29
+ # Cortex-A57 1.17 7.61
30
+ # Denver 0.71 6.02
31
+ #
32
+ # (*) presented for reference/comparison purposes;
33
+
34
+ $flavour = shift;
35
+ $output = shift;
36
+
37
+ $0 =~ m/(.*[\/\\])[^\/\\]+$/; $dir=$1;
38
+ ( $xlate="${dir}arm-xlate.pl" and -f $xlate ) or
39
+ ( $xlate="${dir}../../perlasm/arm-xlate.pl" and -f $xlate) or
40
+ die "can't locate arm-xlate.pl";
41
+
42
+ open OUT,"| \"$^X\" $xlate $flavour $output";
43
+ *STDOUT=*OUT;
44
+
45
+ $Xi="x0"; # argument block
46
+ $Htbl="x1";
47
+ $inp="x2";
48
+ $len="x3";
49
+
50
+ $inc="x12";
51
+
52
+ {
53
+ my ($Xl,$Xm,$Xh,$IN)=map("q$_",(0..3));
54
+ my ($t0,$t1,$t2,$xC2,$H,$Hhl,$H2)=map("q$_",(8..14));
55
+
56
+ $code=<<___;
57
+ #include <openssl/arm_arch.h>
58
+
59
+ .text
60
+ ___
61
+ $code.=<<___ if ($flavour =~ /64/);
62
+ #if !defined(__clang__)
63
+ .arch armv8-a+crypto
64
+ #endif
65
+ ___
66
+ $code.=".fpu neon\n.code 32\n" if ($flavour !~ /64/);
67
+
68
+ ################################################################################
69
+ # void gcm_init_v8(u128 Htable[16],const u64 H[2]);
70
+ #
71
+ # input: 128-bit H - secret parameter E(K,0^128)
72
+ # output: precomputed table filled with degrees of twisted H;
73
+ # H is twisted to handle reverse bitness of GHASH;
74
+ # only few of 16 slots of Htable[16] are used;
75
+ # data is opaque to outside world (which allows to
76
+ # optimize the code independently);
77
+ #
78
+ $code.=<<___;
79
+ .global gcm_init_v8
80
+ .type gcm_init_v8,%function
81
+ .align 4
82
+ gcm_init_v8:
83
+ vld1.64 {$t1},[x1] @ load input H
84
+ vmov.i8 $xC2,#0xe1
85
+ vshl.i64 $xC2,$xC2,#57 @ 0xc2.0
86
+ vext.8 $IN,$t1,$t1,#8
87
+ vshr.u64 $t2,$xC2,#63
88
+ vdup.32 $t1,${t1}[1]
89
+ vext.8 $t0,$t2,$xC2,#8 @ t0=0xc2....01
90
+ vshr.u64 $t2,$IN,#63
91
+ vshr.s32 $t1,$t1,#31 @ broadcast carry bit
92
+ vand $t2,$t2,$t0
93
+ vshl.i64 $IN,$IN,#1
94
+ vext.8 $t2,$t2,$t2,#8
95
+ vand $t0,$t0,$t1
96
+ vorr $IN,$IN,$t2 @ H<<<=1
97
+ veor $H,$IN,$t0 @ twisted H
98
+ vst1.64 {$H},[x0],#16 @ store Htable[0]
99
+
100
+ @ calculate H^2
101
+ vext.8 $t0,$H,$H,#8 @ Karatsuba pre-processing
102
+ vpmull.p64 $Xl,$H,$H
103
+ veor $t0,$t0,$H
104
+ vpmull2.p64 $Xh,$H,$H
105
+ vpmull.p64 $Xm,$t0,$t0
106
+
107
+ vext.8 $t1,$Xl,$Xh,#8 @ Karatsuba post-processing
108
+ veor $t2,$Xl,$Xh
109
+ veor $Xm,$Xm,$t1
110
+ veor $Xm,$Xm,$t2
111
+ vpmull.p64 $t2,$Xl,$xC2 @ 1st phase
112
+
113
+ vmov $Xh#lo,$Xm#hi @ Xh|Xm - 256-bit result
114
+ vmov $Xm#hi,$Xl#lo @ Xm is rotated Xl
115
+ veor $Xl,$Xm,$t2
116
+
117
+ vext.8 $t2,$Xl,$Xl,#8 @ 2nd phase
118
+ vpmull.p64 $Xl,$Xl,$xC2
119
+ veor $t2,$t2,$Xh
120
+ veor $H2,$Xl,$t2
121
+
122
+ vext.8 $t1,$H2,$H2,#8 @ Karatsuba pre-processing
123
+ veor $t1,$t1,$H2
124
+ vext.8 $Hhl,$t0,$t1,#8 @ pack Karatsuba pre-processed
125
+ vst1.64 {$Hhl-$H2},[x0] @ store Htable[1..2]
126
+
127
+ ret
128
+ .size gcm_init_v8,.-gcm_init_v8
129
+ ___
130
+ ################################################################################
131
+ # void gcm_gmult_v8(u64 Xi[2],const u128 Htable[16]);
132
+ #
133
+ # input: Xi - current hash value;
134
+ # Htable - table precomputed in gcm_init_v8;
135
+ # output: Xi - next hash value Xi;
136
+ #
137
+ $code.=<<___;
138
+ .global gcm_gmult_v8
139
+ .type gcm_gmult_v8,%function
140
+ .align 4
141
+ gcm_gmult_v8:
142
+ vld1.64 {$t1},[$Xi] @ load Xi
143
+ vmov.i8 $xC2,#0xe1
144
+ vld1.64 {$H-$Hhl},[$Htbl] @ load twisted H, ...
145
+ vshl.u64 $xC2,$xC2,#57
146
+ #ifndef __ARMEB__
147
+ vrev64.8 $t1,$t1
148
+ #endif
149
+ vext.8 $IN,$t1,$t1,#8
150
+
151
+ vpmull.p64 $Xl,$H,$IN @ H.lo·Xi.lo
152
+ veor $t1,$t1,$IN @ Karatsuba pre-processing
153
+ vpmull2.p64 $Xh,$H,$IN @ H.hi·Xi.hi
154
+ vpmull.p64 $Xm,$Hhl,$t1 @ (H.lo+H.hi)·(Xi.lo+Xi.hi)
155
+
156
+ vext.8 $t1,$Xl,$Xh,#8 @ Karatsuba post-processing
157
+ veor $t2,$Xl,$Xh
158
+ veor $Xm,$Xm,$t1
159
+ veor $Xm,$Xm,$t2
160
+ vpmull.p64 $t2,$Xl,$xC2 @ 1st phase of reduction
161
+
162
+ vmov $Xh#lo,$Xm#hi @ Xh|Xm - 256-bit result
163
+ vmov $Xm#hi,$Xl#lo @ Xm is rotated Xl
164
+ veor $Xl,$Xm,$t2
165
+
166
+ vext.8 $t2,$Xl,$Xl,#8 @ 2nd phase of reduction
167
+ vpmull.p64 $Xl,$Xl,$xC2
168
+ veor $t2,$t2,$Xh
169
+ veor $Xl,$Xl,$t2
170
+
171
+ #ifndef __ARMEB__
172
+ vrev64.8 $Xl,$Xl
173
+ #endif
174
+ vext.8 $Xl,$Xl,$Xl,#8
175
+ vst1.64 {$Xl},[$Xi] @ write out Xi
176
+
177
+ ret
178
+ .size gcm_gmult_v8,.-gcm_gmult_v8
179
+ ___
180
+ ################################################################################
181
+ # void gcm_ghash_v8(u64 Xi[2],const u128 Htable[16],const u8 *inp,size_t len);
182
+ #
183
+ # input: table precomputed in gcm_init_v8;
184
+ # current hash value Xi;
185
+ # pointer to input data;
186
+ # length of input data in bytes, but divisible by block size;
187
+ # output: next hash value Xi;
188
+ #
189
+ $code.=<<___;
190
+ .global gcm_ghash_v8
191
+ .type gcm_ghash_v8,%function
192
+ .align 4
193
+ gcm_ghash_v8:
194
+ ___
195
+ $code.=<<___ if ($flavour !~ /64/);
196
+ vstmdb sp!,{d8-d15} @ 32-bit ABI says so
197
+ ___
198
+ $code.=<<___;
199
+ vld1.64 {$Xl},[$Xi] @ load [rotated] Xi
200
+ @ "[rotated]" means that
201
+ @ loaded value would have
202
+ @ to be rotated in order to
203
+ @ make it appear as in
204
+ @ alorithm specification
205
+ subs $len,$len,#32 @ see if $len is 32 or larger
206
+ mov $inc,#16 @ $inc is used as post-
207
+ @ increment for input pointer;
208
+ @ as loop is modulo-scheduled
209
+ @ $inc is zeroed just in time
210
+ @ to preclude oversteping
211
+ @ inp[len], which means that
212
+ @ last block[s] are actually
213
+ @ loaded twice, but last
214
+ @ copy is not processed
215
+ vld1.64 {$H-$Hhl},[$Htbl],#32 @ load twisted H, ..., H^2
216
+ vmov.i8 $xC2,#0xe1
217
+ vld1.64 {$H2},[$Htbl]
218
+ cclr $inc,eq @ is it time to zero $inc?
219
+ vext.8 $Xl,$Xl,$Xl,#8 @ rotate Xi
220
+ vld1.64 {$t0},[$inp],#16 @ load [rotated] I[0]
221
+ vshl.u64 $xC2,$xC2,#57 @ compose 0xc2.0 constant
222
+ #ifndef __ARMEB__
223
+ vrev64.8 $t0,$t0
224
+ vrev64.8 $Xl,$Xl
225
+ #endif
226
+ vext.8 $IN,$t0,$t0,#8 @ rotate I[0]
227
+ b.lo .Lodd_tail_v8 @ $len was less than 32
228
+ ___
229
+ { my ($Xln,$Xmn,$Xhn,$In) = map("q$_",(4..7));
230
+ #######
231
+ # Xi+2 =[H*(Ii+1 + Xi+1)] mod P =
232
+ # [(H*Ii+1) + (H*Xi+1)] mod P =
233
+ # [(H*Ii+1) + H^2*(Ii+Xi)] mod P
234
+ #
235
+ $code.=<<___;
236
+ vld1.64 {$t1},[$inp],$inc @ load [rotated] I[1]
237
+ #ifndef __ARMEB__
238
+ vrev64.8 $t1,$t1
239
+ #endif
240
+ vext.8 $In,$t1,$t1,#8
241
+ veor $IN,$IN,$Xl @ I[i]^=Xi
242
+ vpmull.p64 $Xln,$H,$In @ H·Ii+1
243
+ veor $t1,$t1,$In @ Karatsuba pre-processing
244
+ vpmull2.p64 $Xhn,$H,$In
245
+ b .Loop_mod2x_v8
246
+
247
+ .align 4
248
+ .Loop_mod2x_v8:
249
+ vext.8 $t2,$IN,$IN,#8
250
+ subs $len,$len,#32 @ is there more data?
251
+ vpmull.p64 $Xl,$H2,$IN @ H^2.lo·Xi.lo
252
+ cclr $inc,lo @ is it time to zero $inc?
253
+
254
+ vpmull.p64 $Xmn,$Hhl,$t1
255
+ veor $t2,$t2,$IN @ Karatsuba pre-processing
256
+ vpmull2.p64 $Xh,$H2,$IN @ H^2.hi·Xi.hi
257
+ veor $Xl,$Xl,$Xln @ accumulate
258
+ vpmull2.p64 $Xm,$Hhl,$t2 @ (H^2.lo+H^2.hi)·(Xi.lo+Xi.hi)
259
+ vld1.64 {$t0},[$inp],$inc @ load [rotated] I[i+2]
260
+
261
+ veor $Xh,$Xh,$Xhn
262
+ cclr $inc,eq @ is it time to zero $inc?
263
+ veor $Xm,$Xm,$Xmn
264
+
265
+ vext.8 $t1,$Xl,$Xh,#8 @ Karatsuba post-processing
266
+ veor $t2,$Xl,$Xh
267
+ veor $Xm,$Xm,$t1
268
+ vld1.64 {$t1},[$inp],$inc @ load [rotated] I[i+3]
269
+ #ifndef __ARMEB__
270
+ vrev64.8 $t0,$t0
271
+ #endif
272
+ veor $Xm,$Xm,$t2
273
+ vpmull.p64 $t2,$Xl,$xC2 @ 1st phase of reduction
274
+
275
+ #ifndef __ARMEB__
276
+ vrev64.8 $t1,$t1
277
+ #endif
278
+ vmov $Xh#lo,$Xm#hi @ Xh|Xm - 256-bit result
279
+ vmov $Xm#hi,$Xl#lo @ Xm is rotated Xl
280
+ vext.8 $In,$t1,$t1,#8
281
+ vext.8 $IN,$t0,$t0,#8
282
+ veor $Xl,$Xm,$t2
283
+ vpmull.p64 $Xln,$H,$In @ H·Ii+1
284
+ veor $IN,$IN,$Xh @ accumulate $IN early
285
+
286
+ vext.8 $t2,$Xl,$Xl,#8 @ 2nd phase of reduction
287
+ vpmull.p64 $Xl,$Xl,$xC2
288
+ veor $IN,$IN,$t2
289
+ veor $t1,$t1,$In @ Karatsuba pre-processing
290
+ veor $IN,$IN,$Xl
291
+ vpmull2.p64 $Xhn,$H,$In
292
+ b.hs .Loop_mod2x_v8 @ there was at least 32 more bytes
293
+
294
+ veor $Xh,$Xh,$t2
295
+ vext.8 $IN,$t0,$t0,#8 @ re-construct $IN
296
+ adds $len,$len,#32 @ re-construct $len
297
+ veor $Xl,$Xl,$Xh @ re-construct $Xl
298
+ b.eq .Ldone_v8 @ is $len zero?
299
+ ___
300
+ }
301
+ $code.=<<___;
302
+ .Lodd_tail_v8:
303
+ vext.8 $t2,$Xl,$Xl,#8
304
+ veor $IN,$IN,$Xl @ inp^=Xi
305
+ veor $t1,$t0,$t2 @ $t1 is rotated inp^Xi
306
+
307
+ vpmull.p64 $Xl,$H,$IN @ H.lo·Xi.lo
308
+ veor $t1,$t1,$IN @ Karatsuba pre-processing
309
+ vpmull2.p64 $Xh,$H,$IN @ H.hi·Xi.hi
310
+ vpmull.p64 $Xm,$Hhl,$t1 @ (H.lo+H.hi)·(Xi.lo+Xi.hi)
311
+
312
+ vext.8 $t1,$Xl,$Xh,#8 @ Karatsuba post-processing
313
+ veor $t2,$Xl,$Xh
314
+ veor $Xm,$Xm,$t1
315
+ veor $Xm,$Xm,$t2
316
+ vpmull.p64 $t2,$Xl,$xC2 @ 1st phase of reduction
317
+
318
+ vmov $Xh#lo,$Xm#hi @ Xh|Xm - 256-bit result
319
+ vmov $Xm#hi,$Xl#lo @ Xm is rotated Xl
320
+ veor $Xl,$Xm,$t2
321
+
322
+ vext.8 $t2,$Xl,$Xl,#8 @ 2nd phase of reduction
323
+ vpmull.p64 $Xl,$Xl,$xC2
324
+ veor $t2,$t2,$Xh
325
+ veor $Xl,$Xl,$t2
326
+
327
+ .Ldone_v8:
328
+ #ifndef __ARMEB__
329
+ vrev64.8 $Xl,$Xl
330
+ #endif
331
+ vext.8 $Xl,$Xl,$Xl,#8
332
+ vst1.64 {$Xl},[$Xi] @ write out Xi
333
+
334
+ ___
335
+ $code.=<<___ if ($flavour !~ /64/);
336
+ vldmia sp!,{d8-d15} @ 32-bit ABI says so
337
+ ___
338
+ $code.=<<___;
339
+ ret
340
+ .size gcm_ghash_v8,.-gcm_ghash_v8
341
+ ___
342
+ }
343
+ $code.=<<___;
344
+ .asciz "GHASH for ARMv8, CRYPTOGAMS by <appro\@openssl.org>"
345
+ .align 2
346
+ ___
347
+
348
+ if ($flavour =~ /64/) { ######## 64-bit code
349
+ sub unvmov {
350
+ my $arg=shift;
351
+
352
+ $arg =~ m/q([0-9]+)#(lo|hi),\s*q([0-9]+)#(lo|hi)/o &&
353
+ sprintf "ins v%d.d[%d],v%d.d[%d]",$1,($2 eq "lo")?0:1,$3,($4 eq "lo")?0:1;
354
+ }
355
+ foreach(split("\n",$code)) {
356
+ s/cclr\s+([wx])([^,]+),\s*([a-z]+)/csel $1$2,$1zr,$1$2,$3/o or
357
+ s/vmov\.i8/movi/o or # fix up legacy mnemonics
358
+ s/vmov\s+(.*)/unvmov($1)/geo or
359
+ s/vext\.8/ext/o or
360
+ s/vshr\.s/sshr\.s/o or
361
+ s/vshr/ushr/o or
362
+ s/^(\s+)v/$1/o or # strip off v prefix
363
+ s/\bbx\s+lr\b/ret/o;
364
+
365
+ s/\bq([0-9]+)\b/"v".($1<8?$1:$1+8).".16b"/geo; # old->new registers
366
+ s/@\s/\/\//o; # old->new style commentary
367
+
368
+ # fix up remainig legacy suffixes
369
+ s/\.[ui]?8(\s)/$1/o;
370
+ s/\.[uis]?32//o and s/\.16b/\.4s/go;
371
+ m/\.p64/o and s/\.16b/\.1q/o; # 1st pmull argument
372
+ m/l\.p64/o and s/\.16b/\.1d/go; # 2nd and 3rd pmull arguments
373
+ s/\.[uisp]?64//o and s/\.16b/\.2d/go;
374
+ s/\.[42]([sd])\[([0-3])\]/\.$1\[$2\]/o;
375
+
376
+ print $_,"\n";
377
+ }
378
+ } else { ######## 32-bit code
379
+ sub unvdup32 {
380
+ my $arg=shift;
381
+
382
+ $arg =~ m/q([0-9]+),\s*q([0-9]+)\[([0-3])\]/o &&
383
+ sprintf "vdup.32 q%d,d%d[%d]",$1,2*$2+($3>>1),$3&1;
384
+ }
385
+ sub unvpmullp64 {
386
+ my ($mnemonic,$arg)=@_;
387
+
388
+ if ($arg =~ m/q([0-9]+),\s*q([0-9]+),\s*q([0-9]+)/o) {
389
+ my $word = 0xf2a00e00|(($1&7)<<13)|(($1&8)<<19)
390
+ |(($2&7)<<17)|(($2&8)<<4)
391
+ |(($3&7)<<1) |(($3&8)<<2);
392
+ $word |= 0x00010001 if ($mnemonic =~ "2");
393
+ # since ARMv7 instructions are always encoded little-endian.
394
+ # correct solution is to use .inst directive, but older
395
+ # assemblers don't implement it:-(
396
+ sprintf ".byte\t0x%02x,0x%02x,0x%02x,0x%02x\t@ %s %s",
397
+ $word&0xff,($word>>8)&0xff,
398
+ ($word>>16)&0xff,($word>>24)&0xff,
399
+ $mnemonic,$arg;
400
+ }
401
+ }
402
+
403
+ foreach(split("\n",$code)) {
404
+ s/\b[wx]([0-9]+)\b/r$1/go; # new->old registers
405
+ s/\bv([0-9])\.[12468]+[bsd]\b/q$1/go; # new->old registers
406
+ s/\/\/\s?/@ /o; # new->old style commentary
407
+
408
+ # fix up remainig new-style suffixes
409
+ s/\],#[0-9]+/]!/o;
410
+
411
+ s/cclr\s+([^,]+),\s*([a-z]+)/mov$2 $1,#0/o or
412
+ s/vdup\.32\s+(.*)/unvdup32($1)/geo or
413
+ s/v?(pmull2?)\.p64\s+(.*)/unvpmullp64($1,$2)/geo or
414
+ s/\bq([0-9]+)#(lo|hi)/sprintf "d%d",2*$1+($2 eq "hi")/geo or
415
+ s/^(\s+)b\./$1b/o or
416
+ s/^(\s+)ret/$1bx\tlr/o;
417
+
418
+ print $_,"\n";
419
+ }
420
+ }
421
+
422
+ close STDOUT; # enforce flush
@@ -0,0 +1,226 @@
1
+ /* ====================================================================
2
+ * Copyright (c) 2008 The OpenSSL Project. All rights reserved.
3
+ *
4
+ * Redistribution and use in source and binary forms, with or without
5
+ * modification, are permitted provided that the following conditions
6
+ * are met:
7
+ *
8
+ * 1. Redistributions of source code must retain the above copyright
9
+ * notice, this list of conditions and the following disclaimer.
10
+ *
11
+ * 2. Redistributions in binary form must reproduce the above copyright
12
+ * notice, this list of conditions and the following disclaimer in
13
+ * the documentation and/or other materials provided with the
14
+ * distribution.
15
+ *
16
+ * 3. All advertising materials mentioning features or use of this
17
+ * software must display the following acknowledgment:
18
+ * "This product includes software developed by the OpenSSL Project
19
+ * for use in the OpenSSL Toolkit. (http://www.openssl.org/)"
20
+ *
21
+ * 4. The names "OpenSSL Toolkit" and "OpenSSL Project" must not be used to
22
+ * endorse or promote products derived from this software without
23
+ * prior written permission. For written permission, please contact
24
+ * openssl-core@openssl.org.
25
+ *
26
+ * 5. Products derived from this software may not be called "OpenSSL"
27
+ * nor may "OpenSSL" appear in their names without prior written
28
+ * permission of the OpenSSL Project.
29
+ *
30
+ * 6. Redistributions of any form whatsoever must retain the following
31
+ * acknowledgment:
32
+ * "This product includes software developed by the OpenSSL Project
33
+ * for use in the OpenSSL Toolkit (http://www.openssl.org/)"
34
+ *
35
+ * THIS SOFTWARE IS PROVIDED BY THE OpenSSL PROJECT ``AS IS'' AND ANY
36
+ * EXPRESSED OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
37
+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
38
+ * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE OpenSSL PROJECT OR
39
+ * ITS CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
40
+ * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
41
+ * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
42
+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
43
+ * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT,
44
+ * STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
45
+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED
46
+ * OF THE POSSIBILITY OF SUCH DAMAGE.
47
+ * ==================================================================== */
48
+
49
+ #include <openssl/type_check.h>
50
+
51
+ #include <assert.h>
52
+ #include <string.h>
53
+
54
+ #include "internal.h"
55
+
56
+
57
+ /* NOTE: the IV/counter CTR mode is big-endian. The code itself
58
+ * is endian-neutral. */
59
+
60
+ /* increment counter (128-bit int) by 1 */
61
+ static void ctr128_inc(uint8_t *counter) {
62
+ uint32_t n = 16;
63
+ uint8_t c;
64
+
65
+ do {
66
+ --n;
67
+ c = counter[n];
68
+ ++c;
69
+ counter[n] = c;
70
+ if (c) {
71
+ return;
72
+ }
73
+ } while (n);
74
+ }
75
+
76
+ OPENSSL_COMPILE_ASSERT((16 % sizeof(size_t)) == 0, bad_size_t_size);
77
+
78
+ /* The input encrypted as though 128bit counter mode is being used. The extra
79
+ * state information to record how much of the 128bit block we have used is
80
+ * contained in *num, and the encrypted counter is kept in ecount_buf. Both
81
+ * *num and ecount_buf must be initialised with zeros before the first call to
82
+ * CRYPTO_ctr128_encrypt().
83
+ *
84
+ * This algorithm assumes that the counter is in the x lower bits of the IV
85
+ * (ivec), and that the application has full control over overflow and the rest
86
+ * of the IV. This implementation takes NO responsibility for checking that
87
+ * the counter doesn't overflow into the rest of the IV when incremented. */
88
+ void CRYPTO_ctr128_encrypt(const uint8_t *in, uint8_t *out, size_t len,
89
+ const void *key, uint8_t ivec[16],
90
+ uint8_t ecount_buf[16], unsigned int *num,
91
+ block128_f block) {
92
+ unsigned int n;
93
+
94
+ assert(key && ecount_buf && num);
95
+ assert(len == 0 || (in && out));
96
+ assert(*num < 16);
97
+
98
+ n = *num;
99
+
100
+ while (n && len) {
101
+ *(out++) = *(in++) ^ ecount_buf[n];
102
+ --len;
103
+ n = (n + 1) % 16;
104
+ }
105
+
106
+ #if STRICT_ALIGNMENT
107
+ if (((size_t)in | (size_t)out | (size_t)ivec) % sizeof(size_t) != 0) {
108
+ size_t l = 0;
109
+ while (l < len) {
110
+ if (n == 0) {
111
+ (*block)(ivec, ecount_buf, key);
112
+ ctr128_inc(ivec);
113
+ }
114
+ out[l] = in[l] ^ ecount_buf[n];
115
+ ++l;
116
+ n = (n + 1) % 16;
117
+ }
118
+
119
+ *num = n;
120
+ return;
121
+ }
122
+ #endif
123
+
124
+ while (len >= 16) {
125
+ (*block)(ivec, ecount_buf, key);
126
+ ctr128_inc(ivec);
127
+ for (; n < 16; n += sizeof(size_t)) {
128
+ *(size_t *)(out + n) = *(size_t *)(in + n) ^ *(size_t *)(ecount_buf + n);
129
+ }
130
+ len -= 16;
131
+ out += 16;
132
+ in += 16;
133
+ n = 0;
134
+ }
135
+ if (len) {
136
+ (*block)(ivec, ecount_buf, key);
137
+ ctr128_inc(ivec);
138
+ while (len--) {
139
+ out[n] = in[n] ^ ecount_buf[n];
140
+ ++n;
141
+ }
142
+ }
143
+ *num = n;
144
+ }
145
+
146
+ /* increment upper 96 bits of 128-bit counter by 1 */
147
+ static void ctr96_inc(uint8_t *counter) {
148
+ uint32_t n = 12;
149
+ uint8_t c;
150
+
151
+ do {
152
+ --n;
153
+ c = counter[n];
154
+ ++c;
155
+ counter[n] = c;
156
+ if (c) {
157
+ return;
158
+ }
159
+ } while (n);
160
+ }
161
+
162
+ void CRYPTO_ctr128_encrypt_ctr32(const uint8_t *in, uint8_t *out,
163
+ size_t len, const void *key,
164
+ uint8_t ivec[16],
165
+ uint8_t ecount_buf[16],
166
+ unsigned int *num, ctr128_f func) {
167
+ unsigned int n, ctr32;
168
+
169
+ assert(key && ecount_buf && num);
170
+ assert(len == 0 || (in && out));
171
+ assert(*num < 16);
172
+
173
+ n = *num;
174
+
175
+ while (n && len) {
176
+ *(out++) = *(in++) ^ ecount_buf[n];
177
+ --len;
178
+ n = (n + 1) % 16;
179
+ }
180
+
181
+ ctr32 = GETU32(ivec + 12);
182
+ while (len >= 16) {
183
+ size_t blocks = len / 16;
184
+ /* 1<<28 is just a not-so-small yet not-so-large number...
185
+ * Below condition is practically never met, but it has to
186
+ * be checked for code correctness. */
187
+ if (sizeof(size_t) > sizeof(unsigned int) && blocks > (1U << 28)) {
188
+ blocks = (1U << 28);
189
+ }
190
+ /* As (*func) operates on 32-bit counter, caller
191
+ * has to handle overflow. 'if' below detects the
192
+ * overflow, which is then handled by limiting the
193
+ * amount of blocks to the exact overflow point... */
194
+ ctr32 += (uint32_t)blocks;
195
+ if (ctr32 < blocks) {
196
+ blocks -= ctr32;
197
+ ctr32 = 0;
198
+ }
199
+ (*func)(in, out, blocks, key, ivec);
200
+ /* (*func) does not update ivec, caller does: */
201
+ PUTU32(ivec + 12, ctr32);
202
+ /* ... overflow was detected, propogate carry. */
203
+ if (ctr32 == 0) {
204
+ ctr96_inc(ivec);
205
+ }
206
+ blocks *= 16;
207
+ len -= blocks;
208
+ out += blocks;
209
+ in += blocks;
210
+ }
211
+ if (len) {
212
+ memset(ecount_buf, 0, 16);
213
+ (*func)(ecount_buf, ecount_buf, 1, key, ivec);
214
+ ++ctr32;
215
+ PUTU32(ivec + 12, ctr32);
216
+ if (ctr32 == 0) {
217
+ ctr96_inc(ivec);
218
+ }
219
+ while (len--) {
220
+ out[n] = in[n] ^ ecount_buf[n];
221
+ ++n;
222
+ }
223
+ }
224
+
225
+ *num = n;
226
+ }