ring-native 0.0.0
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +7 -0
- data/.gitignore +9 -0
- data/Gemfile +3 -0
- data/README.md +22 -0
- data/Rakefile +1 -0
- data/ext/ring/extconf.rb +29 -0
- data/lib/ring/native.rb +8 -0
- data/lib/ring/native/version.rb +5 -0
- data/ring-native.gemspec +25 -0
- data/vendor/ring/BUILDING.md +40 -0
- data/vendor/ring/Cargo.toml +43 -0
- data/vendor/ring/LICENSE +185 -0
- data/vendor/ring/Makefile +35 -0
- data/vendor/ring/PORTING.md +163 -0
- data/vendor/ring/README.md +113 -0
- data/vendor/ring/STYLE.md +197 -0
- data/vendor/ring/appveyor.yml +27 -0
- data/vendor/ring/build.rs +108 -0
- data/vendor/ring/crypto/aes/aes.c +1142 -0
- data/vendor/ring/crypto/aes/aes_test.Windows.vcxproj +25 -0
- data/vendor/ring/crypto/aes/aes_test.cc +93 -0
- data/vendor/ring/crypto/aes/asm/aes-586.pl +2368 -0
- data/vendor/ring/crypto/aes/asm/aes-armv4.pl +1249 -0
- data/vendor/ring/crypto/aes/asm/aes-x86_64.pl +2246 -0
- data/vendor/ring/crypto/aes/asm/aesni-x86.pl +1318 -0
- data/vendor/ring/crypto/aes/asm/aesni-x86_64.pl +2084 -0
- data/vendor/ring/crypto/aes/asm/aesv8-armx.pl +675 -0
- data/vendor/ring/crypto/aes/asm/bsaes-armv7.pl +1364 -0
- data/vendor/ring/crypto/aes/asm/bsaes-x86_64.pl +1565 -0
- data/vendor/ring/crypto/aes/asm/vpaes-x86.pl +841 -0
- data/vendor/ring/crypto/aes/asm/vpaes-x86_64.pl +1116 -0
- data/vendor/ring/crypto/aes/internal.h +87 -0
- data/vendor/ring/crypto/aes/mode_wrappers.c +61 -0
- data/vendor/ring/crypto/bn/add.c +394 -0
- data/vendor/ring/crypto/bn/asm/armv4-mont.pl +694 -0
- data/vendor/ring/crypto/bn/asm/armv8-mont.pl +1503 -0
- data/vendor/ring/crypto/bn/asm/bn-586.pl +774 -0
- data/vendor/ring/crypto/bn/asm/co-586.pl +287 -0
- data/vendor/ring/crypto/bn/asm/rsaz-avx2.pl +1882 -0
- data/vendor/ring/crypto/bn/asm/x86-mont.pl +592 -0
- data/vendor/ring/crypto/bn/asm/x86_64-gcc.c +599 -0
- data/vendor/ring/crypto/bn/asm/x86_64-mont.pl +1393 -0
- data/vendor/ring/crypto/bn/asm/x86_64-mont5.pl +3507 -0
- data/vendor/ring/crypto/bn/bn.c +352 -0
- data/vendor/ring/crypto/bn/bn_asn1.c +74 -0
- data/vendor/ring/crypto/bn/bn_test.Windows.vcxproj +25 -0
- data/vendor/ring/crypto/bn/bn_test.cc +1696 -0
- data/vendor/ring/crypto/bn/cmp.c +200 -0
- data/vendor/ring/crypto/bn/convert.c +433 -0
- data/vendor/ring/crypto/bn/ctx.c +311 -0
- data/vendor/ring/crypto/bn/div.c +594 -0
- data/vendor/ring/crypto/bn/exponentiation.c +1335 -0
- data/vendor/ring/crypto/bn/gcd.c +711 -0
- data/vendor/ring/crypto/bn/generic.c +1019 -0
- data/vendor/ring/crypto/bn/internal.h +316 -0
- data/vendor/ring/crypto/bn/montgomery.c +516 -0
- data/vendor/ring/crypto/bn/mul.c +888 -0
- data/vendor/ring/crypto/bn/prime.c +829 -0
- data/vendor/ring/crypto/bn/random.c +334 -0
- data/vendor/ring/crypto/bn/rsaz_exp.c +262 -0
- data/vendor/ring/crypto/bn/rsaz_exp.h +53 -0
- data/vendor/ring/crypto/bn/shift.c +276 -0
- data/vendor/ring/crypto/bytestring/bytestring_test.Windows.vcxproj +25 -0
- data/vendor/ring/crypto/bytestring/bytestring_test.cc +421 -0
- data/vendor/ring/crypto/bytestring/cbb.c +399 -0
- data/vendor/ring/crypto/bytestring/cbs.c +227 -0
- data/vendor/ring/crypto/bytestring/internal.h +46 -0
- data/vendor/ring/crypto/chacha/chacha_generic.c +140 -0
- data/vendor/ring/crypto/chacha/chacha_vec.c +323 -0
- data/vendor/ring/crypto/chacha/chacha_vec_arm.S +1447 -0
- data/vendor/ring/crypto/chacha/chacha_vec_arm_generate.go +153 -0
- data/vendor/ring/crypto/cipher/cipher_test.Windows.vcxproj +25 -0
- data/vendor/ring/crypto/cipher/e_aes.c +390 -0
- data/vendor/ring/crypto/cipher/e_chacha20poly1305.c +208 -0
- data/vendor/ring/crypto/cipher/internal.h +173 -0
- data/vendor/ring/crypto/cipher/test/aes_128_gcm_tests.txt +543 -0
- data/vendor/ring/crypto/cipher/test/aes_128_key_wrap_tests.txt +9 -0
- data/vendor/ring/crypto/cipher/test/aes_256_gcm_tests.txt +475 -0
- data/vendor/ring/crypto/cipher/test/aes_256_key_wrap_tests.txt +23 -0
- data/vendor/ring/crypto/cipher/test/chacha20_poly1305_old_tests.txt +422 -0
- data/vendor/ring/crypto/cipher/test/chacha20_poly1305_tests.txt +484 -0
- data/vendor/ring/crypto/cipher/test/cipher_test.txt +100 -0
- data/vendor/ring/crypto/constant_time_test.Windows.vcxproj +25 -0
- data/vendor/ring/crypto/constant_time_test.c +304 -0
- data/vendor/ring/crypto/cpu-arm-asm.S +32 -0
- data/vendor/ring/crypto/cpu-arm.c +199 -0
- data/vendor/ring/crypto/cpu-intel.c +261 -0
- data/vendor/ring/crypto/crypto.c +151 -0
- data/vendor/ring/crypto/curve25519/asm/x25519-arm.S +2118 -0
- data/vendor/ring/crypto/curve25519/curve25519.c +4888 -0
- data/vendor/ring/crypto/curve25519/x25519_test.cc +128 -0
- data/vendor/ring/crypto/digest/md32_common.h +181 -0
- data/vendor/ring/crypto/ec/asm/p256-x86_64-asm.pl +2725 -0
- data/vendor/ring/crypto/ec/ec.c +193 -0
- data/vendor/ring/crypto/ec/ec_curves.c +61 -0
- data/vendor/ring/crypto/ec/ec_key.c +228 -0
- data/vendor/ring/crypto/ec/ec_montgomery.c +114 -0
- data/vendor/ring/crypto/ec/example_mul.Windows.vcxproj +25 -0
- data/vendor/ring/crypto/ec/internal.h +243 -0
- data/vendor/ring/crypto/ec/oct.c +253 -0
- data/vendor/ring/crypto/ec/p256-64.c +1794 -0
- data/vendor/ring/crypto/ec/p256-x86_64-table.h +9548 -0
- data/vendor/ring/crypto/ec/p256-x86_64.c +509 -0
- data/vendor/ring/crypto/ec/simple.c +1007 -0
- data/vendor/ring/crypto/ec/util-64.c +183 -0
- data/vendor/ring/crypto/ec/wnaf.c +508 -0
- data/vendor/ring/crypto/ecdh/ecdh.c +155 -0
- data/vendor/ring/crypto/ecdsa/ecdsa.c +304 -0
- data/vendor/ring/crypto/ecdsa/ecdsa_asn1.c +193 -0
- data/vendor/ring/crypto/ecdsa/ecdsa_test.Windows.vcxproj +25 -0
- data/vendor/ring/crypto/ecdsa/ecdsa_test.cc +327 -0
- data/vendor/ring/crypto/header_removed.h +17 -0
- data/vendor/ring/crypto/internal.h +495 -0
- data/vendor/ring/crypto/libring.Windows.vcxproj +101 -0
- data/vendor/ring/crypto/mem.c +98 -0
- data/vendor/ring/crypto/modes/asm/aesni-gcm-x86_64.pl +1045 -0
- data/vendor/ring/crypto/modes/asm/ghash-armv4.pl +517 -0
- data/vendor/ring/crypto/modes/asm/ghash-x86.pl +1393 -0
- data/vendor/ring/crypto/modes/asm/ghash-x86_64.pl +1741 -0
- data/vendor/ring/crypto/modes/asm/ghashv8-armx.pl +422 -0
- data/vendor/ring/crypto/modes/ctr.c +226 -0
- data/vendor/ring/crypto/modes/gcm.c +1206 -0
- data/vendor/ring/crypto/modes/gcm_test.Windows.vcxproj +25 -0
- data/vendor/ring/crypto/modes/gcm_test.c +348 -0
- data/vendor/ring/crypto/modes/internal.h +299 -0
- data/vendor/ring/crypto/perlasm/arm-xlate.pl +170 -0
- data/vendor/ring/crypto/perlasm/readme +100 -0
- data/vendor/ring/crypto/perlasm/x86_64-xlate.pl +1164 -0
- data/vendor/ring/crypto/perlasm/x86asm.pl +292 -0
- data/vendor/ring/crypto/perlasm/x86gas.pl +263 -0
- data/vendor/ring/crypto/perlasm/x86masm.pl +200 -0
- data/vendor/ring/crypto/perlasm/x86nasm.pl +187 -0
- data/vendor/ring/crypto/poly1305/poly1305.c +331 -0
- data/vendor/ring/crypto/poly1305/poly1305_arm.c +301 -0
- data/vendor/ring/crypto/poly1305/poly1305_arm_asm.S +2015 -0
- data/vendor/ring/crypto/poly1305/poly1305_test.Windows.vcxproj +25 -0
- data/vendor/ring/crypto/poly1305/poly1305_test.cc +80 -0
- data/vendor/ring/crypto/poly1305/poly1305_test.txt +52 -0
- data/vendor/ring/crypto/poly1305/poly1305_vec.c +892 -0
- data/vendor/ring/crypto/rand/asm/rdrand-x86_64.pl +75 -0
- data/vendor/ring/crypto/rand/internal.h +32 -0
- data/vendor/ring/crypto/rand/rand.c +189 -0
- data/vendor/ring/crypto/rand/urandom.c +219 -0
- data/vendor/ring/crypto/rand/windows.c +56 -0
- data/vendor/ring/crypto/refcount_c11.c +66 -0
- data/vendor/ring/crypto/refcount_lock.c +53 -0
- data/vendor/ring/crypto/refcount_test.Windows.vcxproj +25 -0
- data/vendor/ring/crypto/refcount_test.c +58 -0
- data/vendor/ring/crypto/rsa/blinding.c +462 -0
- data/vendor/ring/crypto/rsa/internal.h +108 -0
- data/vendor/ring/crypto/rsa/padding.c +300 -0
- data/vendor/ring/crypto/rsa/rsa.c +450 -0
- data/vendor/ring/crypto/rsa/rsa_asn1.c +261 -0
- data/vendor/ring/crypto/rsa/rsa_impl.c +944 -0
- data/vendor/ring/crypto/rsa/rsa_test.Windows.vcxproj +25 -0
- data/vendor/ring/crypto/rsa/rsa_test.cc +437 -0
- data/vendor/ring/crypto/sha/asm/sha-armv8.pl +436 -0
- data/vendor/ring/crypto/sha/asm/sha-x86_64.pl +2390 -0
- data/vendor/ring/crypto/sha/asm/sha256-586.pl +1275 -0
- data/vendor/ring/crypto/sha/asm/sha256-armv4.pl +735 -0
- data/vendor/ring/crypto/sha/asm/sha256-armv8.pl +14 -0
- data/vendor/ring/crypto/sha/asm/sha256-x86_64.pl +14 -0
- data/vendor/ring/crypto/sha/asm/sha512-586.pl +911 -0
- data/vendor/ring/crypto/sha/asm/sha512-armv4.pl +666 -0
- data/vendor/ring/crypto/sha/asm/sha512-armv8.pl +14 -0
- data/vendor/ring/crypto/sha/asm/sha512-x86_64.pl +14 -0
- data/vendor/ring/crypto/sha/sha1.c +271 -0
- data/vendor/ring/crypto/sha/sha256.c +204 -0
- data/vendor/ring/crypto/sha/sha512.c +355 -0
- data/vendor/ring/crypto/test/file_test.cc +326 -0
- data/vendor/ring/crypto/test/file_test.h +181 -0
- data/vendor/ring/crypto/test/malloc.cc +150 -0
- data/vendor/ring/crypto/test/scoped_types.h +95 -0
- data/vendor/ring/crypto/test/test.Windows.vcxproj +35 -0
- data/vendor/ring/crypto/test/test_util.cc +46 -0
- data/vendor/ring/crypto/test/test_util.h +41 -0
- data/vendor/ring/crypto/thread_none.c +55 -0
- data/vendor/ring/crypto/thread_pthread.c +165 -0
- data/vendor/ring/crypto/thread_test.Windows.vcxproj +25 -0
- data/vendor/ring/crypto/thread_test.c +200 -0
- data/vendor/ring/crypto/thread_win.c +282 -0
- data/vendor/ring/examples/checkdigest.rs +103 -0
- data/vendor/ring/include/openssl/aes.h +121 -0
- data/vendor/ring/include/openssl/arm_arch.h +129 -0
- data/vendor/ring/include/openssl/base.h +156 -0
- data/vendor/ring/include/openssl/bn.h +794 -0
- data/vendor/ring/include/openssl/buffer.h +18 -0
- data/vendor/ring/include/openssl/bytestring.h +235 -0
- data/vendor/ring/include/openssl/chacha.h +37 -0
- data/vendor/ring/include/openssl/cmac.h +76 -0
- data/vendor/ring/include/openssl/cpu.h +184 -0
- data/vendor/ring/include/openssl/crypto.h +43 -0
- data/vendor/ring/include/openssl/curve25519.h +88 -0
- data/vendor/ring/include/openssl/ec.h +225 -0
- data/vendor/ring/include/openssl/ec_key.h +129 -0
- data/vendor/ring/include/openssl/ecdh.h +110 -0
- data/vendor/ring/include/openssl/ecdsa.h +156 -0
- data/vendor/ring/include/openssl/err.h +201 -0
- data/vendor/ring/include/openssl/mem.h +101 -0
- data/vendor/ring/include/openssl/obj_mac.h +71 -0
- data/vendor/ring/include/openssl/opensslfeatures.h +68 -0
- data/vendor/ring/include/openssl/opensslv.h +18 -0
- data/vendor/ring/include/openssl/ossl_typ.h +18 -0
- data/vendor/ring/include/openssl/poly1305.h +51 -0
- data/vendor/ring/include/openssl/rand.h +70 -0
- data/vendor/ring/include/openssl/rsa.h +399 -0
- data/vendor/ring/include/openssl/thread.h +133 -0
- data/vendor/ring/include/openssl/type_check.h +71 -0
- data/vendor/ring/mk/Common.props +63 -0
- data/vendor/ring/mk/Windows.props +42 -0
- data/vendor/ring/mk/WindowsTest.props +18 -0
- data/vendor/ring/mk/appveyor.bat +62 -0
- data/vendor/ring/mk/bottom_of_makefile.mk +54 -0
- data/vendor/ring/mk/ring.mk +266 -0
- data/vendor/ring/mk/top_of_makefile.mk +214 -0
- data/vendor/ring/mk/travis.sh +40 -0
- data/vendor/ring/mk/update-travis-yml.py +229 -0
- data/vendor/ring/ring.sln +153 -0
- data/vendor/ring/src/aead.rs +682 -0
- data/vendor/ring/src/agreement.rs +248 -0
- data/vendor/ring/src/c.rs +129 -0
- data/vendor/ring/src/constant_time.rs +37 -0
- data/vendor/ring/src/der.rs +96 -0
- data/vendor/ring/src/digest.rs +690 -0
- data/vendor/ring/src/digest_tests.txt +57 -0
- data/vendor/ring/src/ecc.rs +28 -0
- data/vendor/ring/src/ecc_build.rs +279 -0
- data/vendor/ring/src/ecc_curves.rs +117 -0
- data/vendor/ring/src/ed25519_tests.txt +2579 -0
- data/vendor/ring/src/exe_tests.rs +46 -0
- data/vendor/ring/src/ffi.rs +29 -0
- data/vendor/ring/src/file_test.rs +187 -0
- data/vendor/ring/src/hkdf.rs +153 -0
- data/vendor/ring/src/hkdf_tests.txt +59 -0
- data/vendor/ring/src/hmac.rs +414 -0
- data/vendor/ring/src/hmac_tests.txt +97 -0
- data/vendor/ring/src/input.rs +312 -0
- data/vendor/ring/src/lib.rs +41 -0
- data/vendor/ring/src/pbkdf2.rs +265 -0
- data/vendor/ring/src/pbkdf2_tests.txt +113 -0
- data/vendor/ring/src/polyfill.rs +57 -0
- data/vendor/ring/src/rand.rs +28 -0
- data/vendor/ring/src/signature.rs +314 -0
- data/vendor/ring/third-party/NIST/README.md +9 -0
- data/vendor/ring/third-party/NIST/SHAVS/SHA1LongMsg.rsp +263 -0
- data/vendor/ring/third-party/NIST/SHAVS/SHA1Monte.rsp +309 -0
- data/vendor/ring/third-party/NIST/SHAVS/SHA1ShortMsg.rsp +267 -0
- data/vendor/ring/third-party/NIST/SHAVS/SHA224LongMsg.rsp +263 -0
- data/vendor/ring/third-party/NIST/SHAVS/SHA224Monte.rsp +309 -0
- data/vendor/ring/third-party/NIST/SHAVS/SHA224ShortMsg.rsp +267 -0
- data/vendor/ring/third-party/NIST/SHAVS/SHA256LongMsg.rsp +263 -0
- data/vendor/ring/third-party/NIST/SHAVS/SHA256Monte.rsp +309 -0
- data/vendor/ring/third-party/NIST/SHAVS/SHA256ShortMsg.rsp +267 -0
- data/vendor/ring/third-party/NIST/SHAVS/SHA384LongMsg.rsp +519 -0
- data/vendor/ring/third-party/NIST/SHAVS/SHA384Monte.rsp +309 -0
- data/vendor/ring/third-party/NIST/SHAVS/SHA384ShortMsg.rsp +523 -0
- data/vendor/ring/third-party/NIST/SHAVS/SHA512LongMsg.rsp +519 -0
- data/vendor/ring/third-party/NIST/SHAVS/SHA512Monte.rsp +309 -0
- data/vendor/ring/third-party/NIST/SHAVS/SHA512ShortMsg.rsp +523 -0
- data/vendor/ring/third-party/NIST/sha256sums.txt +1 -0
- metadata +333 -0
@@ -0,0 +1,422 @@
|
|
1
|
+
#!/usr/bin/env perl
|
2
|
+
#
|
3
|
+
# ====================================================================
|
4
|
+
# Written by Andy Polyakov <appro@openssl.org> for the OpenSSL
|
5
|
+
# project. The module is, however, dual licensed under OpenSSL and
|
6
|
+
# CRYPTOGAMS licenses depending on where you obtain it. For further
|
7
|
+
# details see http://www.openssl.org/~appro/cryptogams/.
|
8
|
+
# ====================================================================
|
9
|
+
#
|
10
|
+
# GHASH for ARMv8 Crypto Extension, 64-bit polynomial multiplication.
|
11
|
+
#
|
12
|
+
# June 2014
|
13
|
+
#
|
14
|
+
# Initial version was developed in tight cooperation with Ard
|
15
|
+
# Biesheuvel <ard.biesheuvel@linaro.org> from bits-n-pieces from
|
16
|
+
# other assembly modules. Just like aesv8-armx.pl this module
|
17
|
+
# supports both AArch32 and AArch64 execution modes.
|
18
|
+
#
|
19
|
+
# July 2014
|
20
|
+
#
|
21
|
+
# Implement 2x aggregated reduction [see ghash-x86.pl for background
|
22
|
+
# information].
|
23
|
+
#
|
24
|
+
# Current performance in cycles per processed byte:
|
25
|
+
#
|
26
|
+
# PMULL[2] 32-bit NEON(*)
|
27
|
+
# Apple A7 0.92 5.62
|
28
|
+
# Cortex-A53 1.01 8.39
|
29
|
+
# Cortex-A57 1.17 7.61
|
30
|
+
# Denver 0.71 6.02
|
31
|
+
#
|
32
|
+
# (*) presented for reference/comparison purposes;
|
33
|
+
|
34
|
+
$flavour = shift;
|
35
|
+
$output = shift;
|
36
|
+
|
37
|
+
$0 =~ m/(.*[\/\\])[^\/\\]+$/; $dir=$1;
|
38
|
+
( $xlate="${dir}arm-xlate.pl" and -f $xlate ) or
|
39
|
+
( $xlate="${dir}../../perlasm/arm-xlate.pl" and -f $xlate) or
|
40
|
+
die "can't locate arm-xlate.pl";
|
41
|
+
|
42
|
+
open OUT,"| \"$^X\" $xlate $flavour $output";
|
43
|
+
*STDOUT=*OUT;
|
44
|
+
|
45
|
+
$Xi="x0"; # argument block
|
46
|
+
$Htbl="x1";
|
47
|
+
$inp="x2";
|
48
|
+
$len="x3";
|
49
|
+
|
50
|
+
$inc="x12";
|
51
|
+
|
52
|
+
{
|
53
|
+
my ($Xl,$Xm,$Xh,$IN)=map("q$_",(0..3));
|
54
|
+
my ($t0,$t1,$t2,$xC2,$H,$Hhl,$H2)=map("q$_",(8..14));
|
55
|
+
|
56
|
+
$code=<<___;
|
57
|
+
#include <openssl/arm_arch.h>
|
58
|
+
|
59
|
+
.text
|
60
|
+
___
|
61
|
+
$code.=<<___ if ($flavour =~ /64/);
|
62
|
+
#if !defined(__clang__)
|
63
|
+
.arch armv8-a+crypto
|
64
|
+
#endif
|
65
|
+
___
|
66
|
+
$code.=".fpu neon\n.code 32\n" if ($flavour !~ /64/);
|
67
|
+
|
68
|
+
################################################################################
|
69
|
+
# void gcm_init_v8(u128 Htable[16],const u64 H[2]);
|
70
|
+
#
|
71
|
+
# input: 128-bit H - secret parameter E(K,0^128)
|
72
|
+
# output: precomputed table filled with degrees of twisted H;
|
73
|
+
# H is twisted to handle reverse bitness of GHASH;
|
74
|
+
# only few of 16 slots of Htable[16] are used;
|
75
|
+
# data is opaque to outside world (which allows to
|
76
|
+
# optimize the code independently);
|
77
|
+
#
|
78
|
+
$code.=<<___;
|
79
|
+
.global gcm_init_v8
|
80
|
+
.type gcm_init_v8,%function
|
81
|
+
.align 4
|
82
|
+
gcm_init_v8:
|
83
|
+
vld1.64 {$t1},[x1] @ load input H
|
84
|
+
vmov.i8 $xC2,#0xe1
|
85
|
+
vshl.i64 $xC2,$xC2,#57 @ 0xc2.0
|
86
|
+
vext.8 $IN,$t1,$t1,#8
|
87
|
+
vshr.u64 $t2,$xC2,#63
|
88
|
+
vdup.32 $t1,${t1}[1]
|
89
|
+
vext.8 $t0,$t2,$xC2,#8 @ t0=0xc2....01
|
90
|
+
vshr.u64 $t2,$IN,#63
|
91
|
+
vshr.s32 $t1,$t1,#31 @ broadcast carry bit
|
92
|
+
vand $t2,$t2,$t0
|
93
|
+
vshl.i64 $IN,$IN,#1
|
94
|
+
vext.8 $t2,$t2,$t2,#8
|
95
|
+
vand $t0,$t0,$t1
|
96
|
+
vorr $IN,$IN,$t2 @ H<<<=1
|
97
|
+
veor $H,$IN,$t0 @ twisted H
|
98
|
+
vst1.64 {$H},[x0],#16 @ store Htable[0]
|
99
|
+
|
100
|
+
@ calculate H^2
|
101
|
+
vext.8 $t0,$H,$H,#8 @ Karatsuba pre-processing
|
102
|
+
vpmull.p64 $Xl,$H,$H
|
103
|
+
veor $t0,$t0,$H
|
104
|
+
vpmull2.p64 $Xh,$H,$H
|
105
|
+
vpmull.p64 $Xm,$t0,$t0
|
106
|
+
|
107
|
+
vext.8 $t1,$Xl,$Xh,#8 @ Karatsuba post-processing
|
108
|
+
veor $t2,$Xl,$Xh
|
109
|
+
veor $Xm,$Xm,$t1
|
110
|
+
veor $Xm,$Xm,$t2
|
111
|
+
vpmull.p64 $t2,$Xl,$xC2 @ 1st phase
|
112
|
+
|
113
|
+
vmov $Xh#lo,$Xm#hi @ Xh|Xm - 256-bit result
|
114
|
+
vmov $Xm#hi,$Xl#lo @ Xm is rotated Xl
|
115
|
+
veor $Xl,$Xm,$t2
|
116
|
+
|
117
|
+
vext.8 $t2,$Xl,$Xl,#8 @ 2nd phase
|
118
|
+
vpmull.p64 $Xl,$Xl,$xC2
|
119
|
+
veor $t2,$t2,$Xh
|
120
|
+
veor $H2,$Xl,$t2
|
121
|
+
|
122
|
+
vext.8 $t1,$H2,$H2,#8 @ Karatsuba pre-processing
|
123
|
+
veor $t1,$t1,$H2
|
124
|
+
vext.8 $Hhl,$t0,$t1,#8 @ pack Karatsuba pre-processed
|
125
|
+
vst1.64 {$Hhl-$H2},[x0] @ store Htable[1..2]
|
126
|
+
|
127
|
+
ret
|
128
|
+
.size gcm_init_v8,.-gcm_init_v8
|
129
|
+
___
|
130
|
+
################################################################################
|
131
|
+
# void gcm_gmult_v8(u64 Xi[2],const u128 Htable[16]);
|
132
|
+
#
|
133
|
+
# input: Xi - current hash value;
|
134
|
+
# Htable - table precomputed in gcm_init_v8;
|
135
|
+
# output: Xi - next hash value Xi;
|
136
|
+
#
|
137
|
+
$code.=<<___;
|
138
|
+
.global gcm_gmult_v8
|
139
|
+
.type gcm_gmult_v8,%function
|
140
|
+
.align 4
|
141
|
+
gcm_gmult_v8:
|
142
|
+
vld1.64 {$t1},[$Xi] @ load Xi
|
143
|
+
vmov.i8 $xC2,#0xe1
|
144
|
+
vld1.64 {$H-$Hhl},[$Htbl] @ load twisted H, ...
|
145
|
+
vshl.u64 $xC2,$xC2,#57
|
146
|
+
#ifndef __ARMEB__
|
147
|
+
vrev64.8 $t1,$t1
|
148
|
+
#endif
|
149
|
+
vext.8 $IN,$t1,$t1,#8
|
150
|
+
|
151
|
+
vpmull.p64 $Xl,$H,$IN @ H.lo·Xi.lo
|
152
|
+
veor $t1,$t1,$IN @ Karatsuba pre-processing
|
153
|
+
vpmull2.p64 $Xh,$H,$IN @ H.hi·Xi.hi
|
154
|
+
vpmull.p64 $Xm,$Hhl,$t1 @ (H.lo+H.hi)·(Xi.lo+Xi.hi)
|
155
|
+
|
156
|
+
vext.8 $t1,$Xl,$Xh,#8 @ Karatsuba post-processing
|
157
|
+
veor $t2,$Xl,$Xh
|
158
|
+
veor $Xm,$Xm,$t1
|
159
|
+
veor $Xm,$Xm,$t2
|
160
|
+
vpmull.p64 $t2,$Xl,$xC2 @ 1st phase of reduction
|
161
|
+
|
162
|
+
vmov $Xh#lo,$Xm#hi @ Xh|Xm - 256-bit result
|
163
|
+
vmov $Xm#hi,$Xl#lo @ Xm is rotated Xl
|
164
|
+
veor $Xl,$Xm,$t2
|
165
|
+
|
166
|
+
vext.8 $t2,$Xl,$Xl,#8 @ 2nd phase of reduction
|
167
|
+
vpmull.p64 $Xl,$Xl,$xC2
|
168
|
+
veor $t2,$t2,$Xh
|
169
|
+
veor $Xl,$Xl,$t2
|
170
|
+
|
171
|
+
#ifndef __ARMEB__
|
172
|
+
vrev64.8 $Xl,$Xl
|
173
|
+
#endif
|
174
|
+
vext.8 $Xl,$Xl,$Xl,#8
|
175
|
+
vst1.64 {$Xl},[$Xi] @ write out Xi
|
176
|
+
|
177
|
+
ret
|
178
|
+
.size gcm_gmult_v8,.-gcm_gmult_v8
|
179
|
+
___
|
180
|
+
################################################################################
|
181
|
+
# void gcm_ghash_v8(u64 Xi[2],const u128 Htable[16],const u8 *inp,size_t len);
|
182
|
+
#
|
183
|
+
# input: table precomputed in gcm_init_v8;
|
184
|
+
# current hash value Xi;
|
185
|
+
# pointer to input data;
|
186
|
+
# length of input data in bytes, but divisible by block size;
|
187
|
+
# output: next hash value Xi;
|
188
|
+
#
|
189
|
+
$code.=<<___;
|
190
|
+
.global gcm_ghash_v8
|
191
|
+
.type gcm_ghash_v8,%function
|
192
|
+
.align 4
|
193
|
+
gcm_ghash_v8:
|
194
|
+
___
|
195
|
+
$code.=<<___ if ($flavour !~ /64/);
|
196
|
+
vstmdb sp!,{d8-d15} @ 32-bit ABI says so
|
197
|
+
___
|
198
|
+
$code.=<<___;
|
199
|
+
vld1.64 {$Xl},[$Xi] @ load [rotated] Xi
|
200
|
+
@ "[rotated]" means that
|
201
|
+
@ loaded value would have
|
202
|
+
@ to be rotated in order to
|
203
|
+
@ make it appear as in
|
204
|
+
@ alorithm specification
|
205
|
+
subs $len,$len,#32 @ see if $len is 32 or larger
|
206
|
+
mov $inc,#16 @ $inc is used as post-
|
207
|
+
@ increment for input pointer;
|
208
|
+
@ as loop is modulo-scheduled
|
209
|
+
@ $inc is zeroed just in time
|
210
|
+
@ to preclude oversteping
|
211
|
+
@ inp[len], which means that
|
212
|
+
@ last block[s] are actually
|
213
|
+
@ loaded twice, but last
|
214
|
+
@ copy is not processed
|
215
|
+
vld1.64 {$H-$Hhl},[$Htbl],#32 @ load twisted H, ..., H^2
|
216
|
+
vmov.i8 $xC2,#0xe1
|
217
|
+
vld1.64 {$H2},[$Htbl]
|
218
|
+
cclr $inc,eq @ is it time to zero $inc?
|
219
|
+
vext.8 $Xl,$Xl,$Xl,#8 @ rotate Xi
|
220
|
+
vld1.64 {$t0},[$inp],#16 @ load [rotated] I[0]
|
221
|
+
vshl.u64 $xC2,$xC2,#57 @ compose 0xc2.0 constant
|
222
|
+
#ifndef __ARMEB__
|
223
|
+
vrev64.8 $t0,$t0
|
224
|
+
vrev64.8 $Xl,$Xl
|
225
|
+
#endif
|
226
|
+
vext.8 $IN,$t0,$t0,#8 @ rotate I[0]
|
227
|
+
b.lo .Lodd_tail_v8 @ $len was less than 32
|
228
|
+
___
|
229
|
+
{ my ($Xln,$Xmn,$Xhn,$In) = map("q$_",(4..7));
|
230
|
+
#######
|
231
|
+
# Xi+2 =[H*(Ii+1 + Xi+1)] mod P =
|
232
|
+
# [(H*Ii+1) + (H*Xi+1)] mod P =
|
233
|
+
# [(H*Ii+1) + H^2*(Ii+Xi)] mod P
|
234
|
+
#
|
235
|
+
$code.=<<___;
|
236
|
+
vld1.64 {$t1},[$inp],$inc @ load [rotated] I[1]
|
237
|
+
#ifndef __ARMEB__
|
238
|
+
vrev64.8 $t1,$t1
|
239
|
+
#endif
|
240
|
+
vext.8 $In,$t1,$t1,#8
|
241
|
+
veor $IN,$IN,$Xl @ I[i]^=Xi
|
242
|
+
vpmull.p64 $Xln,$H,$In @ H·Ii+1
|
243
|
+
veor $t1,$t1,$In @ Karatsuba pre-processing
|
244
|
+
vpmull2.p64 $Xhn,$H,$In
|
245
|
+
b .Loop_mod2x_v8
|
246
|
+
|
247
|
+
.align 4
|
248
|
+
.Loop_mod2x_v8:
|
249
|
+
vext.8 $t2,$IN,$IN,#8
|
250
|
+
subs $len,$len,#32 @ is there more data?
|
251
|
+
vpmull.p64 $Xl,$H2,$IN @ H^2.lo·Xi.lo
|
252
|
+
cclr $inc,lo @ is it time to zero $inc?
|
253
|
+
|
254
|
+
vpmull.p64 $Xmn,$Hhl,$t1
|
255
|
+
veor $t2,$t2,$IN @ Karatsuba pre-processing
|
256
|
+
vpmull2.p64 $Xh,$H2,$IN @ H^2.hi·Xi.hi
|
257
|
+
veor $Xl,$Xl,$Xln @ accumulate
|
258
|
+
vpmull2.p64 $Xm,$Hhl,$t2 @ (H^2.lo+H^2.hi)·(Xi.lo+Xi.hi)
|
259
|
+
vld1.64 {$t0},[$inp],$inc @ load [rotated] I[i+2]
|
260
|
+
|
261
|
+
veor $Xh,$Xh,$Xhn
|
262
|
+
cclr $inc,eq @ is it time to zero $inc?
|
263
|
+
veor $Xm,$Xm,$Xmn
|
264
|
+
|
265
|
+
vext.8 $t1,$Xl,$Xh,#8 @ Karatsuba post-processing
|
266
|
+
veor $t2,$Xl,$Xh
|
267
|
+
veor $Xm,$Xm,$t1
|
268
|
+
vld1.64 {$t1},[$inp],$inc @ load [rotated] I[i+3]
|
269
|
+
#ifndef __ARMEB__
|
270
|
+
vrev64.8 $t0,$t0
|
271
|
+
#endif
|
272
|
+
veor $Xm,$Xm,$t2
|
273
|
+
vpmull.p64 $t2,$Xl,$xC2 @ 1st phase of reduction
|
274
|
+
|
275
|
+
#ifndef __ARMEB__
|
276
|
+
vrev64.8 $t1,$t1
|
277
|
+
#endif
|
278
|
+
vmov $Xh#lo,$Xm#hi @ Xh|Xm - 256-bit result
|
279
|
+
vmov $Xm#hi,$Xl#lo @ Xm is rotated Xl
|
280
|
+
vext.8 $In,$t1,$t1,#8
|
281
|
+
vext.8 $IN,$t0,$t0,#8
|
282
|
+
veor $Xl,$Xm,$t2
|
283
|
+
vpmull.p64 $Xln,$H,$In @ H·Ii+1
|
284
|
+
veor $IN,$IN,$Xh @ accumulate $IN early
|
285
|
+
|
286
|
+
vext.8 $t2,$Xl,$Xl,#8 @ 2nd phase of reduction
|
287
|
+
vpmull.p64 $Xl,$Xl,$xC2
|
288
|
+
veor $IN,$IN,$t2
|
289
|
+
veor $t1,$t1,$In @ Karatsuba pre-processing
|
290
|
+
veor $IN,$IN,$Xl
|
291
|
+
vpmull2.p64 $Xhn,$H,$In
|
292
|
+
b.hs .Loop_mod2x_v8 @ there was at least 32 more bytes
|
293
|
+
|
294
|
+
veor $Xh,$Xh,$t2
|
295
|
+
vext.8 $IN,$t0,$t0,#8 @ re-construct $IN
|
296
|
+
adds $len,$len,#32 @ re-construct $len
|
297
|
+
veor $Xl,$Xl,$Xh @ re-construct $Xl
|
298
|
+
b.eq .Ldone_v8 @ is $len zero?
|
299
|
+
___
|
300
|
+
}
|
301
|
+
$code.=<<___;
|
302
|
+
.Lodd_tail_v8:
|
303
|
+
vext.8 $t2,$Xl,$Xl,#8
|
304
|
+
veor $IN,$IN,$Xl @ inp^=Xi
|
305
|
+
veor $t1,$t0,$t2 @ $t1 is rotated inp^Xi
|
306
|
+
|
307
|
+
vpmull.p64 $Xl,$H,$IN @ H.lo·Xi.lo
|
308
|
+
veor $t1,$t1,$IN @ Karatsuba pre-processing
|
309
|
+
vpmull2.p64 $Xh,$H,$IN @ H.hi·Xi.hi
|
310
|
+
vpmull.p64 $Xm,$Hhl,$t1 @ (H.lo+H.hi)·(Xi.lo+Xi.hi)
|
311
|
+
|
312
|
+
vext.8 $t1,$Xl,$Xh,#8 @ Karatsuba post-processing
|
313
|
+
veor $t2,$Xl,$Xh
|
314
|
+
veor $Xm,$Xm,$t1
|
315
|
+
veor $Xm,$Xm,$t2
|
316
|
+
vpmull.p64 $t2,$Xl,$xC2 @ 1st phase of reduction
|
317
|
+
|
318
|
+
vmov $Xh#lo,$Xm#hi @ Xh|Xm - 256-bit result
|
319
|
+
vmov $Xm#hi,$Xl#lo @ Xm is rotated Xl
|
320
|
+
veor $Xl,$Xm,$t2
|
321
|
+
|
322
|
+
vext.8 $t2,$Xl,$Xl,#8 @ 2nd phase of reduction
|
323
|
+
vpmull.p64 $Xl,$Xl,$xC2
|
324
|
+
veor $t2,$t2,$Xh
|
325
|
+
veor $Xl,$Xl,$t2
|
326
|
+
|
327
|
+
.Ldone_v8:
|
328
|
+
#ifndef __ARMEB__
|
329
|
+
vrev64.8 $Xl,$Xl
|
330
|
+
#endif
|
331
|
+
vext.8 $Xl,$Xl,$Xl,#8
|
332
|
+
vst1.64 {$Xl},[$Xi] @ write out Xi
|
333
|
+
|
334
|
+
___
|
335
|
+
$code.=<<___ if ($flavour !~ /64/);
|
336
|
+
vldmia sp!,{d8-d15} @ 32-bit ABI says so
|
337
|
+
___
|
338
|
+
$code.=<<___;
|
339
|
+
ret
|
340
|
+
.size gcm_ghash_v8,.-gcm_ghash_v8
|
341
|
+
___
|
342
|
+
}
|
343
|
+
$code.=<<___;
|
344
|
+
.asciz "GHASH for ARMv8, CRYPTOGAMS by <appro\@openssl.org>"
|
345
|
+
.align 2
|
346
|
+
___
|
347
|
+
|
348
|
+
if ($flavour =~ /64/) { ######## 64-bit code
|
349
|
+
sub unvmov {
|
350
|
+
my $arg=shift;
|
351
|
+
|
352
|
+
$arg =~ m/q([0-9]+)#(lo|hi),\s*q([0-9]+)#(lo|hi)/o &&
|
353
|
+
sprintf "ins v%d.d[%d],v%d.d[%d]",$1,($2 eq "lo")?0:1,$3,($4 eq "lo")?0:1;
|
354
|
+
}
|
355
|
+
foreach(split("\n",$code)) {
|
356
|
+
s/cclr\s+([wx])([^,]+),\s*([a-z]+)/csel $1$2,$1zr,$1$2,$3/o or
|
357
|
+
s/vmov\.i8/movi/o or # fix up legacy mnemonics
|
358
|
+
s/vmov\s+(.*)/unvmov($1)/geo or
|
359
|
+
s/vext\.8/ext/o or
|
360
|
+
s/vshr\.s/sshr\.s/o or
|
361
|
+
s/vshr/ushr/o or
|
362
|
+
s/^(\s+)v/$1/o or # strip off v prefix
|
363
|
+
s/\bbx\s+lr\b/ret/o;
|
364
|
+
|
365
|
+
s/\bq([0-9]+)\b/"v".($1<8?$1:$1+8).".16b"/geo; # old->new registers
|
366
|
+
s/@\s/\/\//o; # old->new style commentary
|
367
|
+
|
368
|
+
# fix up remainig legacy suffixes
|
369
|
+
s/\.[ui]?8(\s)/$1/o;
|
370
|
+
s/\.[uis]?32//o and s/\.16b/\.4s/go;
|
371
|
+
m/\.p64/o and s/\.16b/\.1q/o; # 1st pmull argument
|
372
|
+
m/l\.p64/o and s/\.16b/\.1d/go; # 2nd and 3rd pmull arguments
|
373
|
+
s/\.[uisp]?64//o and s/\.16b/\.2d/go;
|
374
|
+
s/\.[42]([sd])\[([0-3])\]/\.$1\[$2\]/o;
|
375
|
+
|
376
|
+
print $_,"\n";
|
377
|
+
}
|
378
|
+
} else { ######## 32-bit code
|
379
|
+
sub unvdup32 {
|
380
|
+
my $arg=shift;
|
381
|
+
|
382
|
+
$arg =~ m/q([0-9]+),\s*q([0-9]+)\[([0-3])\]/o &&
|
383
|
+
sprintf "vdup.32 q%d,d%d[%d]",$1,2*$2+($3>>1),$3&1;
|
384
|
+
}
|
385
|
+
sub unvpmullp64 {
|
386
|
+
my ($mnemonic,$arg)=@_;
|
387
|
+
|
388
|
+
if ($arg =~ m/q([0-9]+),\s*q([0-9]+),\s*q([0-9]+)/o) {
|
389
|
+
my $word = 0xf2a00e00|(($1&7)<<13)|(($1&8)<<19)
|
390
|
+
|(($2&7)<<17)|(($2&8)<<4)
|
391
|
+
|(($3&7)<<1) |(($3&8)<<2);
|
392
|
+
$word |= 0x00010001 if ($mnemonic =~ "2");
|
393
|
+
# since ARMv7 instructions are always encoded little-endian.
|
394
|
+
# correct solution is to use .inst directive, but older
|
395
|
+
# assemblers don't implement it:-(
|
396
|
+
sprintf ".byte\t0x%02x,0x%02x,0x%02x,0x%02x\t@ %s %s",
|
397
|
+
$word&0xff,($word>>8)&0xff,
|
398
|
+
($word>>16)&0xff,($word>>24)&0xff,
|
399
|
+
$mnemonic,$arg;
|
400
|
+
}
|
401
|
+
}
|
402
|
+
|
403
|
+
foreach(split("\n",$code)) {
|
404
|
+
s/\b[wx]([0-9]+)\b/r$1/go; # new->old registers
|
405
|
+
s/\bv([0-9])\.[12468]+[bsd]\b/q$1/go; # new->old registers
|
406
|
+
s/\/\/\s?/@ /o; # new->old style commentary
|
407
|
+
|
408
|
+
# fix up remainig new-style suffixes
|
409
|
+
s/\],#[0-9]+/]!/o;
|
410
|
+
|
411
|
+
s/cclr\s+([^,]+),\s*([a-z]+)/mov$2 $1,#0/o or
|
412
|
+
s/vdup\.32\s+(.*)/unvdup32($1)/geo or
|
413
|
+
s/v?(pmull2?)\.p64\s+(.*)/unvpmullp64($1,$2)/geo or
|
414
|
+
s/\bq([0-9]+)#(lo|hi)/sprintf "d%d",2*$1+($2 eq "hi")/geo or
|
415
|
+
s/^(\s+)b\./$1b/o or
|
416
|
+
s/^(\s+)ret/$1bx\tlr/o;
|
417
|
+
|
418
|
+
print $_,"\n";
|
419
|
+
}
|
420
|
+
}
|
421
|
+
|
422
|
+
close STDOUT; # enforce flush
|
@@ -0,0 +1,226 @@
|
|
1
|
+
/* ====================================================================
|
2
|
+
* Copyright (c) 2008 The OpenSSL Project. All rights reserved.
|
3
|
+
*
|
4
|
+
* Redistribution and use in source and binary forms, with or without
|
5
|
+
* modification, are permitted provided that the following conditions
|
6
|
+
* are met:
|
7
|
+
*
|
8
|
+
* 1. Redistributions of source code must retain the above copyright
|
9
|
+
* notice, this list of conditions and the following disclaimer.
|
10
|
+
*
|
11
|
+
* 2. Redistributions in binary form must reproduce the above copyright
|
12
|
+
* notice, this list of conditions and the following disclaimer in
|
13
|
+
* the documentation and/or other materials provided with the
|
14
|
+
* distribution.
|
15
|
+
*
|
16
|
+
* 3. All advertising materials mentioning features or use of this
|
17
|
+
* software must display the following acknowledgment:
|
18
|
+
* "This product includes software developed by the OpenSSL Project
|
19
|
+
* for use in the OpenSSL Toolkit. (http://www.openssl.org/)"
|
20
|
+
*
|
21
|
+
* 4. The names "OpenSSL Toolkit" and "OpenSSL Project" must not be used to
|
22
|
+
* endorse or promote products derived from this software without
|
23
|
+
* prior written permission. For written permission, please contact
|
24
|
+
* openssl-core@openssl.org.
|
25
|
+
*
|
26
|
+
* 5. Products derived from this software may not be called "OpenSSL"
|
27
|
+
* nor may "OpenSSL" appear in their names without prior written
|
28
|
+
* permission of the OpenSSL Project.
|
29
|
+
*
|
30
|
+
* 6. Redistributions of any form whatsoever must retain the following
|
31
|
+
* acknowledgment:
|
32
|
+
* "This product includes software developed by the OpenSSL Project
|
33
|
+
* for use in the OpenSSL Toolkit (http://www.openssl.org/)"
|
34
|
+
*
|
35
|
+
* THIS SOFTWARE IS PROVIDED BY THE OpenSSL PROJECT ``AS IS'' AND ANY
|
36
|
+
* EXPRESSED OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
37
|
+
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
|
38
|
+
* PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE OpenSSL PROJECT OR
|
39
|
+
* ITS CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
40
|
+
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
|
41
|
+
* NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
|
42
|
+
* LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
43
|
+
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT,
|
44
|
+
* STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
45
|
+
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED
|
46
|
+
* OF THE POSSIBILITY OF SUCH DAMAGE.
|
47
|
+
* ==================================================================== */
|
48
|
+
|
49
|
+
#include <openssl/type_check.h>
|
50
|
+
|
51
|
+
#include <assert.h>
|
52
|
+
#include <string.h>
|
53
|
+
|
54
|
+
#include "internal.h"
|
55
|
+
|
56
|
+
|
57
|
+
/* NOTE: the IV/counter CTR mode is big-endian. The code itself
|
58
|
+
* is endian-neutral. */
|
59
|
+
|
60
|
+
/* increment counter (128-bit int) by 1 */
|
61
|
+
static void ctr128_inc(uint8_t *counter) {
|
62
|
+
uint32_t n = 16;
|
63
|
+
uint8_t c;
|
64
|
+
|
65
|
+
do {
|
66
|
+
--n;
|
67
|
+
c = counter[n];
|
68
|
+
++c;
|
69
|
+
counter[n] = c;
|
70
|
+
if (c) {
|
71
|
+
return;
|
72
|
+
}
|
73
|
+
} while (n);
|
74
|
+
}
|
75
|
+
|
76
|
+
OPENSSL_COMPILE_ASSERT((16 % sizeof(size_t)) == 0, bad_size_t_size);
|
77
|
+
|
78
|
+
/* The input encrypted as though 128bit counter mode is being used. The extra
|
79
|
+
* state information to record how much of the 128bit block we have used is
|
80
|
+
* contained in *num, and the encrypted counter is kept in ecount_buf. Both
|
81
|
+
* *num and ecount_buf must be initialised with zeros before the first call to
|
82
|
+
* CRYPTO_ctr128_encrypt().
|
83
|
+
*
|
84
|
+
* This algorithm assumes that the counter is in the x lower bits of the IV
|
85
|
+
* (ivec), and that the application has full control over overflow and the rest
|
86
|
+
* of the IV. This implementation takes NO responsibility for checking that
|
87
|
+
* the counter doesn't overflow into the rest of the IV when incremented. */
|
88
|
+
void CRYPTO_ctr128_encrypt(const uint8_t *in, uint8_t *out, size_t len,
|
89
|
+
const void *key, uint8_t ivec[16],
|
90
|
+
uint8_t ecount_buf[16], unsigned int *num,
|
91
|
+
block128_f block) {
|
92
|
+
unsigned int n;
|
93
|
+
|
94
|
+
assert(key && ecount_buf && num);
|
95
|
+
assert(len == 0 || (in && out));
|
96
|
+
assert(*num < 16);
|
97
|
+
|
98
|
+
n = *num;
|
99
|
+
|
100
|
+
while (n && len) {
|
101
|
+
*(out++) = *(in++) ^ ecount_buf[n];
|
102
|
+
--len;
|
103
|
+
n = (n + 1) % 16;
|
104
|
+
}
|
105
|
+
|
106
|
+
#if STRICT_ALIGNMENT
|
107
|
+
if (((size_t)in | (size_t)out | (size_t)ivec) % sizeof(size_t) != 0) {
|
108
|
+
size_t l = 0;
|
109
|
+
while (l < len) {
|
110
|
+
if (n == 0) {
|
111
|
+
(*block)(ivec, ecount_buf, key);
|
112
|
+
ctr128_inc(ivec);
|
113
|
+
}
|
114
|
+
out[l] = in[l] ^ ecount_buf[n];
|
115
|
+
++l;
|
116
|
+
n = (n + 1) % 16;
|
117
|
+
}
|
118
|
+
|
119
|
+
*num = n;
|
120
|
+
return;
|
121
|
+
}
|
122
|
+
#endif
|
123
|
+
|
124
|
+
while (len >= 16) {
|
125
|
+
(*block)(ivec, ecount_buf, key);
|
126
|
+
ctr128_inc(ivec);
|
127
|
+
for (; n < 16; n += sizeof(size_t)) {
|
128
|
+
*(size_t *)(out + n) = *(size_t *)(in + n) ^ *(size_t *)(ecount_buf + n);
|
129
|
+
}
|
130
|
+
len -= 16;
|
131
|
+
out += 16;
|
132
|
+
in += 16;
|
133
|
+
n = 0;
|
134
|
+
}
|
135
|
+
if (len) {
|
136
|
+
(*block)(ivec, ecount_buf, key);
|
137
|
+
ctr128_inc(ivec);
|
138
|
+
while (len--) {
|
139
|
+
out[n] = in[n] ^ ecount_buf[n];
|
140
|
+
++n;
|
141
|
+
}
|
142
|
+
}
|
143
|
+
*num = n;
|
144
|
+
}
|
145
|
+
|
146
|
+
/* increment upper 96 bits of 128-bit counter by 1 */
|
147
|
+
static void ctr96_inc(uint8_t *counter) {
|
148
|
+
uint32_t n = 12;
|
149
|
+
uint8_t c;
|
150
|
+
|
151
|
+
do {
|
152
|
+
--n;
|
153
|
+
c = counter[n];
|
154
|
+
++c;
|
155
|
+
counter[n] = c;
|
156
|
+
if (c) {
|
157
|
+
return;
|
158
|
+
}
|
159
|
+
} while (n);
|
160
|
+
}
|
161
|
+
|
162
|
+
void CRYPTO_ctr128_encrypt_ctr32(const uint8_t *in, uint8_t *out,
|
163
|
+
size_t len, const void *key,
|
164
|
+
uint8_t ivec[16],
|
165
|
+
uint8_t ecount_buf[16],
|
166
|
+
unsigned int *num, ctr128_f func) {
|
167
|
+
unsigned int n, ctr32;
|
168
|
+
|
169
|
+
assert(key && ecount_buf && num);
|
170
|
+
assert(len == 0 || (in && out));
|
171
|
+
assert(*num < 16);
|
172
|
+
|
173
|
+
n = *num;
|
174
|
+
|
175
|
+
while (n && len) {
|
176
|
+
*(out++) = *(in++) ^ ecount_buf[n];
|
177
|
+
--len;
|
178
|
+
n = (n + 1) % 16;
|
179
|
+
}
|
180
|
+
|
181
|
+
ctr32 = GETU32(ivec + 12);
|
182
|
+
while (len >= 16) {
|
183
|
+
size_t blocks = len / 16;
|
184
|
+
/* 1<<28 is just a not-so-small yet not-so-large number...
|
185
|
+
* Below condition is practically never met, but it has to
|
186
|
+
* be checked for code correctness. */
|
187
|
+
if (sizeof(size_t) > sizeof(unsigned int) && blocks > (1U << 28)) {
|
188
|
+
blocks = (1U << 28);
|
189
|
+
}
|
190
|
+
/* As (*func) operates on 32-bit counter, caller
|
191
|
+
* has to handle overflow. 'if' below detects the
|
192
|
+
* overflow, which is then handled by limiting the
|
193
|
+
* amount of blocks to the exact overflow point... */
|
194
|
+
ctr32 += (uint32_t)blocks;
|
195
|
+
if (ctr32 < blocks) {
|
196
|
+
blocks -= ctr32;
|
197
|
+
ctr32 = 0;
|
198
|
+
}
|
199
|
+
(*func)(in, out, blocks, key, ivec);
|
200
|
+
/* (*func) does not update ivec, caller does: */
|
201
|
+
PUTU32(ivec + 12, ctr32);
|
202
|
+
/* ... overflow was detected, propogate carry. */
|
203
|
+
if (ctr32 == 0) {
|
204
|
+
ctr96_inc(ivec);
|
205
|
+
}
|
206
|
+
blocks *= 16;
|
207
|
+
len -= blocks;
|
208
|
+
out += blocks;
|
209
|
+
in += blocks;
|
210
|
+
}
|
211
|
+
if (len) {
|
212
|
+
memset(ecount_buf, 0, 16);
|
213
|
+
(*func)(ecount_buf, ecount_buf, 1, key, ivec);
|
214
|
+
++ctr32;
|
215
|
+
PUTU32(ivec + 12, ctr32);
|
216
|
+
if (ctr32 == 0) {
|
217
|
+
ctr96_inc(ivec);
|
218
|
+
}
|
219
|
+
while (len--) {
|
220
|
+
out[n] = in[n] ^ ecount_buf[n];
|
221
|
+
++n;
|
222
|
+
}
|
223
|
+
}
|
224
|
+
|
225
|
+
*num = n;
|
226
|
+
}
|