zstd-ruby 1.4.5.0 → 1.4.9.0
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/.github/workflows/ruby.yml +35 -0
- data/README.md +2 -2
- data/ext/zstdruby/libzstd/Makefile +237 -138
- data/ext/zstdruby/libzstd/README.md +28 -0
- data/ext/zstdruby/libzstd/common/bitstream.h +25 -16
- data/ext/zstdruby/libzstd/common/compiler.h +118 -4
- data/ext/zstdruby/libzstd/common/cpu.h +1 -3
- data/ext/zstdruby/libzstd/common/debug.c +1 -1
- data/ext/zstdruby/libzstd/common/debug.h +12 -19
- data/ext/zstdruby/libzstd/common/entropy_common.c +189 -43
- data/ext/zstdruby/libzstd/common/error_private.c +2 -1
- data/ext/zstdruby/libzstd/common/error_private.h +2 -2
- data/ext/zstdruby/libzstd/common/fse.h +40 -12
- data/ext/zstdruby/libzstd/common/fse_decompress.c +124 -17
- data/ext/zstdruby/libzstd/common/huf.h +27 -6
- data/ext/zstdruby/libzstd/common/mem.h +67 -94
- data/ext/zstdruby/libzstd/common/pool.c +23 -17
- data/ext/zstdruby/libzstd/common/pool.h +2 -2
- data/ext/zstdruby/libzstd/common/threading.c +6 -5
- data/ext/zstdruby/libzstd/common/xxhash.c +19 -57
- data/ext/zstdruby/libzstd/common/xxhash.h +2 -2
- data/ext/zstdruby/libzstd/common/zstd_common.c +10 -10
- data/ext/zstdruby/libzstd/common/zstd_deps.h +111 -0
- data/ext/zstdruby/libzstd/common/zstd_errors.h +2 -1
- data/ext/zstdruby/libzstd/common/zstd_internal.h +90 -59
- data/ext/zstdruby/libzstd/common/zstd_trace.c +42 -0
- data/ext/zstdruby/libzstd/common/zstd_trace.h +152 -0
- data/ext/zstdruby/libzstd/compress/fse_compress.c +31 -24
- data/ext/zstdruby/libzstd/compress/hist.c +27 -29
- data/ext/zstdruby/libzstd/compress/hist.h +2 -2
- data/ext/zstdruby/libzstd/compress/huf_compress.c +217 -101
- data/ext/zstdruby/libzstd/compress/zstd_compress.c +1495 -478
- data/ext/zstdruby/libzstd/compress/zstd_compress_internal.h +143 -44
- data/ext/zstdruby/libzstd/compress/zstd_compress_literals.c +7 -7
- data/ext/zstdruby/libzstd/compress/zstd_compress_literals.h +1 -1
- data/ext/zstdruby/libzstd/compress/zstd_compress_sequences.c +18 -4
- data/ext/zstdruby/libzstd/compress/zstd_compress_sequences.h +1 -1
- data/ext/zstdruby/libzstd/compress/zstd_compress_superblock.c +25 -21
- data/ext/zstdruby/libzstd/compress/zstd_compress_superblock.h +1 -1
- data/ext/zstdruby/libzstd/compress/zstd_cwksp.h +62 -26
- data/ext/zstdruby/libzstd/compress/zstd_double_fast.c +23 -23
- data/ext/zstdruby/libzstd/compress/zstd_double_fast.h +1 -1
- data/ext/zstdruby/libzstd/compress/zstd_fast.c +21 -21
- data/ext/zstdruby/libzstd/compress/zstd_fast.h +1 -1
- data/ext/zstdruby/libzstd/compress/zstd_lazy.c +352 -78
- data/ext/zstdruby/libzstd/compress/zstd_lazy.h +21 -1
- data/ext/zstdruby/libzstd/compress/zstd_ldm.c +276 -209
- data/ext/zstdruby/libzstd/compress/zstd_ldm.h +8 -2
- data/ext/zstdruby/libzstd/compress/zstd_ldm_geartab.h +103 -0
- data/ext/zstdruby/libzstd/compress/zstd_opt.c +191 -46
- data/ext/zstdruby/libzstd/compress/zstd_opt.h +1 -1
- data/ext/zstdruby/libzstd/compress/zstdmt_compress.c +79 -410
- data/ext/zstdruby/libzstd/compress/zstdmt_compress.h +27 -109
- data/ext/zstdruby/libzstd/decompress/huf_decompress.c +303 -201
- data/ext/zstdruby/libzstd/decompress/zstd_ddict.c +9 -9
- data/ext/zstdruby/libzstd/decompress/zstd_ddict.h +2 -2
- data/ext/zstdruby/libzstd/decompress/zstd_decompress.c +370 -87
- data/ext/zstdruby/libzstd/decompress/zstd_decompress_block.c +153 -45
- data/ext/zstdruby/libzstd/decompress/zstd_decompress_block.h +6 -3
- data/ext/zstdruby/libzstd/decompress/zstd_decompress_internal.h +28 -11
- data/ext/zstdruby/libzstd/deprecated/zbuff.h +1 -1
- data/ext/zstdruby/libzstd/deprecated/zbuff_common.c +1 -1
- data/ext/zstdruby/libzstd/deprecated/zbuff_compress.c +1 -1
- data/ext/zstdruby/libzstd/deprecated/zbuff_decompress.c +1 -1
- data/ext/zstdruby/libzstd/dictBuilder/cover.c +40 -31
- data/ext/zstdruby/libzstd/dictBuilder/cover.h +2 -2
- data/ext/zstdruby/libzstd/dictBuilder/divsufsort.c +1 -1
- data/ext/zstdruby/libzstd/dictBuilder/fastcover.c +26 -25
- data/ext/zstdruby/libzstd/dictBuilder/zdict.c +22 -24
- data/ext/zstdruby/libzstd/dictBuilder/zdict.h +5 -4
- data/ext/zstdruby/libzstd/dll/example/Makefile +1 -1
- data/ext/zstdruby/libzstd/dll/example/README.md +16 -22
- data/ext/zstdruby/libzstd/legacy/zstd_legacy.h +1 -1
- data/ext/zstdruby/libzstd/legacy/zstd_v01.c +6 -2
- data/ext/zstdruby/libzstd/legacy/zstd_v01.h +1 -1
- data/ext/zstdruby/libzstd/legacy/zstd_v02.c +6 -2
- data/ext/zstdruby/libzstd/legacy/zstd_v02.h +1 -1
- data/ext/zstdruby/libzstd/legacy/zstd_v03.c +6 -2
- data/ext/zstdruby/libzstd/legacy/zstd_v03.h +1 -1
- data/ext/zstdruby/libzstd/legacy/zstd_v04.c +7 -3
- data/ext/zstdruby/libzstd/legacy/zstd_v04.h +1 -1
- data/ext/zstdruby/libzstd/legacy/zstd_v05.c +10 -6
- data/ext/zstdruby/libzstd/legacy/zstd_v05.h +1 -1
- data/ext/zstdruby/libzstd/legacy/zstd_v06.c +10 -6
- data/ext/zstdruby/libzstd/legacy/zstd_v06.h +1 -1
- data/ext/zstdruby/libzstd/legacy/zstd_v07.c +10 -6
- data/ext/zstdruby/libzstd/legacy/zstd_v07.h +1 -1
- data/ext/zstdruby/libzstd/libzstd.pc.in +3 -3
- data/ext/zstdruby/libzstd/zstd.h +414 -54
- data/lib/zstd-ruby/version.rb +1 -1
- metadata +7 -3
- data/.travis.yml +0 -14
@@ -143,6 +143,14 @@ The file structure is designed to make this selection manually achievable for an
|
|
143
143
|
Setting this macro will either force to generate the BMI2 dispatcher (1)
|
144
144
|
or prevent it (0). It overrides automatic detection.
|
145
145
|
|
146
|
+
- The build macro `ZSTD_NO_UNUSED_FUNCTIONS` can be defined to hide the definitions of functions
|
147
|
+
that zstd does not use. Not all unused functions are hidden, but they can be if needed.
|
148
|
+
Currently, this macro will hide function definitions in FSE and HUF that use an excessive
|
149
|
+
amount of stack space.
|
150
|
+
|
151
|
+
- The build macro `ZSTD_NO_INTRINSICS` can be defined to disable all explicit intrinsics.
|
152
|
+
Compiler builtins are still used.
|
153
|
+
|
146
154
|
|
147
155
|
#### Windows : using MinGW+MSYS to create DLL
|
148
156
|
|
@@ -160,6 +168,26 @@ file it should be linked with `dll\libzstd.dll`. For example:
|
|
160
168
|
The compiled executable will require ZSTD DLL which is available at `dll\libzstd.dll`.
|
161
169
|
|
162
170
|
|
171
|
+
#### Advanced Build options
|
172
|
+
|
173
|
+
The build system requires a hash function in order to
|
174
|
+
separate object files created with different compilation flags.
|
175
|
+
By default, it tries to use `md5sum` or equivalent.
|
176
|
+
The hash function can be manually switched by setting the `HASH` variable.
|
177
|
+
For example : `make HASH=xxhsum`
|
178
|
+
The hash function needs to generate at least 64-bit using hexadecimal format.
|
179
|
+
When no hash function is found,
|
180
|
+
the Makefile just generates all object files into the same default directory,
|
181
|
+
irrespective of compilation flags.
|
182
|
+
This functionality only matters if `libzstd` is compiled multiple times
|
183
|
+
with different build flags.
|
184
|
+
|
185
|
+
The build directory, where object files are stored
|
186
|
+
can also be manually controlled using variable `BUILD_DIR`,
|
187
|
+
for example `make BUILD_DIR=objectDir/v1`.
|
188
|
+
In which case, the hash function doesn't matter.
|
189
|
+
|
190
|
+
|
163
191
|
#### Deprecated API
|
164
192
|
|
165
193
|
Obsolete API on their way out are stored in directory `lib/deprecated`.
|
@@ -1,7 +1,7 @@
|
|
1
1
|
/* ******************************************************************
|
2
2
|
* bitstream
|
3
3
|
* Part of FSE library
|
4
|
-
* Copyright (c) 2013-
|
4
|
+
* Copyright (c) 2013-2021, Yann Collet, Facebook, Inc.
|
5
5
|
*
|
6
6
|
* You can contact the author at :
|
7
7
|
* - Source repository : https://github.com/Cyan4973/FiniteStateEntropy
|
@@ -17,7 +17,6 @@
|
|
17
17
|
#if defined (__cplusplus)
|
18
18
|
extern "C" {
|
19
19
|
#endif
|
20
|
-
|
21
20
|
/*
|
22
21
|
* This API consists of small unitary functions, which must be inlined for best performance.
|
23
22
|
* Since link-time-optimization is not available for all compilers,
|
@@ -36,10 +35,12 @@ extern "C" {
|
|
36
35
|
/*=========================================
|
37
36
|
* Target specific
|
38
37
|
=========================================*/
|
39
|
-
#
|
40
|
-
#
|
41
|
-
#
|
42
|
-
#
|
38
|
+
#ifndef ZSTD_NO_INTRINSICS
|
39
|
+
# if defined(__BMI__) && defined(__GNUC__)
|
40
|
+
# include <immintrin.h> /* support for bextr (experimental) */
|
41
|
+
# elif defined(__ICCARM__)
|
42
|
+
# include <intrinsics.h>
|
43
|
+
# endif
|
43
44
|
#endif
|
44
45
|
|
45
46
|
#define STREAM_ACCUMULATOR_MIN_32 25
|
@@ -141,8 +142,12 @@ MEM_STATIC unsigned BIT_highbit32 (U32 val)
|
|
141
142
|
assert(val != 0);
|
142
143
|
{
|
143
144
|
# if defined(_MSC_VER) /* Visual */
|
144
|
-
|
145
|
-
|
145
|
+
# if STATIC_BMI2 == 1
|
146
|
+
return _lzcnt_u32(val) ^ 31;
|
147
|
+
# else
|
148
|
+
unsigned long r = 0;
|
149
|
+
return _BitScanReverse(&r, val) ? (unsigned)r : 0;
|
150
|
+
# endif
|
146
151
|
# elif defined(__GNUC__) && (__GNUC__ >= 3) /* Use GCC Intrinsic */
|
147
152
|
return __builtin_clz (val) ^ 31;
|
148
153
|
# elif defined(__ICCARM__) /* IAR Intrinsic */
|
@@ -198,7 +203,7 @@ MEM_STATIC size_t BIT_initCStream(BIT_CStream_t* bitC,
|
|
198
203
|
MEM_STATIC void BIT_addBits(BIT_CStream_t* bitC,
|
199
204
|
size_t value, unsigned nbBits)
|
200
205
|
{
|
201
|
-
|
206
|
+
DEBUG_STATIC_ASSERT(BIT_MASK_SIZE == 32);
|
202
207
|
assert(nbBits < BIT_MASK_SIZE);
|
203
208
|
assert(nbBits + bitC->bitPos < sizeof(bitC->bitContainer) * 8);
|
204
209
|
bitC->bitContainer |= (value & BIT_mask[nbBits]) << bitC->bitPos;
|
@@ -271,7 +276,7 @@ MEM_STATIC size_t BIT_closeCStream(BIT_CStream_t* bitC)
|
|
271
276
|
*/
|
272
277
|
MEM_STATIC size_t BIT_initDStream(BIT_DStream_t* bitD, const void* srcBuffer, size_t srcSize)
|
273
278
|
{
|
274
|
-
if (srcSize < 1) {
|
279
|
+
if (srcSize < 1) { ZSTD_memset(bitD, 0, sizeof(*bitD)); return ERROR(srcSize_wrong); }
|
275
280
|
|
276
281
|
bitD->start = (const char*)srcBuffer;
|
277
282
|
bitD->limitPtr = bitD->start + sizeof(bitD->bitContainer);
|
@@ -317,12 +322,12 @@ MEM_STATIC size_t BIT_initDStream(BIT_DStream_t* bitD, const void* srcBuffer, si
|
|
317
322
|
return srcSize;
|
318
323
|
}
|
319
324
|
|
320
|
-
MEM_STATIC size_t BIT_getUpperBits(size_t bitContainer, U32 const start)
|
325
|
+
MEM_STATIC FORCE_INLINE_ATTR size_t BIT_getUpperBits(size_t bitContainer, U32 const start)
|
321
326
|
{
|
322
327
|
return bitContainer >> start;
|
323
328
|
}
|
324
329
|
|
325
|
-
MEM_STATIC size_t BIT_getMiddleBits(size_t bitContainer, U32 const start, U32 const nbBits)
|
330
|
+
MEM_STATIC FORCE_INLINE_ATTR size_t BIT_getMiddleBits(size_t bitContainer, U32 const start, U32 const nbBits)
|
326
331
|
{
|
327
332
|
U32 const regMask = sizeof(bitContainer)*8 - 1;
|
328
333
|
/* if start > regMask, bitstream is corrupted, and result is undefined */
|
@@ -330,10 +335,14 @@ MEM_STATIC size_t BIT_getMiddleBits(size_t bitContainer, U32 const start, U32 co
|
|
330
335
|
return (bitContainer >> (start & regMask)) & BIT_mask[nbBits];
|
331
336
|
}
|
332
337
|
|
333
|
-
MEM_STATIC size_t BIT_getLowerBits(size_t bitContainer, U32 const nbBits)
|
338
|
+
MEM_STATIC FORCE_INLINE_ATTR size_t BIT_getLowerBits(size_t bitContainer, U32 const nbBits)
|
334
339
|
{
|
340
|
+
#if defined(STATIC_BMI2) && STATIC_BMI2 == 1
|
341
|
+
return _bzhi_u64(bitContainer, nbBits);
|
342
|
+
#else
|
335
343
|
assert(nbBits < BIT_MASK_SIZE);
|
336
344
|
return bitContainer & BIT_mask[nbBits];
|
345
|
+
#endif
|
337
346
|
}
|
338
347
|
|
339
348
|
/*! BIT_lookBits() :
|
@@ -342,7 +351,7 @@ MEM_STATIC size_t BIT_getLowerBits(size_t bitContainer, U32 const nbBits)
|
|
342
351
|
* On 32-bits, maxNbBits==24.
|
343
352
|
* On 64-bits, maxNbBits==56.
|
344
353
|
* @return : value extracted */
|
345
|
-
MEM_STATIC size_t BIT_lookBits(const BIT_DStream_t*
|
354
|
+
MEM_STATIC FORCE_INLINE_ATTR size_t BIT_lookBits(const BIT_DStream_t* bitD, U32 nbBits)
|
346
355
|
{
|
347
356
|
/* arbitrate between double-shift and shift+mask */
|
348
357
|
#if 1
|
@@ -365,7 +374,7 @@ MEM_STATIC size_t BIT_lookBitsFast(const BIT_DStream_t* bitD, U32 nbBits)
|
|
365
374
|
return (bitD->bitContainer << (bitD->bitsConsumed & regMask)) >> (((regMask+1)-nbBits) & regMask);
|
366
375
|
}
|
367
376
|
|
368
|
-
MEM_STATIC void BIT_skipBits(BIT_DStream_t* bitD, U32 nbBits)
|
377
|
+
MEM_STATIC FORCE_INLINE_ATTR void BIT_skipBits(BIT_DStream_t* bitD, U32 nbBits)
|
369
378
|
{
|
370
379
|
bitD->bitsConsumed += nbBits;
|
371
380
|
}
|
@@ -374,7 +383,7 @@ MEM_STATIC void BIT_skipBits(BIT_DStream_t* bitD, U32 nbBits)
|
|
374
383
|
* Read (consume) next n bits from local register and update.
|
375
384
|
* Pay attention to not read more than nbBits contained into local register.
|
376
385
|
* @return : extracted value. */
|
377
|
-
MEM_STATIC size_t BIT_readBits(BIT_DStream_t* bitD, unsigned nbBits)
|
386
|
+
MEM_STATIC FORCE_INLINE_ATTR size_t BIT_readBits(BIT_DStream_t* bitD, unsigned nbBits)
|
378
387
|
{
|
379
388
|
size_t const value = BIT_lookBits(bitD, nbBits);
|
380
389
|
BIT_skipBits(bitD, nbBits);
|
@@ -1,5 +1,5 @@
|
|
1
1
|
/*
|
2
|
-
* Copyright (c) 2016-
|
2
|
+
* Copyright (c) 2016-2021, Yann Collet, Facebook, Inc.
|
3
3
|
* All rights reserved.
|
4
4
|
*
|
5
5
|
* This source code is licensed under both the BSD-style license (found in the
|
@@ -38,6 +38,17 @@
|
|
38
38
|
|
39
39
|
#endif
|
40
40
|
|
41
|
+
/**
|
42
|
+
On MSVC qsort requires that functions passed into it use the __cdecl calling conversion(CC).
|
43
|
+
This explictly marks such functions as __cdecl so that the code will still compile
|
44
|
+
if a CC other than __cdecl has been made the default.
|
45
|
+
*/
|
46
|
+
#if defined(_MSC_VER)
|
47
|
+
# define WIN_CDECL __cdecl
|
48
|
+
#else
|
49
|
+
# define WIN_CDECL
|
50
|
+
#endif
|
51
|
+
|
41
52
|
/**
|
42
53
|
* FORCE_INLINE_TEMPLATE is used to define C "templates", which take constant
|
43
54
|
* parameters. They must be inlined for the compiler to eliminate the constant
|
@@ -79,6 +90,7 @@
|
|
79
90
|
# endif
|
80
91
|
#endif
|
81
92
|
|
93
|
+
|
82
94
|
/* target attribute */
|
83
95
|
#ifndef __has_attribute
|
84
96
|
#define __has_attribute(x) 0 /* Compatibility with non-clang compilers. */
|
@@ -114,12 +126,12 @@
|
|
114
126
|
# include <mmintrin.h> /* https://msdn.microsoft.com/fr-fr/library/84szxsww(v=vs.90).aspx */
|
115
127
|
# define PREFETCH_L1(ptr) _mm_prefetch((const char*)(ptr), _MM_HINT_T0)
|
116
128
|
# define PREFETCH_L2(ptr) _mm_prefetch((const char*)(ptr), _MM_HINT_T1)
|
117
|
-
# elif defined(__aarch64__)
|
118
|
-
# define PREFETCH_L1(ptr) __asm__ __volatile__("prfm pldl1keep, %0" ::"Q"(*(ptr)))
|
119
|
-
# define PREFETCH_L2(ptr) __asm__ __volatile__("prfm pldl2keep, %0" ::"Q"(*(ptr)))
|
120
129
|
# elif defined(__GNUC__) && ( (__GNUC__ >= 4) || ( (__GNUC__ == 3) && (__GNUC_MINOR__ >= 1) ) )
|
121
130
|
# define PREFETCH_L1(ptr) __builtin_prefetch((ptr), 0 /* rw==read */, 3 /* locality */)
|
122
131
|
# define PREFETCH_L2(ptr) __builtin_prefetch((ptr), 0 /* rw==read */, 2 /* locality */)
|
132
|
+
# elif defined(__aarch64__)
|
133
|
+
# define PREFETCH_L1(ptr) __asm__ __volatile__("prfm pldl1keep, %0" ::"Q"(*(ptr)))
|
134
|
+
# define PREFETCH_L2(ptr) __asm__ __volatile__("prfm pldl2keep, %0" ::"Q"(*(ptr)))
|
123
135
|
# else
|
124
136
|
# define PREFETCH_L1(ptr) (void)(ptr) /* disabled */
|
125
137
|
# define PREFETCH_L2(ptr) (void)(ptr) /* disabled */
|
@@ -172,4 +184,106 @@
|
|
172
184
|
# pragma warning(disable : 4324) /* disable: C4324: padded structure */
|
173
185
|
#endif
|
174
186
|
|
187
|
+
/*Like DYNAMIC_BMI2 but for compile time determination of BMI2 support*/
|
188
|
+
#ifndef STATIC_BMI2
|
189
|
+
# if defined(_MSC_VER) && (defined(_M_X64) || defined(_M_I86))
|
190
|
+
# ifdef __AVX2__ //MSVC does not have a BMI2 specific flag, but every CPU that supports AVX2 also supports BMI2
|
191
|
+
# define STATIC_BMI2 1
|
192
|
+
# endif
|
193
|
+
# endif
|
194
|
+
#endif
|
195
|
+
|
196
|
+
#ifndef STATIC_BMI2
|
197
|
+
#define STATIC_BMI2 0
|
198
|
+
#endif
|
199
|
+
|
200
|
+
/* compat. with non-clang compilers */
|
201
|
+
#ifndef __has_builtin
|
202
|
+
# define __has_builtin(x) 0
|
203
|
+
#endif
|
204
|
+
|
205
|
+
/* compat. with non-clang compilers */
|
206
|
+
#ifndef __has_feature
|
207
|
+
# define __has_feature(x) 0
|
208
|
+
#endif
|
209
|
+
|
210
|
+
/* detects whether we are being compiled under msan */
|
211
|
+
#ifndef ZSTD_MEMORY_SANITIZER
|
212
|
+
# if __has_feature(memory_sanitizer)
|
213
|
+
# define ZSTD_MEMORY_SANITIZER 1
|
214
|
+
# else
|
215
|
+
# define ZSTD_MEMORY_SANITIZER 0
|
216
|
+
# endif
|
217
|
+
#endif
|
218
|
+
|
219
|
+
#if ZSTD_MEMORY_SANITIZER
|
220
|
+
/* Not all platforms that support msan provide sanitizers/msan_interface.h.
|
221
|
+
* We therefore declare the functions we need ourselves, rather than trying to
|
222
|
+
* include the header file... */
|
223
|
+
#include <stddef.h> /* size_t */
|
224
|
+
#define ZSTD_DEPS_NEED_STDINT
|
225
|
+
#include "zstd_deps.h" /* intptr_t */
|
226
|
+
|
227
|
+
/* Make memory region fully initialized (without changing its contents). */
|
228
|
+
void __msan_unpoison(const volatile void *a, size_t size);
|
229
|
+
|
230
|
+
/* Make memory region fully uninitialized (without changing its contents).
|
231
|
+
This is a legacy interface that does not update origin information. Use
|
232
|
+
__msan_allocated_memory() instead. */
|
233
|
+
void __msan_poison(const volatile void *a, size_t size);
|
234
|
+
|
235
|
+
/* Returns the offset of the first (at least partially) poisoned byte in the
|
236
|
+
memory range, or -1 if the whole range is good. */
|
237
|
+
intptr_t __msan_test_shadow(const volatile void *x, size_t size);
|
238
|
+
#endif
|
239
|
+
|
240
|
+
/* detects whether we are being compiled under asan */
|
241
|
+
#ifndef ZSTD_ADDRESS_SANITIZER
|
242
|
+
# if __has_feature(address_sanitizer)
|
243
|
+
# define ZSTD_ADDRESS_SANITIZER 1
|
244
|
+
# elif defined(__SANITIZE_ADDRESS__)
|
245
|
+
# define ZSTD_ADDRESS_SANITIZER 1
|
246
|
+
# else
|
247
|
+
# define ZSTD_ADDRESS_SANITIZER 0
|
248
|
+
# endif
|
249
|
+
#endif
|
250
|
+
|
251
|
+
#if ZSTD_ADDRESS_SANITIZER
|
252
|
+
/* Not all platforms that support asan provide sanitizers/asan_interface.h.
|
253
|
+
* We therefore declare the functions we need ourselves, rather than trying to
|
254
|
+
* include the header file... */
|
255
|
+
#include <stddef.h> /* size_t */
|
256
|
+
|
257
|
+
/**
|
258
|
+
* Marks a memory region (<c>[addr, addr+size)</c>) as unaddressable.
|
259
|
+
*
|
260
|
+
* This memory must be previously allocated by your program. Instrumented
|
261
|
+
* code is forbidden from accessing addresses in this region until it is
|
262
|
+
* unpoisoned. This function is not guaranteed to poison the entire region -
|
263
|
+
* it could poison only a subregion of <c>[addr, addr+size)</c> due to ASan
|
264
|
+
* alignment restrictions.
|
265
|
+
*
|
266
|
+
* \note This function is not thread-safe because no two threads can poison or
|
267
|
+
* unpoison memory in the same memory region simultaneously.
|
268
|
+
*
|
269
|
+
* \param addr Start of memory region.
|
270
|
+
* \param size Size of memory region. */
|
271
|
+
void __asan_poison_memory_region(void const volatile *addr, size_t size);
|
272
|
+
|
273
|
+
/**
|
274
|
+
* Marks a memory region (<c>[addr, addr+size)</c>) as addressable.
|
275
|
+
*
|
276
|
+
* This memory must be previously allocated by your program. Accessing
|
277
|
+
* addresses in this region is allowed until this region is poisoned again.
|
278
|
+
* This function could unpoison a super-region of <c>[addr, addr+size)</c> due
|
279
|
+
* to ASan alignment restrictions.
|
280
|
+
*
|
281
|
+
* \note This function is not thread-safe because no two threads can
|
282
|
+
* poison or unpoison memory in the same memory region simultaneously.
|
283
|
+
*
|
284
|
+
* \param addr Start of memory region.
|
285
|
+
* \param size Size of memory region. */
|
286
|
+
void __asan_unpoison_memory_region(void const volatile *addr, size_t size);
|
287
|
+
#endif
|
288
|
+
|
175
289
|
#endif /* ZSTD_COMPILER_H */
|
@@ -1,5 +1,5 @@
|
|
1
1
|
/*
|
2
|
-
* Copyright (c) 2018-
|
2
|
+
* Copyright (c) 2018-2021, Facebook, Inc.
|
3
3
|
* All rights reserved.
|
4
4
|
*
|
5
5
|
* This source code is licensed under both the BSD-style license (found in the
|
@@ -16,8 +16,6 @@
|
|
16
16
|
* https://github.com/facebook/folly/blob/master/folly/CpuId.h
|
17
17
|
*/
|
18
18
|
|
19
|
-
#include <string.h>
|
20
|
-
|
21
19
|
#include "mem.h"
|
22
20
|
|
23
21
|
#ifdef _MSC_VER
|
@@ -1,7 +1,7 @@
|
|
1
1
|
/* ******************************************************************
|
2
2
|
* debug
|
3
3
|
* Part of FSE library
|
4
|
-
* Copyright (c) 2013-
|
4
|
+
* Copyright (c) 2013-2021, Yann Collet, Facebook, Inc.
|
5
5
|
*
|
6
6
|
* You can contact the author at :
|
7
7
|
* - Source repository : https://github.com/Cyan4973/FiniteStateEntropy
|
@@ -1,7 +1,7 @@
|
|
1
1
|
/* ******************************************************************
|
2
2
|
* debug
|
3
3
|
* Part of FSE library
|
4
|
-
* Copyright (c) 2013-
|
4
|
+
* Copyright (c) 2013-2021, Yann Collet, Facebook, Inc.
|
5
5
|
*
|
6
6
|
* You can contact the author at :
|
7
7
|
* - Source repository : https://github.com/Cyan4973/FiniteStateEntropy
|
@@ -51,15 +51,6 @@ extern "C" {
|
|
51
51
|
#endif
|
52
52
|
|
53
53
|
|
54
|
-
/* DEBUGFILE can be defined externally,
|
55
|
-
* typically through compiler command line.
|
56
|
-
* note : currently useless.
|
57
|
-
* Value must be stderr or stdout */
|
58
|
-
#ifndef DEBUGFILE
|
59
|
-
# define DEBUGFILE stderr
|
60
|
-
#endif
|
61
|
-
|
62
|
-
|
63
54
|
/* recommended values for DEBUGLEVEL :
|
64
55
|
* 0 : release mode, no debug, all run-time checks disabled
|
65
56
|
* 1 : enables assert() only, no display
|
@@ -76,7 +67,8 @@ extern "C" {
|
|
76
67
|
*/
|
77
68
|
|
78
69
|
#if (DEBUGLEVEL>=1)
|
79
|
-
#
|
70
|
+
# define ZSTD_DEPS_NEED_ASSERT
|
71
|
+
# include "zstd_deps.h"
|
80
72
|
#else
|
81
73
|
# ifndef assert /* assert may be already defined, due to prior #include <assert.h> */
|
82
74
|
# define assert(condition) ((void)0) /* disable assert (default) */
|
@@ -84,7 +76,8 @@ extern "C" {
|
|
84
76
|
#endif
|
85
77
|
|
86
78
|
#if (DEBUGLEVEL>=2)
|
87
|
-
#
|
79
|
+
# define ZSTD_DEPS_NEED_IO
|
80
|
+
# include "zstd_deps.h"
|
88
81
|
extern int g_debuglevel; /* the variable is only declared,
|
89
82
|
it actually lives in debug.c,
|
90
83
|
and is shared by the whole process.
|
@@ -92,14 +85,14 @@ extern int g_debuglevel; /* the variable is only declared,
|
|
92
85
|
It's useful when enabling very verbose levels
|
93
86
|
on selective conditions (such as position in src) */
|
94
87
|
|
95
|
-
# define RAWLOG(l, ...) {
|
96
|
-
if (l<=g_debuglevel) {
|
97
|
-
|
88
|
+
# define RAWLOG(l, ...) { \
|
89
|
+
if (l<=g_debuglevel) { \
|
90
|
+
ZSTD_DEBUG_PRINT(__VA_ARGS__); \
|
98
91
|
} }
|
99
|
-
# define DEBUGLOG(l, ...) {
|
100
|
-
if (l<=g_debuglevel) {
|
101
|
-
|
102
|
-
|
92
|
+
# define DEBUGLOG(l, ...) { \
|
93
|
+
if (l<=g_debuglevel) { \
|
94
|
+
ZSTD_DEBUG_PRINT(__FILE__ ": " __VA_ARGS__); \
|
95
|
+
ZSTD_DEBUG_PRINT(" \n"); \
|
103
96
|
} }
|
104
97
|
#else
|
105
98
|
# define RAWLOG(l, ...) {} /* disabled */
|
@@ -1,6 +1,6 @@
|
|
1
1
|
/* ******************************************************************
|
2
2
|
* Common functions of New Generation Entropy library
|
3
|
-
* Copyright (c) 2016-
|
3
|
+
* Copyright (c) 2016-2021, Yann Collet, Facebook, Inc.
|
4
4
|
*
|
5
5
|
* You can contact the author at :
|
6
6
|
* - FSE+HUF source repository : https://github.com/Cyan4973/FiniteStateEntropy
|
@@ -38,8 +38,31 @@ const char* HUF_getErrorName(size_t code) { return ERR_getErrorName(code); }
|
|
38
38
|
/*-**************************************************************
|
39
39
|
* FSE NCount encoding-decoding
|
40
40
|
****************************************************************/
|
41
|
-
|
42
|
-
|
41
|
+
static U32 FSE_ctz(U32 val)
|
42
|
+
{
|
43
|
+
assert(val != 0);
|
44
|
+
{
|
45
|
+
# if defined(_MSC_VER) /* Visual */
|
46
|
+
unsigned long r=0;
|
47
|
+
return _BitScanForward(&r, val) ? (unsigned)r : 0;
|
48
|
+
# elif defined(__GNUC__) && (__GNUC__ >= 3) /* GCC Intrinsic */
|
49
|
+
return __builtin_ctz(val);
|
50
|
+
# elif defined(__ICCARM__) /* IAR Intrinsic */
|
51
|
+
return __CTZ(val);
|
52
|
+
# else /* Software version */
|
53
|
+
U32 count = 0;
|
54
|
+
while ((val & 1) == 0) {
|
55
|
+
val >>= 1;
|
56
|
+
++count;
|
57
|
+
}
|
58
|
+
return count;
|
59
|
+
# endif
|
60
|
+
}
|
61
|
+
}
|
62
|
+
|
63
|
+
FORCE_INLINE_TEMPLATE
|
64
|
+
size_t FSE_readNCount_body(short* normalizedCounter, unsigned* maxSVPtr, unsigned* tableLogPtr,
|
65
|
+
const void* headerBuffer, size_t hbSize)
|
43
66
|
{
|
44
67
|
const BYTE* const istart = (const BYTE*) headerBuffer;
|
45
68
|
const BYTE* const iend = istart + hbSize;
|
@@ -50,23 +73,23 @@ size_t FSE_readNCount (short* normalizedCounter, unsigned* maxSVPtr, unsigned* t
|
|
50
73
|
U32 bitStream;
|
51
74
|
int bitCount;
|
52
75
|
unsigned charnum = 0;
|
76
|
+
unsigned const maxSV1 = *maxSVPtr + 1;
|
53
77
|
int previous0 = 0;
|
54
78
|
|
55
|
-
if (hbSize <
|
56
|
-
/* This function only works when hbSize >=
|
57
|
-
char buffer[
|
58
|
-
|
59
|
-
memcpy(buffer, headerBuffer, hbSize);
|
79
|
+
if (hbSize < 8) {
|
80
|
+
/* This function only works when hbSize >= 8 */
|
81
|
+
char buffer[8] = {0};
|
82
|
+
ZSTD_memcpy(buffer, headerBuffer, hbSize);
|
60
83
|
{ size_t const countSize = FSE_readNCount(normalizedCounter, maxSVPtr, tableLogPtr,
|
61
84
|
buffer, sizeof(buffer));
|
62
85
|
if (FSE_isError(countSize)) return countSize;
|
63
86
|
if (countSize > hbSize) return ERROR(corruption_detected);
|
64
87
|
return countSize;
|
65
88
|
} }
|
66
|
-
assert(hbSize >=
|
89
|
+
assert(hbSize >= 8);
|
67
90
|
|
68
91
|
/* init */
|
69
|
-
|
92
|
+
ZSTD_memset(normalizedCounter, 0, (*maxSVPtr+1) * sizeof(normalizedCounter[0])); /* all symbols not present in NCount have a frequency of 0 */
|
70
93
|
bitStream = MEM_readLE32(ip);
|
71
94
|
nbBits = (bitStream & 0xF) + FSE_MIN_TABLELOG; /* extract tableLog */
|
72
95
|
if (nbBits > FSE_TABLELOG_ABSOLUTE_MAX) return ERROR(tableLog_tooLarge);
|
@@ -77,36 +100,58 @@ size_t FSE_readNCount (short* normalizedCounter, unsigned* maxSVPtr, unsigned* t
|
|
77
100
|
threshold = 1<<nbBits;
|
78
101
|
nbBits++;
|
79
102
|
|
80
|
-
|
103
|
+
for (;;) {
|
81
104
|
if (previous0) {
|
82
|
-
|
83
|
-
|
84
|
-
|
85
|
-
|
86
|
-
|
87
|
-
|
105
|
+
/* Count the number of repeats. Each time the
|
106
|
+
* 2-bit repeat code is 0b11 there is another
|
107
|
+
* repeat.
|
108
|
+
* Avoid UB by setting the high bit to 1.
|
109
|
+
*/
|
110
|
+
int repeats = FSE_ctz(~bitStream | 0x80000000) >> 1;
|
111
|
+
while (repeats >= 12) {
|
112
|
+
charnum += 3 * 12;
|
113
|
+
if (LIKELY(ip <= iend-7)) {
|
114
|
+
ip += 3;
|
88
115
|
} else {
|
89
|
-
|
90
|
-
bitCount
|
91
|
-
|
92
|
-
|
93
|
-
|
94
|
-
bitStream
|
95
|
-
bitCount += 2;
|
116
|
+
bitCount -= (int)(8 * (iend - 7 - ip));
|
117
|
+
bitCount &= 31;
|
118
|
+
ip = iend - 4;
|
119
|
+
}
|
120
|
+
bitStream = MEM_readLE32(ip) >> bitCount;
|
121
|
+
repeats = FSE_ctz(~bitStream | 0x80000000) >> 1;
|
96
122
|
}
|
97
|
-
|
123
|
+
charnum += 3 * repeats;
|
124
|
+
bitStream >>= 2 * repeats;
|
125
|
+
bitCount += 2 * repeats;
|
126
|
+
|
127
|
+
/* Add the final repeat which isn't 0b11. */
|
128
|
+
assert((bitStream & 3) < 3);
|
129
|
+
charnum += bitStream & 3;
|
98
130
|
bitCount += 2;
|
99
|
-
|
100
|
-
|
101
|
-
|
131
|
+
|
132
|
+
/* This is an error, but break and return an error
|
133
|
+
* at the end, because returning out of a loop makes
|
134
|
+
* it harder for the compiler to optimize.
|
135
|
+
*/
|
136
|
+
if (charnum >= maxSV1) break;
|
137
|
+
|
138
|
+
/* We don't need to set the normalized count to 0
|
139
|
+
* because we already memset the whole buffer to 0.
|
140
|
+
*/
|
141
|
+
|
142
|
+
if (LIKELY(ip <= iend-7) || (ip + (bitCount>>3) <= iend-4)) {
|
102
143
|
assert((bitCount >> 3) <= 3); /* For first condition to work */
|
103
144
|
ip += bitCount>>3;
|
104
145
|
bitCount &= 7;
|
105
|
-
bitStream = MEM_readLE32(ip) >> bitCount;
|
106
146
|
} else {
|
107
|
-
|
108
|
-
|
109
|
-
|
147
|
+
bitCount -= (int)(8 * (iend - 4 - ip));
|
148
|
+
bitCount &= 31;
|
149
|
+
ip = iend - 4;
|
150
|
+
}
|
151
|
+
bitStream = MEM_readLE32(ip) >> bitCount;
|
152
|
+
}
|
153
|
+
{
|
154
|
+
int const max = (2*threshold-1) - remaining;
|
110
155
|
int count;
|
111
156
|
|
112
157
|
if ((bitStream & (threshold-1)) < (U32)max) {
|
@@ -119,24 +164,43 @@ size_t FSE_readNCount (short* normalizedCounter, unsigned* maxSVPtr, unsigned* t
|
|
119
164
|
}
|
120
165
|
|
121
166
|
count--; /* extra accuracy */
|
122
|
-
|
167
|
+
/* When it matters (small blocks), this is a
|
168
|
+
* predictable branch, because we don't use -1.
|
169
|
+
*/
|
170
|
+
if (count >= 0) {
|
171
|
+
remaining -= count;
|
172
|
+
} else {
|
173
|
+
assert(count == -1);
|
174
|
+
remaining += count;
|
175
|
+
}
|
123
176
|
normalizedCounter[charnum++] = (short)count;
|
124
177
|
previous0 = !count;
|
125
|
-
|
126
|
-
|
127
|
-
|
178
|
+
|
179
|
+
assert(threshold > 1);
|
180
|
+
if (remaining < threshold) {
|
181
|
+
/* This branch can be folded into the
|
182
|
+
* threshold update condition because we
|
183
|
+
* know that threshold > 1.
|
184
|
+
*/
|
185
|
+
if (remaining <= 1) break;
|
186
|
+
nbBits = BIT_highbit32(remaining) + 1;
|
187
|
+
threshold = 1 << (nbBits - 1);
|
128
188
|
}
|
189
|
+
if (charnum >= maxSV1) break;
|
129
190
|
|
130
|
-
if ((ip <= iend-7) || (ip + (bitCount>>3) <= iend-4)) {
|
191
|
+
if (LIKELY(ip <= iend-7) || (ip + (bitCount>>3) <= iend-4)) {
|
131
192
|
ip += bitCount>>3;
|
132
193
|
bitCount &= 7;
|
133
194
|
} else {
|
134
195
|
bitCount -= (int)(8 * (iend - 4 - ip));
|
196
|
+
bitCount &= 31;
|
135
197
|
ip = iend - 4;
|
136
198
|
}
|
137
|
-
bitStream = MEM_readLE32(ip) >>
|
138
|
-
} }
|
199
|
+
bitStream = MEM_readLE32(ip) >> bitCount;
|
200
|
+
} }
|
139
201
|
if (remaining != 1) return ERROR(corruption_detected);
|
202
|
+
/* Only possible when there are too many zeros. */
|
203
|
+
if (charnum > maxSV1) return ERROR(maxSymbolValue_tooSmall);
|
140
204
|
if (bitCount > 32) return ERROR(corruption_detected);
|
141
205
|
*maxSVPtr = charnum-1;
|
142
206
|
|
@@ -144,6 +208,43 @@ size_t FSE_readNCount (short* normalizedCounter, unsigned* maxSVPtr, unsigned* t
|
|
144
208
|
return ip-istart;
|
145
209
|
}
|
146
210
|
|
211
|
+
/* Avoids the FORCE_INLINE of the _body() function. */
|
212
|
+
static size_t FSE_readNCount_body_default(
|
213
|
+
short* normalizedCounter, unsigned* maxSVPtr, unsigned* tableLogPtr,
|
214
|
+
const void* headerBuffer, size_t hbSize)
|
215
|
+
{
|
216
|
+
return FSE_readNCount_body(normalizedCounter, maxSVPtr, tableLogPtr, headerBuffer, hbSize);
|
217
|
+
}
|
218
|
+
|
219
|
+
#if DYNAMIC_BMI2
|
220
|
+
TARGET_ATTRIBUTE("bmi2") static size_t FSE_readNCount_body_bmi2(
|
221
|
+
short* normalizedCounter, unsigned* maxSVPtr, unsigned* tableLogPtr,
|
222
|
+
const void* headerBuffer, size_t hbSize)
|
223
|
+
{
|
224
|
+
return FSE_readNCount_body(normalizedCounter, maxSVPtr, tableLogPtr, headerBuffer, hbSize);
|
225
|
+
}
|
226
|
+
#endif
|
227
|
+
|
228
|
+
size_t FSE_readNCount_bmi2(
|
229
|
+
short* normalizedCounter, unsigned* maxSVPtr, unsigned* tableLogPtr,
|
230
|
+
const void* headerBuffer, size_t hbSize, int bmi2)
|
231
|
+
{
|
232
|
+
#if DYNAMIC_BMI2
|
233
|
+
if (bmi2) {
|
234
|
+
return FSE_readNCount_body_bmi2(normalizedCounter, maxSVPtr, tableLogPtr, headerBuffer, hbSize);
|
235
|
+
}
|
236
|
+
#endif
|
237
|
+
(void)bmi2;
|
238
|
+
return FSE_readNCount_body_default(normalizedCounter, maxSVPtr, tableLogPtr, headerBuffer, hbSize);
|
239
|
+
}
|
240
|
+
|
241
|
+
size_t FSE_readNCount(
|
242
|
+
short* normalizedCounter, unsigned* maxSVPtr, unsigned* tableLogPtr,
|
243
|
+
const void* headerBuffer, size_t hbSize)
|
244
|
+
{
|
245
|
+
return FSE_readNCount_bmi2(normalizedCounter, maxSVPtr, tableLogPtr, headerBuffer, hbSize, /* bmi2 */ 0);
|
246
|
+
}
|
247
|
+
|
147
248
|
|
148
249
|
/*! HUF_readStats() :
|
149
250
|
Read compact Huffman tree, saved by HUF_writeCTable().
|
@@ -155,6 +256,17 @@ size_t FSE_readNCount (short* normalizedCounter, unsigned* maxSVPtr, unsigned* t
|
|
155
256
|
size_t HUF_readStats(BYTE* huffWeight, size_t hwSize, U32* rankStats,
|
156
257
|
U32* nbSymbolsPtr, U32* tableLogPtr,
|
157
258
|
const void* src, size_t srcSize)
|
259
|
+
{
|
260
|
+
U32 wksp[HUF_READ_STATS_WORKSPACE_SIZE_U32];
|
261
|
+
return HUF_readStats_wksp(huffWeight, hwSize, rankStats, nbSymbolsPtr, tableLogPtr, src, srcSize, wksp, sizeof(wksp), /* bmi2 */ 0);
|
262
|
+
}
|
263
|
+
|
264
|
+
FORCE_INLINE_TEMPLATE size_t
|
265
|
+
HUF_readStats_body(BYTE* huffWeight, size_t hwSize, U32* rankStats,
|
266
|
+
U32* nbSymbolsPtr, U32* tableLogPtr,
|
267
|
+
const void* src, size_t srcSize,
|
268
|
+
void* workSpace, size_t wkspSize,
|
269
|
+
int bmi2)
|
158
270
|
{
|
159
271
|
U32 weightTotal;
|
160
272
|
const BYTE* ip = (const BYTE*) src;
|
@@ -163,7 +275,7 @@ size_t HUF_readStats(BYTE* huffWeight, size_t hwSize, U32* rankStats,
|
|
163
275
|
|
164
276
|
if (!srcSize) return ERROR(srcSize_wrong);
|
165
277
|
iSize = ip[0];
|
166
|
-
/*
|
278
|
+
/* ZSTD_memset(huffWeight, 0, hwSize); *//* is not necessary, even though some analyzer complain ... */
|
167
279
|
|
168
280
|
if (iSize >= 128) { /* special header */
|
169
281
|
oSize = iSize - 127;
|
@@ -177,14 +289,14 @@ size_t HUF_readStats(BYTE* huffWeight, size_t hwSize, U32* rankStats,
|
|
177
289
|
huffWeight[n+1] = ip[n/2] & 15;
|
178
290
|
} } }
|
179
291
|
else { /* header compressed with FSE (normal case) */
|
180
|
-
FSE_DTable fseWorkspace[FSE_DTABLE_SIZE_U32(6)]; /* 6 is max possible tableLog for HUF header (maybe even 5, to be tested) */
|
181
292
|
if (iSize+1 > srcSize) return ERROR(srcSize_wrong);
|
182
|
-
|
293
|
+
/* max (hwSize-1) values decoded, as last one is implied */
|
294
|
+
oSize = FSE_decompress_wksp_bmi2(huffWeight, hwSize-1, ip+1, iSize, 6, workSpace, wkspSize, bmi2);
|
183
295
|
if (FSE_isError(oSize)) return oSize;
|
184
296
|
}
|
185
297
|
|
186
298
|
/* collect weight stats */
|
187
|
-
|
299
|
+
ZSTD_memset(rankStats, 0, (HUF_TABLELOG_MAX + 1) * sizeof(U32));
|
188
300
|
weightTotal = 0;
|
189
301
|
{ U32 n; for (n=0; n<oSize; n++) {
|
190
302
|
if (huffWeight[n] >= HUF_TABLELOG_MAX) return ERROR(corruption_detected);
|
@@ -214,3 +326,37 @@ size_t HUF_readStats(BYTE* huffWeight, size_t hwSize, U32* rankStats,
|
|
214
326
|
*nbSymbolsPtr = (U32)(oSize+1);
|
215
327
|
return iSize+1;
|
216
328
|
}
|
329
|
+
|
330
|
+
/* Avoids the FORCE_INLINE of the _body() function. */
|
331
|
+
static size_t HUF_readStats_body_default(BYTE* huffWeight, size_t hwSize, U32* rankStats,
|
332
|
+
U32* nbSymbolsPtr, U32* tableLogPtr,
|
333
|
+
const void* src, size_t srcSize,
|
334
|
+
void* workSpace, size_t wkspSize)
|
335
|
+
{
|
336
|
+
return HUF_readStats_body(huffWeight, hwSize, rankStats, nbSymbolsPtr, tableLogPtr, src, srcSize, workSpace, wkspSize, 0);
|
337
|
+
}
|
338
|
+
|
339
|
+
#if DYNAMIC_BMI2
|
340
|
+
static TARGET_ATTRIBUTE("bmi2") size_t HUF_readStats_body_bmi2(BYTE* huffWeight, size_t hwSize, U32* rankStats,
|
341
|
+
U32* nbSymbolsPtr, U32* tableLogPtr,
|
342
|
+
const void* src, size_t srcSize,
|
343
|
+
void* workSpace, size_t wkspSize)
|
344
|
+
{
|
345
|
+
return HUF_readStats_body(huffWeight, hwSize, rankStats, nbSymbolsPtr, tableLogPtr, src, srcSize, workSpace, wkspSize, 1);
|
346
|
+
}
|
347
|
+
#endif
|
348
|
+
|
349
|
+
size_t HUF_readStats_wksp(BYTE* huffWeight, size_t hwSize, U32* rankStats,
|
350
|
+
U32* nbSymbolsPtr, U32* tableLogPtr,
|
351
|
+
const void* src, size_t srcSize,
|
352
|
+
void* workSpace, size_t wkspSize,
|
353
|
+
int bmi2)
|
354
|
+
{
|
355
|
+
#if DYNAMIC_BMI2
|
356
|
+
if (bmi2) {
|
357
|
+
return HUF_readStats_body_bmi2(huffWeight, hwSize, rankStats, nbSymbolsPtr, tableLogPtr, src, srcSize, workSpace, wkspSize);
|
358
|
+
}
|
359
|
+
#endif
|
360
|
+
(void)bmi2;
|
361
|
+
return HUF_readStats_body_default(huffWeight, hwSize, rankStats, nbSymbolsPtr, tableLogPtr, src, srcSize, workSpace, wkspSize);
|
362
|
+
}
|