zstd-ruby 1.4.4.0 → 1.5.1.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/.github/dependabot.yml +8 -0
- data/.github/workflows/ruby.yml +35 -0
- data/README.md +2 -2
- data/ext/zstdruby/extconf.rb +1 -0
- data/ext/zstdruby/libzstd/BUCK +5 -7
- data/ext/zstdruby/libzstd/Makefile +241 -173
- data/ext/zstdruby/libzstd/README.md +76 -18
- data/ext/zstdruby/libzstd/common/bitstream.h +75 -57
- data/ext/zstdruby/libzstd/common/compiler.h +196 -20
- data/ext/zstdruby/libzstd/common/cpu.h +1 -3
- data/ext/zstdruby/libzstd/common/debug.c +11 -31
- data/ext/zstdruby/libzstd/common/debug.h +22 -49
- data/ext/zstdruby/libzstd/common/entropy_common.c +208 -76
- data/ext/zstdruby/libzstd/common/error_private.c +3 -1
- data/ext/zstdruby/libzstd/common/error_private.h +87 -4
- data/ext/zstdruby/libzstd/common/fse.h +51 -42
- data/ext/zstdruby/libzstd/common/fse_decompress.c +149 -57
- data/ext/zstdruby/libzstd/common/huf.h +60 -54
- data/ext/zstdruby/libzstd/common/mem.h +87 -98
- data/ext/zstdruby/libzstd/common/pool.c +23 -17
- data/ext/zstdruby/libzstd/common/pool.h +3 -3
- data/ext/zstdruby/libzstd/common/portability_macros.h +131 -0
- data/ext/zstdruby/libzstd/common/threading.c +10 -8
- data/ext/zstdruby/libzstd/common/threading.h +4 -3
- data/ext/zstdruby/libzstd/common/xxhash.c +15 -873
- data/ext/zstdruby/libzstd/common/xxhash.h +5572 -191
- data/ext/zstdruby/libzstd/common/zstd_common.c +10 -10
- data/ext/zstdruby/libzstd/common/zstd_deps.h +111 -0
- data/ext/zstdruby/libzstd/common/zstd_internal.h +252 -108
- data/ext/zstdruby/libzstd/common/zstd_trace.h +163 -0
- data/ext/zstdruby/libzstd/compress/clevels.h +134 -0
- data/ext/zstdruby/libzstd/compress/fse_compress.c +105 -85
- data/ext/zstdruby/libzstd/compress/hist.c +41 -63
- data/ext/zstdruby/libzstd/compress/hist.h +13 -33
- data/ext/zstdruby/libzstd/compress/huf_compress.c +831 -259
- data/ext/zstdruby/libzstd/compress/zstd_compress.c +3213 -1007
- data/ext/zstdruby/libzstd/compress/zstd_compress_internal.h +493 -71
- data/ext/zstdruby/libzstd/compress/zstd_compress_literals.c +21 -16
- data/ext/zstdruby/libzstd/compress/zstd_compress_literals.h +4 -2
- data/ext/zstdruby/libzstd/compress/zstd_compress_sequences.c +51 -24
- data/ext/zstdruby/libzstd/compress/zstd_compress_sequences.h +10 -3
- data/ext/zstdruby/libzstd/compress/zstd_compress_superblock.c +573 -0
- data/ext/zstdruby/libzstd/compress/zstd_compress_superblock.h +32 -0
- data/ext/zstdruby/libzstd/compress/zstd_cwksp.h +208 -81
- data/ext/zstdruby/libzstd/compress/zstd_double_fast.c +315 -137
- data/ext/zstdruby/libzstd/compress/zstd_double_fast.h +2 -2
- data/ext/zstdruby/libzstd/compress/zstd_fast.c +319 -128
- data/ext/zstdruby/libzstd/compress/zstd_fast.h +2 -2
- data/ext/zstdruby/libzstd/compress/zstd_lazy.c +1156 -171
- data/ext/zstdruby/libzstd/compress/zstd_lazy.h +59 -1
- data/ext/zstdruby/libzstd/compress/zstd_ldm.c +331 -206
- data/ext/zstdruby/libzstd/compress/zstd_ldm.h +15 -3
- data/ext/zstdruby/libzstd/compress/zstd_ldm_geartab.h +106 -0
- data/ext/zstdruby/libzstd/compress/zstd_opt.c +403 -226
- data/ext/zstdruby/libzstd/compress/zstd_opt.h +1 -1
- data/ext/zstdruby/libzstd/compress/zstdmt_compress.c +188 -453
- data/ext/zstdruby/libzstd/compress/zstdmt_compress.h +32 -114
- data/ext/zstdruby/libzstd/decompress/huf_decompress.c +1065 -410
- data/ext/zstdruby/libzstd/decompress/huf_decompress_amd64.S +571 -0
- data/ext/zstdruby/libzstd/decompress/zstd_ddict.c +20 -16
- data/ext/zstdruby/libzstd/decompress/zstd_ddict.h +3 -3
- data/ext/zstdruby/libzstd/decompress/zstd_decompress.c +691 -230
- data/ext/zstdruby/libzstd/decompress/zstd_decompress_block.c +1072 -323
- data/ext/zstdruby/libzstd/decompress/zstd_decompress_block.h +16 -7
- data/ext/zstdruby/libzstd/decompress/zstd_decompress_internal.h +71 -10
- data/ext/zstdruby/libzstd/deprecated/zbuff.h +3 -3
- data/ext/zstdruby/libzstd/deprecated/zbuff_common.c +2 -2
- data/ext/zstdruby/libzstd/deprecated/zbuff_compress.c +24 -4
- data/ext/zstdruby/libzstd/deprecated/zbuff_decompress.c +1 -1
- data/ext/zstdruby/libzstd/dictBuilder/cover.c +57 -40
- data/ext/zstdruby/libzstd/dictBuilder/cover.h +20 -9
- data/ext/zstdruby/libzstd/dictBuilder/divsufsort.c +1 -1
- data/ext/zstdruby/libzstd/dictBuilder/fastcover.c +54 -35
- data/ext/zstdruby/libzstd/dictBuilder/zdict.c +151 -57
- data/ext/zstdruby/libzstd/dll/example/Makefile +2 -1
- data/ext/zstdruby/libzstd/dll/example/README.md +16 -22
- data/ext/zstdruby/libzstd/legacy/zstd_legacy.h +4 -4
- data/ext/zstdruby/libzstd/legacy/zstd_v01.c +25 -19
- data/ext/zstdruby/libzstd/legacy/zstd_v01.h +1 -1
- data/ext/zstdruby/libzstd/legacy/zstd_v02.c +18 -14
- data/ext/zstdruby/libzstd/legacy/zstd_v02.h +1 -1
- data/ext/zstdruby/libzstd/legacy/zstd_v03.c +18 -14
- data/ext/zstdruby/libzstd/legacy/zstd_v03.h +1 -1
- data/ext/zstdruby/libzstd/legacy/zstd_v04.c +22 -16
- data/ext/zstdruby/libzstd/legacy/zstd_v04.h +1 -1
- data/ext/zstdruby/libzstd/legacy/zstd_v05.c +29 -25
- data/ext/zstdruby/libzstd/legacy/zstd_v05.h +2 -2
- data/ext/zstdruby/libzstd/legacy/zstd_v06.c +29 -25
- data/ext/zstdruby/libzstd/legacy/zstd_v06.h +1 -1
- data/ext/zstdruby/libzstd/legacy/zstd_v07.c +34 -26
- data/ext/zstdruby/libzstd/legacy/zstd_v07.h +1 -1
- data/ext/zstdruby/libzstd/libzstd.mk +185 -0
- data/ext/zstdruby/libzstd/libzstd.pc.in +4 -3
- data/ext/zstdruby/libzstd/modulemap/module.modulemap +4 -0
- data/ext/zstdruby/libzstd/{dictBuilder/zdict.h → zdict.h} +201 -31
- data/ext/zstdruby/libzstd/zstd.h +760 -234
- data/ext/zstdruby/libzstd/{common/zstd_errors.h → zstd_errors.h} +3 -1
- data/ext/zstdruby/zstdruby.c +2 -2
- data/lib/zstd-ruby/version.rb +1 -1
- metadata +20 -9
- data/.travis.yml +0 -14
|
@@ -1,35 +1,15 @@
|
|
|
1
1
|
/* ******************************************************************
|
|
2
|
-
|
|
3
|
-
|
|
4
|
-
|
|
5
|
-
|
|
6
|
-
|
|
7
|
-
|
|
8
|
-
|
|
9
|
-
|
|
10
|
-
|
|
11
|
-
|
|
12
|
-
|
|
13
|
-
notice, this list of conditions and the following disclaimer.
|
|
14
|
-
* Redistributions in binary form must reproduce the above
|
|
15
|
-
copyright notice, this list of conditions and the following disclaimer
|
|
16
|
-
in the documentation and/or other materials provided with the
|
|
17
|
-
distribution.
|
|
18
|
-
|
|
19
|
-
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
20
|
-
"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
21
|
-
LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
22
|
-
A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
23
|
-
OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
24
|
-
SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
25
|
-
LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
26
|
-
DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
27
|
-
THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
28
|
-
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
29
|
-
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
30
|
-
|
|
31
|
-
You can contact the author at :
|
|
32
|
-
- Source repository : https://github.com/Cyan4973/FiniteStateEntropy
|
|
2
|
+
* huff0 huffman codec,
|
|
3
|
+
* part of Finite State Entropy library
|
|
4
|
+
* Copyright (c) Yann Collet, Facebook, Inc.
|
|
5
|
+
*
|
|
6
|
+
* You can contact the author at :
|
|
7
|
+
* - Source repository : https://github.com/Cyan4973/FiniteStateEntropy
|
|
8
|
+
*
|
|
9
|
+
* This source code is licensed under both the BSD-style license (found in the
|
|
10
|
+
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
|
11
|
+
* in the COPYING file in the root directory of this source tree).
|
|
12
|
+
* You may select, at your option, one of the above-listed licenses.
|
|
33
13
|
****************************************************************** */
|
|
34
14
|
|
|
35
15
|
#if defined (__cplusplus)
|
|
@@ -40,7 +20,7 @@ extern "C" {
|
|
|
40
20
|
#define HUF_H_298734234
|
|
41
21
|
|
|
42
22
|
/* *** Dependencies *** */
|
|
43
|
-
#include
|
|
23
|
+
#include "zstd_deps.h" /* size_t */
|
|
44
24
|
|
|
45
25
|
|
|
46
26
|
/* *** library symbols visibility *** */
|
|
@@ -109,9 +89,9 @@ HUF_PUBLIC_API size_t HUF_compress2 (void* dst, size_t dstCapacity,
|
|
|
109
89
|
|
|
110
90
|
/** HUF_compress4X_wksp() :
|
|
111
91
|
* Same as HUF_compress2(), but uses externally allocated `workSpace`.
|
|
112
|
-
* `workspace` must
|
|
113
|
-
#define HUF_WORKSPACE_SIZE (
|
|
114
|
-
#define
|
|
92
|
+
* `workspace` must be at least as large as HUF_WORKSPACE_SIZE */
|
|
93
|
+
#define HUF_WORKSPACE_SIZE ((8 << 10) + 512 /* sorting scratch space */)
|
|
94
|
+
#define HUF_WORKSPACE_SIZE_U64 (HUF_WORKSPACE_SIZE / sizeof(U64))
|
|
115
95
|
HUF_PUBLIC_API size_t HUF_compress4X_wksp (void* dst, size_t dstCapacity,
|
|
116
96
|
const void* src, size_t srcSize,
|
|
117
97
|
unsigned maxSymbolValue, unsigned tableLog,
|
|
@@ -131,14 +111,16 @@ HUF_PUBLIC_API size_t HUF_compress4X_wksp (void* dst, size_t dstCapacity,
|
|
|
131
111
|
|
|
132
112
|
/* *** Dependencies *** */
|
|
133
113
|
#include "mem.h" /* U32 */
|
|
114
|
+
#define FSE_STATIC_LINKING_ONLY
|
|
115
|
+
#include "fse.h"
|
|
134
116
|
|
|
135
117
|
|
|
136
118
|
/* *** Constants *** */
|
|
137
|
-
#define HUF_TABLELOG_MAX 12 /* max runtime value of tableLog (due to static allocation); can be modified up to
|
|
119
|
+
#define HUF_TABLELOG_MAX 12 /* max runtime value of tableLog (due to static allocation); can be modified up to HUF_TABLELOG_ABSOLUTEMAX */
|
|
138
120
|
#define HUF_TABLELOG_DEFAULT 11 /* default tableLog value when none specified */
|
|
139
121
|
#define HUF_SYMBOLVALUE_MAX 255
|
|
140
122
|
|
|
141
|
-
#define HUF_TABLELOG_ABSOLUTEMAX
|
|
123
|
+
#define HUF_TABLELOG_ABSOLUTEMAX 12 /* absolute limit of HUF_MAX_TABLELOG. Beyond that value, code does not work */
|
|
142
124
|
#if (HUF_TABLELOG_MAX > HUF_TABLELOG_ABSOLUTEMAX)
|
|
143
125
|
# error "HUF_TABLELOG_MAX is too large !"
|
|
144
126
|
#endif
|
|
@@ -153,12 +135,12 @@ HUF_PUBLIC_API size_t HUF_compress4X_wksp (void* dst, size_t dstCapacity,
|
|
|
153
135
|
#define HUF_COMPRESSBOUND(size) (HUF_CTABLEBOUND + HUF_BLOCKBOUND(size)) /* Macro version, useful for static allocation */
|
|
154
136
|
|
|
155
137
|
/* static allocation of HUF's Compression Table */
|
|
156
|
-
|
|
157
|
-
|
|
138
|
+
/* this is a private definition, just exposed for allocation and strict aliasing purpose. never EVER access its members directly */
|
|
139
|
+
typedef size_t HUF_CElt; /* consider it an incomplete type */
|
|
140
|
+
#define HUF_CTABLE_SIZE_ST(maxSymbolValue) ((maxSymbolValue)+2) /* Use tables of size_t, for proper alignment */
|
|
141
|
+
#define HUF_CTABLE_SIZE(maxSymbolValue) (HUF_CTABLE_SIZE_ST(maxSymbolValue) * sizeof(size_t))
|
|
158
142
|
#define HUF_CREATE_STATIC_CTABLE(name, maxSymbolValue) \
|
|
159
|
-
|
|
160
|
-
void* name##hv = &(name##hb); \
|
|
161
|
-
HUF_CElt* name = (HUF_CElt*)(name##hv) /* no final ; */
|
|
143
|
+
HUF_CElt name[HUF_CTABLE_SIZE_ST(maxSymbolValue)] /* no final ; */
|
|
162
144
|
|
|
163
145
|
/* static allocation of HUF's DTable */
|
|
164
146
|
typedef U32 HUF_DTable;
|
|
@@ -204,10 +186,13 @@ size_t HUF_decompress4X2_DCtx_wksp(HUF_DTable* dctx, void* dst, size_t dstSize,
|
|
|
204
186
|
* or to save and regenerate 'CTable' using external methods.
|
|
205
187
|
*/
|
|
206
188
|
unsigned HUF_optimalTableLog(unsigned maxTableLog, size_t srcSize, unsigned maxSymbolValue);
|
|
207
|
-
typedef struct HUF_CElt_s HUF_CElt; /* incomplete type */
|
|
208
189
|
size_t HUF_buildCTable (HUF_CElt* CTable, const unsigned* count, unsigned maxSymbolValue, unsigned maxNbBits); /* @return : maxNbBits; CTable and count can overlap. In which case, CTable will overwrite count content */
|
|
209
190
|
size_t HUF_writeCTable (void* dst, size_t maxDstSize, const HUF_CElt* CTable, unsigned maxSymbolValue, unsigned huffLog);
|
|
191
|
+
size_t HUF_writeCTable_wksp(void* dst, size_t maxDstSize, const HUF_CElt* CTable, unsigned maxSymbolValue, unsigned huffLog, void* workspace, size_t workspaceSize);
|
|
210
192
|
size_t HUF_compress4X_usingCTable(void* dst, size_t dstSize, const void* src, size_t srcSize, const HUF_CElt* CTable);
|
|
193
|
+
size_t HUF_compress4X_usingCTable_bmi2(void* dst, size_t dstSize, const void* src, size_t srcSize, const HUF_CElt* CTable, int bmi2);
|
|
194
|
+
size_t HUF_estimateCompressedSize(const HUF_CElt* CTable, const unsigned* count, unsigned maxSymbolValue);
|
|
195
|
+
int HUF_validateCTable(const HUF_CElt* CTable, const unsigned* count, unsigned maxSymbolValue);
|
|
211
196
|
|
|
212
197
|
typedef enum {
|
|
213
198
|
HUF_repeat_none, /**< Cannot use the previous table */
|
|
@@ -218,12 +203,13 @@ typedef enum {
|
|
|
218
203
|
* Same as HUF_compress4X_wksp(), but considers using hufTable if *repeat != HUF_repeat_none.
|
|
219
204
|
* If it uses hufTable it does not modify hufTable or repeat.
|
|
220
205
|
* If it doesn't, it sets *repeat = HUF_repeat_none, and it sets hufTable to the table used.
|
|
221
|
-
* If preferRepeat then the old table will always be used if valid.
|
|
206
|
+
* If preferRepeat then the old table will always be used if valid.
|
|
207
|
+
* If suspectUncompressible then some sampling checks will be run to potentially skip huffman coding */
|
|
222
208
|
size_t HUF_compress4X_repeat(void* dst, size_t dstSize,
|
|
223
209
|
const void* src, size_t srcSize,
|
|
224
210
|
unsigned maxSymbolValue, unsigned tableLog,
|
|
225
211
|
void* workSpace, size_t wkspSize, /**< `workSpace` must be aligned on 4-bytes boundaries, `wkspSize` must be >= HUF_WORKSPACE_SIZE */
|
|
226
|
-
HUF_CElt* hufTable, HUF_repeat* repeat, int preferRepeat, int bmi2);
|
|
212
|
+
HUF_CElt* hufTable, HUF_repeat* repeat, int preferRepeat, int bmi2, unsigned suspectUncompressible);
|
|
227
213
|
|
|
228
214
|
/** HUF_buildCTable_wksp() :
|
|
229
215
|
* Same as HUF_buildCTable(), but using externally allocated scratch buffer.
|
|
@@ -244,15 +230,27 @@ size_t HUF_readStats(BYTE* huffWeight, size_t hwSize,
|
|
|
244
230
|
U32* rankStats, U32* nbSymbolsPtr, U32* tableLogPtr,
|
|
245
231
|
const void* src, size_t srcSize);
|
|
246
232
|
|
|
233
|
+
/*! HUF_readStats_wksp() :
|
|
234
|
+
* Same as HUF_readStats() but takes an external workspace which must be
|
|
235
|
+
* 4-byte aligned and its size must be >= HUF_READ_STATS_WORKSPACE_SIZE.
|
|
236
|
+
* If the CPU has BMI2 support, pass bmi2=1, otherwise pass bmi2=0.
|
|
237
|
+
*/
|
|
238
|
+
#define HUF_READ_STATS_WORKSPACE_SIZE_U32 FSE_DECOMPRESS_WKSP_SIZE_U32(6, HUF_TABLELOG_MAX-1)
|
|
239
|
+
#define HUF_READ_STATS_WORKSPACE_SIZE (HUF_READ_STATS_WORKSPACE_SIZE_U32 * sizeof(unsigned))
|
|
240
|
+
size_t HUF_readStats_wksp(BYTE* huffWeight, size_t hwSize,
|
|
241
|
+
U32* rankStats, U32* nbSymbolsPtr, U32* tableLogPtr,
|
|
242
|
+
const void* src, size_t srcSize,
|
|
243
|
+
void* workspace, size_t wkspSize,
|
|
244
|
+
int bmi2);
|
|
245
|
+
|
|
247
246
|
/** HUF_readCTable() :
|
|
248
247
|
* Loading a CTable saved with HUF_writeCTable() */
|
|
249
|
-
size_t HUF_readCTable (HUF_CElt* CTable, unsigned* maxSymbolValuePtr, const void* src, size_t srcSize);
|
|
248
|
+
size_t HUF_readCTable (HUF_CElt* CTable, unsigned* maxSymbolValuePtr, const void* src, size_t srcSize, unsigned *hasZeroWeights);
|
|
250
249
|
|
|
251
|
-
/**
|
|
250
|
+
/** HUF_getNbBitsFromCTable() :
|
|
252
251
|
* Read nbBits from CTable symbolTable, for symbol `symbolValue` presumed <= HUF_SYMBOLVALUE_MAX
|
|
253
|
-
* Note 1 : is not inlined, as HUF_CElt definition is private
|
|
254
|
-
|
|
255
|
-
U32 HUF_getNbBits(const void* symbolTable, U32 symbolValue);
|
|
252
|
+
* Note 1 : is not inlined, as HUF_CElt definition is private */
|
|
253
|
+
U32 HUF_getNbBitsFromCTable(const HUF_CElt* symbolTable, U32 symbolValue);
|
|
256
254
|
|
|
257
255
|
/*
|
|
258
256
|
* HUF_decompress() does the following:
|
|
@@ -278,7 +276,7 @@ U32 HUF_selectDecoder (size_t dstSize, size_t cSrcSize);
|
|
|
278
276
|
* a required workspace size greater than that specified in the following
|
|
279
277
|
* macro.
|
|
280
278
|
*/
|
|
281
|
-
#define HUF_DECOMPRESS_WORKSPACE_SIZE (2 << 10)
|
|
279
|
+
#define HUF_DECOMPRESS_WORKSPACE_SIZE ((2 << 10) + (1 << 9))
|
|
282
280
|
#define HUF_DECOMPRESS_WORKSPACE_SIZE_U32 (HUF_DECOMPRESS_WORKSPACE_SIZE / sizeof(U32))
|
|
283
281
|
|
|
284
282
|
#ifndef HUF_FORCE_DECOMPRESS_X2
|
|
@@ -304,18 +302,20 @@ size_t HUF_decompress4X2_usingDTable(void* dst, size_t maxDstSize, const void* c
|
|
|
304
302
|
/* ====================== */
|
|
305
303
|
|
|
306
304
|
size_t HUF_compress1X (void* dst, size_t dstSize, const void* src, size_t srcSize, unsigned maxSymbolValue, unsigned tableLog);
|
|
307
|
-
size_t HUF_compress1X_wksp (void* dst, size_t dstSize, const void* src, size_t srcSize, unsigned maxSymbolValue, unsigned tableLog, void* workSpace, size_t wkspSize); /**< `workSpace` must be a table of at least
|
|
305
|
+
size_t HUF_compress1X_wksp (void* dst, size_t dstSize, const void* src, size_t srcSize, unsigned maxSymbolValue, unsigned tableLog, void* workSpace, size_t wkspSize); /**< `workSpace` must be a table of at least HUF_WORKSPACE_SIZE_U64 U64 */
|
|
308
306
|
size_t HUF_compress1X_usingCTable(void* dst, size_t dstSize, const void* src, size_t srcSize, const HUF_CElt* CTable);
|
|
307
|
+
size_t HUF_compress1X_usingCTable_bmi2(void* dst, size_t dstSize, const void* src, size_t srcSize, const HUF_CElt* CTable, int bmi2);
|
|
309
308
|
/** HUF_compress1X_repeat() :
|
|
310
309
|
* Same as HUF_compress1X_wksp(), but considers using hufTable if *repeat != HUF_repeat_none.
|
|
311
310
|
* If it uses hufTable it does not modify hufTable or repeat.
|
|
312
311
|
* If it doesn't, it sets *repeat = HUF_repeat_none, and it sets hufTable to the table used.
|
|
313
|
-
* If preferRepeat then the old table will always be used if valid.
|
|
312
|
+
* If preferRepeat then the old table will always be used if valid.
|
|
313
|
+
* If suspectUncompressible then some sampling checks will be run to potentially skip huffman coding */
|
|
314
314
|
size_t HUF_compress1X_repeat(void* dst, size_t dstSize,
|
|
315
315
|
const void* src, size_t srcSize,
|
|
316
316
|
unsigned maxSymbolValue, unsigned tableLog,
|
|
317
317
|
void* workSpace, size_t wkspSize, /**< `workSpace` must be aligned on 4-bytes boundaries, `wkspSize` must be >= HUF_WORKSPACE_SIZE */
|
|
318
|
-
HUF_CElt* hufTable, HUF_repeat* repeat, int preferRepeat, int bmi2);
|
|
318
|
+
HUF_CElt* hufTable, HUF_repeat* repeat, int preferRepeat, int bmi2, unsigned suspectUncompressible);
|
|
319
319
|
|
|
320
320
|
size_t HUF_decompress1X1 (void* dst, size_t dstSize, const void* cSrc, size_t cSrcSize); /* single-symbol decoder */
|
|
321
321
|
#ifndef HUF_FORCE_DECOMPRESS_X1
|
|
@@ -350,6 +350,12 @@ size_t HUF_decompress1X1_DCtx_wksp_bmi2(HUF_DTable* dctx, void* dst, size_t dstS
|
|
|
350
350
|
#endif
|
|
351
351
|
size_t HUF_decompress4X_usingDTable_bmi2(void* dst, size_t maxDstSize, const void* cSrc, size_t cSrcSize, const HUF_DTable* DTable, int bmi2);
|
|
352
352
|
size_t HUF_decompress4X_hufOnly_wksp_bmi2(HUF_DTable* dctx, void* dst, size_t dstSize, const void* cSrc, size_t cSrcSize, void* workSpace, size_t wkspSize, int bmi2);
|
|
353
|
+
#ifndef HUF_FORCE_DECOMPRESS_X2
|
|
354
|
+
size_t HUF_readDTableX1_wksp_bmi2(HUF_DTable* DTable, const void* src, size_t srcSize, void* workSpace, size_t wkspSize, int bmi2);
|
|
355
|
+
#endif
|
|
356
|
+
#ifndef HUF_FORCE_DECOMPRESS_X1
|
|
357
|
+
size_t HUF_readDTableX2_wksp_bmi2(HUF_DTable* DTable, const void* src, size_t srcSize, void* workSpace, size_t wkspSize, int bmi2);
|
|
358
|
+
#endif
|
|
353
359
|
|
|
354
360
|
#endif /* HUF_STATIC_LINKING_ONLY */
|
|
355
361
|
|
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
/*
|
|
2
|
-
* Copyright (c)
|
|
2
|
+
* Copyright (c) Yann Collet, Facebook, Inc.
|
|
3
3
|
* All rights reserved.
|
|
4
4
|
*
|
|
5
5
|
* This source code is licensed under both the BSD-style license (found in the
|
|
@@ -18,8 +18,10 @@ extern "C" {
|
|
|
18
18
|
/*-****************************************
|
|
19
19
|
* Dependencies
|
|
20
20
|
******************************************/
|
|
21
|
-
#include <stddef.h>
|
|
22
|
-
#include
|
|
21
|
+
#include <stddef.h> /* size_t, ptrdiff_t */
|
|
22
|
+
#include "compiler.h" /* __has_builtin */
|
|
23
|
+
#include "debug.h" /* DEBUG_STATIC_ASSERT */
|
|
24
|
+
#include "zstd_deps.h" /* ZSTD_memcpy */
|
|
23
25
|
|
|
24
26
|
|
|
25
27
|
/*-****************************************
|
|
@@ -39,94 +41,18 @@ extern "C" {
|
|
|
39
41
|
# define MEM_STATIC static /* this version may generate warnings for unused static functions; disable the relevant warning */
|
|
40
42
|
#endif
|
|
41
43
|
|
|
42
|
-
#ifndef __has_builtin
|
|
43
|
-
# define __has_builtin(x) 0 /* compat. with non-clang compilers */
|
|
44
|
-
#endif
|
|
45
|
-
|
|
46
|
-
/* code only tested on 32 and 64 bits systems */
|
|
47
|
-
#define MEM_STATIC_ASSERT(c) { enum { MEM_static_assert = 1/(int)(!!(c)) }; }
|
|
48
|
-
MEM_STATIC void MEM_check(void) { MEM_STATIC_ASSERT((sizeof(size_t)==4) || (sizeof(size_t)==8)); }
|
|
49
|
-
|
|
50
|
-
/* detects whether we are being compiled under msan */
|
|
51
|
-
#if defined (__has_feature)
|
|
52
|
-
# if __has_feature(memory_sanitizer)
|
|
53
|
-
# define MEMORY_SANITIZER 1
|
|
54
|
-
# endif
|
|
55
|
-
#endif
|
|
56
|
-
|
|
57
|
-
#if defined (MEMORY_SANITIZER)
|
|
58
|
-
/* Not all platforms that support msan provide sanitizers/msan_interface.h.
|
|
59
|
-
* We therefore declare the functions we need ourselves, rather than trying to
|
|
60
|
-
* include the header file... */
|
|
61
|
-
|
|
62
|
-
#include <stdint.h> /* intptr_t */
|
|
63
|
-
|
|
64
|
-
/* Make memory region fully initialized (without changing its contents). */
|
|
65
|
-
void __msan_unpoison(const volatile void *a, size_t size);
|
|
66
|
-
|
|
67
|
-
/* Make memory region fully uninitialized (without changing its contents).
|
|
68
|
-
This is a legacy interface that does not update origin information. Use
|
|
69
|
-
__msan_allocated_memory() instead. */
|
|
70
|
-
void __msan_poison(const volatile void *a, size_t size);
|
|
71
|
-
|
|
72
|
-
/* Returns the offset of the first (at least partially) poisoned byte in the
|
|
73
|
-
memory range, or -1 if the whole range is good. */
|
|
74
|
-
intptr_t __msan_test_shadow(const volatile void *x, size_t size);
|
|
75
|
-
#endif
|
|
76
|
-
|
|
77
|
-
/* detects whether we are being compiled under asan */
|
|
78
|
-
#if defined (__has_feature)
|
|
79
|
-
# if __has_feature(address_sanitizer)
|
|
80
|
-
# define ADDRESS_SANITIZER 1
|
|
81
|
-
# endif
|
|
82
|
-
#elif defined(__SANITIZE_ADDRESS__)
|
|
83
|
-
# define ADDRESS_SANITIZER 1
|
|
84
|
-
#endif
|
|
85
|
-
|
|
86
|
-
#if defined (ADDRESS_SANITIZER)
|
|
87
|
-
/* Not all platforms that support asan provide sanitizers/asan_interface.h.
|
|
88
|
-
* We therefore declare the functions we need ourselves, rather than trying to
|
|
89
|
-
* include the header file... */
|
|
90
|
-
|
|
91
|
-
/**
|
|
92
|
-
* Marks a memory region (<c>[addr, addr+size)</c>) as unaddressable.
|
|
93
|
-
*
|
|
94
|
-
* This memory must be previously allocated by your program. Instrumented
|
|
95
|
-
* code is forbidden from accessing addresses in this region until it is
|
|
96
|
-
* unpoisoned. This function is not guaranteed to poison the entire region -
|
|
97
|
-
* it could poison only a subregion of <c>[addr, addr+size)</c> due to ASan
|
|
98
|
-
* alignment restrictions.
|
|
99
|
-
*
|
|
100
|
-
* \note This function is not thread-safe because no two threads can poison or
|
|
101
|
-
* unpoison memory in the same memory region simultaneously.
|
|
102
|
-
*
|
|
103
|
-
* \param addr Start of memory region.
|
|
104
|
-
* \param size Size of memory region. */
|
|
105
|
-
void __asan_poison_memory_region(void const volatile *addr, size_t size);
|
|
106
|
-
|
|
107
|
-
/**
|
|
108
|
-
* Marks a memory region (<c>[addr, addr+size)</c>) as addressable.
|
|
109
|
-
*
|
|
110
|
-
* This memory must be previously allocated by your program. Accessing
|
|
111
|
-
* addresses in this region is allowed until this region is poisoned again.
|
|
112
|
-
* This function could unpoison a super-region of <c>[addr, addr+size)</c> due
|
|
113
|
-
* to ASan alignment restrictions.
|
|
114
|
-
*
|
|
115
|
-
* \note This function is not thread-safe because no two threads can
|
|
116
|
-
* poison or unpoison memory in the same memory region simultaneously.
|
|
117
|
-
*
|
|
118
|
-
* \param addr Start of memory region.
|
|
119
|
-
* \param size Size of memory region. */
|
|
120
|
-
void __asan_unpoison_memory_region(void const volatile *addr, size_t size);
|
|
121
|
-
#endif
|
|
122
|
-
|
|
123
|
-
|
|
124
44
|
/*-**************************************************************
|
|
125
45
|
* Basic Types
|
|
126
46
|
*****************************************************************/
|
|
127
47
|
#if !defined (__VMS) && (defined (__cplusplus) || (defined (__STDC_VERSION__) && (__STDC_VERSION__ >= 199901L) /* C99 */) )
|
|
128
|
-
#
|
|
48
|
+
# if defined(_AIX)
|
|
49
|
+
# include <inttypes.h>
|
|
50
|
+
# else
|
|
51
|
+
# include <stdint.h> /* intptr_t */
|
|
52
|
+
# endif
|
|
129
53
|
typedef uint8_t BYTE;
|
|
54
|
+
typedef uint8_t U8;
|
|
55
|
+
typedef int8_t S8;
|
|
130
56
|
typedef uint16_t U16;
|
|
131
57
|
typedef int16_t S16;
|
|
132
58
|
typedef uint32_t U32;
|
|
@@ -139,6 +65,8 @@ void __asan_unpoison_memory_region(void const volatile *addr, size_t size);
|
|
|
139
65
|
# error "this implementation requires char to be exactly 8-bit type"
|
|
140
66
|
#endif
|
|
141
67
|
typedef unsigned char BYTE;
|
|
68
|
+
typedef unsigned char U8;
|
|
69
|
+
typedef signed char S8;
|
|
142
70
|
#if USHRT_MAX != 65535
|
|
143
71
|
# error "this implementation requires short to be exactly 16-bit type"
|
|
144
72
|
#endif
|
|
@@ -157,7 +85,53 @@ void __asan_unpoison_memory_region(void const volatile *addr, size_t size);
|
|
|
157
85
|
|
|
158
86
|
|
|
159
87
|
/*-**************************************************************
|
|
160
|
-
* Memory I/O
|
|
88
|
+
* Memory I/O API
|
|
89
|
+
*****************************************************************/
|
|
90
|
+
/*=== Static platform detection ===*/
|
|
91
|
+
MEM_STATIC unsigned MEM_32bits(void);
|
|
92
|
+
MEM_STATIC unsigned MEM_64bits(void);
|
|
93
|
+
MEM_STATIC unsigned MEM_isLittleEndian(void);
|
|
94
|
+
|
|
95
|
+
/*=== Native unaligned read/write ===*/
|
|
96
|
+
MEM_STATIC U16 MEM_read16(const void* memPtr);
|
|
97
|
+
MEM_STATIC U32 MEM_read32(const void* memPtr);
|
|
98
|
+
MEM_STATIC U64 MEM_read64(const void* memPtr);
|
|
99
|
+
MEM_STATIC size_t MEM_readST(const void* memPtr);
|
|
100
|
+
|
|
101
|
+
MEM_STATIC void MEM_write16(void* memPtr, U16 value);
|
|
102
|
+
MEM_STATIC void MEM_write32(void* memPtr, U32 value);
|
|
103
|
+
MEM_STATIC void MEM_write64(void* memPtr, U64 value);
|
|
104
|
+
|
|
105
|
+
/*=== Little endian unaligned read/write ===*/
|
|
106
|
+
MEM_STATIC U16 MEM_readLE16(const void* memPtr);
|
|
107
|
+
MEM_STATIC U32 MEM_readLE24(const void* memPtr);
|
|
108
|
+
MEM_STATIC U32 MEM_readLE32(const void* memPtr);
|
|
109
|
+
MEM_STATIC U64 MEM_readLE64(const void* memPtr);
|
|
110
|
+
MEM_STATIC size_t MEM_readLEST(const void* memPtr);
|
|
111
|
+
|
|
112
|
+
MEM_STATIC void MEM_writeLE16(void* memPtr, U16 val);
|
|
113
|
+
MEM_STATIC void MEM_writeLE24(void* memPtr, U32 val);
|
|
114
|
+
MEM_STATIC void MEM_writeLE32(void* memPtr, U32 val32);
|
|
115
|
+
MEM_STATIC void MEM_writeLE64(void* memPtr, U64 val64);
|
|
116
|
+
MEM_STATIC void MEM_writeLEST(void* memPtr, size_t val);
|
|
117
|
+
|
|
118
|
+
/*=== Big endian unaligned read/write ===*/
|
|
119
|
+
MEM_STATIC U32 MEM_readBE32(const void* memPtr);
|
|
120
|
+
MEM_STATIC U64 MEM_readBE64(const void* memPtr);
|
|
121
|
+
MEM_STATIC size_t MEM_readBEST(const void* memPtr);
|
|
122
|
+
|
|
123
|
+
MEM_STATIC void MEM_writeBE32(void* memPtr, U32 val32);
|
|
124
|
+
MEM_STATIC void MEM_writeBE64(void* memPtr, U64 val64);
|
|
125
|
+
MEM_STATIC void MEM_writeBEST(void* memPtr, size_t val);
|
|
126
|
+
|
|
127
|
+
/*=== Byteswap ===*/
|
|
128
|
+
MEM_STATIC U32 MEM_swap32(U32 in);
|
|
129
|
+
MEM_STATIC U64 MEM_swap64(U64 in);
|
|
130
|
+
MEM_STATIC size_t MEM_swapST(size_t in);
|
|
131
|
+
|
|
132
|
+
|
|
133
|
+
/*-**************************************************************
|
|
134
|
+
* Memory I/O Implementation
|
|
161
135
|
*****************************************************************/
|
|
162
136
|
/* MEM_FORCE_MEMORY_ACCESS :
|
|
163
137
|
* By default, access to unaligned memory is controlled by `memcpy()`, which is safe and portable.
|
|
@@ -173,9 +147,7 @@ void __asan_unpoison_memory_region(void const volatile *addr, size_t size);
|
|
|
173
147
|
* Prefer these methods in priority order (0 > 1 > 2)
|
|
174
148
|
*/
|
|
175
149
|
#ifndef MEM_FORCE_MEMORY_ACCESS /* can be defined externally, on command line for example */
|
|
176
|
-
# if defined(
|
|
177
|
-
# define MEM_FORCE_MEMORY_ACCESS 2
|
|
178
|
-
# elif defined(__INTEL_COMPILER) || defined(__GNUC__) || defined(__ICCARM__)
|
|
150
|
+
# if defined(__INTEL_COMPILER) || defined(__GNUC__) || defined(__ICCARM__)
|
|
179
151
|
# define MEM_FORCE_MEMORY_ACCESS 1
|
|
180
152
|
# endif
|
|
181
153
|
#endif
|
|
@@ -185,8 +157,22 @@ MEM_STATIC unsigned MEM_64bits(void) { return sizeof(size_t)==8; }
|
|
|
185
157
|
|
|
186
158
|
MEM_STATIC unsigned MEM_isLittleEndian(void)
|
|
187
159
|
{
|
|
160
|
+
#if defined(__BYTE_ORDER__) && defined(__ORDER_LITTLE_ENDIAN__) && (__BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__)
|
|
161
|
+
return 1;
|
|
162
|
+
#elif defined(__BYTE_ORDER__) && defined(__ORDER_BIG_ENDIAN__) && (__BYTE_ORDER__ == __ORDER_BIG_ENDIAN__)
|
|
163
|
+
return 0;
|
|
164
|
+
#elif defined(__clang__) && __LITTLE_ENDIAN__
|
|
165
|
+
return 1;
|
|
166
|
+
#elif defined(__clang__) && __BIG_ENDIAN__
|
|
167
|
+
return 0;
|
|
168
|
+
#elif defined(_MSC_VER) && (_M_AMD64 || _M_IX86)
|
|
169
|
+
return 1;
|
|
170
|
+
#elif defined(__DMC__) && defined(_M_IX86)
|
|
171
|
+
return 1;
|
|
172
|
+
#else
|
|
188
173
|
const union { U32 u; BYTE c[4]; } one = { 1 }; /* don't use static : performance detrimental */
|
|
189
174
|
return one.c[0];
|
|
175
|
+
#endif
|
|
190
176
|
}
|
|
191
177
|
|
|
192
178
|
#if defined(MEM_FORCE_MEMORY_ACCESS) && (MEM_FORCE_MEMORY_ACCESS==2)
|
|
@@ -236,37 +222,37 @@ MEM_STATIC void MEM_write64(void* memPtr, U64 value) { ((unalign64*)memPtr)->v =
|
|
|
236
222
|
|
|
237
223
|
MEM_STATIC U16 MEM_read16(const void* memPtr)
|
|
238
224
|
{
|
|
239
|
-
U16 val;
|
|
225
|
+
U16 val; ZSTD_memcpy(&val, memPtr, sizeof(val)); return val;
|
|
240
226
|
}
|
|
241
227
|
|
|
242
228
|
MEM_STATIC U32 MEM_read32(const void* memPtr)
|
|
243
229
|
{
|
|
244
|
-
U32 val;
|
|
230
|
+
U32 val; ZSTD_memcpy(&val, memPtr, sizeof(val)); return val;
|
|
245
231
|
}
|
|
246
232
|
|
|
247
233
|
MEM_STATIC U64 MEM_read64(const void* memPtr)
|
|
248
234
|
{
|
|
249
|
-
U64 val;
|
|
235
|
+
U64 val; ZSTD_memcpy(&val, memPtr, sizeof(val)); return val;
|
|
250
236
|
}
|
|
251
237
|
|
|
252
238
|
MEM_STATIC size_t MEM_readST(const void* memPtr)
|
|
253
239
|
{
|
|
254
|
-
size_t val;
|
|
240
|
+
size_t val; ZSTD_memcpy(&val, memPtr, sizeof(val)); return val;
|
|
255
241
|
}
|
|
256
242
|
|
|
257
243
|
MEM_STATIC void MEM_write16(void* memPtr, U16 value)
|
|
258
244
|
{
|
|
259
|
-
|
|
245
|
+
ZSTD_memcpy(memPtr, &value, sizeof(value));
|
|
260
246
|
}
|
|
261
247
|
|
|
262
248
|
MEM_STATIC void MEM_write32(void* memPtr, U32 value)
|
|
263
249
|
{
|
|
264
|
-
|
|
250
|
+
ZSTD_memcpy(memPtr, &value, sizeof(value));
|
|
265
251
|
}
|
|
266
252
|
|
|
267
253
|
MEM_STATIC void MEM_write64(void* memPtr, U64 value)
|
|
268
254
|
{
|
|
269
|
-
|
|
255
|
+
ZSTD_memcpy(memPtr, &value, sizeof(value));
|
|
270
256
|
}
|
|
271
257
|
|
|
272
258
|
#endif /* MEM_FORCE_MEMORY_ACCESS */
|
|
@@ -338,7 +324,7 @@ MEM_STATIC void MEM_writeLE16(void* memPtr, U16 val)
|
|
|
338
324
|
|
|
339
325
|
MEM_STATIC U32 MEM_readLE24(const void* memPtr)
|
|
340
326
|
{
|
|
341
|
-
return MEM_readLE16(memPtr) + (((const BYTE*)memPtr)[2] << 16);
|
|
327
|
+
return (U32)MEM_readLE16(memPtr) + ((U32)(((const BYTE*)memPtr)[2]) << 16);
|
|
342
328
|
}
|
|
343
329
|
|
|
344
330
|
MEM_STATIC void MEM_writeLE24(void* memPtr, U32 val)
|
|
@@ -445,6 +431,9 @@ MEM_STATIC void MEM_writeBEST(void* memPtr, size_t val)
|
|
|
445
431
|
MEM_writeBE64(memPtr, (U64)val);
|
|
446
432
|
}
|
|
447
433
|
|
|
434
|
+
/* code only tested on 32 and 64 bits systems */
|
|
435
|
+
MEM_STATIC void MEM_check(void) { DEBUG_STATIC_ASSERT((sizeof(size_t)==4) || (sizeof(size_t)==8)); }
|
|
436
|
+
|
|
448
437
|
|
|
449
438
|
#if defined (__cplusplus)
|
|
450
439
|
}
|
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
/*
|
|
2
|
-
* Copyright (c)
|
|
2
|
+
* Copyright (c) Yann Collet, Facebook, Inc.
|
|
3
3
|
* All rights reserved.
|
|
4
4
|
*
|
|
5
5
|
* This source code is licensed under both the BSD-style license (found in the
|
|
@@ -10,9 +10,9 @@
|
|
|
10
10
|
|
|
11
11
|
|
|
12
12
|
/* ====== Dependencies ======= */
|
|
13
|
-
#include
|
|
13
|
+
#include "zstd_deps.h" /* size_t */
|
|
14
14
|
#include "debug.h" /* assert */
|
|
15
|
-
#include "zstd_internal.h" /*
|
|
15
|
+
#include "zstd_internal.h" /* ZSTD_customMalloc, ZSTD_customFree */
|
|
16
16
|
#include "pool.h"
|
|
17
17
|
|
|
18
18
|
/* ====== Compiler specifics ====== */
|
|
@@ -105,6 +105,10 @@ static void* POOL_thread(void* opaque) {
|
|
|
105
105
|
assert(0); /* Unreachable */
|
|
106
106
|
}
|
|
107
107
|
|
|
108
|
+
POOL_ctx* ZSTD_createThreadPool(size_t numThreads) {
|
|
109
|
+
return POOL_create (numThreads, 0);
|
|
110
|
+
}
|
|
111
|
+
|
|
108
112
|
POOL_ctx* POOL_create(size_t numThreads, size_t queueSize) {
|
|
109
113
|
return POOL_create_advanced(numThreads, queueSize, ZSTD_defaultCMem);
|
|
110
114
|
}
|
|
@@ -115,14 +119,14 @@ POOL_ctx* POOL_create_advanced(size_t numThreads, size_t queueSize,
|
|
|
115
119
|
/* Check parameters */
|
|
116
120
|
if (!numThreads) { return NULL; }
|
|
117
121
|
/* Allocate the context and zero initialize */
|
|
118
|
-
ctx = (POOL_ctx*)
|
|
122
|
+
ctx = (POOL_ctx*)ZSTD_customCalloc(sizeof(POOL_ctx), customMem);
|
|
119
123
|
if (!ctx) { return NULL; }
|
|
120
124
|
/* Initialize the job queue.
|
|
121
125
|
* It needs one extra space since one space is wasted to differentiate
|
|
122
126
|
* empty and full queues.
|
|
123
127
|
*/
|
|
124
128
|
ctx->queueSize = queueSize + 1;
|
|
125
|
-
ctx->queue = (POOL_job*)
|
|
129
|
+
ctx->queue = (POOL_job*)ZSTD_customMalloc(ctx->queueSize * sizeof(POOL_job), customMem);
|
|
126
130
|
ctx->queueHead = 0;
|
|
127
131
|
ctx->queueTail = 0;
|
|
128
132
|
ctx->numThreadsBusy = 0;
|
|
@@ -136,7 +140,7 @@ POOL_ctx* POOL_create_advanced(size_t numThreads, size_t queueSize,
|
|
|
136
140
|
}
|
|
137
141
|
ctx->shutdown = 0;
|
|
138
142
|
/* Allocate space for the thread handles */
|
|
139
|
-
ctx->threads = (ZSTD_pthread_t*)
|
|
143
|
+
ctx->threads = (ZSTD_pthread_t*)ZSTD_customMalloc(numThreads * sizeof(ZSTD_pthread_t), customMem);
|
|
140
144
|
ctx->threadCapacity = 0;
|
|
141
145
|
ctx->customMem = customMem;
|
|
142
146
|
/* Check for errors */
|
|
@@ -179,12 +183,14 @@ void POOL_free(POOL_ctx *ctx) {
|
|
|
179
183
|
ZSTD_pthread_mutex_destroy(&ctx->queueMutex);
|
|
180
184
|
ZSTD_pthread_cond_destroy(&ctx->queuePushCond);
|
|
181
185
|
ZSTD_pthread_cond_destroy(&ctx->queuePopCond);
|
|
182
|
-
|
|
183
|
-
|
|
184
|
-
|
|
186
|
+
ZSTD_customFree(ctx->queue, ctx->customMem);
|
|
187
|
+
ZSTD_customFree(ctx->threads, ctx->customMem);
|
|
188
|
+
ZSTD_customFree(ctx, ctx->customMem);
|
|
185
189
|
}
|
|
186
190
|
|
|
187
|
-
|
|
191
|
+
void ZSTD_freeThreadPool (ZSTD_threadPool* pool) {
|
|
192
|
+
POOL_free (pool);
|
|
193
|
+
}
|
|
188
194
|
|
|
189
195
|
size_t POOL_sizeof(POOL_ctx *ctx) {
|
|
190
196
|
if (ctx==NULL) return 0; /* supports sizeof NULL */
|
|
@@ -203,11 +209,11 @@ static int POOL_resize_internal(POOL_ctx* ctx, size_t numThreads)
|
|
|
203
209
|
return 0;
|
|
204
210
|
}
|
|
205
211
|
/* numThreads > threadCapacity */
|
|
206
|
-
{ ZSTD_pthread_t* const threadPool = (ZSTD_pthread_t*)
|
|
212
|
+
{ ZSTD_pthread_t* const threadPool = (ZSTD_pthread_t*)ZSTD_customMalloc(numThreads * sizeof(ZSTD_pthread_t), ctx->customMem);
|
|
207
213
|
if (!threadPool) return 1;
|
|
208
214
|
/* replace existing thread pool */
|
|
209
|
-
|
|
210
|
-
|
|
215
|
+
ZSTD_memcpy(threadPool, ctx->threads, ctx->threadCapacity * sizeof(*threadPool));
|
|
216
|
+
ZSTD_customFree(ctx->threads, ctx->customMem);
|
|
211
217
|
ctx->threads = threadPool;
|
|
212
218
|
/* Initialize additional threads */
|
|
213
219
|
{ size_t threadId;
|
|
@@ -301,7 +307,7 @@ int POOL_tryAdd(POOL_ctx* ctx, POOL_function function, void* opaque)
|
|
|
301
307
|
struct POOL_ctx_s {
|
|
302
308
|
int dummy;
|
|
303
309
|
};
|
|
304
|
-
static POOL_ctx
|
|
310
|
+
static POOL_ctx g_poolCtx;
|
|
305
311
|
|
|
306
312
|
POOL_ctx* POOL_create(size_t numThreads, size_t queueSize) {
|
|
307
313
|
return POOL_create_advanced(numThreads, queueSize, ZSTD_defaultCMem);
|
|
@@ -311,11 +317,11 @@ POOL_ctx* POOL_create_advanced(size_t numThreads, size_t queueSize, ZSTD_customM
|
|
|
311
317
|
(void)numThreads;
|
|
312
318
|
(void)queueSize;
|
|
313
319
|
(void)customMem;
|
|
314
|
-
return &
|
|
320
|
+
return &g_poolCtx;
|
|
315
321
|
}
|
|
316
322
|
|
|
317
323
|
void POOL_free(POOL_ctx* ctx) {
|
|
318
|
-
assert(!ctx || ctx == &
|
|
324
|
+
assert(!ctx || ctx == &g_poolCtx);
|
|
319
325
|
(void)ctx;
|
|
320
326
|
}
|
|
321
327
|
|
|
@@ -337,7 +343,7 @@ int POOL_tryAdd(POOL_ctx* ctx, POOL_function function, void* opaque) {
|
|
|
337
343
|
|
|
338
344
|
size_t POOL_sizeof(POOL_ctx* ctx) {
|
|
339
345
|
if (ctx==NULL) return 0; /* supports sizeof NULL */
|
|
340
|
-
assert(ctx == &
|
|
346
|
+
assert(ctx == &g_poolCtx);
|
|
341
347
|
return sizeof(*ctx);
|
|
342
348
|
}
|
|
343
349
|
|
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
/*
|
|
2
|
-
* Copyright (c)
|
|
2
|
+
* Copyright (c) Yann Collet, Facebook, Inc.
|
|
3
3
|
* All rights reserved.
|
|
4
4
|
*
|
|
5
5
|
* This source code is licensed under both the BSD-style license (found in the
|
|
@@ -16,9 +16,9 @@ extern "C" {
|
|
|
16
16
|
#endif
|
|
17
17
|
|
|
18
18
|
|
|
19
|
-
#include
|
|
19
|
+
#include "zstd_deps.h"
|
|
20
20
|
#define ZSTD_STATIC_LINKING_ONLY /* ZSTD_customMem */
|
|
21
|
-
#include "zstd.h"
|
|
21
|
+
#include "../zstd.h"
|
|
22
22
|
|
|
23
23
|
typedef struct POOL_ctx_s POOL_ctx;
|
|
24
24
|
|