zstd-ruby 1.4.4.0 → 1.5.5.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/.gitignore +2 -0
- data/README.md +78 -5
- data/Rakefile +8 -2
- data/ext/zstdruby/common.h +15 -0
- data/ext/zstdruby/extconf.rb +3 -2
- data/ext/zstdruby/libzstd/common/allocations.h +55 -0
- data/ext/zstdruby/libzstd/common/bits.h +200 -0
- data/ext/zstdruby/libzstd/common/bitstream.h +74 -97
- data/ext/zstdruby/libzstd/common/compiler.h +219 -20
- data/ext/zstdruby/libzstd/common/cpu.h +1 -3
- data/ext/zstdruby/libzstd/common/debug.c +11 -31
- data/ext/zstdruby/libzstd/common/debug.h +22 -49
- data/ext/zstdruby/libzstd/common/entropy_common.c +184 -80
- data/ext/zstdruby/libzstd/common/error_private.c +11 -2
- data/ext/zstdruby/libzstd/common/error_private.h +87 -4
- data/ext/zstdruby/libzstd/common/fse.h +47 -116
- data/ext/zstdruby/libzstd/common/fse_decompress.c +127 -127
- data/ext/zstdruby/libzstd/common/huf.h +112 -197
- data/ext/zstdruby/libzstd/common/mem.h +124 -142
- data/ext/zstdruby/libzstd/common/pool.c +54 -27
- data/ext/zstdruby/libzstd/common/pool.h +11 -5
- data/ext/zstdruby/libzstd/common/portability_macros.h +156 -0
- data/ext/zstdruby/libzstd/common/threading.c +78 -22
- data/ext/zstdruby/libzstd/common/threading.h +9 -13
- data/ext/zstdruby/libzstd/common/xxhash.c +15 -873
- data/ext/zstdruby/libzstd/common/xxhash.h +5572 -191
- data/ext/zstdruby/libzstd/common/zstd_common.c +2 -37
- data/ext/zstdruby/libzstd/common/zstd_deps.h +111 -0
- data/ext/zstdruby/libzstd/common/zstd_internal.h +186 -144
- data/ext/zstdruby/libzstd/common/zstd_trace.h +163 -0
- data/ext/zstdruby/libzstd/compress/clevels.h +134 -0
- data/ext/zstdruby/libzstd/compress/fse_compress.c +99 -196
- data/ext/zstdruby/libzstd/compress/hist.c +41 -63
- data/ext/zstdruby/libzstd/compress/hist.h +13 -33
- data/ext/zstdruby/libzstd/compress/huf_compress.c +968 -331
- data/ext/zstdruby/libzstd/compress/zstd_compress.c +4120 -1191
- data/ext/zstdruby/libzstd/compress/zstd_compress_internal.h +688 -159
- data/ext/zstdruby/libzstd/compress/zstd_compress_literals.c +121 -40
- data/ext/zstdruby/libzstd/compress/zstd_compress_literals.h +16 -6
- data/ext/zstdruby/libzstd/compress/zstd_compress_sequences.c +62 -35
- data/ext/zstdruby/libzstd/compress/zstd_compress_sequences.h +10 -3
- data/ext/zstdruby/libzstd/compress/zstd_compress_superblock.c +577 -0
- data/ext/zstdruby/libzstd/compress/zstd_compress_superblock.h +32 -0
- data/ext/zstdruby/libzstd/compress/zstd_cwksp.h +322 -115
- data/ext/zstdruby/libzstd/compress/zstd_double_fast.c +394 -154
- data/ext/zstdruby/libzstd/compress/zstd_double_fast.h +4 -3
- data/ext/zstdruby/libzstd/compress/zstd_fast.c +729 -253
- data/ext/zstdruby/libzstd/compress/zstd_fast.h +4 -3
- data/ext/zstdruby/libzstd/compress/zstd_lazy.c +1289 -247
- data/ext/zstdruby/libzstd/compress/zstd_lazy.h +61 -1
- data/ext/zstdruby/libzstd/compress/zstd_ldm.c +339 -212
- data/ext/zstdruby/libzstd/compress/zstd_ldm.h +15 -3
- data/ext/zstdruby/libzstd/compress/zstd_ldm_geartab.h +106 -0
- data/ext/zstdruby/libzstd/compress/zstd_opt.c +508 -282
- data/ext/zstdruby/libzstd/compress/zstd_opt.h +1 -1
- data/ext/zstdruby/libzstd/compress/zstdmt_compress.c +217 -466
- data/ext/zstdruby/libzstd/compress/zstdmt_compress.h +35 -114
- data/ext/zstdruby/libzstd/decompress/huf_decompress.c +1220 -572
- data/ext/zstdruby/libzstd/decompress/huf_decompress_amd64.S +576 -0
- data/ext/zstdruby/libzstd/decompress/zstd_ddict.c +23 -19
- data/ext/zstdruby/libzstd/decompress/zstd_ddict.h +3 -3
- data/ext/zstdruby/libzstd/decompress/zstd_decompress.c +859 -273
- data/ext/zstdruby/libzstd/decompress/zstd_decompress_block.c +1244 -375
- data/ext/zstdruby/libzstd/decompress/zstd_decompress_block.h +21 -7
- data/ext/zstdruby/libzstd/decompress/zstd_decompress_internal.h +74 -11
- data/ext/zstdruby/libzstd/dictBuilder/cover.c +75 -54
- data/ext/zstdruby/libzstd/dictBuilder/cover.h +20 -9
- data/ext/zstdruby/libzstd/dictBuilder/divsufsort.c +1 -1
- data/ext/zstdruby/libzstd/dictBuilder/fastcover.c +55 -36
- data/ext/zstdruby/libzstd/dictBuilder/zdict.c +126 -110
- data/ext/zstdruby/libzstd/{dictBuilder/zdict.h → zdict.h} +248 -56
- data/ext/zstdruby/libzstd/zstd.h +1277 -306
- data/ext/zstdruby/libzstd/{common/zstd_errors.h → zstd_errors.h} +29 -8
- data/ext/zstdruby/main.c +20 -0
- data/ext/zstdruby/skippable_frame.c +63 -0
- data/ext/zstdruby/streaming_compress.c +177 -0
- data/ext/zstdruby/streaming_compress.h +5 -0
- data/ext/zstdruby/streaming_decompress.c +123 -0
- data/ext/zstdruby/zstdruby.c +114 -32
- data/lib/zstd-ruby/version.rb +1 -1
- data/lib/zstd-ruby.rb +0 -1
- data/zstd-ruby.gemspec +1 -1
- metadata +24 -39
- data/.travis.yml +0 -14
- data/ext/zstdruby/libzstd/.gitignore +0 -3
- data/ext/zstdruby/libzstd/BUCK +0 -234
- data/ext/zstdruby/libzstd/Makefile +0 -289
- data/ext/zstdruby/libzstd/README.md +0 -159
- data/ext/zstdruby/libzstd/deprecated/zbuff.h +0 -214
- data/ext/zstdruby/libzstd/deprecated/zbuff_common.c +0 -26
- data/ext/zstdruby/libzstd/deprecated/zbuff_compress.c +0 -147
- data/ext/zstdruby/libzstd/deprecated/zbuff_decompress.c +0 -75
- data/ext/zstdruby/libzstd/dll/example/Makefile +0 -47
- data/ext/zstdruby/libzstd/dll/example/README.md +0 -69
- data/ext/zstdruby/libzstd/dll/example/build_package.bat +0 -20
- data/ext/zstdruby/libzstd/dll/example/fullbench-dll.sln +0 -25
- data/ext/zstdruby/libzstd/dll/example/fullbench-dll.vcxproj +0 -181
- data/ext/zstdruby/libzstd/legacy/zstd_legacy.h +0 -415
- data/ext/zstdruby/libzstd/legacy/zstd_v01.c +0 -2152
- data/ext/zstdruby/libzstd/legacy/zstd_v01.h +0 -94
- data/ext/zstdruby/libzstd/legacy/zstd_v02.c +0 -3514
- data/ext/zstdruby/libzstd/legacy/zstd_v02.h +0 -93
- data/ext/zstdruby/libzstd/legacy/zstd_v03.c +0 -3156
- data/ext/zstdruby/libzstd/legacy/zstd_v03.h +0 -93
- data/ext/zstdruby/libzstd/legacy/zstd_v04.c +0 -3641
- data/ext/zstdruby/libzstd/legacy/zstd_v04.h +0 -142
- data/ext/zstdruby/libzstd/legacy/zstd_v05.c +0 -4046
- data/ext/zstdruby/libzstd/legacy/zstd_v05.h +0 -162
- data/ext/zstdruby/libzstd/legacy/zstd_v06.c +0 -4150
- data/ext/zstdruby/libzstd/legacy/zstd_v06.h +0 -172
- data/ext/zstdruby/libzstd/legacy/zstd_v07.c +0 -4533
- data/ext/zstdruby/libzstd/legacy/zstd_v07.h +0 -187
- data/ext/zstdruby/libzstd/libzstd.pc.in +0 -15
- data/ext/zstdruby/zstdruby.h +0 -6
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
/*
|
|
2
|
-
* Copyright (c)
|
|
2
|
+
* Copyright (c) Meta Platforms, Inc. and affiliates.
|
|
3
3
|
* All rights reserved.
|
|
4
4
|
*
|
|
5
5
|
* This source code is licensed under both the BSD-style license (found in the
|
|
@@ -18,17 +18,17 @@
|
|
|
18
18
|
/*-*************************************
|
|
19
19
|
* Dependencies
|
|
20
20
|
***************************************/
|
|
21
|
-
#include "zstd_internal.h"
|
|
21
|
+
#include "../common/zstd_internal.h"
|
|
22
22
|
#include "zstd_cwksp.h"
|
|
23
23
|
#ifdef ZSTD_MULTITHREAD
|
|
24
24
|
# include "zstdmt_compress.h"
|
|
25
25
|
#endif
|
|
26
|
+
#include "../common/bits.h" /* ZSTD_highbit32, ZSTD_NbCommonBytes */
|
|
26
27
|
|
|
27
28
|
#if defined (__cplusplus)
|
|
28
29
|
extern "C" {
|
|
29
30
|
#endif
|
|
30
31
|
|
|
31
|
-
|
|
32
32
|
/*-*************************************
|
|
33
33
|
* Constants
|
|
34
34
|
***************************************/
|
|
@@ -64,7 +64,7 @@ typedef struct {
|
|
|
64
64
|
} ZSTD_localDict;
|
|
65
65
|
|
|
66
66
|
typedef struct {
|
|
67
|
-
|
|
67
|
+
HUF_CElt CTable[HUF_CTABLE_SIZE_ST(255)];
|
|
68
68
|
HUF_repeat repeatMode;
|
|
69
69
|
} ZSTD_hufCTables_t;
|
|
70
70
|
|
|
@@ -82,11 +82,82 @@ typedef struct {
|
|
|
82
82
|
ZSTD_fseCTables_t fse;
|
|
83
83
|
} ZSTD_entropyCTables_t;
|
|
84
84
|
|
|
85
|
+
/***********************************************
|
|
86
|
+
* Entropy buffer statistics structs and funcs *
|
|
87
|
+
***********************************************/
|
|
88
|
+
/** ZSTD_hufCTablesMetadata_t :
|
|
89
|
+
* Stores Literals Block Type for a super-block in hType, and
|
|
90
|
+
* huffman tree description in hufDesBuffer.
|
|
91
|
+
* hufDesSize refers to the size of huffman tree description in bytes.
|
|
92
|
+
* This metadata is populated in ZSTD_buildBlockEntropyStats_literals() */
|
|
85
93
|
typedef struct {
|
|
86
|
-
|
|
87
|
-
|
|
94
|
+
symbolEncodingType_e hType;
|
|
95
|
+
BYTE hufDesBuffer[ZSTD_MAX_HUF_HEADER_SIZE];
|
|
96
|
+
size_t hufDesSize;
|
|
97
|
+
} ZSTD_hufCTablesMetadata_t;
|
|
98
|
+
|
|
99
|
+
/** ZSTD_fseCTablesMetadata_t :
|
|
100
|
+
* Stores symbol compression modes for a super-block in {ll, ol, ml}Type, and
|
|
101
|
+
* fse tables in fseTablesBuffer.
|
|
102
|
+
* fseTablesSize refers to the size of fse tables in bytes.
|
|
103
|
+
* This metadata is populated in ZSTD_buildBlockEntropyStats_sequences() */
|
|
104
|
+
typedef struct {
|
|
105
|
+
symbolEncodingType_e llType;
|
|
106
|
+
symbolEncodingType_e ofType;
|
|
107
|
+
symbolEncodingType_e mlType;
|
|
108
|
+
BYTE fseTablesBuffer[ZSTD_MAX_FSE_HEADERS_SIZE];
|
|
109
|
+
size_t fseTablesSize;
|
|
110
|
+
size_t lastCountSize; /* This is to account for bug in 1.3.4. More detail in ZSTD_entropyCompressSeqStore_internal() */
|
|
111
|
+
} ZSTD_fseCTablesMetadata_t;
|
|
112
|
+
|
|
113
|
+
typedef struct {
|
|
114
|
+
ZSTD_hufCTablesMetadata_t hufMetadata;
|
|
115
|
+
ZSTD_fseCTablesMetadata_t fseMetadata;
|
|
116
|
+
} ZSTD_entropyCTablesMetadata_t;
|
|
117
|
+
|
|
118
|
+
/** ZSTD_buildBlockEntropyStats() :
|
|
119
|
+
* Builds entropy for the block.
|
|
120
|
+
* @return : 0 on success or error code */
|
|
121
|
+
size_t ZSTD_buildBlockEntropyStats(
|
|
122
|
+
const seqStore_t* seqStorePtr,
|
|
123
|
+
const ZSTD_entropyCTables_t* prevEntropy,
|
|
124
|
+
ZSTD_entropyCTables_t* nextEntropy,
|
|
125
|
+
const ZSTD_CCtx_params* cctxParams,
|
|
126
|
+
ZSTD_entropyCTablesMetadata_t* entropyMetadata,
|
|
127
|
+
void* workspace, size_t wkspSize);
|
|
128
|
+
|
|
129
|
+
/*********************************
|
|
130
|
+
* Compression internals structs *
|
|
131
|
+
*********************************/
|
|
132
|
+
|
|
133
|
+
typedef struct {
|
|
134
|
+
U32 off; /* Offset sumtype code for the match, using ZSTD_storeSeq() format */
|
|
135
|
+
U32 len; /* Raw length of match */
|
|
88
136
|
} ZSTD_match_t;
|
|
89
137
|
|
|
138
|
+
typedef struct {
|
|
139
|
+
U32 offset; /* Offset of sequence */
|
|
140
|
+
U32 litLength; /* Length of literals prior to match */
|
|
141
|
+
U32 matchLength; /* Raw length of match */
|
|
142
|
+
} rawSeq;
|
|
143
|
+
|
|
144
|
+
typedef struct {
|
|
145
|
+
rawSeq* seq; /* The start of the sequences */
|
|
146
|
+
size_t pos; /* The index in seq where reading stopped. pos <= size. */
|
|
147
|
+
size_t posInSequence; /* The position within the sequence at seq[pos] where reading
|
|
148
|
+
stopped. posInSequence <= seq[pos].litLength + seq[pos].matchLength */
|
|
149
|
+
size_t size; /* The number of sequences. <= capacity. */
|
|
150
|
+
size_t capacity; /* The capacity starting from `seq` pointer */
|
|
151
|
+
} rawSeqStore_t;
|
|
152
|
+
|
|
153
|
+
typedef struct {
|
|
154
|
+
U32 idx; /* Index in array of ZSTD_Sequence */
|
|
155
|
+
U32 posInSequence; /* Position within sequence at idx */
|
|
156
|
+
size_t posInSrc; /* Number of bytes given by sequences provided so far */
|
|
157
|
+
} ZSTD_sequencePosition;
|
|
158
|
+
|
|
159
|
+
UNUSED_ATTR static const rawSeqStore_t kNullRawSeqStore = {NULL, 0, 0, 0, 0};
|
|
160
|
+
|
|
90
161
|
typedef struct {
|
|
91
162
|
int price;
|
|
92
163
|
U32 off;
|
|
@@ -116,7 +187,7 @@ typedef struct {
|
|
|
116
187
|
U32 offCodeSumBasePrice; /* to compare to log2(offreq) */
|
|
117
188
|
ZSTD_OptPrice_e priceType; /* prices can be determined dynamically, or follow a pre-defined cost structure */
|
|
118
189
|
const ZSTD_entropyCTables_t* symbolCosts; /* pre-calculated dictionary statistics */
|
|
119
|
-
|
|
190
|
+
ZSTD_paramSwitch_e literalCompressionMode;
|
|
120
191
|
} optState_t;
|
|
121
192
|
|
|
122
193
|
typedef struct {
|
|
@@ -125,14 +196,23 @@ typedef struct {
|
|
|
125
196
|
} ZSTD_compressedBlockState_t;
|
|
126
197
|
|
|
127
198
|
typedef struct {
|
|
128
|
-
BYTE const* nextSrc;
|
|
129
|
-
BYTE const* base;
|
|
130
|
-
BYTE const* dictBase;
|
|
131
|
-
U32 dictLimit;
|
|
132
|
-
U32 lowLimit;
|
|
199
|
+
BYTE const* nextSrc; /* next block here to continue on current prefix */
|
|
200
|
+
BYTE const* base; /* All regular indexes relative to this position */
|
|
201
|
+
BYTE const* dictBase; /* extDict indexes relative to this position */
|
|
202
|
+
U32 dictLimit; /* below that point, need extDict */
|
|
203
|
+
U32 lowLimit; /* below that point, no more valid data */
|
|
204
|
+
U32 nbOverflowCorrections; /* Number of times overflow correction has run since
|
|
205
|
+
* ZSTD_window_init(). Useful for debugging coredumps
|
|
206
|
+
* and for ZSTD_WINDOW_OVERFLOW_CORRECT_FREQUENTLY.
|
|
207
|
+
*/
|
|
133
208
|
} ZSTD_window_t;
|
|
134
209
|
|
|
210
|
+
#define ZSTD_WINDOW_START_INDEX 2
|
|
211
|
+
|
|
135
212
|
typedef struct ZSTD_matchState_t ZSTD_matchState_t;
|
|
213
|
+
|
|
214
|
+
#define ZSTD_ROW_HASH_CACHE_SIZE 8 /* Size of prefetching hash cache for row-based matchfinder */
|
|
215
|
+
|
|
136
216
|
struct ZSTD_matchState_t {
|
|
137
217
|
ZSTD_window_t window; /* State for window round buffer management */
|
|
138
218
|
U32 loadedDictEnd; /* index of end of dictionary, within context's referential.
|
|
@@ -144,12 +224,38 @@ struct ZSTD_matchState_t {
|
|
|
144
224
|
*/
|
|
145
225
|
U32 nextToUpdate; /* index from which to continue table update */
|
|
146
226
|
U32 hashLog3; /* dispatch table for matches of len==3 : larger == faster, more memory */
|
|
227
|
+
|
|
228
|
+
U32 rowHashLog; /* For row-based matchfinder: Hashlog based on nb of rows in the hashTable.*/
|
|
229
|
+
BYTE* tagTable; /* For row-based matchFinder: A row-based table containing the hashes and head index. */
|
|
230
|
+
U32 hashCache[ZSTD_ROW_HASH_CACHE_SIZE]; /* For row-based matchFinder: a cache of hashes to improve speed */
|
|
231
|
+
U64 hashSalt; /* For row-based matchFinder: salts the hash for re-use of tag table */
|
|
232
|
+
U32 hashSaltEntropy; /* For row-based matchFinder: collects entropy for salt generation */
|
|
233
|
+
|
|
147
234
|
U32* hashTable;
|
|
148
235
|
U32* hashTable3;
|
|
149
236
|
U32* chainTable;
|
|
237
|
+
|
|
238
|
+
U32 forceNonContiguous; /* Non-zero if we should force non-contiguous load for the next window update. */
|
|
239
|
+
|
|
240
|
+
int dedicatedDictSearch; /* Indicates whether this matchState is using the
|
|
241
|
+
* dedicated dictionary search structure.
|
|
242
|
+
*/
|
|
150
243
|
optState_t opt; /* optimal parser state */
|
|
151
244
|
const ZSTD_matchState_t* dictMatchState;
|
|
152
245
|
ZSTD_compressionParameters cParams;
|
|
246
|
+
const rawSeqStore_t* ldmSeqStore;
|
|
247
|
+
|
|
248
|
+
/* Controls prefetching in some dictMatchState matchfinders.
|
|
249
|
+
* This behavior is controlled from the cctx ms.
|
|
250
|
+
* This parameter has no effect in the cdict ms. */
|
|
251
|
+
int prefetchCDictTables;
|
|
252
|
+
|
|
253
|
+
/* When == 0, lazy match finders insert every position.
|
|
254
|
+
* When != 0, lazy match finders only insert positions they search.
|
|
255
|
+
* This allows them to skip much faster over incompressible data,
|
|
256
|
+
* at a small cost to compression ratio.
|
|
257
|
+
*/
|
|
258
|
+
int lazySkipping;
|
|
153
259
|
};
|
|
154
260
|
|
|
155
261
|
typedef struct {
|
|
@@ -163,16 +269,26 @@ typedef struct {
|
|
|
163
269
|
U32 checksum;
|
|
164
270
|
} ldmEntry_t;
|
|
165
271
|
|
|
272
|
+
typedef struct {
|
|
273
|
+
BYTE const* split;
|
|
274
|
+
U32 hash;
|
|
275
|
+
U32 checksum;
|
|
276
|
+
ldmEntry_t* bucket;
|
|
277
|
+
} ldmMatchCandidate_t;
|
|
278
|
+
|
|
279
|
+
#define LDM_BATCH_SIZE 64
|
|
280
|
+
|
|
166
281
|
typedef struct {
|
|
167
282
|
ZSTD_window_t window; /* State for the window round buffer management */
|
|
168
283
|
ldmEntry_t* hashTable;
|
|
284
|
+
U32 loadedDictEnd;
|
|
169
285
|
BYTE* bucketOffsets; /* Next position in bucket to insert entry */
|
|
170
|
-
|
|
171
|
-
|
|
286
|
+
size_t splitIndices[LDM_BATCH_SIZE];
|
|
287
|
+
ldmMatchCandidate_t matchCandidates[LDM_BATCH_SIZE];
|
|
172
288
|
} ldmState_t;
|
|
173
289
|
|
|
174
290
|
typedef struct {
|
|
175
|
-
|
|
291
|
+
ZSTD_paramSwitch_e enableLdm; /* ZSTD_ps_enable to enable LDM. ZSTD_ps_auto by default */
|
|
176
292
|
U32 hashLog; /* Log size of hashTable */
|
|
177
293
|
U32 bucketSizeLog; /* Log bucket size for collision resolution, at most 8 */
|
|
178
294
|
U32 minMatchLength; /* Minimum match length */
|
|
@@ -180,19 +296,6 @@ typedef struct {
|
|
|
180
296
|
U32 windowLog; /* Window log for the LDM */
|
|
181
297
|
} ldmParams_t;
|
|
182
298
|
|
|
183
|
-
typedef struct {
|
|
184
|
-
U32 offset;
|
|
185
|
-
U32 litLength;
|
|
186
|
-
U32 matchLength;
|
|
187
|
-
} rawSeq;
|
|
188
|
-
|
|
189
|
-
typedef struct {
|
|
190
|
-
rawSeq* seq; /* The start of the sequences */
|
|
191
|
-
size_t pos; /* The position where reading stopped. <= size. */
|
|
192
|
-
size_t size; /* The number of sequences. <= capacity. */
|
|
193
|
-
size_t capacity; /* The capacity starting from `seq` pointer */
|
|
194
|
-
} rawSeqStore_t;
|
|
195
|
-
|
|
196
299
|
typedef struct {
|
|
197
300
|
int collectSequences;
|
|
198
301
|
ZSTD_Sequence* seqStart;
|
|
@@ -216,7 +319,7 @@ struct ZSTD_CCtx_params_s {
|
|
|
216
319
|
* There is no guarantee that hint is close to actual source size */
|
|
217
320
|
|
|
218
321
|
ZSTD_dictAttachPref_e attachDictPref;
|
|
219
|
-
|
|
322
|
+
ZSTD_paramSwitch_e literalCompressionMode;
|
|
220
323
|
|
|
221
324
|
/* Multithreading: used to pass parameters to mtctx */
|
|
222
325
|
int nbWorkers;
|
|
@@ -227,17 +330,94 @@ struct ZSTD_CCtx_params_s {
|
|
|
227
330
|
/* Long distance matching parameters */
|
|
228
331
|
ldmParams_t ldmParams;
|
|
229
332
|
|
|
333
|
+
/* Dedicated dict search algorithm trigger */
|
|
334
|
+
int enableDedicatedDictSearch;
|
|
335
|
+
|
|
336
|
+
/* Input/output buffer modes */
|
|
337
|
+
ZSTD_bufferMode_e inBufferMode;
|
|
338
|
+
ZSTD_bufferMode_e outBufferMode;
|
|
339
|
+
|
|
340
|
+
/* Sequence compression API */
|
|
341
|
+
ZSTD_sequenceFormat_e blockDelimiters;
|
|
342
|
+
int validateSequences;
|
|
343
|
+
|
|
344
|
+
/* Block splitting */
|
|
345
|
+
ZSTD_paramSwitch_e useBlockSplitter;
|
|
346
|
+
|
|
347
|
+
/* Param for deciding whether to use row-based matchfinder */
|
|
348
|
+
ZSTD_paramSwitch_e useRowMatchFinder;
|
|
349
|
+
|
|
350
|
+
/* Always load a dictionary in ext-dict mode (not prefix mode)? */
|
|
351
|
+
int deterministicRefPrefix;
|
|
352
|
+
|
|
230
353
|
/* Internal use, for createCCtxParams() and freeCCtxParams() only */
|
|
231
354
|
ZSTD_customMem customMem;
|
|
355
|
+
|
|
356
|
+
/* Controls prefetching in some dictMatchState matchfinders */
|
|
357
|
+
ZSTD_paramSwitch_e prefetchCDictTables;
|
|
358
|
+
|
|
359
|
+
/* Controls whether zstd will fall back to an internal matchfinder
|
|
360
|
+
* if the external matchfinder returns an error code. */
|
|
361
|
+
int enableMatchFinderFallback;
|
|
362
|
+
|
|
363
|
+
/* Indicates whether an external matchfinder has been referenced.
|
|
364
|
+
* Users can't set this externally.
|
|
365
|
+
* It is set internally in ZSTD_registerSequenceProducer(). */
|
|
366
|
+
int useSequenceProducer;
|
|
367
|
+
|
|
368
|
+
/* Adjust the max block size*/
|
|
369
|
+
size_t maxBlockSize;
|
|
370
|
+
|
|
371
|
+
/* Controls repcode search in external sequence parsing */
|
|
372
|
+
ZSTD_paramSwitch_e searchForExternalRepcodes;
|
|
232
373
|
}; /* typedef'd to ZSTD_CCtx_params within "zstd.h" */
|
|
233
374
|
|
|
375
|
+
#define COMPRESS_SEQUENCES_WORKSPACE_SIZE (sizeof(unsigned) * (MaxSeq + 2))
|
|
376
|
+
#define ENTROPY_WORKSPACE_SIZE (HUF_WORKSPACE_SIZE + COMPRESS_SEQUENCES_WORKSPACE_SIZE)
|
|
377
|
+
|
|
378
|
+
/**
|
|
379
|
+
* Indicates whether this compression proceeds directly from user-provided
|
|
380
|
+
* source buffer to user-provided destination buffer (ZSTDb_not_buffered), or
|
|
381
|
+
* whether the context needs to buffer the input/output (ZSTDb_buffered).
|
|
382
|
+
*/
|
|
383
|
+
typedef enum {
|
|
384
|
+
ZSTDb_not_buffered,
|
|
385
|
+
ZSTDb_buffered
|
|
386
|
+
} ZSTD_buffered_policy_e;
|
|
387
|
+
|
|
388
|
+
/**
|
|
389
|
+
* Struct that contains all elements of block splitter that should be allocated
|
|
390
|
+
* in a wksp.
|
|
391
|
+
*/
|
|
392
|
+
#define ZSTD_MAX_NB_BLOCK_SPLITS 196
|
|
393
|
+
typedef struct {
|
|
394
|
+
seqStore_t fullSeqStoreChunk;
|
|
395
|
+
seqStore_t firstHalfSeqStore;
|
|
396
|
+
seqStore_t secondHalfSeqStore;
|
|
397
|
+
seqStore_t currSeqStore;
|
|
398
|
+
seqStore_t nextSeqStore;
|
|
399
|
+
|
|
400
|
+
U32 partitions[ZSTD_MAX_NB_BLOCK_SPLITS];
|
|
401
|
+
ZSTD_entropyCTablesMetadata_t entropyMetadata;
|
|
402
|
+
} ZSTD_blockSplitCtx;
|
|
403
|
+
|
|
404
|
+
/* Context for block-level external matchfinder API */
|
|
405
|
+
typedef struct {
|
|
406
|
+
void* mState;
|
|
407
|
+
ZSTD_sequenceProducer_F* mFinder;
|
|
408
|
+
ZSTD_Sequence* seqBuffer;
|
|
409
|
+
size_t seqBufferCapacity;
|
|
410
|
+
} ZSTD_externalMatchCtx;
|
|
411
|
+
|
|
234
412
|
struct ZSTD_CCtx_s {
|
|
235
413
|
ZSTD_compressionStage_e stage;
|
|
236
414
|
int cParamsChanged; /* == 1 if cParams(except wlog) or compression level are changed in requestedParams. Triggers transmission of new params to ZSTDMT (if available) then reset to 0. */
|
|
237
415
|
int bmi2; /* == 1 if the CPU supports BMI2 and 0 otherwise. CPU support is determined dynamically once per context lifetime. */
|
|
238
416
|
ZSTD_CCtx_params requestedParams;
|
|
239
417
|
ZSTD_CCtx_params appliedParams;
|
|
418
|
+
ZSTD_CCtx_params simpleApiParams; /* Param storage used by the simple API - not sticky. Must only be used in top-level simple API functions for storage. */
|
|
240
419
|
U32 dictID;
|
|
420
|
+
size_t dictContentSize;
|
|
241
421
|
|
|
242
422
|
ZSTD_cwksp workspace; /* manages buffer for dynamic allocations */
|
|
243
423
|
size_t blockSize;
|
|
@@ -246,9 +426,11 @@ struct ZSTD_CCtx_s {
|
|
|
246
426
|
unsigned long long producedCSize;
|
|
247
427
|
XXH64_state_t xxhState;
|
|
248
428
|
ZSTD_customMem customMem;
|
|
429
|
+
ZSTD_threadPool* pool;
|
|
249
430
|
size_t staticSize;
|
|
250
431
|
SeqCollector seqCollector;
|
|
251
432
|
int isFirstBlock;
|
|
433
|
+
int initialized;
|
|
252
434
|
|
|
253
435
|
seqStore_t seqStore; /* sequences storage ptrs */
|
|
254
436
|
ldmState_t ldmState; /* long distance matching state */
|
|
@@ -256,7 +438,10 @@ struct ZSTD_CCtx_s {
|
|
|
256
438
|
size_t maxNbLdmSequences;
|
|
257
439
|
rawSeqStore_t externSeqStore; /* Mutable reference to external sequences */
|
|
258
440
|
ZSTD_blockState_t blockState;
|
|
259
|
-
U32* entropyWorkspace; /* entropy workspace of
|
|
441
|
+
U32* entropyWorkspace; /* entropy workspace of ENTROPY_WORKSPACE_SIZE bytes */
|
|
442
|
+
|
|
443
|
+
/* Whether we are streaming or not */
|
|
444
|
+
ZSTD_buffered_policy_e bufferedPolicy;
|
|
260
445
|
|
|
261
446
|
/* streaming */
|
|
262
447
|
char* inBuff;
|
|
@@ -271,6 +456,11 @@ struct ZSTD_CCtx_s {
|
|
|
271
456
|
ZSTD_cStreamStage streamStage;
|
|
272
457
|
U32 frameEnded;
|
|
273
458
|
|
|
459
|
+
/* Stable in/out buffer verification */
|
|
460
|
+
ZSTD_inBuffer expectedInBuffer;
|
|
461
|
+
size_t stableIn_notConsumed; /* nb bytes within stable input buffer that are said to be consumed but are not */
|
|
462
|
+
size_t expectedOutBufferSize;
|
|
463
|
+
|
|
274
464
|
/* Dictionary */
|
|
275
465
|
ZSTD_localDict localDict;
|
|
276
466
|
const ZSTD_CDict* cdict;
|
|
@@ -280,17 +470,53 @@ struct ZSTD_CCtx_s {
|
|
|
280
470
|
#ifdef ZSTD_MULTITHREAD
|
|
281
471
|
ZSTDMT_CCtx* mtctx;
|
|
282
472
|
#endif
|
|
283
|
-
};
|
|
284
473
|
|
|
285
|
-
|
|
474
|
+
/* Tracing */
|
|
475
|
+
#if ZSTD_TRACE
|
|
476
|
+
ZSTD_TraceCtx traceCtx;
|
|
477
|
+
#endif
|
|
286
478
|
|
|
287
|
-
|
|
479
|
+
/* Workspace for block splitter */
|
|
480
|
+
ZSTD_blockSplitCtx blockSplitCtx;
|
|
288
481
|
|
|
482
|
+
/* Workspace for external matchfinder */
|
|
483
|
+
ZSTD_externalMatchCtx externalMatchCtx;
|
|
484
|
+
};
|
|
485
|
+
|
|
486
|
+
typedef enum { ZSTD_dtlm_fast, ZSTD_dtlm_full } ZSTD_dictTableLoadMethod_e;
|
|
487
|
+
typedef enum { ZSTD_tfp_forCCtx, ZSTD_tfp_forCDict } ZSTD_tableFillPurpose_e;
|
|
488
|
+
|
|
489
|
+
typedef enum {
|
|
490
|
+
ZSTD_noDict = 0,
|
|
491
|
+
ZSTD_extDict = 1,
|
|
492
|
+
ZSTD_dictMatchState = 2,
|
|
493
|
+
ZSTD_dedicatedDictSearch = 3
|
|
494
|
+
} ZSTD_dictMode_e;
|
|
495
|
+
|
|
496
|
+
typedef enum {
|
|
497
|
+
ZSTD_cpm_noAttachDict = 0, /* Compression with ZSTD_noDict or ZSTD_extDict.
|
|
498
|
+
* In this mode we use both the srcSize and the dictSize
|
|
499
|
+
* when selecting and adjusting parameters.
|
|
500
|
+
*/
|
|
501
|
+
ZSTD_cpm_attachDict = 1, /* Compression with ZSTD_dictMatchState or ZSTD_dedicatedDictSearch.
|
|
502
|
+
* In this mode we only take the srcSize into account when selecting
|
|
503
|
+
* and adjusting parameters.
|
|
504
|
+
*/
|
|
505
|
+
ZSTD_cpm_createCDict = 2, /* Creating a CDict.
|
|
506
|
+
* In this mode we take both the source size and the dictionary size
|
|
507
|
+
* into account when selecting and adjusting the parameters.
|
|
508
|
+
*/
|
|
509
|
+
ZSTD_cpm_unknown = 3 /* ZSTD_getCParams, ZSTD_getParams, ZSTD_adjustParams.
|
|
510
|
+
* We don't know what these parameters are for. We default to the legacy
|
|
511
|
+
* behavior of taking both the source size and the dict size into account
|
|
512
|
+
* when selecting and adjusting parameters.
|
|
513
|
+
*/
|
|
514
|
+
} ZSTD_cParamMode_e;
|
|
289
515
|
|
|
290
516
|
typedef size_t (*ZSTD_blockCompressor) (
|
|
291
517
|
ZSTD_matchState_t* bs, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
|
292
518
|
void const* src, size_t srcSize);
|
|
293
|
-
ZSTD_blockCompressor ZSTD_selectBlockCompressor(ZSTD_strategy strat, ZSTD_dictMode_e dictMode);
|
|
519
|
+
ZSTD_blockCompressor ZSTD_selectBlockCompressor(ZSTD_strategy strat, ZSTD_paramSwitch_e rowMatchfinderMode, ZSTD_dictMode_e dictMode);
|
|
294
520
|
|
|
295
521
|
|
|
296
522
|
MEM_STATIC U32 ZSTD_LLcode(U32 litLength)
|
|
@@ -336,6 +562,33 @@ MEM_STATIC int ZSTD_cParam_withinBounds(ZSTD_cParameter cParam, int value)
|
|
|
336
562
|
return 1;
|
|
337
563
|
}
|
|
338
564
|
|
|
565
|
+
/* ZSTD_noCompressBlock() :
|
|
566
|
+
* Writes uncompressed block to dst buffer from given src.
|
|
567
|
+
* Returns the size of the block */
|
|
568
|
+
MEM_STATIC size_t
|
|
569
|
+
ZSTD_noCompressBlock(void* dst, size_t dstCapacity, const void* src, size_t srcSize, U32 lastBlock)
|
|
570
|
+
{
|
|
571
|
+
U32 const cBlockHeader24 = lastBlock + (((U32)bt_raw)<<1) + (U32)(srcSize << 3);
|
|
572
|
+
DEBUGLOG(5, "ZSTD_noCompressBlock (srcSize=%zu, dstCapacity=%zu)", srcSize, dstCapacity);
|
|
573
|
+
RETURN_ERROR_IF(srcSize + ZSTD_blockHeaderSize > dstCapacity,
|
|
574
|
+
dstSize_tooSmall, "dst buf too small for uncompressed block");
|
|
575
|
+
MEM_writeLE24(dst, cBlockHeader24);
|
|
576
|
+
ZSTD_memcpy((BYTE*)dst + ZSTD_blockHeaderSize, src, srcSize);
|
|
577
|
+
return ZSTD_blockHeaderSize + srcSize;
|
|
578
|
+
}
|
|
579
|
+
|
|
580
|
+
MEM_STATIC size_t
|
|
581
|
+
ZSTD_rleCompressBlock(void* dst, size_t dstCapacity, BYTE src, size_t srcSize, U32 lastBlock)
|
|
582
|
+
{
|
|
583
|
+
BYTE* const op = (BYTE*)dst;
|
|
584
|
+
U32 const cBlockHeader = lastBlock + (((U32)bt_rle)<<1) + (U32)(srcSize << 3);
|
|
585
|
+
RETURN_ERROR_IF(dstCapacity < 4, dstSize_tooSmall, "");
|
|
586
|
+
MEM_writeLE24(op, cBlockHeader);
|
|
587
|
+
op[3] = src;
|
|
588
|
+
return 4;
|
|
589
|
+
}
|
|
590
|
+
|
|
591
|
+
|
|
339
592
|
/* ZSTD_minGain() :
|
|
340
593
|
* minimum compression required
|
|
341
594
|
* to generate a compress block or a compressed literals section.
|
|
@@ -344,16 +597,33 @@ MEM_STATIC size_t ZSTD_minGain(size_t srcSize, ZSTD_strategy strat)
|
|
|
344
597
|
{
|
|
345
598
|
U32 const minlog = (strat>=ZSTD_btultra) ? (U32)(strat) - 1 : 6;
|
|
346
599
|
ZSTD_STATIC_ASSERT(ZSTD_btultra == 8);
|
|
347
|
-
assert(ZSTD_cParam_withinBounds(ZSTD_c_strategy, strat));
|
|
600
|
+
assert(ZSTD_cParam_withinBounds(ZSTD_c_strategy, (int)strat));
|
|
348
601
|
return (srcSize >> minlog) + 2;
|
|
349
602
|
}
|
|
350
603
|
|
|
604
|
+
MEM_STATIC int ZSTD_literalsCompressionIsDisabled(const ZSTD_CCtx_params* cctxParams)
|
|
605
|
+
{
|
|
606
|
+
switch (cctxParams->literalCompressionMode) {
|
|
607
|
+
case ZSTD_ps_enable:
|
|
608
|
+
return 0;
|
|
609
|
+
case ZSTD_ps_disable:
|
|
610
|
+
return 1;
|
|
611
|
+
default:
|
|
612
|
+
assert(0 /* impossible: pre-validated */);
|
|
613
|
+
ZSTD_FALLTHROUGH;
|
|
614
|
+
case ZSTD_ps_auto:
|
|
615
|
+
return (cctxParams->cParams.strategy == ZSTD_fast) && (cctxParams->cParams.targetLength > 0);
|
|
616
|
+
}
|
|
617
|
+
}
|
|
618
|
+
|
|
351
619
|
/*! ZSTD_safecopyLiterals() :
|
|
352
620
|
* memcpy() function that won't read beyond more than WILDCOPY_OVERLENGTH bytes past ilimit_w.
|
|
353
621
|
* Only called when the sequence ends past ilimit_w, so it only needs to be optimized for single
|
|
354
622
|
* large copies.
|
|
355
623
|
*/
|
|
356
|
-
static void
|
|
624
|
+
static void
|
|
625
|
+
ZSTD_safecopyLiterals(BYTE* op, BYTE const* ip, BYTE const* const iend, BYTE const* ilimit_w)
|
|
626
|
+
{
|
|
357
627
|
assert(iend > ilimit_w);
|
|
358
628
|
if (ip <= ilimit_w) {
|
|
359
629
|
ZSTD_wildcopy(op, ip, ilimit_w - ip, ZSTD_no_overlap);
|
|
@@ -363,14 +633,28 @@ static void ZSTD_safecopyLiterals(BYTE* op, BYTE const* ip, BYTE const* const ie
|
|
|
363
633
|
while (ip < iend) *op++ = *ip++;
|
|
364
634
|
}
|
|
365
635
|
|
|
636
|
+
|
|
637
|
+
#define REPCODE1_TO_OFFBASE REPCODE_TO_OFFBASE(1)
|
|
638
|
+
#define REPCODE2_TO_OFFBASE REPCODE_TO_OFFBASE(2)
|
|
639
|
+
#define REPCODE3_TO_OFFBASE REPCODE_TO_OFFBASE(3)
|
|
640
|
+
#define REPCODE_TO_OFFBASE(r) (assert((r)>=1), assert((r)<=ZSTD_REP_NUM), (r)) /* accepts IDs 1,2,3 */
|
|
641
|
+
#define OFFSET_TO_OFFBASE(o) (assert((o)>0), o + ZSTD_REP_NUM)
|
|
642
|
+
#define OFFBASE_IS_OFFSET(o) ((o) > ZSTD_REP_NUM)
|
|
643
|
+
#define OFFBASE_IS_REPCODE(o) ( 1 <= (o) && (o) <= ZSTD_REP_NUM)
|
|
644
|
+
#define OFFBASE_TO_OFFSET(o) (assert(OFFBASE_IS_OFFSET(o)), (o) - ZSTD_REP_NUM)
|
|
645
|
+
#define OFFBASE_TO_REPCODE(o) (assert(OFFBASE_IS_REPCODE(o)), (o)) /* returns ID 1,2,3 */
|
|
646
|
+
|
|
366
647
|
/*! ZSTD_storeSeq() :
|
|
367
|
-
* Store a sequence (litlen, litPtr,
|
|
368
|
-
*
|
|
369
|
-
*
|
|
370
|
-
* Allowed to
|
|
648
|
+
* Store a sequence (litlen, litPtr, offBase and matchLength) into seqStore_t.
|
|
649
|
+
* @offBase : Users should employ macros REPCODE_TO_OFFBASE() and OFFSET_TO_OFFBASE().
|
|
650
|
+
* @matchLength : must be >= MINMATCH
|
|
651
|
+
* Allowed to over-read literals up to litLimit.
|
|
371
652
|
*/
|
|
372
|
-
HINT_INLINE UNUSED_ATTR
|
|
373
|
-
|
|
653
|
+
HINT_INLINE UNUSED_ATTR void
|
|
654
|
+
ZSTD_storeSeq(seqStore_t* seqStorePtr,
|
|
655
|
+
size_t litLength, const BYTE* literals, const BYTE* litLimit,
|
|
656
|
+
U32 offBase,
|
|
657
|
+
size_t matchLength)
|
|
374
658
|
{
|
|
375
659
|
BYTE const* const litLimit_w = litLimit - WILDCOPY_OVERLENGTH;
|
|
376
660
|
BYTE const* const litEnd = literals + litLength;
|
|
@@ -378,8 +662,8 @@ void ZSTD_storeSeq(seqStore_t* seqStorePtr, size_t litLength, const BYTE* litera
|
|
|
378
662
|
static const BYTE* g_start = NULL;
|
|
379
663
|
if (g_start==NULL) g_start = (const BYTE*)literals; /* note : index only works for compression within a single segment */
|
|
380
664
|
{ U32 const pos = (U32)((const BYTE*)literals - g_start);
|
|
381
|
-
DEBUGLOG(6, "Cpos%7u :%3u literals, match%4u bytes at
|
|
382
|
-
pos, (U32)litLength, (U32)
|
|
665
|
+
DEBUGLOG(6, "Cpos%7u :%3u literals, match%4u bytes at offBase%7u",
|
|
666
|
+
pos, (U32)litLength, (U32)matchLength, (U32)offBase);
|
|
383
667
|
}
|
|
384
668
|
#endif
|
|
385
669
|
assert((size_t)(seqStorePtr->sequences - seqStorePtr->sequencesStart) < seqStorePtr->maxNbSeq);
|
|
@@ -389,9 +673,9 @@ void ZSTD_storeSeq(seqStore_t* seqStorePtr, size_t litLength, const BYTE* litera
|
|
|
389
673
|
assert(literals + litLength <= litLimit);
|
|
390
674
|
if (litEnd <= litLimit_w) {
|
|
391
675
|
/* Common case we can use wildcopy.
|
|
392
|
-
|
|
393
|
-
|
|
394
|
-
|
|
676
|
+
* First copy 16 bytes, because literals are likely short.
|
|
677
|
+
*/
|
|
678
|
+
ZSTD_STATIC_ASSERT(WILDCOPY_OVERLENGTH >= 16);
|
|
395
679
|
ZSTD_copy16(seqStorePtr->lit, literals);
|
|
396
680
|
if (litLength > 16) {
|
|
397
681
|
ZSTD_wildcopy(seqStorePtr->lit+16, literals+16, (ptrdiff_t)litLength-16, ZSTD_no_overlap);
|
|
@@ -403,99 +687,70 @@ void ZSTD_storeSeq(seqStore_t* seqStorePtr, size_t litLength, const BYTE* litera
|
|
|
403
687
|
|
|
404
688
|
/* literal Length */
|
|
405
689
|
if (litLength>0xFFFF) {
|
|
406
|
-
assert(seqStorePtr->
|
|
407
|
-
seqStorePtr->
|
|
690
|
+
assert(seqStorePtr->longLengthType == ZSTD_llt_none); /* there can only be a single long length */
|
|
691
|
+
seqStorePtr->longLengthType = ZSTD_llt_literalLength;
|
|
408
692
|
seqStorePtr->longLengthPos = (U32)(seqStorePtr->sequences - seqStorePtr->sequencesStart);
|
|
409
693
|
}
|
|
410
694
|
seqStorePtr->sequences[0].litLength = (U16)litLength;
|
|
411
695
|
|
|
412
696
|
/* match offset */
|
|
413
|
-
seqStorePtr->sequences[0].
|
|
697
|
+
seqStorePtr->sequences[0].offBase = offBase;
|
|
414
698
|
|
|
415
699
|
/* match Length */
|
|
416
|
-
|
|
417
|
-
|
|
418
|
-
|
|
419
|
-
|
|
700
|
+
assert(matchLength >= MINMATCH);
|
|
701
|
+
{ size_t const mlBase = matchLength - MINMATCH;
|
|
702
|
+
if (mlBase>0xFFFF) {
|
|
703
|
+
assert(seqStorePtr->longLengthType == ZSTD_llt_none); /* there can only be a single long length */
|
|
704
|
+
seqStorePtr->longLengthType = ZSTD_llt_matchLength;
|
|
705
|
+
seqStorePtr->longLengthPos = (U32)(seqStorePtr->sequences - seqStorePtr->sequencesStart);
|
|
706
|
+
}
|
|
707
|
+
seqStorePtr->sequences[0].mlBase = (U16)mlBase;
|
|
420
708
|
}
|
|
421
|
-
seqStorePtr->sequences[0].matchLength = (U16)mlBase;
|
|
422
709
|
|
|
423
710
|
seqStorePtr->sequences++;
|
|
424
711
|
}
|
|
425
712
|
|
|
426
|
-
|
|
427
|
-
|
|
428
|
-
*
|
|
429
|
-
|
|
430
|
-
|
|
431
|
-
|
|
432
|
-
|
|
433
|
-
|
|
434
|
-
|
|
435
|
-
|
|
436
|
-
|
|
437
|
-
|
|
438
|
-
|
|
439
|
-
|
|
440
|
-
|
|
441
|
-
|
|
442
|
-
|
|
443
|
-
|
|
444
|
-
|
|
445
|
-
|
|
446
|
-
2, 6, 5, 5, 3, 4, 5, 6,
|
|
447
|
-
7, 1, 2, 4, 6, 4, 4, 5,
|
|
448
|
-
7, 2, 6, 5, 7, 6, 7, 7 };
|
|
449
|
-
return DeBruijnBytePos[((U64)((val & -(long long)val) * 0x0218A392CDABBD3FULL)) >> 58];
|
|
450
|
-
# endif
|
|
451
|
-
} else { /* 32 bits */
|
|
452
|
-
# if defined(_MSC_VER)
|
|
453
|
-
unsigned long r=0;
|
|
454
|
-
_BitScanForward( &r, (U32)val );
|
|
455
|
-
return (unsigned)(r>>3);
|
|
456
|
-
# elif defined(__GNUC__) && (__GNUC__ >= 3)
|
|
457
|
-
return (__builtin_ctz((U32)val) >> 3);
|
|
458
|
-
# else
|
|
459
|
-
static const int DeBruijnBytePos[32] = { 0, 0, 3, 0, 3, 1, 3, 0,
|
|
460
|
-
3, 2, 2, 1, 3, 2, 0, 1,
|
|
461
|
-
3, 3, 1, 2, 2, 2, 2, 0,
|
|
462
|
-
3, 1, 2, 0, 1, 0, 1, 1 };
|
|
463
|
-
return DeBruijnBytePos[((U32)((val & -(S32)val) * 0x077CB531U)) >> 27];
|
|
464
|
-
# endif
|
|
713
|
+
/* ZSTD_updateRep() :
|
|
714
|
+
* updates in-place @rep (array of repeat offsets)
|
|
715
|
+
* @offBase : sum-type, using numeric representation of ZSTD_storeSeq()
|
|
716
|
+
*/
|
|
717
|
+
MEM_STATIC void
|
|
718
|
+
ZSTD_updateRep(U32 rep[ZSTD_REP_NUM], U32 const offBase, U32 const ll0)
|
|
719
|
+
{
|
|
720
|
+
if (OFFBASE_IS_OFFSET(offBase)) { /* full offset */
|
|
721
|
+
rep[2] = rep[1];
|
|
722
|
+
rep[1] = rep[0];
|
|
723
|
+
rep[0] = OFFBASE_TO_OFFSET(offBase);
|
|
724
|
+
} else { /* repcode */
|
|
725
|
+
U32 const repCode = OFFBASE_TO_REPCODE(offBase) - 1 + ll0;
|
|
726
|
+
if (repCode > 0) { /* note : if repCode==0, no change */
|
|
727
|
+
U32 const currentOffset = (repCode==ZSTD_REP_NUM) ? (rep[0] - 1) : rep[repCode];
|
|
728
|
+
rep[2] = (repCode >= 2) ? rep[1] : rep[2];
|
|
729
|
+
rep[1] = rep[0];
|
|
730
|
+
rep[0] = currentOffset;
|
|
731
|
+
} else { /* repCode == 0 */
|
|
732
|
+
/* nothing to do */
|
|
465
733
|
}
|
|
466
|
-
}
|
|
467
|
-
|
|
468
|
-
|
|
469
|
-
|
|
470
|
-
|
|
471
|
-
|
|
472
|
-
|
|
473
|
-
|
|
474
|
-
|
|
475
|
-
|
|
476
|
-
|
|
477
|
-
|
|
478
|
-
|
|
479
|
-
|
|
480
|
-
return r;
|
|
481
|
-
# endif
|
|
482
|
-
} else { /* 32 bits */
|
|
483
|
-
# if defined(_MSC_VER)
|
|
484
|
-
unsigned long r = 0;
|
|
485
|
-
_BitScanReverse( &r, (unsigned long)val );
|
|
486
|
-
return (unsigned)(r>>3);
|
|
487
|
-
# elif defined(__GNUC__) && (__GNUC__ >= 3)
|
|
488
|
-
return (__builtin_clz((U32)val) >> 3);
|
|
489
|
-
# else
|
|
490
|
-
unsigned r;
|
|
491
|
-
if (!(val>>16)) { r=2; val>>=8; } else { r=0; val>>=24; }
|
|
492
|
-
r += (!val);
|
|
493
|
-
return r;
|
|
494
|
-
# endif
|
|
495
|
-
} }
|
|
734
|
+
}
|
|
735
|
+
}
|
|
736
|
+
|
|
737
|
+
typedef struct repcodes_s {
|
|
738
|
+
U32 rep[3];
|
|
739
|
+
} repcodes_t;
|
|
740
|
+
|
|
741
|
+
MEM_STATIC repcodes_t
|
|
742
|
+
ZSTD_newRep(U32 const rep[ZSTD_REP_NUM], U32 const offBase, U32 const ll0)
|
|
743
|
+
{
|
|
744
|
+
repcodes_t newReps;
|
|
745
|
+
ZSTD_memcpy(&newReps, rep, sizeof(newReps));
|
|
746
|
+
ZSTD_updateRep(newReps.rep, offBase, ll0);
|
|
747
|
+
return newReps;
|
|
496
748
|
}
|
|
497
749
|
|
|
498
750
|
|
|
751
|
+
/*-*************************************
|
|
752
|
+
* Match length counter
|
|
753
|
+
***************************************/
|
|
499
754
|
MEM_STATIC size_t ZSTD_count(const BYTE* pIn, const BYTE* pMatch, const BYTE* const pInLimit)
|
|
500
755
|
{
|
|
501
756
|
const BYTE* const pStart = pIn;
|
|
@@ -541,31 +796,43 @@ ZSTD_count_2segments(const BYTE* ip, const BYTE* match,
|
|
|
541
796
|
* Hashes
|
|
542
797
|
***************************************/
|
|
543
798
|
static const U32 prime3bytes = 506832829U;
|
|
544
|
-
static U32 ZSTD_hash3(U32 u, U32 h) { return ((u << (32-24)) * prime3bytes) >> (32-h) ; }
|
|
545
|
-
MEM_STATIC size_t ZSTD_hash3Ptr(const void* ptr, U32 h) { return ZSTD_hash3(MEM_readLE32(ptr), h); } /* only in zstd_opt.h */
|
|
799
|
+
static U32 ZSTD_hash3(U32 u, U32 h, U32 s) { assert(h <= 32); return (((u << (32-24)) * prime3bytes) ^ s) >> (32-h) ; }
|
|
800
|
+
MEM_STATIC size_t ZSTD_hash3Ptr(const void* ptr, U32 h) { return ZSTD_hash3(MEM_readLE32(ptr), h, 0); } /* only in zstd_opt.h */
|
|
801
|
+
MEM_STATIC size_t ZSTD_hash3PtrS(const void* ptr, U32 h, U32 s) { return ZSTD_hash3(MEM_readLE32(ptr), h, s); }
|
|
546
802
|
|
|
547
803
|
static const U32 prime4bytes = 2654435761U;
|
|
548
|
-
static U32 ZSTD_hash4(U32 u, U32 h) { return (u * prime4bytes) >> (32-h) ; }
|
|
549
|
-
static size_t ZSTD_hash4Ptr(const void* ptr, U32 h) { return ZSTD_hash4(
|
|
804
|
+
static U32 ZSTD_hash4(U32 u, U32 h, U32 s) { assert(h <= 32); return ((u * prime4bytes) ^ s) >> (32-h) ; }
|
|
805
|
+
static size_t ZSTD_hash4Ptr(const void* ptr, U32 h) { return ZSTD_hash4(MEM_readLE32(ptr), h, 0); }
|
|
806
|
+
static size_t ZSTD_hash4PtrS(const void* ptr, U32 h, U32 s) { return ZSTD_hash4(MEM_readLE32(ptr), h, s); }
|
|
550
807
|
|
|
551
808
|
static const U64 prime5bytes = 889523592379ULL;
|
|
552
|
-
static size_t ZSTD_hash5(U64 u, U32 h) { return (size_t)(((u << (64-40)) * prime5bytes) >> (64-h)) ; }
|
|
553
|
-
static size_t ZSTD_hash5Ptr(const void* p, U32 h) { return ZSTD_hash5(MEM_readLE64(p), h); }
|
|
809
|
+
static size_t ZSTD_hash5(U64 u, U32 h, U64 s) { assert(h <= 64); return (size_t)((((u << (64-40)) * prime5bytes) ^ s) >> (64-h)) ; }
|
|
810
|
+
static size_t ZSTD_hash5Ptr(const void* p, U32 h) { return ZSTD_hash5(MEM_readLE64(p), h, 0); }
|
|
811
|
+
static size_t ZSTD_hash5PtrS(const void* p, U32 h, U64 s) { return ZSTD_hash5(MEM_readLE64(p), h, s); }
|
|
554
812
|
|
|
555
813
|
static const U64 prime6bytes = 227718039650203ULL;
|
|
556
|
-
static size_t ZSTD_hash6(U64 u, U32 h) { return (size_t)(((u << (64-48)) * prime6bytes) >> (64-h)) ; }
|
|
557
|
-
static size_t ZSTD_hash6Ptr(const void* p, U32 h) { return ZSTD_hash6(MEM_readLE64(p), h); }
|
|
814
|
+
static size_t ZSTD_hash6(U64 u, U32 h, U64 s) { assert(h <= 64); return (size_t)((((u << (64-48)) * prime6bytes) ^ s) >> (64-h)) ; }
|
|
815
|
+
static size_t ZSTD_hash6Ptr(const void* p, U32 h) { return ZSTD_hash6(MEM_readLE64(p), h, 0); }
|
|
816
|
+
static size_t ZSTD_hash6PtrS(const void* p, U32 h, U64 s) { return ZSTD_hash6(MEM_readLE64(p), h, s); }
|
|
558
817
|
|
|
559
818
|
static const U64 prime7bytes = 58295818150454627ULL;
|
|
560
|
-
static size_t ZSTD_hash7(U64 u, U32 h) { return (size_t)(((u << (64-56)) * prime7bytes) >> (64-h)) ; }
|
|
561
|
-
static size_t ZSTD_hash7Ptr(const void* p, U32 h) { return ZSTD_hash7(MEM_readLE64(p), h); }
|
|
819
|
+
static size_t ZSTD_hash7(U64 u, U32 h, U64 s) { assert(h <= 64); return (size_t)((((u << (64-56)) * prime7bytes) ^ s) >> (64-h)) ; }
|
|
820
|
+
static size_t ZSTD_hash7Ptr(const void* p, U32 h) { return ZSTD_hash7(MEM_readLE64(p), h, 0); }
|
|
821
|
+
static size_t ZSTD_hash7PtrS(const void* p, U32 h, U64 s) { return ZSTD_hash7(MEM_readLE64(p), h, s); }
|
|
562
822
|
|
|
563
823
|
static const U64 prime8bytes = 0xCF1BBCDCB7A56463ULL;
|
|
564
|
-
static size_t ZSTD_hash8(U64 u, U32 h) { return (size_t)(((u) * prime8bytes) >> (64-h)) ; }
|
|
565
|
-
static size_t ZSTD_hash8Ptr(const void* p, U32 h) { return ZSTD_hash8(MEM_readLE64(p), h); }
|
|
824
|
+
static size_t ZSTD_hash8(U64 u, U32 h, U64 s) { assert(h <= 64); return (size_t)((((u) * prime8bytes) ^ s) >> (64-h)) ; }
|
|
825
|
+
static size_t ZSTD_hash8Ptr(const void* p, U32 h) { return ZSTD_hash8(MEM_readLE64(p), h, 0); }
|
|
826
|
+
static size_t ZSTD_hash8PtrS(const void* p, U32 h, U64 s) { return ZSTD_hash8(MEM_readLE64(p), h, s); }
|
|
827
|
+
|
|
566
828
|
|
|
567
|
-
MEM_STATIC
|
|
829
|
+
MEM_STATIC FORCE_INLINE_ATTR
|
|
830
|
+
size_t ZSTD_hashPtr(const void* p, U32 hBits, U32 mls)
|
|
568
831
|
{
|
|
832
|
+
/* Although some of these hashes do support hBits up to 64, some do not.
|
|
833
|
+
* To be on the safe side, always avoid hBits > 32. */
|
|
834
|
+
assert(hBits <= 32);
|
|
835
|
+
|
|
569
836
|
switch(mls)
|
|
570
837
|
{
|
|
571
838
|
default:
|
|
@@ -577,6 +844,24 @@ MEM_STATIC size_t ZSTD_hashPtr(const void* p, U32 hBits, U32 mls)
|
|
|
577
844
|
}
|
|
578
845
|
}
|
|
579
846
|
|
|
847
|
+
MEM_STATIC FORCE_INLINE_ATTR
|
|
848
|
+
size_t ZSTD_hashPtrSalted(const void* p, U32 hBits, U32 mls, const U64 hashSalt) {
|
|
849
|
+
/* Although some of these hashes do support hBits up to 64, some do not.
|
|
850
|
+
* To be on the safe side, always avoid hBits > 32. */
|
|
851
|
+
assert(hBits <= 32);
|
|
852
|
+
|
|
853
|
+
switch(mls)
|
|
854
|
+
{
|
|
855
|
+
default:
|
|
856
|
+
case 4: return ZSTD_hash4PtrS(p, hBits, (U32)hashSalt);
|
|
857
|
+
case 5: return ZSTD_hash5PtrS(p, hBits, hashSalt);
|
|
858
|
+
case 6: return ZSTD_hash6PtrS(p, hBits, hashSalt);
|
|
859
|
+
case 7: return ZSTD_hash7PtrS(p, hBits, hashSalt);
|
|
860
|
+
case 8: return ZSTD_hash8PtrS(p, hBits, hashSalt);
|
|
861
|
+
}
|
|
862
|
+
}
|
|
863
|
+
|
|
864
|
+
|
|
580
865
|
/** ZSTD_ipow() :
|
|
581
866
|
* Return base^exponent.
|
|
582
867
|
*/
|
|
@@ -661,6 +946,13 @@ MEM_STATIC void ZSTD_window_clear(ZSTD_window_t* window)
|
|
|
661
946
|
window->dictLimit = end;
|
|
662
947
|
}
|
|
663
948
|
|
|
949
|
+
MEM_STATIC U32 ZSTD_window_isEmpty(ZSTD_window_t const window)
|
|
950
|
+
{
|
|
951
|
+
return window.dictLimit == ZSTD_WINDOW_START_INDEX &&
|
|
952
|
+
window.lowLimit == ZSTD_WINDOW_START_INDEX &&
|
|
953
|
+
(window.nextSrc - window.base) == ZSTD_WINDOW_START_INDEX;
|
|
954
|
+
}
|
|
955
|
+
|
|
664
956
|
/**
|
|
665
957
|
* ZSTD_window_hasExtDict():
|
|
666
958
|
* Returns non-zero if the window has a non-empty extDict.
|
|
@@ -680,20 +972,76 @@ MEM_STATIC ZSTD_dictMode_e ZSTD_matchState_dictMode(const ZSTD_matchState_t *ms)
|
|
|
680
972
|
return ZSTD_window_hasExtDict(ms->window) ?
|
|
681
973
|
ZSTD_extDict :
|
|
682
974
|
ms->dictMatchState != NULL ?
|
|
683
|
-
ZSTD_dictMatchState :
|
|
975
|
+
(ms->dictMatchState->dedicatedDictSearch ? ZSTD_dedicatedDictSearch : ZSTD_dictMatchState) :
|
|
684
976
|
ZSTD_noDict;
|
|
685
977
|
}
|
|
686
978
|
|
|
979
|
+
/* Defining this macro to non-zero tells zstd to run the overflow correction
|
|
980
|
+
* code much more frequently. This is very inefficient, and should only be
|
|
981
|
+
* used for tests and fuzzers.
|
|
982
|
+
*/
|
|
983
|
+
#ifndef ZSTD_WINDOW_OVERFLOW_CORRECT_FREQUENTLY
|
|
984
|
+
# ifdef FUZZING_BUILD_MODE_UNSAFE_FOR_PRODUCTION
|
|
985
|
+
# define ZSTD_WINDOW_OVERFLOW_CORRECT_FREQUENTLY 1
|
|
986
|
+
# else
|
|
987
|
+
# define ZSTD_WINDOW_OVERFLOW_CORRECT_FREQUENTLY 0
|
|
988
|
+
# endif
|
|
989
|
+
#endif
|
|
990
|
+
|
|
991
|
+
/**
|
|
992
|
+
* ZSTD_window_canOverflowCorrect():
|
|
993
|
+
* Returns non-zero if the indices are large enough for overflow correction
|
|
994
|
+
* to work correctly without impacting compression ratio.
|
|
995
|
+
*/
|
|
996
|
+
MEM_STATIC U32 ZSTD_window_canOverflowCorrect(ZSTD_window_t const window,
|
|
997
|
+
U32 cycleLog,
|
|
998
|
+
U32 maxDist,
|
|
999
|
+
U32 loadedDictEnd,
|
|
1000
|
+
void const* src)
|
|
1001
|
+
{
|
|
1002
|
+
U32 const cycleSize = 1u << cycleLog;
|
|
1003
|
+
U32 const curr = (U32)((BYTE const*)src - window.base);
|
|
1004
|
+
U32 const minIndexToOverflowCorrect = cycleSize
|
|
1005
|
+
+ MAX(maxDist, cycleSize)
|
|
1006
|
+
+ ZSTD_WINDOW_START_INDEX;
|
|
1007
|
+
|
|
1008
|
+
/* Adjust the min index to backoff the overflow correction frequency,
|
|
1009
|
+
* so we don't waste too much CPU in overflow correction. If this
|
|
1010
|
+
* computation overflows we don't really care, we just need to make
|
|
1011
|
+
* sure it is at least minIndexToOverflowCorrect.
|
|
1012
|
+
*/
|
|
1013
|
+
U32 const adjustment = window.nbOverflowCorrections + 1;
|
|
1014
|
+
U32 const adjustedIndex = MAX(minIndexToOverflowCorrect * adjustment,
|
|
1015
|
+
minIndexToOverflowCorrect);
|
|
1016
|
+
U32 const indexLargeEnough = curr > adjustedIndex;
|
|
1017
|
+
|
|
1018
|
+
/* Only overflow correct early if the dictionary is invalidated already,
|
|
1019
|
+
* so we don't hurt compression ratio.
|
|
1020
|
+
*/
|
|
1021
|
+
U32 const dictionaryInvalidated = curr > maxDist + loadedDictEnd;
|
|
1022
|
+
|
|
1023
|
+
return indexLargeEnough && dictionaryInvalidated;
|
|
1024
|
+
}
|
|
1025
|
+
|
|
687
1026
|
/**
|
|
688
1027
|
* ZSTD_window_needOverflowCorrection():
|
|
689
1028
|
* Returns non-zero if the indices are getting too large and need overflow
|
|
690
1029
|
* protection.
|
|
691
1030
|
*/
|
|
692
1031
|
MEM_STATIC U32 ZSTD_window_needOverflowCorrection(ZSTD_window_t const window,
|
|
1032
|
+
U32 cycleLog,
|
|
1033
|
+
U32 maxDist,
|
|
1034
|
+
U32 loadedDictEnd,
|
|
1035
|
+
void const* src,
|
|
693
1036
|
void const* srcEnd)
|
|
694
1037
|
{
|
|
695
|
-
U32 const
|
|
696
|
-
|
|
1038
|
+
U32 const curr = (U32)((BYTE const*)srcEnd - window.base);
|
|
1039
|
+
if (ZSTD_WINDOW_OVERFLOW_CORRECT_FREQUENTLY) {
|
|
1040
|
+
if (ZSTD_window_canOverflowCorrect(window, cycleLog, maxDist, loadedDictEnd, src)) {
|
|
1041
|
+
return 1;
|
|
1042
|
+
}
|
|
1043
|
+
}
|
|
1044
|
+
return curr > ZSTD_CURRENT_MAX;
|
|
697
1045
|
}
|
|
698
1046
|
|
|
699
1047
|
/**
|
|
@@ -704,7 +1052,6 @@ MEM_STATIC U32 ZSTD_window_needOverflowCorrection(ZSTD_window_t const window,
|
|
|
704
1052
|
*
|
|
705
1053
|
* The least significant cycleLog bits of the indices must remain the same,
|
|
706
1054
|
* which may be 0. Every index up to maxDist in the past must be valid.
|
|
707
|
-
* NOTE: (maxDist & cycleMask) must be zero.
|
|
708
1055
|
*/
|
|
709
1056
|
MEM_STATIC U32 ZSTD_window_correctOverflow(ZSTD_window_t* window, U32 cycleLog,
|
|
710
1057
|
U32 maxDist, void const* src)
|
|
@@ -728,19 +1075,51 @@ MEM_STATIC U32 ZSTD_window_correctOverflow(ZSTD_window_t* window, U32 cycleLog,
|
|
|
728
1075
|
* 3. (cctx->lowLimit + 1<<windowLog) < 1<<32:
|
|
729
1076
|
* windowLog <= 31 ==> 3<<29 + 1<<windowLog < 7<<29 < 1<<32.
|
|
730
1077
|
*/
|
|
731
|
-
U32 const
|
|
732
|
-
U32 const
|
|
733
|
-
U32 const
|
|
734
|
-
U32 const
|
|
735
|
-
|
|
736
|
-
|
|
737
|
-
|
|
738
|
-
|
|
1078
|
+
U32 const cycleSize = 1u << cycleLog;
|
|
1079
|
+
U32 const cycleMask = cycleSize - 1;
|
|
1080
|
+
U32 const curr = (U32)((BYTE const*)src - window->base);
|
|
1081
|
+
U32 const currentCycle = curr & cycleMask;
|
|
1082
|
+
/* Ensure newCurrent - maxDist >= ZSTD_WINDOW_START_INDEX. */
|
|
1083
|
+
U32 const currentCycleCorrection = currentCycle < ZSTD_WINDOW_START_INDEX
|
|
1084
|
+
? MAX(cycleSize, ZSTD_WINDOW_START_INDEX)
|
|
1085
|
+
: 0;
|
|
1086
|
+
U32 const newCurrent = currentCycle
|
|
1087
|
+
+ currentCycleCorrection
|
|
1088
|
+
+ MAX(maxDist, cycleSize);
|
|
1089
|
+
U32 const correction = curr - newCurrent;
|
|
1090
|
+
/* maxDist must be a power of two so that:
|
|
1091
|
+
* (newCurrent & cycleMask) == (curr & cycleMask)
|
|
1092
|
+
* This is required to not corrupt the chains / binary tree.
|
|
1093
|
+
*/
|
|
1094
|
+
assert((maxDist & (maxDist - 1)) == 0);
|
|
1095
|
+
assert((curr & cycleMask) == (newCurrent & cycleMask));
|
|
1096
|
+
assert(curr > newCurrent);
|
|
1097
|
+
if (!ZSTD_WINDOW_OVERFLOW_CORRECT_FREQUENTLY) {
|
|
1098
|
+
/* Loose bound, should be around 1<<29 (see above) */
|
|
1099
|
+
assert(correction > 1<<28);
|
|
1100
|
+
}
|
|
739
1101
|
|
|
740
1102
|
window->base += correction;
|
|
741
1103
|
window->dictBase += correction;
|
|
742
|
-
window->lowLimit
|
|
743
|
-
|
|
1104
|
+
if (window->lowLimit < correction + ZSTD_WINDOW_START_INDEX) {
|
|
1105
|
+
window->lowLimit = ZSTD_WINDOW_START_INDEX;
|
|
1106
|
+
} else {
|
|
1107
|
+
window->lowLimit -= correction;
|
|
1108
|
+
}
|
|
1109
|
+
if (window->dictLimit < correction + ZSTD_WINDOW_START_INDEX) {
|
|
1110
|
+
window->dictLimit = ZSTD_WINDOW_START_INDEX;
|
|
1111
|
+
} else {
|
|
1112
|
+
window->dictLimit -= correction;
|
|
1113
|
+
}
|
|
1114
|
+
|
|
1115
|
+
/* Ensure we can still reference the full window. */
|
|
1116
|
+
assert(newCurrent >= maxDist);
|
|
1117
|
+
assert(newCurrent - maxDist >= ZSTD_WINDOW_START_INDEX);
|
|
1118
|
+
/* Ensure that lowLimit and dictLimit didn't underflow. */
|
|
1119
|
+
assert(window->lowLimit <= newCurrent);
|
|
1120
|
+
assert(window->dictLimit <= newCurrent);
|
|
1121
|
+
|
|
1122
|
+
++window->nbOverflowCorrections;
|
|
744
1123
|
|
|
745
1124
|
DEBUGLOG(4, "Correction of 0x%x bytes to lowLimit=0x%x", correction,
|
|
746
1125
|
window->lowLimit);
|
|
@@ -830,10 +1209,15 @@ ZSTD_checkDictValidity(const ZSTD_window_t* window,
|
|
|
830
1209
|
(unsigned)blockEndIdx, (unsigned)maxDist, (unsigned)loadedDictEnd);
|
|
831
1210
|
assert(blockEndIdx >= loadedDictEnd);
|
|
832
1211
|
|
|
833
|
-
if (blockEndIdx > loadedDictEnd + maxDist) {
|
|
1212
|
+
if (blockEndIdx > loadedDictEnd + maxDist || loadedDictEnd != window->dictLimit) {
|
|
834
1213
|
/* On reaching window size, dictionaries are invalidated.
|
|
835
1214
|
* For simplification, if window size is reached anywhere within next block,
|
|
836
1215
|
* the dictionary is invalidated for the full block.
|
|
1216
|
+
*
|
|
1217
|
+
* We also have to invalidate the dictionary if ZSTD_window_update() has detected
|
|
1218
|
+
* non-contiguous segments, which means that loadedDictEnd != window->dictLimit.
|
|
1219
|
+
* loadedDictEnd may be 0, if forceWindow is true, but in that case we never use
|
|
1220
|
+
* dictMatchState, so setting it to NULL is not a problem.
|
|
837
1221
|
*/
|
|
838
1222
|
DEBUGLOG(6, "invalidating dictionary for current block (distance > windowSize)");
|
|
839
1223
|
*loadedDictEndPtr = 0;
|
|
@@ -844,6 +1228,17 @@ ZSTD_checkDictValidity(const ZSTD_window_t* window,
|
|
|
844
1228
|
} } }
|
|
845
1229
|
}
|
|
846
1230
|
|
|
1231
|
+
MEM_STATIC void ZSTD_window_init(ZSTD_window_t* window) {
|
|
1232
|
+
ZSTD_memset(window, 0, sizeof(*window));
|
|
1233
|
+
window->base = (BYTE const*)" ";
|
|
1234
|
+
window->dictBase = (BYTE const*)" ";
|
|
1235
|
+
ZSTD_STATIC_ASSERT(ZSTD_DUBT_UNSORTED_MARK < ZSTD_WINDOW_START_INDEX); /* Start above ZSTD_DUBT_UNSORTED_MARK */
|
|
1236
|
+
window->dictLimit = ZSTD_WINDOW_START_INDEX; /* start from >0, so that 1st position is valid */
|
|
1237
|
+
window->lowLimit = ZSTD_WINDOW_START_INDEX; /* it ensures first and later CCtx usages compress the same */
|
|
1238
|
+
window->nextSrc = window->base + ZSTD_WINDOW_START_INDEX; /* see issue #1241 */
|
|
1239
|
+
window->nbOverflowCorrections = 0;
|
|
1240
|
+
}
|
|
1241
|
+
|
|
847
1242
|
/**
|
|
848
1243
|
* ZSTD_window_update():
|
|
849
1244
|
* Updates the window by appending [src, src + srcSize) to the window.
|
|
@@ -852,13 +1247,18 @@ ZSTD_checkDictValidity(const ZSTD_window_t* window,
|
|
|
852
1247
|
* Returns non-zero if the segment is contiguous.
|
|
853
1248
|
*/
|
|
854
1249
|
MEM_STATIC U32 ZSTD_window_update(ZSTD_window_t* window,
|
|
855
|
-
void const* src, size_t srcSize
|
|
1250
|
+
void const* src, size_t srcSize,
|
|
1251
|
+
int forceNonContiguous)
|
|
856
1252
|
{
|
|
857
1253
|
BYTE const* const ip = (BYTE const*)src;
|
|
858
1254
|
U32 contiguous = 1;
|
|
859
1255
|
DEBUGLOG(5, "ZSTD_window_update");
|
|
1256
|
+
if (srcSize == 0)
|
|
1257
|
+
return contiguous;
|
|
1258
|
+
assert(window->base != NULL);
|
|
1259
|
+
assert(window->dictBase != NULL);
|
|
860
1260
|
/* Check if blocks follow each other */
|
|
861
|
-
if (src != window->nextSrc) {
|
|
1261
|
+
if (src != window->nextSrc || forceNonContiguous) {
|
|
862
1262
|
/* not contiguous */
|
|
863
1263
|
size_t const distanceFromBase = (size_t)(window->nextSrc - window->base);
|
|
864
1264
|
DEBUGLOG(5, "Non contiguous blocks, new segment starts at %u", window->dictLimit);
|
|
@@ -867,7 +1267,7 @@ MEM_STATIC U32 ZSTD_window_update(ZSTD_window_t* window,
|
|
|
867
1267
|
window->dictLimit = (U32)distanceFromBase;
|
|
868
1268
|
window->dictBase = window->base;
|
|
869
1269
|
window->base = ip - distanceFromBase;
|
|
870
|
-
|
|
1270
|
+
/* ms->nextToUpdate = window->dictLimit; */
|
|
871
1271
|
if (window->dictLimit - window->lowLimit < HASH_READ_SIZE) window->lowLimit = window->dictLimit; /* too small extDict */
|
|
872
1272
|
contiguous = 0;
|
|
873
1273
|
}
|
|
@@ -883,12 +1283,35 @@ MEM_STATIC U32 ZSTD_window_update(ZSTD_window_t* window,
|
|
|
883
1283
|
return contiguous;
|
|
884
1284
|
}
|
|
885
1285
|
|
|
886
|
-
|
|
1286
|
+
/**
|
|
1287
|
+
* Returns the lowest allowed match index. It may either be in the ext-dict or the prefix.
|
|
1288
|
+
*/
|
|
1289
|
+
MEM_STATIC U32 ZSTD_getLowestMatchIndex(const ZSTD_matchState_t* ms, U32 curr, unsigned windowLog)
|
|
1290
|
+
{
|
|
1291
|
+
U32 const maxDistance = 1U << windowLog;
|
|
1292
|
+
U32 const lowestValid = ms->window.lowLimit;
|
|
1293
|
+
U32 const withinWindow = (curr - lowestValid > maxDistance) ? curr - maxDistance : lowestValid;
|
|
1294
|
+
U32 const isDictionary = (ms->loadedDictEnd != 0);
|
|
1295
|
+
/* When using a dictionary the entire dictionary is valid if a single byte of the dictionary
|
|
1296
|
+
* is within the window. We invalidate the dictionary (and set loadedDictEnd to 0) when it isn't
|
|
1297
|
+
* valid for the entire block. So this check is sufficient to find the lowest valid match index.
|
|
1298
|
+
*/
|
|
1299
|
+
U32 const matchLowest = isDictionary ? lowestValid : withinWindow;
|
|
1300
|
+
return matchLowest;
|
|
1301
|
+
}
|
|
1302
|
+
|
|
1303
|
+
/**
|
|
1304
|
+
* Returns the lowest allowed match index in the prefix.
|
|
1305
|
+
*/
|
|
1306
|
+
MEM_STATIC U32 ZSTD_getLowestPrefixIndex(const ZSTD_matchState_t* ms, U32 curr, unsigned windowLog)
|
|
887
1307
|
{
|
|
888
1308
|
U32 const maxDistance = 1U << windowLog;
|
|
889
|
-
U32 const lowestValid = ms->window.
|
|
890
|
-
U32 const withinWindow = (
|
|
1309
|
+
U32 const lowestValid = ms->window.dictLimit;
|
|
1310
|
+
U32 const withinWindow = (curr - lowestValid > maxDistance) ? curr - maxDistance : lowestValid;
|
|
891
1311
|
U32 const isDictionary = (ms->loadedDictEnd != 0);
|
|
1312
|
+
/* When computing the lowest prefix index we need to take the dictionary into account to handle
|
|
1313
|
+
* the edge case where the dictionary and the source are contiguous in memory.
|
|
1314
|
+
*/
|
|
892
1315
|
U32 const matchLowest = isDictionary ? lowestValid : withinWindow;
|
|
893
1316
|
return matchLowest;
|
|
894
1317
|
}
|
|
@@ -926,11 +1349,61 @@ MEM_STATIC void ZSTD_debugTable(const U32* table, U32 max)
|
|
|
926
1349
|
|
|
927
1350
|
#endif
|
|
928
1351
|
|
|
1352
|
+
/* Short Cache */
|
|
1353
|
+
|
|
1354
|
+
/* Normally, zstd matchfinders follow this flow:
|
|
1355
|
+
* 1. Compute hash at ip
|
|
1356
|
+
* 2. Load index from hashTable[hash]
|
|
1357
|
+
* 3. Check if *ip == *(base + index)
|
|
1358
|
+
* In dictionary compression, loading *(base + index) is often an L2 or even L3 miss.
|
|
1359
|
+
*
|
|
1360
|
+
* Short cache is an optimization which allows us to avoid step 3 most of the time
|
|
1361
|
+
* when the data doesn't actually match. With short cache, the flow becomes:
|
|
1362
|
+
* 1. Compute (hash, currentTag) at ip. currentTag is an 8-bit independent hash at ip.
|
|
1363
|
+
* 2. Load (index, matchTag) from hashTable[hash]. See ZSTD_writeTaggedIndex to understand how this works.
|
|
1364
|
+
* 3. Only if currentTag == matchTag, check *ip == *(base + index). Otherwise, continue.
|
|
1365
|
+
*
|
|
1366
|
+
* Currently, short cache is only implemented in CDict hashtables. Thus, its use is limited to
|
|
1367
|
+
* dictMatchState matchfinders.
|
|
1368
|
+
*/
|
|
1369
|
+
#define ZSTD_SHORT_CACHE_TAG_BITS 8
|
|
1370
|
+
#define ZSTD_SHORT_CACHE_TAG_MASK ((1u << ZSTD_SHORT_CACHE_TAG_BITS) - 1)
|
|
1371
|
+
|
|
1372
|
+
/* Helper function for ZSTD_fillHashTable and ZSTD_fillDoubleHashTable.
|
|
1373
|
+
* Unpacks hashAndTag into (hash, tag), then packs (index, tag) into hashTable[hash]. */
|
|
1374
|
+
MEM_STATIC void ZSTD_writeTaggedIndex(U32* const hashTable, size_t hashAndTag, U32 index) {
|
|
1375
|
+
size_t const hash = hashAndTag >> ZSTD_SHORT_CACHE_TAG_BITS;
|
|
1376
|
+
U32 const tag = (U32)(hashAndTag & ZSTD_SHORT_CACHE_TAG_MASK);
|
|
1377
|
+
assert(index >> (32 - ZSTD_SHORT_CACHE_TAG_BITS) == 0);
|
|
1378
|
+
hashTable[hash] = (index << ZSTD_SHORT_CACHE_TAG_BITS) | tag;
|
|
1379
|
+
}
|
|
1380
|
+
|
|
1381
|
+
/* Helper function for short cache matchfinders.
|
|
1382
|
+
* Unpacks tag1 and tag2 from lower bits of packedTag1 and packedTag2, then checks if the tags match. */
|
|
1383
|
+
MEM_STATIC int ZSTD_comparePackedTags(size_t packedTag1, size_t packedTag2) {
|
|
1384
|
+
U32 const tag1 = packedTag1 & ZSTD_SHORT_CACHE_TAG_MASK;
|
|
1385
|
+
U32 const tag2 = packedTag2 & ZSTD_SHORT_CACHE_TAG_MASK;
|
|
1386
|
+
return tag1 == tag2;
|
|
1387
|
+
}
|
|
929
1388
|
|
|
930
1389
|
#if defined (__cplusplus)
|
|
931
1390
|
}
|
|
932
1391
|
#endif
|
|
933
1392
|
|
|
1393
|
+
/* ===============================================================
|
|
1394
|
+
* Shared internal declarations
|
|
1395
|
+
* These prototypes may be called from sources not in lib/compress
|
|
1396
|
+
* =============================================================== */
|
|
1397
|
+
|
|
1398
|
+
/* ZSTD_loadCEntropy() :
|
|
1399
|
+
* dict : must point at beginning of a valid zstd dictionary.
|
|
1400
|
+
* return : size of dictionary header (size of magic number + dict ID + entropy tables)
|
|
1401
|
+
* assumptions : magic number supposed already checked
|
|
1402
|
+
* and dictSize >= 8 */
|
|
1403
|
+
size_t ZSTD_loadCEntropy(ZSTD_compressedBlockState_t* bs, void* workspace,
|
|
1404
|
+
const void* const dict, size_t dictSize);
|
|
1405
|
+
|
|
1406
|
+
void ZSTD_reset_compressedBlockState(ZSTD_compressedBlockState_t* bs);
|
|
934
1407
|
|
|
935
1408
|
/* ==============================================================
|
|
936
1409
|
* Private declarations
|
|
@@ -940,9 +1413,10 @@ MEM_STATIC void ZSTD_debugTable(const U32* table, U32 max)
|
|
|
940
1413
|
/* ZSTD_getCParamsFromCCtxParams() :
|
|
941
1414
|
* cParams are built depending on compressionLevel, src size hints,
|
|
942
1415
|
* LDM and manually set compression parameters.
|
|
1416
|
+
* Note: srcSizeHint == 0 means 0!
|
|
943
1417
|
*/
|
|
944
1418
|
ZSTD_compressionParameters ZSTD_getCParamsFromCCtxParams(
|
|
945
|
-
const ZSTD_CCtx_params* CCtxParams, U64 srcSizeHint, size_t dictSize);
|
|
1419
|
+
const ZSTD_CCtx_params* CCtxParams, U64 srcSizeHint, size_t dictSize, ZSTD_cParamMode_e mode);
|
|
946
1420
|
|
|
947
1421
|
/*! ZSTD_initCStream_internal() :
|
|
948
1422
|
* Private use only. Init streaming operation.
|
|
@@ -999,5 +1473,60 @@ size_t ZSTD_writeLastEmptyBlock(void* dst, size_t dstCapacity);
|
|
|
999
1473
|
*/
|
|
1000
1474
|
size_t ZSTD_referenceExternalSequences(ZSTD_CCtx* cctx, rawSeq* seq, size_t nbSeq);
|
|
1001
1475
|
|
|
1476
|
+
/** ZSTD_cycleLog() :
|
|
1477
|
+
* condition for correct operation : hashLog > 1 */
|
|
1478
|
+
U32 ZSTD_cycleLog(U32 hashLog, ZSTD_strategy strat);
|
|
1479
|
+
|
|
1480
|
+
/** ZSTD_CCtx_trace() :
|
|
1481
|
+
* Trace the end of a compression call.
|
|
1482
|
+
*/
|
|
1483
|
+
void ZSTD_CCtx_trace(ZSTD_CCtx* cctx, size_t extraCSize);
|
|
1484
|
+
|
|
1485
|
+
/* Returns 0 on success, and a ZSTD_error otherwise. This function scans through an array of
|
|
1486
|
+
* ZSTD_Sequence, storing the sequences it finds, until it reaches a block delimiter.
|
|
1487
|
+
* Note that the block delimiter must include the last literals of the block.
|
|
1488
|
+
*/
|
|
1489
|
+
size_t
|
|
1490
|
+
ZSTD_copySequencesToSeqStoreExplicitBlockDelim(ZSTD_CCtx* cctx,
|
|
1491
|
+
ZSTD_sequencePosition* seqPos,
|
|
1492
|
+
const ZSTD_Sequence* const inSeqs, size_t inSeqsSize,
|
|
1493
|
+
const void* src, size_t blockSize, ZSTD_paramSwitch_e externalRepSearch);
|
|
1494
|
+
|
|
1495
|
+
/* Returns the number of bytes to move the current read position back by.
|
|
1496
|
+
* Only non-zero if we ended up splitting a sequence.
|
|
1497
|
+
* Otherwise, it may return a ZSTD error if something went wrong.
|
|
1498
|
+
*
|
|
1499
|
+
* This function will attempt to scan through blockSize bytes
|
|
1500
|
+
* represented by the sequences in @inSeqs,
|
|
1501
|
+
* storing any (partial) sequences.
|
|
1502
|
+
*
|
|
1503
|
+
* Occasionally, we may want to change the actual number of bytes we consumed from inSeqs to
|
|
1504
|
+
* avoid splitting a match, or to avoid splitting a match such that it would produce a match
|
|
1505
|
+
* smaller than MINMATCH. In this case, we return the number of bytes that we didn't read from this block.
|
|
1506
|
+
*/
|
|
1507
|
+
size_t
|
|
1508
|
+
ZSTD_copySequencesToSeqStoreNoBlockDelim(ZSTD_CCtx* cctx, ZSTD_sequencePosition* seqPos,
|
|
1509
|
+
const ZSTD_Sequence* const inSeqs, size_t inSeqsSize,
|
|
1510
|
+
const void* src, size_t blockSize, ZSTD_paramSwitch_e externalRepSearch);
|
|
1511
|
+
|
|
1512
|
+
|
|
1513
|
+
/* ===============================================================
|
|
1514
|
+
* Deprecated definitions that are still used internally to avoid
|
|
1515
|
+
* deprecation warnings. These functions are exactly equivalent to
|
|
1516
|
+
* their public variants, but avoid the deprecation warnings.
|
|
1517
|
+
* =============================================================== */
|
|
1518
|
+
|
|
1519
|
+
size_t ZSTD_compressBegin_usingCDict_deprecated(ZSTD_CCtx* cctx, const ZSTD_CDict* cdict);
|
|
1520
|
+
|
|
1521
|
+
size_t ZSTD_compressContinue_public(ZSTD_CCtx* cctx,
|
|
1522
|
+
void* dst, size_t dstCapacity,
|
|
1523
|
+
const void* src, size_t srcSize);
|
|
1524
|
+
|
|
1525
|
+
size_t ZSTD_compressEnd_public(ZSTD_CCtx* cctx,
|
|
1526
|
+
void* dst, size_t dstCapacity,
|
|
1527
|
+
const void* src, size_t srcSize);
|
|
1528
|
+
|
|
1529
|
+
size_t ZSTD_compressBlock_deprecated(ZSTD_CCtx* cctx, void* dst, size_t dstCapacity, const void* src, size_t srcSize);
|
|
1530
|
+
|
|
1002
1531
|
|
|
1003
1532
|
#endif /* ZSTD_COMPRESS_H */
|