extzstd 0.3.2 → 0.3.3
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/README.md +3 -3
- data/contrib/zstd/CHANGELOG +188 -1
- data/contrib/zstd/CONTRIBUTING.md +157 -74
- data/contrib/zstd/LICENSE +4 -4
- data/contrib/zstd/Makefile +81 -58
- data/contrib/zstd/Package.swift +36 -0
- data/contrib/zstd/README.md +59 -35
- data/contrib/zstd/TESTING.md +2 -3
- data/contrib/zstd/appveyor.yml +49 -136
- data/contrib/zstd/lib/BUCK +5 -7
- data/contrib/zstd/lib/Makefile +87 -181
- data/contrib/zstd/lib/README.md +23 -6
- data/contrib/zstd/lib/common/allocations.h +55 -0
- data/contrib/zstd/lib/common/bits.h +200 -0
- data/contrib/zstd/lib/common/bitstream.h +33 -59
- data/contrib/zstd/lib/common/compiler.h +115 -45
- data/contrib/zstd/lib/common/cpu.h +1 -1
- data/contrib/zstd/lib/common/debug.c +1 -1
- data/contrib/zstd/lib/common/debug.h +1 -1
- data/contrib/zstd/lib/common/entropy_common.c +15 -37
- data/contrib/zstd/lib/common/error_private.c +9 -2
- data/contrib/zstd/lib/common/error_private.h +82 -3
- data/contrib/zstd/lib/common/fse.h +9 -85
- data/contrib/zstd/lib/common/fse_decompress.c +29 -111
- data/contrib/zstd/lib/common/huf.h +84 -172
- data/contrib/zstd/lib/common/mem.h +58 -49
- data/contrib/zstd/lib/common/pool.c +37 -16
- data/contrib/zstd/lib/common/pool.h +9 -3
- data/contrib/zstd/lib/common/portability_macros.h +156 -0
- data/contrib/zstd/lib/common/threading.c +68 -14
- data/contrib/zstd/lib/common/threading.h +5 -10
- data/contrib/zstd/lib/common/xxhash.c +7 -809
- data/contrib/zstd/lib/common/xxhash.h +5568 -167
- data/contrib/zstd/lib/common/zstd_common.c +1 -36
- data/contrib/zstd/lib/common/zstd_deps.h +1 -1
- data/contrib/zstd/lib/common/zstd_internal.h +64 -150
- data/contrib/zstd/lib/common/zstd_trace.h +163 -0
- data/contrib/zstd/lib/compress/clevels.h +134 -0
- data/contrib/zstd/lib/compress/fse_compress.c +69 -150
- data/contrib/zstd/lib/compress/hist.c +1 -1
- data/contrib/zstd/lib/compress/hist.h +1 -1
- data/contrib/zstd/lib/compress/huf_compress.c +773 -251
- data/contrib/zstd/lib/compress/zstd_compress.c +2650 -826
- data/contrib/zstd/lib/compress/zstd_compress_internal.h +509 -180
- data/contrib/zstd/lib/compress/zstd_compress_literals.c +117 -40
- data/contrib/zstd/lib/compress/zstd_compress_literals.h +16 -6
- data/contrib/zstd/lib/compress/zstd_compress_sequences.c +28 -19
- data/contrib/zstd/lib/compress/zstd_compress_sequences.h +1 -1
- data/contrib/zstd/lib/compress/zstd_compress_superblock.c +33 -305
- data/contrib/zstd/lib/compress/zstd_compress_superblock.h +1 -1
- data/contrib/zstd/lib/compress/zstd_cwksp.h +266 -85
- data/contrib/zstd/lib/compress/zstd_double_fast.c +369 -132
- data/contrib/zstd/lib/compress/zstd_double_fast.h +3 -2
- data/contrib/zstd/lib/compress/zstd_fast.c +722 -258
- data/contrib/zstd/lib/compress/zstd_fast.h +3 -2
- data/contrib/zstd/lib/compress/zstd_lazy.c +1105 -360
- data/contrib/zstd/lib/compress/zstd_lazy.h +41 -1
- data/contrib/zstd/lib/compress/zstd_ldm.c +272 -208
- data/contrib/zstd/lib/compress/zstd_ldm.h +3 -2
- data/contrib/zstd/lib/compress/zstd_ldm_geartab.h +106 -0
- data/contrib/zstd/lib/compress/zstd_opt.c +324 -197
- data/contrib/zstd/lib/compress/zstd_opt.h +1 -1
- data/contrib/zstd/lib/compress/zstdmt_compress.c +109 -53
- data/contrib/zstd/lib/compress/zstdmt_compress.h +9 -6
- data/contrib/zstd/lib/decompress/huf_decompress.c +1071 -539
- data/contrib/zstd/lib/decompress/huf_decompress_amd64.S +576 -0
- data/contrib/zstd/lib/decompress/zstd_ddict.c +4 -4
- data/contrib/zstd/lib/decompress/zstd_ddict.h +1 -1
- data/contrib/zstd/lib/decompress/zstd_decompress.c +507 -82
- data/contrib/zstd/lib/decompress/zstd_decompress_block.c +962 -310
- data/contrib/zstd/lib/decompress/zstd_decompress_block.h +14 -3
- data/contrib/zstd/lib/decompress/zstd_decompress_internal.h +54 -6
- data/contrib/zstd/lib/deprecated/zbuff.h +1 -1
- data/contrib/zstd/lib/deprecated/zbuff_common.c +1 -1
- data/contrib/zstd/lib/deprecated/zbuff_compress.c +24 -4
- data/contrib/zstd/lib/deprecated/zbuff_decompress.c +3 -1
- data/contrib/zstd/lib/dictBuilder/cover.c +44 -32
- data/contrib/zstd/lib/dictBuilder/cover.h +6 -5
- data/contrib/zstd/lib/dictBuilder/divsufsort.c +1 -1
- data/contrib/zstd/lib/dictBuilder/fastcover.c +24 -16
- data/contrib/zstd/lib/dictBuilder/zdict.c +88 -95
- data/contrib/zstd/lib/legacy/zstd_legacy.h +8 -1
- data/contrib/zstd/lib/legacy/zstd_v01.c +16 -53
- data/contrib/zstd/lib/legacy/zstd_v01.h +1 -1
- data/contrib/zstd/lib/legacy/zstd_v02.c +24 -69
- data/contrib/zstd/lib/legacy/zstd_v02.h +1 -1
- data/contrib/zstd/lib/legacy/zstd_v03.c +25 -72
- data/contrib/zstd/lib/legacy/zstd_v03.h +1 -1
- data/contrib/zstd/lib/legacy/zstd_v04.c +23 -69
- data/contrib/zstd/lib/legacy/zstd_v04.h +1 -1
- data/contrib/zstd/lib/legacy/zstd_v05.c +35 -85
- data/contrib/zstd/lib/legacy/zstd_v05.h +1 -1
- data/contrib/zstd/lib/legacy/zstd_v06.c +42 -87
- data/contrib/zstd/lib/legacy/zstd_v06.h +1 -1
- data/contrib/zstd/lib/legacy/zstd_v07.c +35 -82
- data/contrib/zstd/lib/legacy/zstd_v07.h +1 -1
- data/contrib/zstd/lib/libzstd.mk +214 -0
- data/contrib/zstd/lib/libzstd.pc.in +4 -3
- data/contrib/zstd/lib/module.modulemap +35 -0
- data/contrib/zstd/lib/{dictBuilder/zdict.h → zdict.h} +202 -33
- data/contrib/zstd/lib/zstd.h +922 -293
- data/contrib/zstd/lib/{common/zstd_errors.h → zstd_errors.h} +27 -8
- data/ext/extconf.rb +7 -6
- data/ext/extzstd.c +13 -10
- data/ext/libzstd_conf.h +0 -1
- data/ext/zstd_decompress_asm.S +1 -0
- metadata +16 -5
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
/*
|
|
2
|
-
* Copyright (c)
|
|
2
|
+
* Copyright (c) Meta Platforms, Inc. and affiliates.
|
|
3
3
|
* All rights reserved.
|
|
4
4
|
*
|
|
5
5
|
* This source code is licensed under both the BSD-style license (found in the
|
|
@@ -11,13 +11,126 @@
|
|
|
11
11
|
#include "zstd_ldm.h"
|
|
12
12
|
|
|
13
13
|
#include "../common/debug.h"
|
|
14
|
+
#include "../common/xxhash.h"
|
|
14
15
|
#include "zstd_fast.h" /* ZSTD_fillHashTable() */
|
|
15
16
|
#include "zstd_double_fast.h" /* ZSTD_fillDoubleHashTable() */
|
|
17
|
+
#include "zstd_ldm_geartab.h"
|
|
16
18
|
|
|
17
19
|
#define LDM_BUCKET_SIZE_LOG 3
|
|
18
20
|
#define LDM_MIN_MATCH_LENGTH 64
|
|
19
21
|
#define LDM_HASH_RLOG 7
|
|
20
|
-
|
|
22
|
+
|
|
23
|
+
typedef struct {
|
|
24
|
+
U64 rolling;
|
|
25
|
+
U64 stopMask;
|
|
26
|
+
} ldmRollingHashState_t;
|
|
27
|
+
|
|
28
|
+
/** ZSTD_ldm_gear_init():
|
|
29
|
+
*
|
|
30
|
+
* Initializes the rolling hash state such that it will honor the
|
|
31
|
+
* settings in params. */
|
|
32
|
+
static void ZSTD_ldm_gear_init(ldmRollingHashState_t* state, ldmParams_t const* params)
|
|
33
|
+
{
|
|
34
|
+
unsigned maxBitsInMask = MIN(params->minMatchLength, 64);
|
|
35
|
+
unsigned hashRateLog = params->hashRateLog;
|
|
36
|
+
|
|
37
|
+
state->rolling = ~(U32)0;
|
|
38
|
+
|
|
39
|
+
/* The choice of the splitting criterion is subject to two conditions:
|
|
40
|
+
* 1. it has to trigger on average every 2^(hashRateLog) bytes;
|
|
41
|
+
* 2. ideally, it has to depend on a window of minMatchLength bytes.
|
|
42
|
+
*
|
|
43
|
+
* In the gear hash algorithm, bit n depends on the last n bytes;
|
|
44
|
+
* so in order to obtain a good quality splitting criterion it is
|
|
45
|
+
* preferable to use bits with high weight.
|
|
46
|
+
*
|
|
47
|
+
* To match condition 1 we use a mask with hashRateLog bits set
|
|
48
|
+
* and, because of the previous remark, we make sure these bits
|
|
49
|
+
* have the highest possible weight while still respecting
|
|
50
|
+
* condition 2.
|
|
51
|
+
*/
|
|
52
|
+
if (hashRateLog > 0 && hashRateLog <= maxBitsInMask) {
|
|
53
|
+
state->stopMask = (((U64)1 << hashRateLog) - 1) << (maxBitsInMask - hashRateLog);
|
|
54
|
+
} else {
|
|
55
|
+
/* In this degenerate case we simply honor the hash rate. */
|
|
56
|
+
state->stopMask = ((U64)1 << hashRateLog) - 1;
|
|
57
|
+
}
|
|
58
|
+
}
|
|
59
|
+
|
|
60
|
+
/** ZSTD_ldm_gear_reset()
|
|
61
|
+
* Feeds [data, data + minMatchLength) into the hash without registering any
|
|
62
|
+
* splits. This effectively resets the hash state. This is used when skipping
|
|
63
|
+
* over data, either at the beginning of a block, or skipping sections.
|
|
64
|
+
*/
|
|
65
|
+
static void ZSTD_ldm_gear_reset(ldmRollingHashState_t* state,
|
|
66
|
+
BYTE const* data, size_t minMatchLength)
|
|
67
|
+
{
|
|
68
|
+
U64 hash = state->rolling;
|
|
69
|
+
size_t n = 0;
|
|
70
|
+
|
|
71
|
+
#define GEAR_ITER_ONCE() do { \
|
|
72
|
+
hash = (hash << 1) + ZSTD_ldm_gearTab[data[n] & 0xff]; \
|
|
73
|
+
n += 1; \
|
|
74
|
+
} while (0)
|
|
75
|
+
while (n + 3 < minMatchLength) {
|
|
76
|
+
GEAR_ITER_ONCE();
|
|
77
|
+
GEAR_ITER_ONCE();
|
|
78
|
+
GEAR_ITER_ONCE();
|
|
79
|
+
GEAR_ITER_ONCE();
|
|
80
|
+
}
|
|
81
|
+
while (n < minMatchLength) {
|
|
82
|
+
GEAR_ITER_ONCE();
|
|
83
|
+
}
|
|
84
|
+
#undef GEAR_ITER_ONCE
|
|
85
|
+
}
|
|
86
|
+
|
|
87
|
+
/** ZSTD_ldm_gear_feed():
|
|
88
|
+
*
|
|
89
|
+
* Registers in the splits array all the split points found in the first
|
|
90
|
+
* size bytes following the data pointer. This function terminates when
|
|
91
|
+
* either all the data has been processed or LDM_BATCH_SIZE splits are
|
|
92
|
+
* present in the splits array.
|
|
93
|
+
*
|
|
94
|
+
* Precondition: The splits array must not be full.
|
|
95
|
+
* Returns: The number of bytes processed. */
|
|
96
|
+
static size_t ZSTD_ldm_gear_feed(ldmRollingHashState_t* state,
|
|
97
|
+
BYTE const* data, size_t size,
|
|
98
|
+
size_t* splits, unsigned* numSplits)
|
|
99
|
+
{
|
|
100
|
+
size_t n;
|
|
101
|
+
U64 hash, mask;
|
|
102
|
+
|
|
103
|
+
hash = state->rolling;
|
|
104
|
+
mask = state->stopMask;
|
|
105
|
+
n = 0;
|
|
106
|
+
|
|
107
|
+
#define GEAR_ITER_ONCE() do { \
|
|
108
|
+
hash = (hash << 1) + ZSTD_ldm_gearTab[data[n] & 0xff]; \
|
|
109
|
+
n += 1; \
|
|
110
|
+
if (UNLIKELY((hash & mask) == 0)) { \
|
|
111
|
+
splits[*numSplits] = n; \
|
|
112
|
+
*numSplits += 1; \
|
|
113
|
+
if (*numSplits == LDM_BATCH_SIZE) \
|
|
114
|
+
goto done; \
|
|
115
|
+
} \
|
|
116
|
+
} while (0)
|
|
117
|
+
|
|
118
|
+
while (n + 3 < size) {
|
|
119
|
+
GEAR_ITER_ONCE();
|
|
120
|
+
GEAR_ITER_ONCE();
|
|
121
|
+
GEAR_ITER_ONCE();
|
|
122
|
+
GEAR_ITER_ONCE();
|
|
123
|
+
}
|
|
124
|
+
while (n < size) {
|
|
125
|
+
GEAR_ITER_ONCE();
|
|
126
|
+
}
|
|
127
|
+
|
|
128
|
+
#undef GEAR_ITER_ONCE
|
|
129
|
+
|
|
130
|
+
done:
|
|
131
|
+
state->rolling = hash;
|
|
132
|
+
return n;
|
|
133
|
+
}
|
|
21
134
|
|
|
22
135
|
void ZSTD_ldm_adjustParameters(ldmParams_t* params,
|
|
23
136
|
ZSTD_compressionParameters const* cParams)
|
|
@@ -46,47 +159,12 @@ size_t ZSTD_ldm_getTableSize(ldmParams_t params)
|
|
|
46
159
|
size_t const ldmBucketSize = ((size_t)1) << (params.hashLog - ldmBucketSizeLog);
|
|
47
160
|
size_t const totalSize = ZSTD_cwksp_alloc_size(ldmBucketSize)
|
|
48
161
|
+ ZSTD_cwksp_alloc_size(ldmHSize * sizeof(ldmEntry_t));
|
|
49
|
-
return params.enableLdm ? totalSize : 0;
|
|
162
|
+
return params.enableLdm == ZSTD_ps_enable ? totalSize : 0;
|
|
50
163
|
}
|
|
51
164
|
|
|
52
165
|
size_t ZSTD_ldm_getMaxNbSeq(ldmParams_t params, size_t maxChunkSize)
|
|
53
166
|
{
|
|
54
|
-
return params.enableLdm ? (maxChunkSize / params.minMatchLength) : 0;
|
|
55
|
-
}
|
|
56
|
-
|
|
57
|
-
/** ZSTD_ldm_getSmallHash() :
|
|
58
|
-
* numBits should be <= 32
|
|
59
|
-
* If numBits==0, returns 0.
|
|
60
|
-
* @return : the most significant numBits of value. */
|
|
61
|
-
static U32 ZSTD_ldm_getSmallHash(U64 value, U32 numBits)
|
|
62
|
-
{
|
|
63
|
-
assert(numBits <= 32);
|
|
64
|
-
return numBits == 0 ? 0 : (U32)(value >> (64 - numBits));
|
|
65
|
-
}
|
|
66
|
-
|
|
67
|
-
/** ZSTD_ldm_getChecksum() :
|
|
68
|
-
* numBitsToDiscard should be <= 32
|
|
69
|
-
* @return : the next most significant 32 bits after numBitsToDiscard */
|
|
70
|
-
static U32 ZSTD_ldm_getChecksum(U64 hash, U32 numBitsToDiscard)
|
|
71
|
-
{
|
|
72
|
-
assert(numBitsToDiscard <= 32);
|
|
73
|
-
return (hash >> (64 - 32 - numBitsToDiscard)) & 0xFFFFFFFF;
|
|
74
|
-
}
|
|
75
|
-
|
|
76
|
-
/** ZSTD_ldm_getTag() ;
|
|
77
|
-
* Given the hash, returns the most significant numTagBits bits
|
|
78
|
-
* after (32 + hbits) bits.
|
|
79
|
-
*
|
|
80
|
-
* If there are not enough bits remaining, return the last
|
|
81
|
-
* numTagBits bits. */
|
|
82
|
-
static U32 ZSTD_ldm_getTag(U64 hash, U32 hbits, U32 numTagBits)
|
|
83
|
-
{
|
|
84
|
-
assert(numTagBits < 32 && hbits <= 32);
|
|
85
|
-
if (32 - hbits < numTagBits) {
|
|
86
|
-
return hash & (((U32)1 << numTagBits) - 1);
|
|
87
|
-
} else {
|
|
88
|
-
return (hash >> (32 - hbits - numTagBits)) & (((U32)1 << numTagBits) - 1);
|
|
89
|
-
}
|
|
167
|
+
return params.enableLdm == ZSTD_ps_enable ? (maxChunkSize / params.minMatchLength) : 0;
|
|
90
168
|
}
|
|
91
169
|
|
|
92
170
|
/** ZSTD_ldm_getBucket() :
|
|
@@ -103,38 +181,12 @@ static void ZSTD_ldm_insertEntry(ldmState_t* ldmState,
|
|
|
103
181
|
size_t const hash, const ldmEntry_t entry,
|
|
104
182
|
ldmParams_t const ldmParams)
|
|
105
183
|
{
|
|
106
|
-
BYTE* const
|
|
107
|
-
|
|
108
|
-
|
|
109
|
-
|
|
110
|
-
|
|
184
|
+
BYTE* const pOffset = ldmState->bucketOffsets + hash;
|
|
185
|
+
unsigned const offset = *pOffset;
|
|
186
|
+
|
|
187
|
+
*(ZSTD_ldm_getBucket(ldmState, hash, ldmParams) + offset) = entry;
|
|
188
|
+
*pOffset = (BYTE)((offset + 1) & ((1u << ldmParams.bucketSizeLog) - 1));
|
|
111
189
|
|
|
112
|
-
/** ZSTD_ldm_makeEntryAndInsertByTag() :
|
|
113
|
-
*
|
|
114
|
-
* Gets the small hash, checksum, and tag from the rollingHash.
|
|
115
|
-
*
|
|
116
|
-
* If the tag matches (1 << ldmParams.hashRateLog)-1, then
|
|
117
|
-
* creates an ldmEntry from the offset, and inserts it into the hash table.
|
|
118
|
-
*
|
|
119
|
-
* hBits is the length of the small hash, which is the most significant hBits
|
|
120
|
-
* of rollingHash. The checksum is the next 32 most significant bits, followed
|
|
121
|
-
* by ldmParams.hashRateLog bits that make up the tag. */
|
|
122
|
-
static void ZSTD_ldm_makeEntryAndInsertByTag(ldmState_t* ldmState,
|
|
123
|
-
U64 const rollingHash,
|
|
124
|
-
U32 const hBits,
|
|
125
|
-
U32 const offset,
|
|
126
|
-
ldmParams_t const ldmParams)
|
|
127
|
-
{
|
|
128
|
-
U32 const tag = ZSTD_ldm_getTag(rollingHash, hBits, ldmParams.hashRateLog);
|
|
129
|
-
U32 const tagMask = ((U32)1 << ldmParams.hashRateLog) - 1;
|
|
130
|
-
if (tag == tagMask) {
|
|
131
|
-
U32 const hash = ZSTD_ldm_getSmallHash(rollingHash, hBits);
|
|
132
|
-
U32 const checksum = ZSTD_ldm_getChecksum(rollingHash, hBits);
|
|
133
|
-
ldmEntry_t entry;
|
|
134
|
-
entry.offset = offset;
|
|
135
|
-
entry.checksum = checksum;
|
|
136
|
-
ZSTD_ldm_insertEntry(ldmState, hash, entry, ldmParams);
|
|
137
|
-
}
|
|
138
190
|
}
|
|
139
191
|
|
|
140
192
|
/** ZSTD_ldm_countBackwardsMatch() :
|
|
@@ -190,11 +242,11 @@ static size_t ZSTD_ldm_fillFastTables(ZSTD_matchState_t* ms,
|
|
|
190
242
|
switch(ms->cParams.strategy)
|
|
191
243
|
{
|
|
192
244
|
case ZSTD_fast:
|
|
193
|
-
ZSTD_fillHashTable(ms, iend, ZSTD_dtlm_fast);
|
|
245
|
+
ZSTD_fillHashTable(ms, iend, ZSTD_dtlm_fast, ZSTD_tfp_forCCtx);
|
|
194
246
|
break;
|
|
195
247
|
|
|
196
248
|
case ZSTD_dfast:
|
|
197
|
-
ZSTD_fillDoubleHashTable(ms, iend, ZSTD_dtlm_fast);
|
|
249
|
+
ZSTD_fillDoubleHashTable(ms, iend, ZSTD_dtlm_fast, ZSTD_tfp_forCCtx);
|
|
198
250
|
break;
|
|
199
251
|
|
|
200
252
|
case ZSTD_greedy:
|
|
@@ -212,43 +264,42 @@ static size_t ZSTD_ldm_fillFastTables(ZSTD_matchState_t* ms,
|
|
|
212
264
|
return 0;
|
|
213
265
|
}
|
|
214
266
|
|
|
215
|
-
/** ZSTD_ldm_fillLdmHashTable() :
|
|
216
|
-
*
|
|
217
|
-
* Fills hashTable from (lastHashed + 1) to iend (non-inclusive).
|
|
218
|
-
* lastHash is the rolling hash that corresponds to lastHashed.
|
|
219
|
-
*
|
|
220
|
-
* Returns the rolling hash corresponding to position iend-1. */
|
|
221
|
-
static U64 ZSTD_ldm_fillLdmHashTable(ldmState_t* state,
|
|
222
|
-
U64 lastHash, const BYTE* lastHashed,
|
|
223
|
-
const BYTE* iend, const BYTE* base,
|
|
224
|
-
U32 hBits, ldmParams_t const ldmParams)
|
|
225
|
-
{
|
|
226
|
-
U64 rollingHash = lastHash;
|
|
227
|
-
const BYTE* cur = lastHashed + 1;
|
|
228
|
-
|
|
229
|
-
while (cur < iend) {
|
|
230
|
-
rollingHash = ZSTD_rollingHash_rotate(rollingHash, cur[-1],
|
|
231
|
-
cur[ldmParams.minMatchLength-1],
|
|
232
|
-
state->hashPower);
|
|
233
|
-
ZSTD_ldm_makeEntryAndInsertByTag(state,
|
|
234
|
-
rollingHash, hBits,
|
|
235
|
-
(U32)(cur - base), ldmParams);
|
|
236
|
-
++cur;
|
|
237
|
-
}
|
|
238
|
-
return rollingHash;
|
|
239
|
-
}
|
|
240
|
-
|
|
241
267
|
void ZSTD_ldm_fillHashTable(
|
|
242
|
-
ldmState_t*
|
|
268
|
+
ldmState_t* ldmState, const BYTE* ip,
|
|
243
269
|
const BYTE* iend, ldmParams_t const* params)
|
|
244
270
|
{
|
|
271
|
+
U32 const minMatchLength = params->minMatchLength;
|
|
272
|
+
U32 const hBits = params->hashLog - params->bucketSizeLog;
|
|
273
|
+
BYTE const* const base = ldmState->window.base;
|
|
274
|
+
BYTE const* const istart = ip;
|
|
275
|
+
ldmRollingHashState_t hashState;
|
|
276
|
+
size_t* const splits = ldmState->splitIndices;
|
|
277
|
+
unsigned numSplits;
|
|
278
|
+
|
|
245
279
|
DEBUGLOG(5, "ZSTD_ldm_fillHashTable");
|
|
246
|
-
|
|
247
|
-
|
|
248
|
-
|
|
249
|
-
|
|
250
|
-
|
|
251
|
-
|
|
280
|
+
|
|
281
|
+
ZSTD_ldm_gear_init(&hashState, params);
|
|
282
|
+
while (ip < iend) {
|
|
283
|
+
size_t hashed;
|
|
284
|
+
unsigned n;
|
|
285
|
+
|
|
286
|
+
numSplits = 0;
|
|
287
|
+
hashed = ZSTD_ldm_gear_feed(&hashState, ip, iend - ip, splits, &numSplits);
|
|
288
|
+
|
|
289
|
+
for (n = 0; n < numSplits; n++) {
|
|
290
|
+
if (ip + splits[n] >= istart + minMatchLength) {
|
|
291
|
+
BYTE const* const split = ip + splits[n] - minMatchLength;
|
|
292
|
+
U64 const xxhash = XXH64(split, minMatchLength, 0);
|
|
293
|
+
U32 const hash = (U32)(xxhash & (((U32)1 << hBits) - 1));
|
|
294
|
+
ldmEntry_t entry;
|
|
295
|
+
|
|
296
|
+
entry.offset = (U32)(split - base);
|
|
297
|
+
entry.checksum = (U32)(xxhash >> 32);
|
|
298
|
+
ZSTD_ldm_insertEntry(ldmState, hash, entry, *params);
|
|
299
|
+
}
|
|
300
|
+
}
|
|
301
|
+
|
|
302
|
+
ip += hashed;
|
|
252
303
|
}
|
|
253
304
|
}
|
|
254
305
|
|
|
@@ -274,11 +325,8 @@ static size_t ZSTD_ldm_generateSequences_internal(
|
|
|
274
325
|
/* LDM parameters */
|
|
275
326
|
int const extDict = ZSTD_window_hasExtDict(ldmState->window);
|
|
276
327
|
U32 const minMatchLength = params->minMatchLength;
|
|
277
|
-
|
|
328
|
+
U32 const entsPerBucket = 1U << params->bucketSizeLog;
|
|
278
329
|
U32 const hBits = params->hashLog - params->bucketSizeLog;
|
|
279
|
-
U32 const ldmBucketSize = 1U << params->bucketSizeLog;
|
|
280
|
-
U32 const hashRateLog = params->hashRateLog;
|
|
281
|
-
U32 const ldmTagMask = (1U << params->hashRateLog) - 1;
|
|
282
330
|
/* Prefix and extDict parameters */
|
|
283
331
|
U32 const dictLimit = ldmState->window.dictLimit;
|
|
284
332
|
U32 const lowestIndex = extDict ? ldmState->window.lowLimit : dictLimit;
|
|
@@ -290,45 +338,69 @@ static size_t ZSTD_ldm_generateSequences_internal(
|
|
|
290
338
|
/* Input bounds */
|
|
291
339
|
BYTE const* const istart = (BYTE const*)src;
|
|
292
340
|
BYTE const* const iend = istart + srcSize;
|
|
293
|
-
BYTE const* const ilimit = iend -
|
|
341
|
+
BYTE const* const ilimit = iend - HASH_READ_SIZE;
|
|
294
342
|
/* Input positions */
|
|
295
343
|
BYTE const* anchor = istart;
|
|
296
344
|
BYTE const* ip = istart;
|
|
297
|
-
/* Rolling hash */
|
|
298
|
-
|
|
299
|
-
|
|
300
|
-
|
|
301
|
-
|
|
302
|
-
|
|
303
|
-
|
|
304
|
-
|
|
305
|
-
|
|
306
|
-
|
|
307
|
-
|
|
308
|
-
|
|
309
|
-
|
|
310
|
-
|
|
311
|
-
|
|
345
|
+
/* Rolling hash state */
|
|
346
|
+
ldmRollingHashState_t hashState;
|
|
347
|
+
/* Arrays for staged-processing */
|
|
348
|
+
size_t* const splits = ldmState->splitIndices;
|
|
349
|
+
ldmMatchCandidate_t* const candidates = ldmState->matchCandidates;
|
|
350
|
+
unsigned numSplits;
|
|
351
|
+
|
|
352
|
+
if (srcSize < minMatchLength)
|
|
353
|
+
return iend - anchor;
|
|
354
|
+
|
|
355
|
+
/* Initialize the rolling hash state with the first minMatchLength bytes */
|
|
356
|
+
ZSTD_ldm_gear_init(&hashState, params);
|
|
357
|
+
ZSTD_ldm_gear_reset(&hashState, ip, minMatchLength);
|
|
358
|
+
ip += minMatchLength;
|
|
359
|
+
|
|
360
|
+
while (ip < ilimit) {
|
|
361
|
+
size_t hashed;
|
|
362
|
+
unsigned n;
|
|
363
|
+
|
|
364
|
+
numSplits = 0;
|
|
365
|
+
hashed = ZSTD_ldm_gear_feed(&hashState, ip, ilimit - ip,
|
|
366
|
+
splits, &numSplits);
|
|
367
|
+
|
|
368
|
+
for (n = 0; n < numSplits; n++) {
|
|
369
|
+
BYTE const* const split = ip + splits[n] - minMatchLength;
|
|
370
|
+
U64 const xxhash = XXH64(split, minMatchLength, 0);
|
|
371
|
+
U32 const hash = (U32)(xxhash & (((U32)1 << hBits) - 1));
|
|
372
|
+
|
|
373
|
+
candidates[n].split = split;
|
|
374
|
+
candidates[n].hash = hash;
|
|
375
|
+
candidates[n].checksum = (U32)(xxhash >> 32);
|
|
376
|
+
candidates[n].bucket = ZSTD_ldm_getBucket(ldmState, hash, *params);
|
|
377
|
+
PREFETCH_L1(candidates[n].bucket);
|
|
312
378
|
}
|
|
313
|
-
lastHashed = ip;
|
|
314
379
|
|
|
315
|
-
|
|
316
|
-
|
|
317
|
-
|
|
318
|
-
|
|
319
|
-
|
|
380
|
+
for (n = 0; n < numSplits; n++) {
|
|
381
|
+
size_t forwardMatchLength = 0, backwardMatchLength = 0,
|
|
382
|
+
bestMatchLength = 0, mLength;
|
|
383
|
+
U32 offset;
|
|
384
|
+
BYTE const* const split = candidates[n].split;
|
|
385
|
+
U32 const checksum = candidates[n].checksum;
|
|
386
|
+
U32 const hash = candidates[n].hash;
|
|
387
|
+
ldmEntry_t* const bucket = candidates[n].bucket;
|
|
388
|
+
ldmEntry_t const* cur;
|
|
389
|
+
ldmEntry_t const* bestEntry = NULL;
|
|
390
|
+
ldmEntry_t newEntry;
|
|
391
|
+
|
|
392
|
+
newEntry.offset = (U32)(split - base);
|
|
393
|
+
newEntry.checksum = checksum;
|
|
394
|
+
|
|
395
|
+
/* If a split point would generate a sequence overlapping with
|
|
396
|
+
* the previous one, we merely register it in the hash table and
|
|
397
|
+
* move on */
|
|
398
|
+
if (split < anchor) {
|
|
399
|
+
ZSTD_ldm_insertEntry(ldmState, hash, newEntry, *params);
|
|
400
|
+
continue;
|
|
401
|
+
}
|
|
320
402
|
|
|
321
|
-
|
|
322
|
-
{
|
|
323
|
-
ldmEntry_t* const bucket =
|
|
324
|
-
ZSTD_ldm_getBucket(ldmState,
|
|
325
|
-
ZSTD_ldm_getSmallHash(rollingHash, hBits),
|
|
326
|
-
*params);
|
|
327
|
-
ldmEntry_t* cur;
|
|
328
|
-
size_t bestMatchLength = 0;
|
|
329
|
-
U32 const checksum = ZSTD_ldm_getChecksum(rollingHash, hBits);
|
|
330
|
-
|
|
331
|
-
for (cur = bucket; cur < bucket + ldmBucketSize; ++cur) {
|
|
403
|
+
for (cur = bucket; cur < bucket + entsPerBucket; cur++) {
|
|
332
404
|
size_t curForwardMatchLength, curBackwardMatchLength,
|
|
333
405
|
curTotalMatchLength;
|
|
334
406
|
if (cur->checksum != checksum || cur->offset <= lowestIndex) {
|
|
@@ -342,31 +414,23 @@ static size_t ZSTD_ldm_generateSequences_internal(
|
|
|
342
414
|
cur->offset < dictLimit ? dictEnd : iend;
|
|
343
415
|
BYTE const* const lowMatchPtr =
|
|
344
416
|
cur->offset < dictLimit ? dictStart : lowPrefixPtr;
|
|
345
|
-
|
|
346
|
-
|
|
347
|
-
ip, pMatch, iend,
|
|
348
|
-
matchEnd, lowPrefixPtr);
|
|
417
|
+
curForwardMatchLength =
|
|
418
|
+
ZSTD_count_2segments(split, pMatch, iend, matchEnd, lowPrefixPtr);
|
|
349
419
|
if (curForwardMatchLength < minMatchLength) {
|
|
350
420
|
continue;
|
|
351
421
|
}
|
|
352
|
-
curBackwardMatchLength =
|
|
353
|
-
|
|
354
|
-
pMatch, lowMatchPtr,
|
|
355
|
-
dictStart, dictEnd);
|
|
356
|
-
curTotalMatchLength = curForwardMatchLength +
|
|
357
|
-
curBackwardMatchLength;
|
|
422
|
+
curBackwardMatchLength = ZSTD_ldm_countBackwardsMatch_2segments(
|
|
423
|
+
split, anchor, pMatch, lowMatchPtr, dictStart, dictEnd);
|
|
358
424
|
} else { /* !extDict */
|
|
359
425
|
BYTE const* const pMatch = base + cur->offset;
|
|
360
|
-
curForwardMatchLength = ZSTD_count(
|
|
426
|
+
curForwardMatchLength = ZSTD_count(split, pMatch, iend);
|
|
361
427
|
if (curForwardMatchLength < minMatchLength) {
|
|
362
428
|
continue;
|
|
363
429
|
}
|
|
364
430
|
curBackwardMatchLength =
|
|
365
|
-
ZSTD_ldm_countBackwardsMatch(
|
|
366
|
-
lowPrefixPtr);
|
|
367
|
-
curTotalMatchLength = curForwardMatchLength +
|
|
368
|
-
curBackwardMatchLength;
|
|
431
|
+
ZSTD_ldm_countBackwardsMatch(split, anchor, pMatch, lowPrefixPtr);
|
|
369
432
|
}
|
|
433
|
+
curTotalMatchLength = curForwardMatchLength + curBackwardMatchLength;
|
|
370
434
|
|
|
371
435
|
if (curTotalMatchLength > bestMatchLength) {
|
|
372
436
|
bestMatchLength = curTotalMatchLength;
|
|
@@ -375,57 +439,54 @@ static size_t ZSTD_ldm_generateSequences_internal(
|
|
|
375
439
|
bestEntry = cur;
|
|
376
440
|
}
|
|
377
441
|
}
|
|
378
|
-
}
|
|
379
442
|
|
|
380
|
-
|
|
381
|
-
|
|
382
|
-
|
|
383
|
-
|
|
384
|
-
|
|
385
|
-
|
|
386
|
-
continue;
|
|
387
|
-
}
|
|
388
|
-
|
|
389
|
-
/* Match found */
|
|
390
|
-
mLength = forwardMatchLength + backwardMatchLength;
|
|
391
|
-
ip -= backwardMatchLength;
|
|
443
|
+
/* No match found -- insert an entry into the hash table
|
|
444
|
+
* and process the next candidate match */
|
|
445
|
+
if (bestEntry == NULL) {
|
|
446
|
+
ZSTD_ldm_insertEntry(ldmState, hash, newEntry, *params);
|
|
447
|
+
continue;
|
|
448
|
+
}
|
|
392
449
|
|
|
393
|
-
|
|
394
|
-
|
|
395
|
-
|
|
396
|
-
|
|
397
|
-
|
|
398
|
-
|
|
399
|
-
|
|
400
|
-
|
|
401
|
-
|
|
402
|
-
|
|
403
|
-
|
|
404
|
-
|
|
405
|
-
|
|
406
|
-
|
|
407
|
-
seq->offset = offset;
|
|
408
|
-
rawSeqStore->size++;
|
|
409
|
-
}
|
|
450
|
+
/* Match found */
|
|
451
|
+
offset = (U32)(split - base) - bestEntry->offset;
|
|
452
|
+
mLength = forwardMatchLength + backwardMatchLength;
|
|
453
|
+
{
|
|
454
|
+
rawSeq* const seq = rawSeqStore->seq + rawSeqStore->size;
|
|
455
|
+
|
|
456
|
+
/* Out of sequence storage */
|
|
457
|
+
if (rawSeqStore->size == rawSeqStore->capacity)
|
|
458
|
+
return ERROR(dstSize_tooSmall);
|
|
459
|
+
seq->litLength = (U32)(split - backwardMatchLength - anchor);
|
|
460
|
+
seq->matchLength = (U32)mLength;
|
|
461
|
+
seq->offset = offset;
|
|
462
|
+
rawSeqStore->size++;
|
|
463
|
+
}
|
|
410
464
|
|
|
411
|
-
|
|
412
|
-
|
|
413
|
-
|
|
414
|
-
*params);
|
|
465
|
+
/* Insert the current entry into the hash table --- it must be
|
|
466
|
+
* done after the previous block to avoid clobbering bestEntry */
|
|
467
|
+
ZSTD_ldm_insertEntry(ldmState, hash, newEntry, *params);
|
|
415
468
|
|
|
416
|
-
|
|
469
|
+
anchor = split + forwardMatchLength;
|
|
417
470
|
|
|
418
|
-
|
|
419
|
-
|
|
420
|
-
|
|
421
|
-
|
|
422
|
-
|
|
423
|
-
|
|
424
|
-
|
|
471
|
+
/* If we find a match that ends after the data that we've hashed
|
|
472
|
+
* then we have a repeating, overlapping, pattern. E.g. all zeros.
|
|
473
|
+
* If one repetition of the pattern matches our `stopMask` then all
|
|
474
|
+
* repetitions will. We don't need to insert them all into out table,
|
|
475
|
+
* only the first one. So skip over overlapping matches.
|
|
476
|
+
* This is a major speed boost (20x) for compressing a single byte
|
|
477
|
+
* repeated, when that byte ends up in the table.
|
|
478
|
+
*/
|
|
479
|
+
if (anchor > ip + hashed) {
|
|
480
|
+
ZSTD_ldm_gear_reset(&hashState, anchor - minMatchLength, minMatchLength);
|
|
481
|
+
/* Continue the outer loop at anchor (ip + hashed == anchor). */
|
|
482
|
+
ip = anchor - hashed;
|
|
483
|
+
break;
|
|
484
|
+
}
|
|
425
485
|
}
|
|
426
|
-
|
|
427
|
-
|
|
486
|
+
|
|
487
|
+
ip += hashed;
|
|
428
488
|
}
|
|
489
|
+
|
|
429
490
|
return iend - anchor;
|
|
430
491
|
}
|
|
431
492
|
|
|
@@ -474,7 +535,7 @@ size_t ZSTD_ldm_generateSequences(
|
|
|
474
535
|
|
|
475
536
|
assert(chunkStart < iend);
|
|
476
537
|
/* 1. Perform overflow correction if necessary. */
|
|
477
|
-
if (ZSTD_window_needOverflowCorrection(ldmState->window, chunkEnd)) {
|
|
538
|
+
if (ZSTD_window_needOverflowCorrection(ldmState->window, 0, maxDist, ldmState->loadedDictEnd, chunkStart, chunkEnd)) {
|
|
478
539
|
U32 const ldmHSize = 1U << params->hashLog;
|
|
479
540
|
U32 const correction = ZSTD_window_correctOverflow(
|
|
480
541
|
&ldmState->window, /* cycleLog */ 0, maxDist, chunkStart);
|
|
@@ -488,7 +549,7 @@ size_t ZSTD_ldm_generateSequences(
|
|
|
488
549
|
* the window through early invalidation.
|
|
489
550
|
* TODO: * Test the chunk size.
|
|
490
551
|
* * Try invalidation after the sequence generation and test the
|
|
491
|
-
*
|
|
552
|
+
* offset against maxDist directly.
|
|
492
553
|
*
|
|
493
554
|
* NOTE: Because of dictionaries + sequence splitting we MUST make sure
|
|
494
555
|
* that any offset used is valid at the END of the sequence, since it may
|
|
@@ -518,7 +579,9 @@ size_t ZSTD_ldm_generateSequences(
|
|
|
518
579
|
return 0;
|
|
519
580
|
}
|
|
520
581
|
|
|
521
|
-
void
|
|
582
|
+
void
|
|
583
|
+
ZSTD_ldm_skipSequences(rawSeqStore_t* rawSeqStore, size_t srcSize, U32 const minMatch)
|
|
584
|
+
{
|
|
522
585
|
while (srcSize > 0 && rawSeqStore->pos < rawSeqStore->size) {
|
|
523
586
|
rawSeq* seq = rawSeqStore->seq + rawSeqStore->pos;
|
|
524
587
|
if (srcSize <= seq->litLength) {
|
|
@@ -596,12 +659,13 @@ void ZSTD_ldm_skipRawSeqStoreBytes(rawSeqStore_t* rawSeqStore, size_t nbBytes) {
|
|
|
596
659
|
|
|
597
660
|
size_t ZSTD_ldm_blockCompress(rawSeqStore_t* rawSeqStore,
|
|
598
661
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
|
662
|
+
ZSTD_paramSwitch_e useRowMatchFinder,
|
|
599
663
|
void const* src, size_t srcSize)
|
|
600
664
|
{
|
|
601
665
|
const ZSTD_compressionParameters* const cParams = &ms->cParams;
|
|
602
666
|
unsigned const minMatch = cParams->minMatch;
|
|
603
667
|
ZSTD_blockCompressor const blockCompressor =
|
|
604
|
-
ZSTD_selectBlockCompressor(cParams->strategy, ZSTD_matchState_dictMode(ms));
|
|
668
|
+
ZSTD_selectBlockCompressor(cParams->strategy, useRowMatchFinder, ZSTD_matchState_dictMode(ms));
|
|
605
669
|
/* Input bounds */
|
|
606
670
|
BYTE const* const istart = (BYTE const*)src;
|
|
607
671
|
BYTE const* const iend = istart + srcSize;
|
|
@@ -620,7 +684,7 @@ size_t ZSTD_ldm_blockCompress(rawSeqStore_t* rawSeqStore,
|
|
|
620
684
|
|
|
621
685
|
assert(rawSeqStore->pos <= rawSeqStore->size);
|
|
622
686
|
assert(rawSeqStore->size <= rawSeqStore->capacity);
|
|
623
|
-
/* Loop through each sequence and apply the block compressor to the
|
|
687
|
+
/* Loop through each sequence and apply the block compressor to the literals */
|
|
624
688
|
while (rawSeqStore->pos < rawSeqStore->size && ip < iend) {
|
|
625
689
|
/* maybeSplitSequence updates rawSeqStore->pos */
|
|
626
690
|
rawSeq const sequence = maybeSplitSequence(rawSeqStore,
|
|
@@ -647,8 +711,8 @@ size_t ZSTD_ldm_blockCompress(rawSeqStore_t* rawSeqStore,
|
|
|
647
711
|
rep[0] = sequence.offset;
|
|
648
712
|
/* Store the sequence */
|
|
649
713
|
ZSTD_storeSeq(seqStore, newLitLength, ip - newLitLength, iend,
|
|
650
|
-
sequence.offset
|
|
651
|
-
sequence.matchLength
|
|
714
|
+
OFFSET_TO_OFFBASE(sequence.offset),
|
|
715
|
+
sequence.matchLength);
|
|
652
716
|
ip += sequence.matchLength;
|
|
653
717
|
}
|
|
654
718
|
}
|
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
/*
|
|
2
|
-
* Copyright (c)
|
|
2
|
+
* Copyright (c) Meta Platforms, Inc. and affiliates.
|
|
3
3
|
* All rights reserved.
|
|
4
4
|
*
|
|
5
5
|
* This source code is licensed under both the BSD-style license (found in the
|
|
@@ -66,6 +66,7 @@ size_t ZSTD_ldm_generateSequences(
|
|
|
66
66
|
*/
|
|
67
67
|
size_t ZSTD_ldm_blockCompress(rawSeqStore_t* rawSeqStore,
|
|
68
68
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
|
69
|
+
ZSTD_paramSwitch_e useRowMatchFinder,
|
|
69
70
|
void const* src, size_t srcSize);
|
|
70
71
|
|
|
71
72
|
/**
|
|
@@ -73,7 +74,7 @@ size_t ZSTD_ldm_blockCompress(rawSeqStore_t* rawSeqStore,
|
|
|
73
74
|
*
|
|
74
75
|
* Skip past `srcSize` bytes worth of sequences in `rawSeqStore`.
|
|
75
76
|
* Avoids emitting matches less than `minMatch` bytes.
|
|
76
|
-
* Must be called for data
|
|
77
|
+
* Must be called for data that is not passed to ZSTD_ldm_blockCompress().
|
|
77
78
|
*/
|
|
78
79
|
void ZSTD_ldm_skipSequences(rawSeqStore_t* rawSeqStore, size_t srcSize,
|
|
79
80
|
U32 const minMatch);
|