zstdlib 0.4.0-x64-mingw32 → 0.5.0-x64-mingw32
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/CHANGES.md +6 -1
- data/README.md +1 -1
- data/ext/zstdlib/extconf.rb +2 -2
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/bitstream.h +3 -2
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/compiler.h +14 -2
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/cpu.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/debug.c +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/debug.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/entropy_common.c +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/error_private.c +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/error_private.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/fse.h +1 -1
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/fse_decompress.c +2 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/huf.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/mem.h +73 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/pool.c +7 -3
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/pool.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/threading.c +46 -1
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/threading.h +32 -1
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/xxhash.c +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/xxhash.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/zstd_common.c +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/zstd_errors.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/zstd_internal.h +32 -55
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/fse_compress.c +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/hist.c +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/hist.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/huf_compress.c +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_compress.c +633 -436
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_compress_internal.h +54 -12
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_compress_literals.c +10 -5
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_compress_literals.h +1 -1
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_compress_sequences.c +3 -3
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_compress_sequences.h +1 -1
- data/ext/zstdlib/zstd-1.4.4/lib/compress/zstd_cwksp.h +535 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_double_fast.c +9 -9
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_double_fast.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_fast.c +30 -39
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_fast.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_lazy.c +5 -5
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_lazy.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_ldm.c +4 -4
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_ldm.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_opt.c +1 -1
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_opt.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstdmt_compress.c +32 -26
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstdmt_compress.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/decompress/huf_decompress.c +2 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/decompress/zstd_ddict.c +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/decompress/zstd_ddict.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/decompress/zstd_decompress.c +14 -16
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/decompress/zstd_decompress_block.c +144 -146
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/decompress/zstd_decompress_block.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/decompress/zstd_decompress_internal.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/zstd.h +161 -59
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/zlibWrapper/gzclose.c +1 -1
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/zlibWrapper/gzcompatibility.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/zlibWrapper/gzguts.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/zlibWrapper/gzlib.c +9 -9
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/zlibWrapper/gzread.c +16 -8
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/zlibWrapper/gzwrite.c +8 -8
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/zlibWrapper/zstd_zlibwrapper.c +15 -11
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/zlibWrapper/zstd_zlibwrapper.h +0 -0
- data/lib/2.2/zstdlib.so +0 -0
- data/lib/2.3/zstdlib.so +0 -0
- data/lib/2.4/zstdlib.so +0 -0
- data/lib/2.5/zstdlib.so +0 -0
- data/lib/2.6/zstdlib.so +0 -0
- metadata +62 -61
@@ -148,7 +148,7 @@ size_t ZSTD_compressBlock_doubleFast_generic(
|
|
148
148
|
const BYTE* repMatchEnd = repIndex < prefixLowestIndex ? dictEnd : iend;
|
149
149
|
mLength = ZSTD_count_2segments(ip+1+4, repMatch+4, iend, repMatchEnd, prefixLowest) + 4;
|
150
150
|
ip++;
|
151
|
-
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, 0, mLength-MINMATCH);
|
151
|
+
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, 0, mLength-MINMATCH);
|
152
152
|
goto _match_stored;
|
153
153
|
}
|
154
154
|
|
@@ -157,7 +157,7 @@ size_t ZSTD_compressBlock_doubleFast_generic(
|
|
157
157
|
&& ((offset_1 > 0) & (MEM_read32(ip+1-offset_1) == MEM_read32(ip+1)))) {
|
158
158
|
mLength = ZSTD_count(ip+1+4, ip+1+4-offset_1, iend) + 4;
|
159
159
|
ip++;
|
160
|
-
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, 0, mLength-MINMATCH);
|
160
|
+
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, 0, mLength-MINMATCH);
|
161
161
|
goto _match_stored;
|
162
162
|
}
|
163
163
|
|
@@ -247,7 +247,7 @@ _match_found:
|
|
247
247
|
offset_2 = offset_1;
|
248
248
|
offset_1 = offset;
|
249
249
|
|
250
|
-
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, offset + ZSTD_REP_MOVE, mLength-MINMATCH);
|
250
|
+
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, offset + ZSTD_REP_MOVE, mLength-MINMATCH);
|
251
251
|
|
252
252
|
_match_stored:
|
253
253
|
/* match found */
|
@@ -278,7 +278,7 @@ _match_stored:
|
|
278
278
|
const BYTE* const repEnd2 = repIndex2 < prefixLowestIndex ? dictEnd : iend;
|
279
279
|
size_t const repLength2 = ZSTD_count_2segments(ip+4, repMatch2+4, iend, repEnd2, prefixLowest) + 4;
|
280
280
|
U32 tmpOffset = offset_2; offset_2 = offset_1; offset_1 = tmpOffset; /* swap offset_2 <=> offset_1 */
|
281
|
-
ZSTD_storeSeq(seqStore, 0, anchor, 0, repLength2-MINMATCH);
|
281
|
+
ZSTD_storeSeq(seqStore, 0, anchor, iend, 0, repLength2-MINMATCH);
|
282
282
|
hashSmall[ZSTD_hashPtr(ip, hBitsS, mls)] = current2;
|
283
283
|
hashLong[ZSTD_hashPtr(ip, hBitsL, 8)] = current2;
|
284
284
|
ip += repLength2;
|
@@ -297,7 +297,7 @@ _match_stored:
|
|
297
297
|
U32 const tmpOff = offset_2; offset_2 = offset_1; offset_1 = tmpOff; /* swap offset_2 <=> offset_1 */
|
298
298
|
hashSmall[ZSTD_hashPtr(ip, hBitsS, mls)] = (U32)(ip-base);
|
299
299
|
hashLong[ZSTD_hashPtr(ip, hBitsL, 8)] = (U32)(ip-base);
|
300
|
-
ZSTD_storeSeq(seqStore, 0, anchor, 0, rLength-MINMATCH);
|
300
|
+
ZSTD_storeSeq(seqStore, 0, anchor, iend, 0, rLength-MINMATCH);
|
301
301
|
ip += rLength;
|
302
302
|
anchor = ip;
|
303
303
|
continue; /* faster when present ... (?) */
|
@@ -411,7 +411,7 @@ static size_t ZSTD_compressBlock_doubleFast_extDict_generic(
|
|
411
411
|
const BYTE* repMatchEnd = repIndex < prefixStartIndex ? dictEnd : iend;
|
412
412
|
mLength = ZSTD_count_2segments(ip+1+4, repMatch+4, iend, repMatchEnd, prefixStart) + 4;
|
413
413
|
ip++;
|
414
|
-
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, 0, mLength-MINMATCH);
|
414
|
+
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, 0, mLength-MINMATCH);
|
415
415
|
} else {
|
416
416
|
if ((matchLongIndex > dictStartIndex) && (MEM_read64(matchLong) == MEM_read64(ip))) {
|
417
417
|
const BYTE* const matchEnd = matchLongIndex < prefixStartIndex ? dictEnd : iend;
|
@@ -422,7 +422,7 @@ static size_t ZSTD_compressBlock_doubleFast_extDict_generic(
|
|
422
422
|
while (((ip>anchor) & (matchLong>lowMatchPtr)) && (ip[-1] == matchLong[-1])) { ip--; matchLong--; mLength++; } /* catch up */
|
423
423
|
offset_2 = offset_1;
|
424
424
|
offset_1 = offset;
|
425
|
-
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, offset + ZSTD_REP_MOVE, mLength-MINMATCH);
|
425
|
+
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, offset + ZSTD_REP_MOVE, mLength-MINMATCH);
|
426
426
|
|
427
427
|
} else if ((matchIndex > dictStartIndex) && (MEM_read32(match) == MEM_read32(ip))) {
|
428
428
|
size_t const h3 = ZSTD_hashPtr(ip+1, hBitsL, 8);
|
@@ -447,7 +447,7 @@ static size_t ZSTD_compressBlock_doubleFast_extDict_generic(
|
|
447
447
|
}
|
448
448
|
offset_2 = offset_1;
|
449
449
|
offset_1 = offset;
|
450
|
-
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, offset + ZSTD_REP_MOVE, mLength-MINMATCH);
|
450
|
+
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, offset + ZSTD_REP_MOVE, mLength-MINMATCH);
|
451
451
|
|
452
452
|
} else {
|
453
453
|
ip += ((ip-anchor) >> kSearchStrength) + 1;
|
@@ -479,7 +479,7 @@ static size_t ZSTD_compressBlock_doubleFast_extDict_generic(
|
|
479
479
|
const BYTE* const repEnd2 = repIndex2 < prefixStartIndex ? dictEnd : iend;
|
480
480
|
size_t const repLength2 = ZSTD_count_2segments(ip+4, repMatch2+4, iend, repEnd2, prefixStart) + 4;
|
481
481
|
U32 const tmpOffset = offset_2; offset_2 = offset_1; offset_1 = tmpOffset; /* swap offset_2 <=> offset_1 */
|
482
|
-
ZSTD_storeSeq(seqStore, 0, anchor, 0, repLength2-MINMATCH);
|
482
|
+
ZSTD_storeSeq(seqStore, 0, anchor, iend, 0, repLength2-MINMATCH);
|
483
483
|
hashSmall[ZSTD_hashPtr(ip, hBitsS, mls)] = current2;
|
484
484
|
hashLong[ZSTD_hashPtr(ip, hBitsL, 8)] = current2;
|
485
485
|
ip += repLength2;
|
File without changes
|
@@ -8,7 +8,7 @@
|
|
8
8
|
* You may select, at your option, one of the above-listed licenses.
|
9
9
|
*/
|
10
10
|
|
11
|
-
#include "zstd_compress_internal.h"
|
11
|
+
#include "zstd_compress_internal.h" /* ZSTD_hashPtr, ZSTD_count, ZSTD_storeSeq */
|
12
12
|
#include "zstd_fast.h"
|
13
13
|
|
14
14
|
|
@@ -43,8 +43,8 @@ void ZSTD_fillHashTable(ZSTD_matchState_t* ms,
|
|
43
43
|
}
|
44
44
|
|
45
45
|
|
46
|
-
FORCE_INLINE_TEMPLATE
|
47
|
-
|
46
|
+
FORCE_INLINE_TEMPLATE size_t
|
47
|
+
ZSTD_compressBlock_fast_generic(
|
48
48
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
49
49
|
void const* src, size_t srcSize,
|
50
50
|
U32 const mls)
|
@@ -74,8 +74,7 @@ size_t ZSTD_compressBlock_fast_generic(
|
|
74
74
|
DEBUGLOG(5, "ZSTD_compressBlock_fast_generic");
|
75
75
|
ip0 += (ip0 == prefixStart);
|
76
76
|
ip1 = ip0 + 1;
|
77
|
-
{
|
78
|
-
U32 const maxRep = (U32)(ip0 - prefixStart);
|
77
|
+
{ U32 const maxRep = (U32)(ip0 - prefixStart);
|
79
78
|
if (offset_2 > maxRep) offsetSaved = offset_2, offset_2 = 0;
|
80
79
|
if (offset_1 > maxRep) offsetSaved = offset_1, offset_1 = 0;
|
81
80
|
}
|
@@ -118,8 +117,7 @@ size_t ZSTD_compressBlock_fast_generic(
|
|
118
117
|
match0 = match1;
|
119
118
|
goto _offset;
|
120
119
|
}
|
121
|
-
{
|
122
|
-
size_t const step = ((ip0-anchor) >> (kSearchStrength - 1)) + stepSize;
|
120
|
+
{ size_t const step = ((size_t)(ip0-anchor) >> (kSearchStrength - 1)) + stepSize;
|
123
121
|
assert(step >= 2);
|
124
122
|
ip0 += step;
|
125
123
|
ip1 += step;
|
@@ -138,7 +136,7 @@ _offset: /* Requires: ip0, match0 */
|
|
138
136
|
_match: /* Requires: ip0, match0, offcode */
|
139
137
|
/* Count the forward length */
|
140
138
|
mLength += ZSTD_count(ip0+mLength+4, match0+mLength+4, iend) + 4;
|
141
|
-
ZSTD_storeSeq(seqStore, ip0-anchor, anchor, offcode, mLength-MINMATCH);
|
139
|
+
ZSTD_storeSeq(seqStore, (size_t)(ip0-anchor), anchor, iend, offcode, mLength-MINMATCH);
|
142
140
|
/* match found */
|
143
141
|
ip0 += mLength;
|
144
142
|
anchor = ip0;
|
@@ -150,16 +148,15 @@ _match: /* Requires: ip0, match0, offcode */
|
|
150
148
|
hashTable[ZSTD_hashPtr(base+current0+2, hlog, mls)] = current0+2; /* here because current+2 could be > iend-8 */
|
151
149
|
hashTable[ZSTD_hashPtr(ip0-2, hlog, mls)] = (U32)(ip0-2-base);
|
152
150
|
|
153
|
-
while ( (ip0 <= ilimit)
|
154
|
-
&& ( (offset_2
|
155
|
-
& (MEM_read32(ip0) == MEM_read32(ip0 - offset_2)) )) {
|
151
|
+
while ( ((ip0 <= ilimit) & (offset_2>0)) /* offset_2==0 means offset_2 is invalidated */
|
152
|
+
&& (MEM_read32(ip0) == MEM_read32(ip0 - offset_2)) ) {
|
156
153
|
/* store sequence */
|
157
154
|
size_t const rLength = ZSTD_count(ip0+4, ip0+4-offset_2, iend) + 4;
|
158
|
-
U32 const tmpOff = offset_2; offset_2 = offset_1; offset_1 = tmpOff;
|
155
|
+
{ U32 const tmpOff = offset_2; offset_2 = offset_1; offset_1 = tmpOff; } /* swap offset_2 <=> offset_1 */
|
159
156
|
hashTable[ZSTD_hashPtr(ip0, hlog, mls)] = (U32)(ip0-base);
|
160
157
|
ip0 += rLength;
|
161
158
|
ip1 = ip0 + 1;
|
162
|
-
ZSTD_storeSeq(seqStore, 0
|
159
|
+
ZSTD_storeSeq(seqStore, 0 /*litLen*/, anchor, iend, 0 /*offCode*/, rLength-MINMATCH);
|
163
160
|
anchor = ip0;
|
164
161
|
continue; /* faster when present (confirmed on gcc-8) ... (?) */
|
165
162
|
}
|
@@ -179,8 +176,7 @@ size_t ZSTD_compressBlock_fast(
|
|
179
176
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
180
177
|
void const* src, size_t srcSize)
|
181
178
|
{
|
182
|
-
|
183
|
-
U32 const mls = cParams->minMatch;
|
179
|
+
U32 const mls = ms->cParams.minMatch;
|
184
180
|
assert(ms->dictMatchState == NULL);
|
185
181
|
switch(mls)
|
186
182
|
{
|
@@ -265,7 +261,7 @@ size_t ZSTD_compressBlock_fast_dictMatchState_generic(
|
|
265
261
|
const BYTE* const repMatchEnd = repIndex < prefixStartIndex ? dictEnd : iend;
|
266
262
|
mLength = ZSTD_count_2segments(ip+1+4, repMatch+4, iend, repMatchEnd, prefixStart) + 4;
|
267
263
|
ip++;
|
268
|
-
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, 0, mLength-MINMATCH);
|
264
|
+
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, 0, mLength-MINMATCH);
|
269
265
|
} else if ( (matchIndex <= prefixStartIndex) ) {
|
270
266
|
size_t const dictHash = ZSTD_hashPtr(ip, dictHLog, mls);
|
271
267
|
U32 const dictMatchIndex = dictHashTable[dictHash];
|
@@ -285,7 +281,7 @@ size_t ZSTD_compressBlock_fast_dictMatchState_generic(
|
|
285
281
|
} /* catch up */
|
286
282
|
offset_2 = offset_1;
|
287
283
|
offset_1 = offset;
|
288
|
-
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, offset + ZSTD_REP_MOVE, mLength-MINMATCH);
|
284
|
+
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, offset + ZSTD_REP_MOVE, mLength-MINMATCH);
|
289
285
|
}
|
290
286
|
} else if (MEM_read32(match) != MEM_read32(ip)) {
|
291
287
|
/* it's not a match, and we're not going to check the dictionary */
|
@@ -300,7 +296,7 @@ size_t ZSTD_compressBlock_fast_dictMatchState_generic(
|
|
300
296
|
&& (ip[-1] == match[-1])) { ip--; match--; mLength++; } /* catch up */
|
301
297
|
offset_2 = offset_1;
|
302
298
|
offset_1 = offset;
|
303
|
-
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, offset + ZSTD_REP_MOVE, mLength-MINMATCH);
|
299
|
+
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, offset + ZSTD_REP_MOVE, mLength-MINMATCH);
|
304
300
|
}
|
305
301
|
|
306
302
|
/* match found */
|
@@ -325,7 +321,7 @@ size_t ZSTD_compressBlock_fast_dictMatchState_generic(
|
|
325
321
|
const BYTE* const repEnd2 = repIndex2 < prefixStartIndex ? dictEnd : iend;
|
326
322
|
size_t const repLength2 = ZSTD_count_2segments(ip+4, repMatch2+4, iend, repEnd2, prefixStart) + 4;
|
327
323
|
U32 tmpOffset = offset_2; offset_2 = offset_1; offset_1 = tmpOffset; /* swap offset_2 <=> offset_1 */
|
328
|
-
ZSTD_storeSeq(seqStore, 0, anchor, 0, repLength2-MINMATCH);
|
324
|
+
ZSTD_storeSeq(seqStore, 0, anchor, iend, 0, repLength2-MINMATCH);
|
329
325
|
hashTable[ZSTD_hashPtr(ip, hlog, mls)] = current2;
|
330
326
|
ip += repLength2;
|
331
327
|
anchor = ip;
|
@@ -348,8 +344,7 @@ size_t ZSTD_compressBlock_fast_dictMatchState(
|
|
348
344
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
349
345
|
void const* src, size_t srcSize)
|
350
346
|
{
|
351
|
-
|
352
|
-
U32 const mls = cParams->minMatch;
|
347
|
+
U32 const mls = ms->cParams.minMatch;
|
353
348
|
assert(ms->dictMatchState != NULL);
|
354
349
|
switch(mls)
|
355
350
|
{
|
@@ -408,16 +403,17 @@ static size_t ZSTD_compressBlock_fast_extDict_generic(
|
|
408
403
|
const U32 repIndex = current + 1 - offset_1;
|
409
404
|
const BYTE* const repBase = repIndex < prefixStartIndex ? dictBase : base;
|
410
405
|
const BYTE* const repMatch = repBase + repIndex;
|
411
|
-
size_t mLength;
|
412
406
|
hashTable[h] = current; /* update hash table */
|
413
407
|
assert(offset_1 <= current +1); /* check repIndex */
|
414
408
|
|
415
409
|
if ( (((U32)((prefixStartIndex-1) - repIndex) >= 3) /* intentional underflow */ & (repIndex > dictStartIndex))
|
416
410
|
&& (MEM_read32(repMatch) == MEM_read32(ip+1)) ) {
|
417
411
|
const BYTE* const repMatchEnd = repIndex < prefixStartIndex ? dictEnd : iend;
|
418
|
-
|
412
|
+
size_t const rLength = ZSTD_count_2segments(ip+1 +4, repMatch +4, iend, repMatchEnd, prefixStart) + 4;
|
419
413
|
ip++;
|
420
|
-
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, 0,
|
414
|
+
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, 0, rLength-MINMATCH);
|
415
|
+
ip += rLength;
|
416
|
+
anchor = ip;
|
421
417
|
} else {
|
422
418
|
if ( (matchIndex < dictStartIndex) ||
|
423
419
|
(MEM_read32(match) != MEM_read32(ip)) ) {
|
@@ -427,19 +423,15 @@ static size_t ZSTD_compressBlock_fast_extDict_generic(
|
|
427
423
|
}
|
428
424
|
{ const BYTE* const matchEnd = matchIndex < prefixStartIndex ? dictEnd : iend;
|
429
425
|
const BYTE* const lowMatchPtr = matchIndex < prefixStartIndex ? dictStart : prefixStart;
|
430
|
-
U32 offset;
|
431
|
-
mLength = ZSTD_count_2segments(ip+4, match+4, iend, matchEnd, prefixStart) + 4;
|
426
|
+
U32 const offset = current - matchIndex;
|
427
|
+
size_t mLength = ZSTD_count_2segments(ip+4, match+4, iend, matchEnd, prefixStart) + 4;
|
432
428
|
while (((ip>anchor) & (match>lowMatchPtr)) && (ip[-1] == match[-1])) { ip--; match--; mLength++; } /* catch up */
|
433
|
-
|
434
|
-
|
435
|
-
|
436
|
-
|
429
|
+
offset_2 = offset_1; offset_1 = offset; /* update offset history */
|
430
|
+
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, offset + ZSTD_REP_MOVE, mLength-MINMATCH);
|
431
|
+
ip += mLength;
|
432
|
+
anchor = ip;
|
437
433
|
} }
|
438
434
|
|
439
|
-
/* found a match : store it */
|
440
|
-
ip += mLength;
|
441
|
-
anchor = ip;
|
442
|
-
|
443
435
|
if (ip <= ilimit) {
|
444
436
|
/* Fill Table */
|
445
437
|
hashTable[ZSTD_hashPtr(base+current+2, hlog, mls)] = current+2;
|
@@ -448,13 +440,13 @@ static size_t ZSTD_compressBlock_fast_extDict_generic(
|
|
448
440
|
while (ip <= ilimit) {
|
449
441
|
U32 const current2 = (U32)(ip-base);
|
450
442
|
U32 const repIndex2 = current2 - offset_2;
|
451
|
-
const BYTE* repMatch2 = repIndex2 < prefixStartIndex ? dictBase + repIndex2 : base + repIndex2;
|
443
|
+
const BYTE* const repMatch2 = repIndex2 < prefixStartIndex ? dictBase + repIndex2 : base + repIndex2;
|
452
444
|
if ( (((U32)((prefixStartIndex-1) - repIndex2) >= 3) & (repIndex2 > dictStartIndex)) /* intentional overflow */
|
453
445
|
&& (MEM_read32(repMatch2) == MEM_read32(ip)) ) {
|
454
446
|
const BYTE* const repEnd2 = repIndex2 < prefixStartIndex ? dictEnd : iend;
|
455
447
|
size_t const repLength2 = ZSTD_count_2segments(ip+4, repMatch2+4, iend, repEnd2, prefixStart) + 4;
|
456
|
-
U32 const tmpOffset = offset_2; offset_2 = offset_1; offset_1 = tmpOffset;
|
457
|
-
ZSTD_storeSeq(seqStore, 0
|
448
|
+
{ U32 const tmpOffset = offset_2; offset_2 = offset_1; offset_1 = tmpOffset; } /* swap offset_2 <=> offset_1 */
|
449
|
+
ZSTD_storeSeq(seqStore, 0 /*litlen*/, anchor, iend, 0 /*offcode*/, repLength2-MINMATCH);
|
458
450
|
hashTable[ZSTD_hashPtr(ip, hlog, mls)] = current2;
|
459
451
|
ip += repLength2;
|
460
452
|
anchor = ip;
|
@@ -476,8 +468,7 @@ size_t ZSTD_compressBlock_fast_extDict(
|
|
476
468
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
477
469
|
void const* src, size_t srcSize)
|
478
470
|
{
|
479
|
-
|
480
|
-
U32 const mls = cParams->minMatch;
|
471
|
+
U32 const mls = ms->cParams.minMatch;
|
481
472
|
switch(mls)
|
482
473
|
{
|
483
474
|
default: /* includes case 3 */
|
File without changes
|
@@ -810,7 +810,7 @@ ZSTD_compressBlock_lazy_generic(
|
|
810
810
|
/* store sequence */
|
811
811
|
_storeSequence:
|
812
812
|
{ size_t const litLength = start - anchor;
|
813
|
-
ZSTD_storeSeq(seqStore, litLength, anchor, (U32)offset, matchLength-MINMATCH);
|
813
|
+
ZSTD_storeSeq(seqStore, litLength, anchor, iend, (U32)offset, matchLength-MINMATCH);
|
814
814
|
anchor = ip = start + matchLength;
|
815
815
|
}
|
816
816
|
|
@@ -828,7 +828,7 @@ _storeSequence:
|
|
828
828
|
const BYTE* const repEnd2 = repIndex < prefixLowestIndex ? dictEnd : iend;
|
829
829
|
matchLength = ZSTD_count_2segments(ip+4, repMatch+4, iend, repEnd2, prefixLowest) + 4;
|
830
830
|
offset = offset_2; offset_2 = offset_1; offset_1 = (U32)offset; /* swap offset_2 <=> offset_1 */
|
831
|
-
ZSTD_storeSeq(seqStore, 0, anchor, 0, matchLength-MINMATCH);
|
831
|
+
ZSTD_storeSeq(seqStore, 0, anchor, iend, 0, matchLength-MINMATCH);
|
832
832
|
ip += matchLength;
|
833
833
|
anchor = ip;
|
834
834
|
continue;
|
@@ -843,7 +843,7 @@ _storeSequence:
|
|
843
843
|
/* store sequence */
|
844
844
|
matchLength = ZSTD_count(ip+4, ip+4-offset_2, iend) + 4;
|
845
845
|
offset = offset_2; offset_2 = offset_1; offset_1 = (U32)offset; /* swap repcodes */
|
846
|
-
ZSTD_storeSeq(seqStore, 0, anchor, 0, matchLength-MINMATCH);
|
846
|
+
ZSTD_storeSeq(seqStore, 0, anchor, iend, 0, matchLength-MINMATCH);
|
847
847
|
ip += matchLength;
|
848
848
|
anchor = ip;
|
849
849
|
continue; /* faster when present ... (?) */
|
@@ -1051,7 +1051,7 @@ size_t ZSTD_compressBlock_lazy_extDict_generic(
|
|
1051
1051
|
/* store sequence */
|
1052
1052
|
_storeSequence:
|
1053
1053
|
{ size_t const litLength = start - anchor;
|
1054
|
-
ZSTD_storeSeq(seqStore, litLength, anchor, (U32)offset, matchLength-MINMATCH);
|
1054
|
+
ZSTD_storeSeq(seqStore, litLength, anchor, iend, (U32)offset, matchLength-MINMATCH);
|
1055
1055
|
anchor = ip = start + matchLength;
|
1056
1056
|
}
|
1057
1057
|
|
@@ -1066,7 +1066,7 @@ _storeSequence:
|
|
1066
1066
|
const BYTE* const repEnd = repIndex < dictLimit ? dictEnd : iend;
|
1067
1067
|
matchLength = ZSTD_count_2segments(ip+4, repMatch+4, iend, repEnd, prefixStart) + 4;
|
1068
1068
|
offset = offset_2; offset_2 = offset_1; offset_1 = (U32)offset; /* swap offset history */
|
1069
|
-
ZSTD_storeSeq(seqStore, 0, anchor, 0, matchLength-MINMATCH);
|
1069
|
+
ZSTD_storeSeq(seqStore, 0, anchor, iend, 0, matchLength-MINMATCH);
|
1070
1070
|
ip += matchLength;
|
1071
1071
|
anchor = ip;
|
1072
1072
|
continue; /* faster when present ... (?) */
|
File without changes
|
@@ -49,9 +49,9 @@ size_t ZSTD_ldm_getTableSize(ldmParams_t params)
|
|
49
49
|
{
|
50
50
|
size_t const ldmHSize = ((size_t)1) << params.hashLog;
|
51
51
|
size_t const ldmBucketSizeLog = MIN(params.bucketSizeLog, params.hashLog);
|
52
|
-
size_t const ldmBucketSize =
|
53
|
-
|
54
|
-
|
52
|
+
size_t const ldmBucketSize = ((size_t)1) << (params.hashLog - ldmBucketSizeLog);
|
53
|
+
size_t const totalSize = ZSTD_cwksp_alloc_size(ldmBucketSize)
|
54
|
+
+ ZSTD_cwksp_alloc_size(ldmHSize * sizeof(ldmEntry_t));
|
55
55
|
return params.enableLdm ? totalSize : 0;
|
56
56
|
}
|
57
57
|
|
@@ -583,7 +583,7 @@ size_t ZSTD_ldm_blockCompress(rawSeqStore_t* rawSeqStore,
|
|
583
583
|
rep[i] = rep[i-1];
|
584
584
|
rep[0] = sequence.offset;
|
585
585
|
/* Store the sequence */
|
586
|
-
ZSTD_storeSeq(seqStore, newLitLength, ip - newLitLength,
|
586
|
+
ZSTD_storeSeq(seqStore, newLitLength, ip - newLitLength, iend,
|
587
587
|
sequence.offset + ZSTD_REP_MOVE,
|
588
588
|
sequence.matchLength - MINMATCH);
|
589
589
|
ip += sequence.matchLength;
|
File without changes
|
@@ -1098,7 +1098,7 @@ _shortestPath: /* cur, last_pos, best_mlen, best_off have to be set */
|
|
1098
1098
|
|
1099
1099
|
assert(anchor + llen <= iend);
|
1100
1100
|
ZSTD_updateStats(optStatePtr, llen, anchor, offCode, mlen);
|
1101
|
-
ZSTD_storeSeq(seqStore, llen, anchor, offCode, mlen-MINMATCH);
|
1101
|
+
ZSTD_storeSeq(seqStore, llen, anchor, iend, offCode, mlen-MINMATCH);
|
1102
1102
|
anchor += advance;
|
1103
1103
|
ip = anchor;
|
1104
1104
|
} }
|
File without changes
|
@@ -668,7 +668,7 @@ static void ZSTDMT_compressionJob(void* jobDescription)
|
|
668
668
|
|
669
669
|
/* init */
|
670
670
|
if (job->cdict) {
|
671
|
-
size_t const initError = ZSTD_compressBegin_advanced_internal(cctx, NULL, 0, ZSTD_dct_auto, ZSTD_dtlm_fast, job->cdict, jobParams, job->fullFrameSize);
|
671
|
+
size_t const initError = ZSTD_compressBegin_advanced_internal(cctx, NULL, 0, ZSTD_dct_auto, ZSTD_dtlm_fast, job->cdict, &jobParams, job->fullFrameSize);
|
672
672
|
assert(job->firstJob); /* only allowed for first job */
|
673
673
|
if (ZSTD_isError(initError)) JOB_ERROR(initError);
|
674
674
|
} else { /* srcStart points at reloaded section */
|
@@ -680,7 +680,7 @@ static void ZSTDMT_compressionJob(void* jobDescription)
|
|
680
680
|
job->prefix.start, job->prefix.size, ZSTD_dct_rawContent, /* load dictionary in "content-only" mode (no header analysis) */
|
681
681
|
ZSTD_dtlm_fast,
|
682
682
|
NULL, /*cdict*/
|
683
|
-
jobParams, pledgedSrcSize);
|
683
|
+
&jobParams, pledgedSrcSize);
|
684
684
|
if (ZSTD_isError(initError)) JOB_ERROR(initError);
|
685
685
|
} }
|
686
686
|
|
@@ -927,12 +927,18 @@ static void ZSTDMT_releaseAllJobResources(ZSTDMT_CCtx* mtctx)
|
|
927
927
|
unsigned jobID;
|
928
928
|
DEBUGLOG(3, "ZSTDMT_releaseAllJobResources");
|
929
929
|
for (jobID=0; jobID <= mtctx->jobIDMask; jobID++) {
|
930
|
+
/* Copy the mutex/cond out */
|
931
|
+
ZSTD_pthread_mutex_t const mutex = mtctx->jobs[jobID].job_mutex;
|
932
|
+
ZSTD_pthread_cond_t const cond = mtctx->jobs[jobID].job_cond;
|
933
|
+
|
930
934
|
DEBUGLOG(4, "job%02u: release dst address %08X", jobID, (U32)(size_t)mtctx->jobs[jobID].dstBuff.start);
|
931
935
|
ZSTDMT_releaseBuffer(mtctx->bufPool, mtctx->jobs[jobID].dstBuff);
|
932
|
-
|
933
|
-
|
936
|
+
|
937
|
+
/* Clear the job description, but keep the mutex/cond */
|
938
|
+
memset(&mtctx->jobs[jobID], 0, sizeof(mtctx->jobs[jobID]));
|
939
|
+
mtctx->jobs[jobID].job_mutex = mutex;
|
940
|
+
mtctx->jobs[jobID].job_cond = cond;
|
934
941
|
}
|
935
|
-
memset(mtctx->jobs, 0, (mtctx->jobIDMask+1)*sizeof(ZSTDMT_jobDescription));
|
936
942
|
mtctx->inBuff.buffer = g_nullBuffer;
|
937
943
|
mtctx->inBuff.filled = 0;
|
938
944
|
mtctx->allJobsCompleted = 1;
|
@@ -1028,9 +1034,9 @@ size_t ZSTDMT_getMTCtxParameter(ZSTDMT_CCtx* mtctx, ZSTDMT_parameter parameter,
|
|
1028
1034
|
|
1029
1035
|
/* Sets parameters relevant to the compression job,
|
1030
1036
|
* initializing others to default values. */
|
1031
|
-
static ZSTD_CCtx_params ZSTDMT_initJobCCtxParams(ZSTD_CCtx_params
|
1037
|
+
static ZSTD_CCtx_params ZSTDMT_initJobCCtxParams(const ZSTD_CCtx_params* params)
|
1032
1038
|
{
|
1033
|
-
ZSTD_CCtx_params jobParams = params;
|
1039
|
+
ZSTD_CCtx_params jobParams = *params;
|
1034
1040
|
/* Clear parameters related to multithreading */
|
1035
1041
|
jobParams.forceWindow = 0;
|
1036
1042
|
jobParams.nbWorkers = 0;
|
@@ -1151,16 +1157,16 @@ size_t ZSTDMT_toFlushNow(ZSTDMT_CCtx* mtctx)
|
|
1151
1157
|
/* ===== Multi-threaded compression ===== */
|
1152
1158
|
/* ------------------------------------------ */
|
1153
1159
|
|
1154
|
-
static unsigned ZSTDMT_computeTargetJobLog(ZSTD_CCtx_params
|
1160
|
+
static unsigned ZSTDMT_computeTargetJobLog(const ZSTD_CCtx_params* params)
|
1155
1161
|
{
|
1156
1162
|
unsigned jobLog;
|
1157
|
-
if (params
|
1163
|
+
if (params->ldmParams.enableLdm) {
|
1158
1164
|
/* In Long Range Mode, the windowLog is typically oversized.
|
1159
1165
|
* In which case, it's preferable to determine the jobSize
|
1160
1166
|
* based on chainLog instead. */
|
1161
|
-
jobLog = MAX(21, params
|
1167
|
+
jobLog = MAX(21, params->cParams.chainLog + 4);
|
1162
1168
|
} else {
|
1163
|
-
jobLog = MAX(20, params
|
1169
|
+
jobLog = MAX(20, params->cParams.windowLog + 2);
|
1164
1170
|
}
|
1165
1171
|
return MIN(jobLog, (unsigned)ZSTDMT_JOBLOG_MAX);
|
1166
1172
|
}
|
@@ -1193,27 +1199,27 @@ static int ZSTDMT_overlapLog(int ovlog, ZSTD_strategy strat)
|
|
1193
1199
|
return ovlog;
|
1194
1200
|
}
|
1195
1201
|
|
1196
|
-
static size_t ZSTDMT_computeOverlapSize(ZSTD_CCtx_params
|
1202
|
+
static size_t ZSTDMT_computeOverlapSize(const ZSTD_CCtx_params* params)
|
1197
1203
|
{
|
1198
|
-
int const overlapRLog = 9 - ZSTDMT_overlapLog(params
|
1199
|
-
int ovLog = (overlapRLog >= 8) ? 0 : (params
|
1204
|
+
int const overlapRLog = 9 - ZSTDMT_overlapLog(params->overlapLog, params->cParams.strategy);
|
1205
|
+
int ovLog = (overlapRLog >= 8) ? 0 : (params->cParams.windowLog - overlapRLog);
|
1200
1206
|
assert(0 <= overlapRLog && overlapRLog <= 8);
|
1201
|
-
if (params
|
1207
|
+
if (params->ldmParams.enableLdm) {
|
1202
1208
|
/* In Long Range Mode, the windowLog is typically oversized.
|
1203
1209
|
* In which case, it's preferable to determine the jobSize
|
1204
1210
|
* based on chainLog instead.
|
1205
1211
|
* Then, ovLog becomes a fraction of the jobSize, rather than windowSize */
|
1206
|
-
ovLog = MIN(params
|
1212
|
+
ovLog = MIN(params->cParams.windowLog, ZSTDMT_computeTargetJobLog(params) - 2)
|
1207
1213
|
- overlapRLog;
|
1208
1214
|
}
|
1209
1215
|
assert(0 <= ovLog && ovLog <= ZSTD_WINDOWLOG_MAX);
|
1210
|
-
DEBUGLOG(4, "overlapLog : %i", params
|
1216
|
+
DEBUGLOG(4, "overlapLog : %i", params->overlapLog);
|
1211
1217
|
DEBUGLOG(4, "overlap size : %i", 1 << ovLog);
|
1212
1218
|
return (ovLog==0) ? 0 : (size_t)1 << ovLog;
|
1213
1219
|
}
|
1214
1220
|
|
1215
1221
|
static unsigned
|
1216
|
-
ZSTDMT_computeNbJobs(ZSTD_CCtx_params params, size_t srcSize, unsigned nbWorkers)
|
1222
|
+
ZSTDMT_computeNbJobs(const ZSTD_CCtx_params* params, size_t srcSize, unsigned nbWorkers)
|
1217
1223
|
{
|
1218
1224
|
assert(nbWorkers>0);
|
1219
1225
|
{ size_t const jobSizeTarget = (size_t)1 << ZSTDMT_computeTargetJobLog(params);
|
@@ -1236,9 +1242,9 @@ static size_t ZSTDMT_compress_advanced_internal(
|
|
1236
1242
|
const ZSTD_CDict* cdict,
|
1237
1243
|
ZSTD_CCtx_params params)
|
1238
1244
|
{
|
1239
|
-
ZSTD_CCtx_params const jobParams = ZSTDMT_initJobCCtxParams(params);
|
1240
|
-
size_t const overlapSize = ZSTDMT_computeOverlapSize(params);
|
1241
|
-
unsigned const nbJobs = ZSTDMT_computeNbJobs(params, srcSize, params.nbWorkers);
|
1245
|
+
ZSTD_CCtx_params const jobParams = ZSTDMT_initJobCCtxParams(¶ms);
|
1246
|
+
size_t const overlapSize = ZSTDMT_computeOverlapSize(¶ms);
|
1247
|
+
unsigned const nbJobs = ZSTDMT_computeNbJobs(¶ms, srcSize, params.nbWorkers);
|
1242
1248
|
size_t const proposedJobSize = (srcSize + (nbJobs-1)) / nbJobs;
|
1243
1249
|
size_t const avgJobSize = (((proposedJobSize-1) & 0x1FFFF) < 0x7FFF) ? proposedJobSize + 0xFFFF : proposedJobSize; /* avoid too small last block */
|
1244
1250
|
const char* const srcStart = (const char*)src;
|
@@ -1256,7 +1262,7 @@ static size_t ZSTDMT_compress_advanced_internal(
|
|
1256
1262
|
ZSTD_CCtx* const cctx = mtctx->cctxPool->cctx[0];
|
1257
1263
|
DEBUGLOG(4, "ZSTDMT_compress_advanced_internal: fallback to single-thread mode");
|
1258
1264
|
if (cdict) return ZSTD_compress_usingCDict_advanced(cctx, dst, dstCapacity, src, srcSize, cdict, jobParams.fParams);
|
1259
|
-
return ZSTD_compress_advanced_internal(cctx, dst, dstCapacity, src, srcSize, NULL, 0, jobParams);
|
1265
|
+
return ZSTD_compress_advanced_internal(cctx, dst, dstCapacity, src, srcSize, NULL, 0, &jobParams);
|
1260
1266
|
}
|
1261
1267
|
|
1262
1268
|
assert(avgJobSize >= 256 KB); /* condition for ZSTD_compressBound(A) + ZSTD_compressBound(B) <= ZSTD_compressBound(A+B), required to compress directly into Dst (no additional buffer) */
|
@@ -1404,12 +1410,12 @@ size_t ZSTDMT_initCStream_internal(
|
|
1404
1410
|
|
1405
1411
|
mtctx->singleBlockingThread = (pledgedSrcSize <= ZSTDMT_JOBSIZE_MIN); /* do not trigger multi-threading when srcSize is too small */
|
1406
1412
|
if (mtctx->singleBlockingThread) {
|
1407
|
-
ZSTD_CCtx_params const singleThreadParams = ZSTDMT_initJobCCtxParams(params);
|
1413
|
+
ZSTD_CCtx_params const singleThreadParams = ZSTDMT_initJobCCtxParams(¶ms);
|
1408
1414
|
DEBUGLOG(5, "ZSTDMT_initCStream_internal: switch to single blocking thread mode");
|
1409
1415
|
assert(singleThreadParams.nbWorkers == 0);
|
1410
1416
|
return ZSTD_initCStream_internal(mtctx->cctxPool->cctx[0],
|
1411
1417
|
dict, dictSize, cdict,
|
1412
|
-
singleThreadParams, pledgedSrcSize);
|
1418
|
+
&singleThreadParams, pledgedSrcSize);
|
1413
1419
|
}
|
1414
1420
|
|
1415
1421
|
DEBUGLOG(4, "ZSTDMT_initCStream_internal: %u workers", params.nbWorkers);
|
@@ -1435,11 +1441,11 @@ size_t ZSTDMT_initCStream_internal(
|
|
1435
1441
|
mtctx->cdict = cdict;
|
1436
1442
|
}
|
1437
1443
|
|
1438
|
-
mtctx->targetPrefixSize = ZSTDMT_computeOverlapSize(params);
|
1444
|
+
mtctx->targetPrefixSize = ZSTDMT_computeOverlapSize(¶ms);
|
1439
1445
|
DEBUGLOG(4, "overlapLog=%i => %u KB", params.overlapLog, (U32)(mtctx->targetPrefixSize>>10));
|
1440
1446
|
mtctx->targetSectionSize = params.jobSize;
|
1441
1447
|
if (mtctx->targetSectionSize == 0) {
|
1442
|
-
mtctx->targetSectionSize = 1ULL << ZSTDMT_computeTargetJobLog(params);
|
1448
|
+
mtctx->targetSectionSize = 1ULL << ZSTDMT_computeTargetJobLog(¶ms);
|
1443
1449
|
}
|
1444
1450
|
assert(mtctx->targetSectionSize <= (size_t)ZSTDMT_JOBSIZE_MAX);
|
1445
1451
|
|