zstdlib 0.4.0-x86-mingw32 → 0.5.0-x86-mingw32

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (70) hide show
  1. checksums.yaml +4 -4
  2. data/CHANGES.md +6 -1
  3. data/README.md +1 -1
  4. data/ext/zstdlib/extconf.rb +2 -2
  5. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/bitstream.h +3 -2
  6. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/compiler.h +14 -2
  7. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/cpu.h +0 -0
  8. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/debug.c +0 -0
  9. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/debug.h +0 -0
  10. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/entropy_common.c +0 -0
  11. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/error_private.c +0 -0
  12. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/error_private.h +0 -0
  13. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/fse.h +1 -1
  14. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/fse_decompress.c +2 -0
  15. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/huf.h +0 -0
  16. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/mem.h +73 -0
  17. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/pool.c +7 -3
  18. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/pool.h +0 -0
  19. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/threading.c +46 -1
  20. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/threading.h +32 -1
  21. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/xxhash.c +0 -0
  22. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/xxhash.h +0 -0
  23. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/zstd_common.c +0 -0
  24. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/zstd_errors.h +0 -0
  25. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/zstd_internal.h +32 -55
  26. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/fse_compress.c +0 -0
  27. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/hist.c +0 -0
  28. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/hist.h +0 -0
  29. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/huf_compress.c +0 -0
  30. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_compress.c +633 -436
  31. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_compress_internal.h +54 -12
  32. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_compress_literals.c +10 -5
  33. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_compress_literals.h +1 -1
  34. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_compress_sequences.c +3 -3
  35. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_compress_sequences.h +1 -1
  36. data/ext/zstdlib/zstd-1.4.4/lib/compress/zstd_cwksp.h +535 -0
  37. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_double_fast.c +9 -9
  38. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_double_fast.h +0 -0
  39. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_fast.c +30 -39
  40. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_fast.h +0 -0
  41. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_lazy.c +5 -5
  42. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_lazy.h +0 -0
  43. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_ldm.c +4 -4
  44. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_ldm.h +0 -0
  45. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_opt.c +1 -1
  46. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_opt.h +0 -0
  47. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstdmt_compress.c +32 -26
  48. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstdmt_compress.h +0 -0
  49. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/decompress/huf_decompress.c +2 -0
  50. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/decompress/zstd_ddict.c +0 -0
  51. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/decompress/zstd_ddict.h +0 -0
  52. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/decompress/zstd_decompress.c +14 -16
  53. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/decompress/zstd_decompress_block.c +144 -146
  54. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/decompress/zstd_decompress_block.h +0 -0
  55. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/decompress/zstd_decompress_internal.h +0 -0
  56. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/zstd.h +161 -59
  57. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/zlibWrapper/gzclose.c +1 -1
  58. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/zlibWrapper/gzcompatibility.h +0 -0
  59. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/zlibWrapper/gzguts.h +0 -0
  60. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/zlibWrapper/gzlib.c +9 -9
  61. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/zlibWrapper/gzread.c +16 -8
  62. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/zlibWrapper/gzwrite.c +8 -8
  63. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/zlibWrapper/zstd_zlibwrapper.c +15 -11
  64. data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/zlibWrapper/zstd_zlibwrapper.h +0 -0
  65. data/lib/2.2/zstdlib.so +0 -0
  66. data/lib/2.3/zstdlib.so +0 -0
  67. data/lib/2.4/zstdlib.so +0 -0
  68. data/lib/2.5/zstdlib.so +0 -0
  69. data/lib/2.6/zstdlib.so +0 -0
  70. metadata +62 -61
@@ -148,7 +148,7 @@ size_t ZSTD_compressBlock_doubleFast_generic(
148
148
  const BYTE* repMatchEnd = repIndex < prefixLowestIndex ? dictEnd : iend;
149
149
  mLength = ZSTD_count_2segments(ip+1+4, repMatch+4, iend, repMatchEnd, prefixLowest) + 4;
150
150
  ip++;
151
- ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, 0, mLength-MINMATCH);
151
+ ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, 0, mLength-MINMATCH);
152
152
  goto _match_stored;
153
153
  }
154
154
 
@@ -157,7 +157,7 @@ size_t ZSTD_compressBlock_doubleFast_generic(
157
157
  && ((offset_1 > 0) & (MEM_read32(ip+1-offset_1) == MEM_read32(ip+1)))) {
158
158
  mLength = ZSTD_count(ip+1+4, ip+1+4-offset_1, iend) + 4;
159
159
  ip++;
160
- ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, 0, mLength-MINMATCH);
160
+ ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, 0, mLength-MINMATCH);
161
161
  goto _match_stored;
162
162
  }
163
163
 
@@ -247,7 +247,7 @@ _match_found:
247
247
  offset_2 = offset_1;
248
248
  offset_1 = offset;
249
249
 
250
- ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, offset + ZSTD_REP_MOVE, mLength-MINMATCH);
250
+ ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, offset + ZSTD_REP_MOVE, mLength-MINMATCH);
251
251
 
252
252
  _match_stored:
253
253
  /* match found */
@@ -278,7 +278,7 @@ _match_stored:
278
278
  const BYTE* const repEnd2 = repIndex2 < prefixLowestIndex ? dictEnd : iend;
279
279
  size_t const repLength2 = ZSTD_count_2segments(ip+4, repMatch2+4, iend, repEnd2, prefixLowest) + 4;
280
280
  U32 tmpOffset = offset_2; offset_2 = offset_1; offset_1 = tmpOffset; /* swap offset_2 <=> offset_1 */
281
- ZSTD_storeSeq(seqStore, 0, anchor, 0, repLength2-MINMATCH);
281
+ ZSTD_storeSeq(seqStore, 0, anchor, iend, 0, repLength2-MINMATCH);
282
282
  hashSmall[ZSTD_hashPtr(ip, hBitsS, mls)] = current2;
283
283
  hashLong[ZSTD_hashPtr(ip, hBitsL, 8)] = current2;
284
284
  ip += repLength2;
@@ -297,7 +297,7 @@ _match_stored:
297
297
  U32 const tmpOff = offset_2; offset_2 = offset_1; offset_1 = tmpOff; /* swap offset_2 <=> offset_1 */
298
298
  hashSmall[ZSTD_hashPtr(ip, hBitsS, mls)] = (U32)(ip-base);
299
299
  hashLong[ZSTD_hashPtr(ip, hBitsL, 8)] = (U32)(ip-base);
300
- ZSTD_storeSeq(seqStore, 0, anchor, 0, rLength-MINMATCH);
300
+ ZSTD_storeSeq(seqStore, 0, anchor, iend, 0, rLength-MINMATCH);
301
301
  ip += rLength;
302
302
  anchor = ip;
303
303
  continue; /* faster when present ... (?) */
@@ -411,7 +411,7 @@ static size_t ZSTD_compressBlock_doubleFast_extDict_generic(
411
411
  const BYTE* repMatchEnd = repIndex < prefixStartIndex ? dictEnd : iend;
412
412
  mLength = ZSTD_count_2segments(ip+1+4, repMatch+4, iend, repMatchEnd, prefixStart) + 4;
413
413
  ip++;
414
- ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, 0, mLength-MINMATCH);
414
+ ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, 0, mLength-MINMATCH);
415
415
  } else {
416
416
  if ((matchLongIndex > dictStartIndex) && (MEM_read64(matchLong) == MEM_read64(ip))) {
417
417
  const BYTE* const matchEnd = matchLongIndex < prefixStartIndex ? dictEnd : iend;
@@ -422,7 +422,7 @@ static size_t ZSTD_compressBlock_doubleFast_extDict_generic(
422
422
  while (((ip>anchor) & (matchLong>lowMatchPtr)) && (ip[-1] == matchLong[-1])) { ip--; matchLong--; mLength++; } /* catch up */
423
423
  offset_2 = offset_1;
424
424
  offset_1 = offset;
425
- ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, offset + ZSTD_REP_MOVE, mLength-MINMATCH);
425
+ ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, offset + ZSTD_REP_MOVE, mLength-MINMATCH);
426
426
 
427
427
  } else if ((matchIndex > dictStartIndex) && (MEM_read32(match) == MEM_read32(ip))) {
428
428
  size_t const h3 = ZSTD_hashPtr(ip+1, hBitsL, 8);
@@ -447,7 +447,7 @@ static size_t ZSTD_compressBlock_doubleFast_extDict_generic(
447
447
  }
448
448
  offset_2 = offset_1;
449
449
  offset_1 = offset;
450
- ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, offset + ZSTD_REP_MOVE, mLength-MINMATCH);
450
+ ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, offset + ZSTD_REP_MOVE, mLength-MINMATCH);
451
451
 
452
452
  } else {
453
453
  ip += ((ip-anchor) >> kSearchStrength) + 1;
@@ -479,7 +479,7 @@ static size_t ZSTD_compressBlock_doubleFast_extDict_generic(
479
479
  const BYTE* const repEnd2 = repIndex2 < prefixStartIndex ? dictEnd : iend;
480
480
  size_t const repLength2 = ZSTD_count_2segments(ip+4, repMatch2+4, iend, repEnd2, prefixStart) + 4;
481
481
  U32 const tmpOffset = offset_2; offset_2 = offset_1; offset_1 = tmpOffset; /* swap offset_2 <=> offset_1 */
482
- ZSTD_storeSeq(seqStore, 0, anchor, 0, repLength2-MINMATCH);
482
+ ZSTD_storeSeq(seqStore, 0, anchor, iend, 0, repLength2-MINMATCH);
483
483
  hashSmall[ZSTD_hashPtr(ip, hBitsS, mls)] = current2;
484
484
  hashLong[ZSTD_hashPtr(ip, hBitsL, 8)] = current2;
485
485
  ip += repLength2;
@@ -8,7 +8,7 @@
8
8
  * You may select, at your option, one of the above-listed licenses.
9
9
  */
10
10
 
11
- #include "zstd_compress_internal.h"
11
+ #include "zstd_compress_internal.h" /* ZSTD_hashPtr, ZSTD_count, ZSTD_storeSeq */
12
12
  #include "zstd_fast.h"
13
13
 
14
14
 
@@ -43,8 +43,8 @@ void ZSTD_fillHashTable(ZSTD_matchState_t* ms,
43
43
  }
44
44
 
45
45
 
46
- FORCE_INLINE_TEMPLATE
47
- size_t ZSTD_compressBlock_fast_generic(
46
+ FORCE_INLINE_TEMPLATE size_t
47
+ ZSTD_compressBlock_fast_generic(
48
48
  ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
49
49
  void const* src, size_t srcSize,
50
50
  U32 const mls)
@@ -74,8 +74,7 @@ size_t ZSTD_compressBlock_fast_generic(
74
74
  DEBUGLOG(5, "ZSTD_compressBlock_fast_generic");
75
75
  ip0 += (ip0 == prefixStart);
76
76
  ip1 = ip0 + 1;
77
- {
78
- U32 const maxRep = (U32)(ip0 - prefixStart);
77
+ { U32 const maxRep = (U32)(ip0 - prefixStart);
79
78
  if (offset_2 > maxRep) offsetSaved = offset_2, offset_2 = 0;
80
79
  if (offset_1 > maxRep) offsetSaved = offset_1, offset_1 = 0;
81
80
  }
@@ -118,8 +117,7 @@ size_t ZSTD_compressBlock_fast_generic(
118
117
  match0 = match1;
119
118
  goto _offset;
120
119
  }
121
- {
122
- size_t const step = ((ip0-anchor) >> (kSearchStrength - 1)) + stepSize;
120
+ { size_t const step = ((size_t)(ip0-anchor) >> (kSearchStrength - 1)) + stepSize;
123
121
  assert(step >= 2);
124
122
  ip0 += step;
125
123
  ip1 += step;
@@ -138,7 +136,7 @@ _offset: /* Requires: ip0, match0 */
138
136
  _match: /* Requires: ip0, match0, offcode */
139
137
  /* Count the forward length */
140
138
  mLength += ZSTD_count(ip0+mLength+4, match0+mLength+4, iend) + 4;
141
- ZSTD_storeSeq(seqStore, ip0-anchor, anchor, offcode, mLength-MINMATCH);
139
+ ZSTD_storeSeq(seqStore, (size_t)(ip0-anchor), anchor, iend, offcode, mLength-MINMATCH);
142
140
  /* match found */
143
141
  ip0 += mLength;
144
142
  anchor = ip0;
@@ -150,16 +148,15 @@ _match: /* Requires: ip0, match0, offcode */
150
148
  hashTable[ZSTD_hashPtr(base+current0+2, hlog, mls)] = current0+2; /* here because current+2 could be > iend-8 */
151
149
  hashTable[ZSTD_hashPtr(ip0-2, hlog, mls)] = (U32)(ip0-2-base);
152
150
 
153
- while ( (ip0 <= ilimit)
154
- && ( (offset_2>0)
155
- & (MEM_read32(ip0) == MEM_read32(ip0 - offset_2)) )) {
151
+ while ( ((ip0 <= ilimit) & (offset_2>0)) /* offset_2==0 means offset_2 is invalidated */
152
+ && (MEM_read32(ip0) == MEM_read32(ip0 - offset_2)) ) {
156
153
  /* store sequence */
157
154
  size_t const rLength = ZSTD_count(ip0+4, ip0+4-offset_2, iend) + 4;
158
- U32 const tmpOff = offset_2; offset_2 = offset_1; offset_1 = tmpOff; /* swap offset_2 <=> offset_1 */
155
+ { U32 const tmpOff = offset_2; offset_2 = offset_1; offset_1 = tmpOff; } /* swap offset_2 <=> offset_1 */
159
156
  hashTable[ZSTD_hashPtr(ip0, hlog, mls)] = (U32)(ip0-base);
160
157
  ip0 += rLength;
161
158
  ip1 = ip0 + 1;
162
- ZSTD_storeSeq(seqStore, 0, anchor, 0, rLength-MINMATCH);
159
+ ZSTD_storeSeq(seqStore, 0 /*litLen*/, anchor, iend, 0 /*offCode*/, rLength-MINMATCH);
163
160
  anchor = ip0;
164
161
  continue; /* faster when present (confirmed on gcc-8) ... (?) */
165
162
  }
@@ -179,8 +176,7 @@ size_t ZSTD_compressBlock_fast(
179
176
  ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
180
177
  void const* src, size_t srcSize)
181
178
  {
182
- ZSTD_compressionParameters const* cParams = &ms->cParams;
183
- U32 const mls = cParams->minMatch;
179
+ U32 const mls = ms->cParams.minMatch;
184
180
  assert(ms->dictMatchState == NULL);
185
181
  switch(mls)
186
182
  {
@@ -265,7 +261,7 @@ size_t ZSTD_compressBlock_fast_dictMatchState_generic(
265
261
  const BYTE* const repMatchEnd = repIndex < prefixStartIndex ? dictEnd : iend;
266
262
  mLength = ZSTD_count_2segments(ip+1+4, repMatch+4, iend, repMatchEnd, prefixStart) + 4;
267
263
  ip++;
268
- ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, 0, mLength-MINMATCH);
264
+ ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, 0, mLength-MINMATCH);
269
265
  } else if ( (matchIndex <= prefixStartIndex) ) {
270
266
  size_t const dictHash = ZSTD_hashPtr(ip, dictHLog, mls);
271
267
  U32 const dictMatchIndex = dictHashTable[dictHash];
@@ -285,7 +281,7 @@ size_t ZSTD_compressBlock_fast_dictMatchState_generic(
285
281
  } /* catch up */
286
282
  offset_2 = offset_1;
287
283
  offset_1 = offset;
288
- ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, offset + ZSTD_REP_MOVE, mLength-MINMATCH);
284
+ ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, offset + ZSTD_REP_MOVE, mLength-MINMATCH);
289
285
  }
290
286
  } else if (MEM_read32(match) != MEM_read32(ip)) {
291
287
  /* it's not a match, and we're not going to check the dictionary */
@@ -300,7 +296,7 @@ size_t ZSTD_compressBlock_fast_dictMatchState_generic(
300
296
  && (ip[-1] == match[-1])) { ip--; match--; mLength++; } /* catch up */
301
297
  offset_2 = offset_1;
302
298
  offset_1 = offset;
303
- ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, offset + ZSTD_REP_MOVE, mLength-MINMATCH);
299
+ ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, offset + ZSTD_REP_MOVE, mLength-MINMATCH);
304
300
  }
305
301
 
306
302
  /* match found */
@@ -325,7 +321,7 @@ size_t ZSTD_compressBlock_fast_dictMatchState_generic(
325
321
  const BYTE* const repEnd2 = repIndex2 < prefixStartIndex ? dictEnd : iend;
326
322
  size_t const repLength2 = ZSTD_count_2segments(ip+4, repMatch2+4, iend, repEnd2, prefixStart) + 4;
327
323
  U32 tmpOffset = offset_2; offset_2 = offset_1; offset_1 = tmpOffset; /* swap offset_2 <=> offset_1 */
328
- ZSTD_storeSeq(seqStore, 0, anchor, 0, repLength2-MINMATCH);
324
+ ZSTD_storeSeq(seqStore, 0, anchor, iend, 0, repLength2-MINMATCH);
329
325
  hashTable[ZSTD_hashPtr(ip, hlog, mls)] = current2;
330
326
  ip += repLength2;
331
327
  anchor = ip;
@@ -348,8 +344,7 @@ size_t ZSTD_compressBlock_fast_dictMatchState(
348
344
  ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
349
345
  void const* src, size_t srcSize)
350
346
  {
351
- ZSTD_compressionParameters const* cParams = &ms->cParams;
352
- U32 const mls = cParams->minMatch;
347
+ U32 const mls = ms->cParams.minMatch;
353
348
  assert(ms->dictMatchState != NULL);
354
349
  switch(mls)
355
350
  {
@@ -408,16 +403,17 @@ static size_t ZSTD_compressBlock_fast_extDict_generic(
408
403
  const U32 repIndex = current + 1 - offset_1;
409
404
  const BYTE* const repBase = repIndex < prefixStartIndex ? dictBase : base;
410
405
  const BYTE* const repMatch = repBase + repIndex;
411
- size_t mLength;
412
406
  hashTable[h] = current; /* update hash table */
413
407
  assert(offset_1 <= current +1); /* check repIndex */
414
408
 
415
409
  if ( (((U32)((prefixStartIndex-1) - repIndex) >= 3) /* intentional underflow */ & (repIndex > dictStartIndex))
416
410
  && (MEM_read32(repMatch) == MEM_read32(ip+1)) ) {
417
411
  const BYTE* const repMatchEnd = repIndex < prefixStartIndex ? dictEnd : iend;
418
- mLength = ZSTD_count_2segments(ip+1 +4, repMatch +4, iend, repMatchEnd, prefixStart) + 4;
412
+ size_t const rLength = ZSTD_count_2segments(ip+1 +4, repMatch +4, iend, repMatchEnd, prefixStart) + 4;
419
413
  ip++;
420
- ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, 0, mLength-MINMATCH);
414
+ ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, 0, rLength-MINMATCH);
415
+ ip += rLength;
416
+ anchor = ip;
421
417
  } else {
422
418
  if ( (matchIndex < dictStartIndex) ||
423
419
  (MEM_read32(match) != MEM_read32(ip)) ) {
@@ -427,19 +423,15 @@ static size_t ZSTD_compressBlock_fast_extDict_generic(
427
423
  }
428
424
  { const BYTE* const matchEnd = matchIndex < prefixStartIndex ? dictEnd : iend;
429
425
  const BYTE* const lowMatchPtr = matchIndex < prefixStartIndex ? dictStart : prefixStart;
430
- U32 offset;
431
- mLength = ZSTD_count_2segments(ip+4, match+4, iend, matchEnd, prefixStart) + 4;
426
+ U32 const offset = current - matchIndex;
427
+ size_t mLength = ZSTD_count_2segments(ip+4, match+4, iend, matchEnd, prefixStart) + 4;
432
428
  while (((ip>anchor) & (match>lowMatchPtr)) && (ip[-1] == match[-1])) { ip--; match--; mLength++; } /* catch up */
433
- offset = current - matchIndex;
434
- offset_2 = offset_1;
435
- offset_1 = offset;
436
- ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, offset + ZSTD_REP_MOVE, mLength-MINMATCH);
429
+ offset_2 = offset_1; offset_1 = offset; /* update offset history */
430
+ ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, offset + ZSTD_REP_MOVE, mLength-MINMATCH);
431
+ ip += mLength;
432
+ anchor = ip;
437
433
  } }
438
434
 
439
- /* found a match : store it */
440
- ip += mLength;
441
- anchor = ip;
442
-
443
435
  if (ip <= ilimit) {
444
436
  /* Fill Table */
445
437
  hashTable[ZSTD_hashPtr(base+current+2, hlog, mls)] = current+2;
@@ -448,13 +440,13 @@ static size_t ZSTD_compressBlock_fast_extDict_generic(
448
440
  while (ip <= ilimit) {
449
441
  U32 const current2 = (U32)(ip-base);
450
442
  U32 const repIndex2 = current2 - offset_2;
451
- const BYTE* repMatch2 = repIndex2 < prefixStartIndex ? dictBase + repIndex2 : base + repIndex2;
443
+ const BYTE* const repMatch2 = repIndex2 < prefixStartIndex ? dictBase + repIndex2 : base + repIndex2;
452
444
  if ( (((U32)((prefixStartIndex-1) - repIndex2) >= 3) & (repIndex2 > dictStartIndex)) /* intentional overflow */
453
445
  && (MEM_read32(repMatch2) == MEM_read32(ip)) ) {
454
446
  const BYTE* const repEnd2 = repIndex2 < prefixStartIndex ? dictEnd : iend;
455
447
  size_t const repLength2 = ZSTD_count_2segments(ip+4, repMatch2+4, iend, repEnd2, prefixStart) + 4;
456
- U32 const tmpOffset = offset_2; offset_2 = offset_1; offset_1 = tmpOffset; /* swap offset_2 <=> offset_1 */
457
- ZSTD_storeSeq(seqStore, 0, anchor, 0, repLength2-MINMATCH);
448
+ { U32 const tmpOffset = offset_2; offset_2 = offset_1; offset_1 = tmpOffset; } /* swap offset_2 <=> offset_1 */
449
+ ZSTD_storeSeq(seqStore, 0 /*litlen*/, anchor, iend, 0 /*offcode*/, repLength2-MINMATCH);
458
450
  hashTable[ZSTD_hashPtr(ip, hlog, mls)] = current2;
459
451
  ip += repLength2;
460
452
  anchor = ip;
@@ -476,8 +468,7 @@ size_t ZSTD_compressBlock_fast_extDict(
476
468
  ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
477
469
  void const* src, size_t srcSize)
478
470
  {
479
- ZSTD_compressionParameters const* cParams = &ms->cParams;
480
- U32 const mls = cParams->minMatch;
471
+ U32 const mls = ms->cParams.minMatch;
481
472
  switch(mls)
482
473
  {
483
474
  default: /* includes case 3 */
@@ -810,7 +810,7 @@ ZSTD_compressBlock_lazy_generic(
810
810
  /* store sequence */
811
811
  _storeSequence:
812
812
  { size_t const litLength = start - anchor;
813
- ZSTD_storeSeq(seqStore, litLength, anchor, (U32)offset, matchLength-MINMATCH);
813
+ ZSTD_storeSeq(seqStore, litLength, anchor, iend, (U32)offset, matchLength-MINMATCH);
814
814
  anchor = ip = start + matchLength;
815
815
  }
816
816
 
@@ -828,7 +828,7 @@ _storeSequence:
828
828
  const BYTE* const repEnd2 = repIndex < prefixLowestIndex ? dictEnd : iend;
829
829
  matchLength = ZSTD_count_2segments(ip+4, repMatch+4, iend, repEnd2, prefixLowest) + 4;
830
830
  offset = offset_2; offset_2 = offset_1; offset_1 = (U32)offset; /* swap offset_2 <=> offset_1 */
831
- ZSTD_storeSeq(seqStore, 0, anchor, 0, matchLength-MINMATCH);
831
+ ZSTD_storeSeq(seqStore, 0, anchor, iend, 0, matchLength-MINMATCH);
832
832
  ip += matchLength;
833
833
  anchor = ip;
834
834
  continue;
@@ -843,7 +843,7 @@ _storeSequence:
843
843
  /* store sequence */
844
844
  matchLength = ZSTD_count(ip+4, ip+4-offset_2, iend) + 4;
845
845
  offset = offset_2; offset_2 = offset_1; offset_1 = (U32)offset; /* swap repcodes */
846
- ZSTD_storeSeq(seqStore, 0, anchor, 0, matchLength-MINMATCH);
846
+ ZSTD_storeSeq(seqStore, 0, anchor, iend, 0, matchLength-MINMATCH);
847
847
  ip += matchLength;
848
848
  anchor = ip;
849
849
  continue; /* faster when present ... (?) */
@@ -1051,7 +1051,7 @@ size_t ZSTD_compressBlock_lazy_extDict_generic(
1051
1051
  /* store sequence */
1052
1052
  _storeSequence:
1053
1053
  { size_t const litLength = start - anchor;
1054
- ZSTD_storeSeq(seqStore, litLength, anchor, (U32)offset, matchLength-MINMATCH);
1054
+ ZSTD_storeSeq(seqStore, litLength, anchor, iend, (U32)offset, matchLength-MINMATCH);
1055
1055
  anchor = ip = start + matchLength;
1056
1056
  }
1057
1057
 
@@ -1066,7 +1066,7 @@ _storeSequence:
1066
1066
  const BYTE* const repEnd = repIndex < dictLimit ? dictEnd : iend;
1067
1067
  matchLength = ZSTD_count_2segments(ip+4, repMatch+4, iend, repEnd, prefixStart) + 4;
1068
1068
  offset = offset_2; offset_2 = offset_1; offset_1 = (U32)offset; /* swap offset history */
1069
- ZSTD_storeSeq(seqStore, 0, anchor, 0, matchLength-MINMATCH);
1069
+ ZSTD_storeSeq(seqStore, 0, anchor, iend, 0, matchLength-MINMATCH);
1070
1070
  ip += matchLength;
1071
1071
  anchor = ip;
1072
1072
  continue; /* faster when present ... (?) */
@@ -49,9 +49,9 @@ size_t ZSTD_ldm_getTableSize(ldmParams_t params)
49
49
  {
50
50
  size_t const ldmHSize = ((size_t)1) << params.hashLog;
51
51
  size_t const ldmBucketSizeLog = MIN(params.bucketSizeLog, params.hashLog);
52
- size_t const ldmBucketSize =
53
- ((size_t)1) << (params.hashLog - ldmBucketSizeLog);
54
- size_t const totalSize = ldmBucketSize + ldmHSize * sizeof(ldmEntry_t);
52
+ size_t const ldmBucketSize = ((size_t)1) << (params.hashLog - ldmBucketSizeLog);
53
+ size_t const totalSize = ZSTD_cwksp_alloc_size(ldmBucketSize)
54
+ + ZSTD_cwksp_alloc_size(ldmHSize * sizeof(ldmEntry_t));
55
55
  return params.enableLdm ? totalSize : 0;
56
56
  }
57
57
 
@@ -583,7 +583,7 @@ size_t ZSTD_ldm_blockCompress(rawSeqStore_t* rawSeqStore,
583
583
  rep[i] = rep[i-1];
584
584
  rep[0] = sequence.offset;
585
585
  /* Store the sequence */
586
- ZSTD_storeSeq(seqStore, newLitLength, ip - newLitLength,
586
+ ZSTD_storeSeq(seqStore, newLitLength, ip - newLitLength, iend,
587
587
  sequence.offset + ZSTD_REP_MOVE,
588
588
  sequence.matchLength - MINMATCH);
589
589
  ip += sequence.matchLength;
@@ -1098,7 +1098,7 @@ _shortestPath: /* cur, last_pos, best_mlen, best_off have to be set */
1098
1098
 
1099
1099
  assert(anchor + llen <= iend);
1100
1100
  ZSTD_updateStats(optStatePtr, llen, anchor, offCode, mlen);
1101
- ZSTD_storeSeq(seqStore, llen, anchor, offCode, mlen-MINMATCH);
1101
+ ZSTD_storeSeq(seqStore, llen, anchor, iend, offCode, mlen-MINMATCH);
1102
1102
  anchor += advance;
1103
1103
  ip = anchor;
1104
1104
  } }
@@ -668,7 +668,7 @@ static void ZSTDMT_compressionJob(void* jobDescription)
668
668
 
669
669
  /* init */
670
670
  if (job->cdict) {
671
- size_t const initError = ZSTD_compressBegin_advanced_internal(cctx, NULL, 0, ZSTD_dct_auto, ZSTD_dtlm_fast, job->cdict, jobParams, job->fullFrameSize);
671
+ size_t const initError = ZSTD_compressBegin_advanced_internal(cctx, NULL, 0, ZSTD_dct_auto, ZSTD_dtlm_fast, job->cdict, &jobParams, job->fullFrameSize);
672
672
  assert(job->firstJob); /* only allowed for first job */
673
673
  if (ZSTD_isError(initError)) JOB_ERROR(initError);
674
674
  } else { /* srcStart points at reloaded section */
@@ -680,7 +680,7 @@ static void ZSTDMT_compressionJob(void* jobDescription)
680
680
  job->prefix.start, job->prefix.size, ZSTD_dct_rawContent, /* load dictionary in "content-only" mode (no header analysis) */
681
681
  ZSTD_dtlm_fast,
682
682
  NULL, /*cdict*/
683
- jobParams, pledgedSrcSize);
683
+ &jobParams, pledgedSrcSize);
684
684
  if (ZSTD_isError(initError)) JOB_ERROR(initError);
685
685
  } }
686
686
 
@@ -927,12 +927,18 @@ static void ZSTDMT_releaseAllJobResources(ZSTDMT_CCtx* mtctx)
927
927
  unsigned jobID;
928
928
  DEBUGLOG(3, "ZSTDMT_releaseAllJobResources");
929
929
  for (jobID=0; jobID <= mtctx->jobIDMask; jobID++) {
930
+ /* Copy the mutex/cond out */
931
+ ZSTD_pthread_mutex_t const mutex = mtctx->jobs[jobID].job_mutex;
932
+ ZSTD_pthread_cond_t const cond = mtctx->jobs[jobID].job_cond;
933
+
930
934
  DEBUGLOG(4, "job%02u: release dst address %08X", jobID, (U32)(size_t)mtctx->jobs[jobID].dstBuff.start);
931
935
  ZSTDMT_releaseBuffer(mtctx->bufPool, mtctx->jobs[jobID].dstBuff);
932
- mtctx->jobs[jobID].dstBuff = g_nullBuffer;
933
- mtctx->jobs[jobID].cSize = 0;
936
+
937
+ /* Clear the job description, but keep the mutex/cond */
938
+ memset(&mtctx->jobs[jobID], 0, sizeof(mtctx->jobs[jobID]));
939
+ mtctx->jobs[jobID].job_mutex = mutex;
940
+ mtctx->jobs[jobID].job_cond = cond;
934
941
  }
935
- memset(mtctx->jobs, 0, (mtctx->jobIDMask+1)*sizeof(ZSTDMT_jobDescription));
936
942
  mtctx->inBuff.buffer = g_nullBuffer;
937
943
  mtctx->inBuff.filled = 0;
938
944
  mtctx->allJobsCompleted = 1;
@@ -1028,9 +1034,9 @@ size_t ZSTDMT_getMTCtxParameter(ZSTDMT_CCtx* mtctx, ZSTDMT_parameter parameter,
1028
1034
 
1029
1035
  /* Sets parameters relevant to the compression job,
1030
1036
  * initializing others to default values. */
1031
- static ZSTD_CCtx_params ZSTDMT_initJobCCtxParams(ZSTD_CCtx_params const params)
1037
+ static ZSTD_CCtx_params ZSTDMT_initJobCCtxParams(const ZSTD_CCtx_params* params)
1032
1038
  {
1033
- ZSTD_CCtx_params jobParams = params;
1039
+ ZSTD_CCtx_params jobParams = *params;
1034
1040
  /* Clear parameters related to multithreading */
1035
1041
  jobParams.forceWindow = 0;
1036
1042
  jobParams.nbWorkers = 0;
@@ -1151,16 +1157,16 @@ size_t ZSTDMT_toFlushNow(ZSTDMT_CCtx* mtctx)
1151
1157
  /* ===== Multi-threaded compression ===== */
1152
1158
  /* ------------------------------------------ */
1153
1159
 
1154
- static unsigned ZSTDMT_computeTargetJobLog(ZSTD_CCtx_params const params)
1160
+ static unsigned ZSTDMT_computeTargetJobLog(const ZSTD_CCtx_params* params)
1155
1161
  {
1156
1162
  unsigned jobLog;
1157
- if (params.ldmParams.enableLdm) {
1163
+ if (params->ldmParams.enableLdm) {
1158
1164
  /* In Long Range Mode, the windowLog is typically oversized.
1159
1165
  * In which case, it's preferable to determine the jobSize
1160
1166
  * based on chainLog instead. */
1161
- jobLog = MAX(21, params.cParams.chainLog + 4);
1167
+ jobLog = MAX(21, params->cParams.chainLog + 4);
1162
1168
  } else {
1163
- jobLog = MAX(20, params.cParams.windowLog + 2);
1169
+ jobLog = MAX(20, params->cParams.windowLog + 2);
1164
1170
  }
1165
1171
  return MIN(jobLog, (unsigned)ZSTDMT_JOBLOG_MAX);
1166
1172
  }
@@ -1193,27 +1199,27 @@ static int ZSTDMT_overlapLog(int ovlog, ZSTD_strategy strat)
1193
1199
  return ovlog;
1194
1200
  }
1195
1201
 
1196
- static size_t ZSTDMT_computeOverlapSize(ZSTD_CCtx_params const params)
1202
+ static size_t ZSTDMT_computeOverlapSize(const ZSTD_CCtx_params* params)
1197
1203
  {
1198
- int const overlapRLog = 9 - ZSTDMT_overlapLog(params.overlapLog, params.cParams.strategy);
1199
- int ovLog = (overlapRLog >= 8) ? 0 : (params.cParams.windowLog - overlapRLog);
1204
+ int const overlapRLog = 9 - ZSTDMT_overlapLog(params->overlapLog, params->cParams.strategy);
1205
+ int ovLog = (overlapRLog >= 8) ? 0 : (params->cParams.windowLog - overlapRLog);
1200
1206
  assert(0 <= overlapRLog && overlapRLog <= 8);
1201
- if (params.ldmParams.enableLdm) {
1207
+ if (params->ldmParams.enableLdm) {
1202
1208
  /* In Long Range Mode, the windowLog is typically oversized.
1203
1209
  * In which case, it's preferable to determine the jobSize
1204
1210
  * based on chainLog instead.
1205
1211
  * Then, ovLog becomes a fraction of the jobSize, rather than windowSize */
1206
- ovLog = MIN(params.cParams.windowLog, ZSTDMT_computeTargetJobLog(params) - 2)
1212
+ ovLog = MIN(params->cParams.windowLog, ZSTDMT_computeTargetJobLog(params) - 2)
1207
1213
  - overlapRLog;
1208
1214
  }
1209
1215
  assert(0 <= ovLog && ovLog <= ZSTD_WINDOWLOG_MAX);
1210
- DEBUGLOG(4, "overlapLog : %i", params.overlapLog);
1216
+ DEBUGLOG(4, "overlapLog : %i", params->overlapLog);
1211
1217
  DEBUGLOG(4, "overlap size : %i", 1 << ovLog);
1212
1218
  return (ovLog==0) ? 0 : (size_t)1 << ovLog;
1213
1219
  }
1214
1220
 
1215
1221
  static unsigned
1216
- ZSTDMT_computeNbJobs(ZSTD_CCtx_params params, size_t srcSize, unsigned nbWorkers)
1222
+ ZSTDMT_computeNbJobs(const ZSTD_CCtx_params* params, size_t srcSize, unsigned nbWorkers)
1217
1223
  {
1218
1224
  assert(nbWorkers>0);
1219
1225
  { size_t const jobSizeTarget = (size_t)1 << ZSTDMT_computeTargetJobLog(params);
@@ -1236,9 +1242,9 @@ static size_t ZSTDMT_compress_advanced_internal(
1236
1242
  const ZSTD_CDict* cdict,
1237
1243
  ZSTD_CCtx_params params)
1238
1244
  {
1239
- ZSTD_CCtx_params const jobParams = ZSTDMT_initJobCCtxParams(params);
1240
- size_t const overlapSize = ZSTDMT_computeOverlapSize(params);
1241
- unsigned const nbJobs = ZSTDMT_computeNbJobs(params, srcSize, params.nbWorkers);
1245
+ ZSTD_CCtx_params const jobParams = ZSTDMT_initJobCCtxParams(&params);
1246
+ size_t const overlapSize = ZSTDMT_computeOverlapSize(&params);
1247
+ unsigned const nbJobs = ZSTDMT_computeNbJobs(&params, srcSize, params.nbWorkers);
1242
1248
  size_t const proposedJobSize = (srcSize + (nbJobs-1)) / nbJobs;
1243
1249
  size_t const avgJobSize = (((proposedJobSize-1) & 0x1FFFF) < 0x7FFF) ? proposedJobSize + 0xFFFF : proposedJobSize; /* avoid too small last block */
1244
1250
  const char* const srcStart = (const char*)src;
@@ -1256,7 +1262,7 @@ static size_t ZSTDMT_compress_advanced_internal(
1256
1262
  ZSTD_CCtx* const cctx = mtctx->cctxPool->cctx[0];
1257
1263
  DEBUGLOG(4, "ZSTDMT_compress_advanced_internal: fallback to single-thread mode");
1258
1264
  if (cdict) return ZSTD_compress_usingCDict_advanced(cctx, dst, dstCapacity, src, srcSize, cdict, jobParams.fParams);
1259
- return ZSTD_compress_advanced_internal(cctx, dst, dstCapacity, src, srcSize, NULL, 0, jobParams);
1265
+ return ZSTD_compress_advanced_internal(cctx, dst, dstCapacity, src, srcSize, NULL, 0, &jobParams);
1260
1266
  }
1261
1267
 
1262
1268
  assert(avgJobSize >= 256 KB); /* condition for ZSTD_compressBound(A) + ZSTD_compressBound(B) <= ZSTD_compressBound(A+B), required to compress directly into Dst (no additional buffer) */
@@ -1404,12 +1410,12 @@ size_t ZSTDMT_initCStream_internal(
1404
1410
 
1405
1411
  mtctx->singleBlockingThread = (pledgedSrcSize <= ZSTDMT_JOBSIZE_MIN); /* do not trigger multi-threading when srcSize is too small */
1406
1412
  if (mtctx->singleBlockingThread) {
1407
- ZSTD_CCtx_params const singleThreadParams = ZSTDMT_initJobCCtxParams(params);
1413
+ ZSTD_CCtx_params const singleThreadParams = ZSTDMT_initJobCCtxParams(&params);
1408
1414
  DEBUGLOG(5, "ZSTDMT_initCStream_internal: switch to single blocking thread mode");
1409
1415
  assert(singleThreadParams.nbWorkers == 0);
1410
1416
  return ZSTD_initCStream_internal(mtctx->cctxPool->cctx[0],
1411
1417
  dict, dictSize, cdict,
1412
- singleThreadParams, pledgedSrcSize);
1418
+ &singleThreadParams, pledgedSrcSize);
1413
1419
  }
1414
1420
 
1415
1421
  DEBUGLOG(4, "ZSTDMT_initCStream_internal: %u workers", params.nbWorkers);
@@ -1435,11 +1441,11 @@ size_t ZSTDMT_initCStream_internal(
1435
1441
  mtctx->cdict = cdict;
1436
1442
  }
1437
1443
 
1438
- mtctx->targetPrefixSize = ZSTDMT_computeOverlapSize(params);
1444
+ mtctx->targetPrefixSize = ZSTDMT_computeOverlapSize(&params);
1439
1445
  DEBUGLOG(4, "overlapLog=%i => %u KB", params.overlapLog, (U32)(mtctx->targetPrefixSize>>10));
1440
1446
  mtctx->targetSectionSize = params.jobSize;
1441
1447
  if (mtctx->targetSectionSize == 0) {
1442
- mtctx->targetSectionSize = 1ULL << ZSTDMT_computeTargetJobLog(params);
1448
+ mtctx->targetSectionSize = 1ULL << ZSTDMT_computeTargetJobLog(&params);
1443
1449
  }
1444
1450
  assert(mtctx->targetSectionSize <= (size_t)ZSTDMT_JOBSIZE_MAX);
1445
1451