zstdlib 0.2.0-x64-mingw32 → 0.3.0-x64-mingw32

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (68) hide show
  1. checksums.yaml +4 -4
  2. data/CHANGES.md +5 -0
  3. data/ext/zstdlib/extconf.rb +1 -1
  4. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/bitstream.h +0 -0
  5. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/compiler.h +7 -0
  6. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/cpu.h +0 -0
  7. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/debug.c +0 -0
  8. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/debug.h +0 -0
  9. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/entropy_common.c +0 -0
  10. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/error_private.c +0 -0
  11. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/error_private.h +0 -0
  12. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/fse.h +0 -0
  13. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/fse_decompress.c +0 -0
  14. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/huf.h +0 -0
  15. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/mem.h +0 -0
  16. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/pool.c +0 -0
  17. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/pool.h +0 -0
  18. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/threading.c +0 -0
  19. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/threading.h +0 -0
  20. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/xxhash.c +0 -0
  21. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/xxhash.h +0 -0
  22. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/zstd_common.c +0 -0
  23. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/zstd_errors.h +0 -0
  24. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/zstd_internal.h +58 -6
  25. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/fse_compress.c +0 -0
  26. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/hist.c +0 -0
  27. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/hist.h +0 -0
  28. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/huf_compress.c +0 -0
  29. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/zstd_compress.c +178 -691
  30. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/zstd_compress_internal.h +98 -30
  31. data/ext/zstdlib/zstd-1.4.2/lib/compress/zstd_compress_literals.c +149 -0
  32. data/ext/zstdlib/zstd-1.4.2/lib/compress/zstd_compress_literals.h +29 -0
  33. data/ext/zstdlib/zstd-1.4.2/lib/compress/zstd_compress_sequences.c +415 -0
  34. data/ext/zstdlib/zstd-1.4.2/lib/compress/zstd_compress_sequences.h +47 -0
  35. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/zstd_double_fast.c +56 -36
  36. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/zstd_double_fast.h +0 -0
  37. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/zstd_fast.c +35 -14
  38. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/zstd_fast.h +0 -0
  39. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/zstd_lazy.c +10 -5
  40. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/zstd_lazy.h +0 -0
  41. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/zstd_ldm.c +1 -1
  42. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/zstd_ldm.h +0 -0
  43. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/zstd_opt.c +45 -32
  44. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/zstd_opt.h +0 -0
  45. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/zstdmt_compress.c +18 -7
  46. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/zstdmt_compress.h +1 -0
  47. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/decompress/huf_decompress.c +0 -0
  48. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/decompress/zstd_ddict.c +0 -0
  49. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/decompress/zstd_ddict.h +0 -0
  50. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/decompress/zstd_decompress.c +14 -9
  51. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/decompress/zstd_decompress_block.c +20 -9
  52. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/decompress/zstd_decompress_block.h +0 -0
  53. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/decompress/zstd_decompress_internal.h +0 -0
  54. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/zstd.h +53 -21
  55. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/zlibWrapper/gzclose.c +0 -0
  56. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/zlibWrapper/gzcompatibility.h +0 -0
  57. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/zlibWrapper/gzguts.h +0 -0
  58. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/zlibWrapper/gzlib.c +0 -0
  59. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/zlibWrapper/gzread.c +0 -0
  60. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/zlibWrapper/gzwrite.c +0 -0
  61. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/zlibWrapper/zstd_zlibwrapper.c +0 -0
  62. data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/zlibWrapper/zstd_zlibwrapper.h +0 -0
  63. data/lib/2.2/zstdlib.so +0 -0
  64. data/lib/2.3/zstdlib.so +0 -0
  65. data/lib/2.4/zstdlib.so +0 -0
  66. data/lib/2.5/zstdlib.so +0 -0
  67. data/lib/2.6/zstdlib.so +0 -0
  68. metadata +61 -57
@@ -447,7 +447,7 @@ size_t ZSTD_ldm_generateSequences(
447
447
  if (ZSTD_window_needOverflowCorrection(ldmState->window, chunkEnd)) {
448
448
  U32 const ldmHSize = 1U << params->hashLog;
449
449
  U32 const correction = ZSTD_window_correctOverflow(
450
- &ldmState->window, /* cycleLog */ 0, maxDist, src);
450
+ &ldmState->window, /* cycleLog */ 0, maxDist, chunkStart);
451
451
  ZSTD_ldm_reduceTable(ldmState->hashTable, ldmHSize, correction);
452
452
  }
453
453
  /* 2. We enforce the maximum offset allowed.
@@ -255,13 +255,13 @@ static U32 ZSTD_litLengthPrice(U32 const litLength, const optState_t* const optP
255
255
  * to provide a cost which is directly comparable to a match ending at same position */
256
256
  static int ZSTD_litLengthContribution(U32 const litLength, const optState_t* const optPtr, int optLevel)
257
257
  {
258
- if (optPtr->priceType >= zop_predef) return WEIGHT(litLength, optLevel);
258
+ if (optPtr->priceType >= zop_predef) return (int)WEIGHT(litLength, optLevel);
259
259
 
260
260
  /* dynamic statistics */
261
261
  { U32 const llCode = ZSTD_LLcode(litLength);
262
- int const contribution = (LL_bits[llCode] * BITCOST_MULTIPLIER)
263
- + WEIGHT(optPtr->litLengthFreq[0], optLevel) /* note: log2litLengthSum cancel out */
264
- - WEIGHT(optPtr->litLengthFreq[llCode], optLevel);
262
+ int const contribution = (int)(LL_bits[llCode] * BITCOST_MULTIPLIER)
263
+ + (int)WEIGHT(optPtr->litLengthFreq[0], optLevel) /* note: log2litLengthSum cancel out */
264
+ - (int)WEIGHT(optPtr->litLengthFreq[llCode], optLevel);
265
265
  #if 1
266
266
  return contribution;
267
267
  #else
@@ -278,7 +278,7 @@ static int ZSTD_literalsContribution(const BYTE* const literals, U32 const litLe
278
278
  const optState_t* const optPtr,
279
279
  int optLevel)
280
280
  {
281
- int const contribution = ZSTD_rawLiteralsCost(literals, litLength, optPtr, optLevel)
281
+ int const contribution = (int)ZSTD_rawLiteralsCost(literals, litLength, optPtr, optLevel)
282
282
  + ZSTD_litLengthContribution(litLength, optPtr, optLevel);
283
283
  return contribution;
284
284
  }
@@ -372,13 +372,15 @@ MEM_STATIC U32 ZSTD_readMINMATCH(const void* memPtr, U32 length)
372
372
 
373
373
  /* Update hashTable3 up to ip (excluded)
374
374
  Assumption : always within prefix (i.e. not within extDict) */
375
- static U32 ZSTD_insertAndFindFirstIndexHash3 (ZSTD_matchState_t* ms, const BYTE* const ip)
375
+ static U32 ZSTD_insertAndFindFirstIndexHash3 (ZSTD_matchState_t* ms,
376
+ U32* nextToUpdate3,
377
+ const BYTE* const ip)
376
378
  {
377
379
  U32* const hashTable3 = ms->hashTable3;
378
380
  U32 const hashLog3 = ms->hashLog3;
379
381
  const BYTE* const base = ms->window.base;
380
- U32 idx = ms->nextToUpdate3;
381
- U32 const target = ms->nextToUpdate3 = (U32)(ip - base);
382
+ U32 idx = *nextToUpdate3;
383
+ U32 const target = (U32)(ip - base);
382
384
  size_t const hash3 = ZSTD_hash3Ptr(ip, hashLog3);
383
385
  assert(hashLog3 > 0);
384
386
 
@@ -387,6 +389,7 @@ static U32 ZSTD_insertAndFindFirstIndexHash3 (ZSTD_matchState_t* ms, const BYTE*
387
389
  idx++;
388
390
  }
389
391
 
392
+ *nextToUpdate3 = target;
390
393
  return hashTable3[hash3];
391
394
  }
392
395
 
@@ -503,9 +506,11 @@ static U32 ZSTD_insertBt1(
503
506
  } }
504
507
 
505
508
  *smallerPtr = *largerPtr = 0;
506
- if (bestLength > 384) return MIN(192, (U32)(bestLength - 384)); /* speed optimization */
507
- assert(matchEndIdx > current + 8);
508
- return matchEndIdx - (current + 8);
509
+ { U32 positions = 0;
510
+ if (bestLength > 384) positions = MIN(192, (U32)(bestLength - 384)); /* speed optimization */
511
+ assert(matchEndIdx > current + 8);
512
+ return MAX(positions, matchEndIdx - (current + 8));
513
+ }
509
514
  }
510
515
 
511
516
  FORCE_INLINE_TEMPLATE
@@ -520,8 +525,13 @@ void ZSTD_updateTree_internal(
520
525
  DEBUGLOG(6, "ZSTD_updateTree_internal, from %u to %u (dictMode:%u)",
521
526
  idx, target, dictMode);
522
527
 
523
- while(idx < target)
524
- idx += ZSTD_insertBt1(ms, base+idx, iend, mls, dictMode == ZSTD_extDict);
528
+ while(idx < target) {
529
+ U32 const forward = ZSTD_insertBt1(ms, base+idx, iend, mls, dictMode == ZSTD_extDict);
530
+ assert(idx < (U32)(idx + forward));
531
+ idx += forward;
532
+ }
533
+ assert((size_t)(ip - base) <= (size_t)(U32)(-1));
534
+ assert((size_t)(iend - base) <= (size_t)(U32)(-1));
525
535
  ms->nextToUpdate = target;
526
536
  }
527
537
 
@@ -531,16 +541,18 @@ void ZSTD_updateTree(ZSTD_matchState_t* ms, const BYTE* ip, const BYTE* iend) {
531
541
 
532
542
  FORCE_INLINE_TEMPLATE
533
543
  U32 ZSTD_insertBtAndGetAllMatches (
544
+ ZSTD_match_t* matches, /* store result (found matches) in this table (presumed large enough) */
534
545
  ZSTD_matchState_t* ms,
546
+ U32* nextToUpdate3,
535
547
  const BYTE* const ip, const BYTE* const iLimit, const ZSTD_dictMode_e dictMode,
536
- U32 rep[ZSTD_REP_NUM],
548
+ const U32 rep[ZSTD_REP_NUM],
537
549
  U32 const ll0, /* tells if associated literal length is 0 or not. This value must be 0 or 1 */
538
- ZSTD_match_t* matches,
539
550
  const U32 lengthToBeat,
540
551
  U32 const mls /* template */)
541
552
  {
542
553
  const ZSTD_compressionParameters* const cParams = &ms->cParams;
543
554
  U32 const sufficient_len = MIN(cParams->targetLength, ZSTD_OPT_NUM -1);
555
+ U32 const maxDistance = 1U << cParams->windowLog;
544
556
  const BYTE* const base = ms->window.base;
545
557
  U32 const current = (U32)(ip-base);
546
558
  U32 const hashLog = cParams->hashLog;
@@ -556,8 +568,9 @@ U32 ZSTD_insertBtAndGetAllMatches (
556
568
  U32 const dictLimit = ms->window.dictLimit;
557
569
  const BYTE* const dictEnd = dictBase + dictLimit;
558
570
  const BYTE* const prefixStart = base + dictLimit;
559
- U32 const btLow = btMask >= current ? 0 : current - btMask;
560
- U32 const windowLow = ms->window.lowLimit;
571
+ U32 const btLow = (btMask >= current) ? 0 : current - btMask;
572
+ U32 const windowValid = ms->window.lowLimit;
573
+ U32 const windowLow = ((current - windowValid) > maxDistance) ? current - maxDistance : windowValid;
561
574
  U32 const matchLow = windowLow ? windowLow : 1;
562
575
  U32* smallerPtr = bt + 2*(current&btMask);
563
576
  U32* largerPtr = bt + 2*(current&btMask) + 1;
@@ -627,7 +640,7 @@ U32 ZSTD_insertBtAndGetAllMatches (
627
640
 
628
641
  /* HC3 match finder */
629
642
  if ((mls == 3) /*static*/ && (bestLength < mls)) {
630
- U32 const matchIndex3 = ZSTD_insertAndFindFirstIndexHash3(ms, ip);
643
+ U32 const matchIndex3 = ZSTD_insertAndFindFirstIndexHash3(ms, nextToUpdate3, ip);
631
644
  if ((matchIndex3 >= matchLow)
632
645
  & (current - matchIndex3 < (1<<18)) /*heuristic : longer distance likely too expensive*/ ) {
633
646
  size_t mlen;
@@ -653,9 +666,7 @@ U32 ZSTD_insertBtAndGetAllMatches (
653
666
  (ip+mlen == iLimit) ) { /* best possible length */
654
667
  ms->nextToUpdate = current+1; /* skip insertion */
655
668
  return 1;
656
- }
657
- }
658
- }
669
+ } } }
659
670
  /* no dictMatchState lookup: dicts don't have a populated HC3 table */
660
671
  }
661
672
 
@@ -760,10 +771,13 @@ U32 ZSTD_insertBtAndGetAllMatches (
760
771
 
761
772
 
762
773
  FORCE_INLINE_TEMPLATE U32 ZSTD_BtGetAllMatches (
774
+ ZSTD_match_t* matches, /* store result (match found, increasing size) in this table */
763
775
  ZSTD_matchState_t* ms,
776
+ U32* nextToUpdate3,
764
777
  const BYTE* ip, const BYTE* const iHighLimit, const ZSTD_dictMode_e dictMode,
765
- U32 rep[ZSTD_REP_NUM], U32 const ll0,
766
- ZSTD_match_t* matches, U32 const lengthToBeat)
778
+ const U32 rep[ZSTD_REP_NUM],
779
+ U32 const ll0,
780
+ U32 const lengthToBeat)
767
781
  {
768
782
  const ZSTD_compressionParameters* const cParams = &ms->cParams;
769
783
  U32 const matchLengthSearch = cParams->minMatch;
@@ -772,12 +786,12 @@ FORCE_INLINE_TEMPLATE U32 ZSTD_BtGetAllMatches (
772
786
  ZSTD_updateTree_internal(ms, ip, iHighLimit, matchLengthSearch, dictMode);
773
787
  switch(matchLengthSearch)
774
788
  {
775
- case 3 : return ZSTD_insertBtAndGetAllMatches(ms, ip, iHighLimit, dictMode, rep, ll0, matches, lengthToBeat, 3);
789
+ case 3 : return ZSTD_insertBtAndGetAllMatches(matches, ms, nextToUpdate3, ip, iHighLimit, dictMode, rep, ll0, lengthToBeat, 3);
776
790
  default :
777
- case 4 : return ZSTD_insertBtAndGetAllMatches(ms, ip, iHighLimit, dictMode, rep, ll0, matches, lengthToBeat, 4);
778
- case 5 : return ZSTD_insertBtAndGetAllMatches(ms, ip, iHighLimit, dictMode, rep, ll0, matches, lengthToBeat, 5);
791
+ case 4 : return ZSTD_insertBtAndGetAllMatches(matches, ms, nextToUpdate3, ip, iHighLimit, dictMode, rep, ll0, lengthToBeat, 4);
792
+ case 5 : return ZSTD_insertBtAndGetAllMatches(matches, ms, nextToUpdate3, ip, iHighLimit, dictMode, rep, ll0, lengthToBeat, 5);
779
793
  case 7 :
780
- case 6 : return ZSTD_insertBtAndGetAllMatches(ms, ip, iHighLimit, dictMode, rep, ll0, matches, lengthToBeat, 6);
794
+ case 6 : return ZSTD_insertBtAndGetAllMatches(matches, ms, nextToUpdate3, ip, iHighLimit, dictMode, rep, ll0, lengthToBeat, 6);
781
795
  }
782
796
  }
783
797
 
@@ -853,6 +867,7 @@ ZSTD_compressBlock_opt_generic(ZSTD_matchState_t* ms,
853
867
 
854
868
  U32 const sufficient_len = MIN(cParams->targetLength, ZSTD_OPT_NUM -1);
855
869
  U32 const minMatch = (cParams->minMatch == 3) ? 3 : 4;
870
+ U32 nextToUpdate3 = ms->nextToUpdate;
856
871
 
857
872
  ZSTD_optimal_t* const opt = optStatePtr->priceTable;
858
873
  ZSTD_match_t* const matches = optStatePtr->matchTable;
@@ -862,7 +877,6 @@ ZSTD_compressBlock_opt_generic(ZSTD_matchState_t* ms,
862
877
  DEBUGLOG(5, "ZSTD_compressBlock_opt_generic: current=%u, prefix=%u, nextToUpdate=%u",
863
878
  (U32)(ip - base), ms->window.dictLimit, ms->nextToUpdate);
864
879
  assert(optLevel <= 2);
865
- ms->nextToUpdate3 = ms->nextToUpdate;
866
880
  ZSTD_rescaleFreqs(optStatePtr, (const BYTE*)src, srcSize, optLevel);
867
881
  ip += (ip==prefixStart);
868
882
 
@@ -873,7 +887,7 @@ ZSTD_compressBlock_opt_generic(ZSTD_matchState_t* ms,
873
887
  /* find first match */
874
888
  { U32 const litlen = (U32)(ip - anchor);
875
889
  U32 const ll0 = !litlen;
876
- U32 const nbMatches = ZSTD_BtGetAllMatches(ms, ip, iend, dictMode, rep, ll0, matches, minMatch);
890
+ U32 const nbMatches = ZSTD_BtGetAllMatches(matches, ms, &nextToUpdate3, ip, iend, dictMode, rep, ll0, minMatch);
877
891
  if (!nbMatches) { ip++; continue; }
878
892
 
879
893
  /* initialize opt[0] */
@@ -970,7 +984,7 @@ ZSTD_compressBlock_opt_generic(ZSTD_matchState_t* ms,
970
984
  U32 const litlen = (opt[cur].mlen == 0) ? opt[cur].litlen : 0;
971
985
  U32 const previousPrice = opt[cur].price;
972
986
  U32 const basePrice = previousPrice + ZSTD_litLengthPrice(0, optStatePtr, optLevel);
973
- U32 const nbMatches = ZSTD_BtGetAllMatches(ms, inr, iend, dictMode, opt[cur].rep, ll0, matches, minMatch);
987
+ U32 const nbMatches = ZSTD_BtGetAllMatches(matches, ms, &nextToUpdate3, inr, iend, dictMode, opt[cur].rep, ll0, minMatch);
974
988
  U32 matchNb;
975
989
  if (!nbMatches) {
976
990
  DEBUGLOG(7, "rPos:%u : no match found", cur);
@@ -1094,7 +1108,7 @@ _shortestPath: /* cur, last_pos, best_mlen, best_off have to be set */
1094
1108
  } /* while (ip < ilimit) */
1095
1109
 
1096
1110
  /* Return the last literals size */
1097
- return iend - anchor;
1111
+ return (size_t)(iend - anchor);
1098
1112
  }
1099
1113
 
1100
1114
 
@@ -1158,7 +1172,6 @@ ZSTD_initStats_ultra(ZSTD_matchState_t* ms,
1158
1172
  ms->window.dictLimit += (U32)srcSize;
1159
1173
  ms->window.lowLimit = ms->window.dictLimit;
1160
1174
  ms->nextToUpdate = ms->window.dictLimit;
1161
- ms->nextToUpdate3 = ms->window.dictLimit;
1162
1175
 
1163
1176
  /* re-inforce weight of collected statistics */
1164
1177
  ZSTD_upscaleStats(&ms->opt);
@@ -1129,9 +1129,14 @@ size_t ZSTDMT_toFlushNow(ZSTDMT_CCtx* mtctx)
1129
1129
  size_t const produced = ZSTD_isError(cResult) ? 0 : cResult;
1130
1130
  size_t const flushed = ZSTD_isError(cResult) ? 0 : jobPtr->dstFlushed;
1131
1131
  assert(flushed <= produced);
1132
+ assert(jobPtr->consumed <= jobPtr->src.size);
1132
1133
  toFlush = produced - flushed;
1133
- if (toFlush==0 && (jobPtr->consumed >= jobPtr->src.size)) {
1134
- /* doneJobID is not-fully-flushed, but toFlush==0 : doneJobID should be compressing some more data */
1134
+ /* if toFlush==0, nothing is available to flush.
1135
+ * However, jobID is expected to still be active:
1136
+ * if jobID was already completed and fully flushed,
1137
+ * ZSTDMT_flushProduced() should have already moved onto next job.
1138
+ * Therefore, some input has not yet been consumed. */
1139
+ if (toFlush==0) {
1135
1140
  assert(jobPtr->consumed < jobPtr->src.size);
1136
1141
  }
1137
1142
  }
@@ -1148,12 +1153,16 @@ size_t ZSTDMT_toFlushNow(ZSTDMT_CCtx* mtctx)
1148
1153
 
1149
1154
  static unsigned ZSTDMT_computeTargetJobLog(ZSTD_CCtx_params const params)
1150
1155
  {
1151
- if (params.ldmParams.enableLdm)
1156
+ unsigned jobLog;
1157
+ if (params.ldmParams.enableLdm) {
1152
1158
  /* In Long Range Mode, the windowLog is typically oversized.
1153
1159
  * In which case, it's preferable to determine the jobSize
1154
1160
  * based on chainLog instead. */
1155
- return MAX(21, params.cParams.chainLog + 4);
1156
- return MAX(20, params.cParams.windowLog + 2);
1161
+ jobLog = MAX(21, params.cParams.chainLog + 4);
1162
+ } else {
1163
+ jobLog = MAX(20, params.cParams.windowLog + 2);
1164
+ }
1165
+ return MIN(jobLog, (unsigned)ZSTDMT_JOBLOG_MAX);
1157
1166
  }
1158
1167
 
1159
1168
  static int ZSTDMT_overlapLog_default(ZSTD_strategy strat)
@@ -1197,7 +1206,7 @@ static size_t ZSTDMT_computeOverlapSize(ZSTD_CCtx_params const params)
1197
1206
  ovLog = MIN(params.cParams.windowLog, ZSTDMT_computeTargetJobLog(params) - 2)
1198
1207
  - overlapRLog;
1199
1208
  }
1200
- assert(0 <= ovLog && ovLog <= 30);
1209
+ assert(0 <= ovLog && ovLog <= ZSTD_WINDOWLOG_MAX);
1201
1210
  DEBUGLOG(4, "overlapLog : %i", params.overlapLog);
1202
1211
  DEBUGLOG(4, "overlap size : %i", 1 << ovLog);
1203
1212
  return (ovLog==0) ? 0 : (size_t)1 << ovLog;
@@ -1391,7 +1400,7 @@ size_t ZSTDMT_initCStream_internal(
1391
1400
  FORWARD_IF_ERROR( ZSTDMT_resize(mtctx, params.nbWorkers) );
1392
1401
 
1393
1402
  if (params.jobSize != 0 && params.jobSize < ZSTDMT_JOBSIZE_MIN) params.jobSize = ZSTDMT_JOBSIZE_MIN;
1394
- if (params.jobSize > (size_t)ZSTDMT_JOBSIZE_MAX) params.jobSize = ZSTDMT_JOBSIZE_MAX;
1403
+ if (params.jobSize > (size_t)ZSTDMT_JOBSIZE_MAX) params.jobSize = (size_t)ZSTDMT_JOBSIZE_MAX;
1395
1404
 
1396
1405
  mtctx->singleBlockingThread = (pledgedSrcSize <= ZSTDMT_JOBSIZE_MIN); /* do not trigger multi-threading when srcSize is too small */
1397
1406
  if (mtctx->singleBlockingThread) {
@@ -1432,6 +1441,8 @@ size_t ZSTDMT_initCStream_internal(
1432
1441
  if (mtctx->targetSectionSize == 0) {
1433
1442
  mtctx->targetSectionSize = 1ULL << ZSTDMT_computeTargetJobLog(params);
1434
1443
  }
1444
+ assert(mtctx->targetSectionSize <= (size_t)ZSTDMT_JOBSIZE_MAX);
1445
+
1435
1446
  if (params.rsyncable) {
1436
1447
  /* Aim for the targetsectionSize as the average job size. */
1437
1448
  U32 const jobSizeMB = (U32)(mtctx->targetSectionSize >> 20);
@@ -50,6 +50,7 @@
50
50
  #ifndef ZSTDMT_JOBSIZE_MIN
51
51
  # define ZSTDMT_JOBSIZE_MIN (1 MB)
52
52
  #endif
53
+ #define ZSTDMT_JOBLOG_MAX (MEM_32bits() ? 29 : 30)
53
54
  #define ZSTDMT_JOBSIZE_MAX (MEM_32bits() ? (512 MB) : (1024 MB))
54
55
 
55
56
 
@@ -360,8 +360,11 @@ static size_t readSkippableFrameSize(void const* src, size_t srcSize)
360
360
  sizeU32 = MEM_readLE32((BYTE const*)src + ZSTD_FRAMEIDSIZE);
361
361
  RETURN_ERROR_IF((U32)(sizeU32 + ZSTD_SKIPPABLEHEADERSIZE) < sizeU32,
362
362
  frameParameter_unsupported);
363
-
364
- return skippableHeaderSize + sizeU32;
363
+ {
364
+ size_t const skippableSize = skippableHeaderSize + sizeU32;
365
+ RETURN_ERROR_IF(skippableSize > srcSize, srcSize_wrong);
366
+ return skippableSize;
367
+ }
365
368
  }
366
369
 
367
370
  /** ZSTD_findDecompressedSize() :
@@ -378,11 +381,10 @@ unsigned long long ZSTD_findDecompressedSize(const void* src, size_t srcSize)
378
381
 
379
382
  if ((magicNumber & ZSTD_MAGIC_SKIPPABLE_MASK) == ZSTD_MAGIC_SKIPPABLE_START) {
380
383
  size_t const skippableSize = readSkippableFrameSize(src, srcSize);
381
- if (ZSTD_isError(skippableSize))
382
- return skippableSize;
383
- if (srcSize < skippableSize) {
384
+ if (ZSTD_isError(skippableSize)) {
384
385
  return ZSTD_CONTENTSIZE_ERROR;
385
386
  }
387
+ assert(skippableSize <= srcSize);
386
388
 
387
389
  src = (const BYTE *)src + skippableSize;
388
390
  srcSize -= skippableSize;
@@ -467,6 +469,8 @@ static ZSTD_frameSizeInfo ZSTD_findFrameSizeInfo(const void* src, size_t srcSize
467
469
  if ((srcSize >= ZSTD_SKIPPABLEHEADERSIZE)
468
470
  && (MEM_readLE32(src) & ZSTD_MAGIC_SKIPPABLE_MASK) == ZSTD_MAGIC_SKIPPABLE_START) {
469
471
  frameSizeInfo.compressedSize = readSkippableFrameSize(src, srcSize);
472
+ assert(ZSTD_isError(frameSizeInfo.compressedSize) ||
473
+ frameSizeInfo.compressedSize <= srcSize);
470
474
  return frameSizeInfo;
471
475
  } else {
472
476
  const BYTE* ip = (const BYTE*)src;
@@ -529,7 +533,6 @@ size_t ZSTD_findFrameCompressedSize(const void *src, size_t srcSize)
529
533
  return frameSizeInfo.compressedSize;
530
534
  }
531
535
 
532
-
533
536
  /** ZSTD_decompressBound() :
534
537
  * compatible with legacy mode
535
538
  * `src` must point to the start of a ZSTD frame or a skippeable frame
@@ -546,6 +549,7 @@ unsigned long long ZSTD_decompressBound(const void* src, size_t srcSize)
546
549
  unsigned long long const decompressedBound = frameSizeInfo.decompressedBound;
547
550
  if (ZSTD_isError(compressedSize) || decompressedBound == ZSTD_CONTENTSIZE_ERROR)
548
551
  return ZSTD_CONTENTSIZE_ERROR;
552
+ assert(srcSize >= compressedSize);
549
553
  src = (const BYTE*)src + compressedSize;
550
554
  srcSize -= compressedSize;
551
555
  bound += decompressedBound;
@@ -738,9 +742,8 @@ static size_t ZSTD_decompressMultiFrame(ZSTD_DCtx* dctx,
738
742
  (unsigned)magicNumber, ZSTD_MAGICNUMBER);
739
743
  if ((magicNumber & ZSTD_MAGIC_SKIPPABLE_MASK) == ZSTD_MAGIC_SKIPPABLE_START) {
740
744
  size_t const skippableSize = readSkippableFrameSize(src, srcSize);
741
- if (ZSTD_isError(skippableSize))
742
- return skippableSize;
743
- RETURN_ERROR_IF(srcSize < skippableSize, srcSize_wrong);
745
+ FORWARD_IF_ERROR(skippableSize);
746
+ assert(skippableSize <= srcSize);
744
747
 
745
748
  src = (const BYTE *)src + skippableSize;
746
749
  srcSize -= skippableSize;
@@ -906,6 +909,7 @@ size_t ZSTD_decompressContinue(ZSTD_DCtx* dctx, void* dst, size_t dstCapacity, c
906
909
  { blockProperties_t bp;
907
910
  size_t const cBlockSize = ZSTD_getcBlockSize(src, ZSTD_blockHeaderSize, &bp);
908
911
  if (ZSTD_isError(cBlockSize)) return cBlockSize;
912
+ RETURN_ERROR_IF(cBlockSize > dctx->fParams.blockSizeMax, corruption_detected, "Block Size Exceeds Maximum");
909
913
  dctx->expected = cBlockSize;
910
914
  dctx->bType = bp.blockType;
911
915
  dctx->rleSize = bp.origSize;
@@ -950,6 +954,7 @@ size_t ZSTD_decompressContinue(ZSTD_DCtx* dctx, void* dst, size_t dstCapacity, c
950
954
  RETURN_ERROR(corruption_detected);
951
955
  }
952
956
  if (ZSTD_isError(rSize)) return rSize;
957
+ RETURN_ERROR_IF(rSize > dctx->fParams.blockSizeMax, corruption_detected, "Decompressed Block Size Exceeds Maximum");
953
958
  DEBUGLOG(5, "ZSTD_decompressContinue: decoded size from block : %u", (unsigned)rSize);
954
959
  dctx->decodedSize += rSize;
955
960
  if (dctx->fParams.checksumFlag) XXH64_update(&dctx->xxhState, dst, rSize);
@@ -505,7 +505,7 @@ size_t ZSTD_decodeSeqHeaders(ZSTD_DCtx* dctx, int* nbSeqPtr,
505
505
  *nbSeqPtr = nbSeq;
506
506
 
507
507
  /* FSE table descriptors */
508
- RETURN_ERROR_IF(ip+4 > iend, srcSize_wrong); /* minimum possible size */
508
+ RETURN_ERROR_IF(ip+1 > iend, srcSize_wrong); /* minimum possible size: 1 byte for symbol encoding types */
509
509
  { symbolEncodingType_e const LLtype = (symbolEncodingType_e)(*ip >> 6);
510
510
  symbolEncodingType_e const OFtype = (symbolEncodingType_e)((*ip >> 4) & 3);
511
511
  symbolEncodingType_e const MLtype = (symbolEncodingType_e)((*ip >> 2) & 3);
@@ -637,9 +637,10 @@ size_t ZSTD_execSequence(BYTE* op,
637
637
  if (oLitEnd>oend_w) return ZSTD_execSequenceLast7(op, oend, sequence, litPtr, litLimit, prefixStart, virtualStart, dictEnd);
638
638
 
639
639
  /* copy Literals */
640
- ZSTD_copy8(op, *litPtr);
641
640
  if (sequence.litLength > 8)
642
- ZSTD_wildcopy(op+8, (*litPtr)+8, sequence.litLength - 8); /* note : since oLitEnd <= oend-WILDCOPY_OVERLENGTH, no risk of overwrite beyond oend */
641
+ ZSTD_wildcopy_16min(op, (*litPtr), sequence.litLength, ZSTD_no_overlap); /* note : since oLitEnd <= oend-WILDCOPY_OVERLENGTH, no risk of overwrite beyond oend */
642
+ else
643
+ ZSTD_copy8(op, *litPtr);
643
644
  op = oLitEnd;
644
645
  *litPtr = iLitEnd; /* update for next sequence */
645
646
 
@@ -686,13 +687,13 @@ size_t ZSTD_execSequence(BYTE* op,
686
687
 
687
688
  if (oMatchEnd > oend-(16-MINMATCH)) {
688
689
  if (op < oend_w) {
689
- ZSTD_wildcopy(op, match, oend_w - op);
690
+ ZSTD_wildcopy(op, match, oend_w - op, ZSTD_overlap_src_before_dst);
690
691
  match += oend_w - op;
691
692
  op = oend_w;
692
693
  }
693
694
  while (op < oMatchEnd) *op++ = *match++;
694
695
  } else {
695
- ZSTD_wildcopy(op, match, (ptrdiff_t)sequence.matchLength-8); /* works even if matchLength < 8 */
696
+ ZSTD_wildcopy(op, match, (ptrdiff_t)sequence.matchLength-8, ZSTD_overlap_src_before_dst); /* works even if matchLength < 8 */
696
697
  }
697
698
  return sequenceLength;
698
699
  }
@@ -717,9 +718,11 @@ size_t ZSTD_execSequenceLong(BYTE* op,
717
718
  if (oLitEnd > oend_w) return ZSTD_execSequenceLast7(op, oend, sequence, litPtr, litLimit, prefixStart, dictStart, dictEnd);
718
719
 
719
720
  /* copy Literals */
720
- ZSTD_copy8(op, *litPtr); /* note : op <= oLitEnd <= oend_w == oend - 8 */
721
721
  if (sequence.litLength > 8)
722
- ZSTD_wildcopy(op+8, (*litPtr)+8, sequence.litLength - 8); /* note : since oLitEnd <= oend-WILDCOPY_OVERLENGTH, no risk of overwrite beyond oend */
722
+ ZSTD_wildcopy_16min(op, *litPtr, sequence.litLength, ZSTD_no_overlap); /* note : since oLitEnd <= oend-WILDCOPY_OVERLENGTH, no risk of overwrite beyond oend */
723
+ else
724
+ ZSTD_copy8(op, *litPtr); /* note : op <= oLitEnd <= oend_w == oend - 8 */
725
+
723
726
  op = oLitEnd;
724
727
  *litPtr = iLitEnd; /* update for next sequence */
725
728
 
@@ -766,13 +769,13 @@ size_t ZSTD_execSequenceLong(BYTE* op,
766
769
 
767
770
  if (oMatchEnd > oend-(16-MINMATCH)) {
768
771
  if (op < oend_w) {
769
- ZSTD_wildcopy(op, match, oend_w - op);
772
+ ZSTD_wildcopy(op, match, oend_w - op, ZSTD_overlap_src_before_dst);
770
773
  match += oend_w - op;
771
774
  op = oend_w;
772
775
  }
773
776
  while (op < oMatchEnd) *op++ = *match++;
774
777
  } else {
775
- ZSTD_wildcopy(op, match, (ptrdiff_t)sequence.matchLength-8); /* works even if matchLength < 8 */
778
+ ZSTD_wildcopy(op, match, (ptrdiff_t)sequence.matchLength-8, ZSTD_overlap_src_before_dst); /* works even if matchLength < 8 */
776
779
  }
777
780
  return sequenceLength;
778
781
  }
@@ -889,6 +892,7 @@ ZSTD_decodeSequence(seqState_t* seqState, const ZSTD_longOffset_e longOffsets)
889
892
  }
890
893
 
891
894
  FORCE_INLINE_TEMPLATE size_t
895
+ DONT_VECTORIZE
892
896
  ZSTD_decompressSequences_body( ZSTD_DCtx* dctx,
893
897
  void* dst, size_t maxDstSize,
894
898
  const void* seqStart, size_t seqSize, int nbSeq,
@@ -918,6 +922,11 @@ ZSTD_decompressSequences_body( ZSTD_DCtx* dctx,
918
922
  ZSTD_initFseState(&seqState.stateOffb, &seqState.DStream, dctx->OFTptr);
919
923
  ZSTD_initFseState(&seqState.stateML, &seqState.DStream, dctx->MLTptr);
920
924
 
925
+ ZSTD_STATIC_ASSERT(
926
+ BIT_DStream_unfinished < BIT_DStream_completed &&
927
+ BIT_DStream_endOfBuffer < BIT_DStream_completed &&
928
+ BIT_DStream_completed < BIT_DStream_overflow);
929
+
921
930
  for ( ; (BIT_reloadDStream(&(seqState.DStream)) <= BIT_DStream_completed) && nbSeq ; ) {
922
931
  nbSeq--;
923
932
  { seq_t const sequence = ZSTD_decodeSequence(&seqState, isLongOffset);
@@ -930,6 +939,7 @@ ZSTD_decompressSequences_body( ZSTD_DCtx* dctx,
930
939
  /* check if reached exact end */
931
940
  DEBUGLOG(5, "ZSTD_decompressSequences_body: after decode loop, remaining nbSeq : %i", nbSeq);
932
941
  RETURN_ERROR_IF(nbSeq, corruption_detected);
942
+ RETURN_ERROR_IF(BIT_reloadDStream(&seqState.DStream) < BIT_DStream_completed, corruption_detected);
933
943
  /* save reps for next block */
934
944
  { U32 i; for (i=0; i<ZSTD_REP_NUM; i++) dctx->entropy.rep[i] = (U32)(seqState.prevOffset[i]); }
935
945
  }
@@ -1131,6 +1141,7 @@ ZSTD_decompressSequencesLong_default(ZSTD_DCtx* dctx,
1131
1141
 
1132
1142
  #ifndef ZSTD_FORCE_DECOMPRESS_SEQUENCES_LONG
1133
1143
  static TARGET_ATTRIBUTE("bmi2") size_t
1144
+ DONT_VECTORIZE
1134
1145
  ZSTD_decompressSequences_bmi2(ZSTD_DCtx* dctx,
1135
1146
  void* dst, size_t maxDstSize,
1136
1147
  const void* seqStart, size_t seqSize, int nbSeq,