zstd-ruby 1.5.2.2 → 1.5.4.0

Sign up to get free protection for your applications and to get access to all the features.
Files changed (75) hide show
  1. checksums.yaml +4 -4
  2. data/README.md +15 -3
  3. data/ext/zstdruby/common.h +7 -0
  4. data/ext/zstdruby/libzstd/common/bits.h +175 -0
  5. data/ext/zstdruby/libzstd/common/bitstream.h +18 -59
  6. data/ext/zstdruby/libzstd/common/compiler.h +22 -3
  7. data/ext/zstdruby/libzstd/common/cpu.h +1 -1
  8. data/ext/zstdruby/libzstd/common/debug.c +1 -1
  9. data/ext/zstdruby/libzstd/common/debug.h +1 -1
  10. data/ext/zstdruby/libzstd/common/entropy_common.c +12 -40
  11. data/ext/zstdruby/libzstd/common/error_private.c +9 -2
  12. data/ext/zstdruby/libzstd/common/error_private.h +1 -1
  13. data/ext/zstdruby/libzstd/common/fse.h +5 -83
  14. data/ext/zstdruby/libzstd/common/fse_decompress.c +7 -99
  15. data/ext/zstdruby/libzstd/common/huf.h +65 -156
  16. data/ext/zstdruby/libzstd/common/mem.h +39 -46
  17. data/ext/zstdruby/libzstd/common/pool.c +26 -10
  18. data/ext/zstdruby/libzstd/common/pool.h +7 -1
  19. data/ext/zstdruby/libzstd/common/portability_macros.h +22 -3
  20. data/ext/zstdruby/libzstd/common/threading.c +68 -14
  21. data/ext/zstdruby/libzstd/common/threading.h +5 -10
  22. data/ext/zstdruby/libzstd/common/xxhash.c +2 -2
  23. data/ext/zstdruby/libzstd/common/xxhash.h +8 -8
  24. data/ext/zstdruby/libzstd/common/zstd_common.c +1 -1
  25. data/ext/zstdruby/libzstd/common/zstd_deps.h +1 -1
  26. data/ext/zstdruby/libzstd/common/zstd_internal.h +17 -113
  27. data/ext/zstdruby/libzstd/common/zstd_trace.h +3 -3
  28. data/ext/zstdruby/libzstd/compress/clevels.h +1 -1
  29. data/ext/zstdruby/libzstd/compress/fse_compress.c +7 -124
  30. data/ext/zstdruby/libzstd/compress/hist.c +1 -1
  31. data/ext/zstdruby/libzstd/compress/hist.h +1 -1
  32. data/ext/zstdruby/libzstd/compress/huf_compress.c +234 -169
  33. data/ext/zstdruby/libzstd/compress/zstd_compress.c +1055 -455
  34. data/ext/zstdruby/libzstd/compress/zstd_compress_internal.h +165 -145
  35. data/ext/zstdruby/libzstd/compress/zstd_compress_literals.c +115 -39
  36. data/ext/zstdruby/libzstd/compress/zstd_compress_literals.h +16 -8
  37. data/ext/zstdruby/libzstd/compress/zstd_compress_sequences.c +3 -3
  38. data/ext/zstdruby/libzstd/compress/zstd_compress_sequences.h +1 -1
  39. data/ext/zstdruby/libzstd/compress/zstd_compress_superblock.c +25 -21
  40. data/ext/zstdruby/libzstd/compress/zstd_compress_superblock.h +1 -1
  41. data/ext/zstdruby/libzstd/compress/zstd_cwksp.h +5 -3
  42. data/ext/zstdruby/libzstd/compress/zstd_double_fast.c +95 -33
  43. data/ext/zstdruby/libzstd/compress/zstd_double_fast.h +3 -2
  44. data/ext/zstdruby/libzstd/compress/zstd_fast.c +433 -148
  45. data/ext/zstdruby/libzstd/compress/zstd_fast.h +3 -2
  46. data/ext/zstdruby/libzstd/compress/zstd_lazy.c +306 -283
  47. data/ext/zstdruby/libzstd/compress/zstd_lazy.h +4 -2
  48. data/ext/zstdruby/libzstd/compress/zstd_ldm.c +5 -5
  49. data/ext/zstdruby/libzstd/compress/zstd_ldm.h +1 -1
  50. data/ext/zstdruby/libzstd/compress/zstd_ldm_geartab.h +1 -1
  51. data/ext/zstdruby/libzstd/compress/zstd_opt.c +104 -80
  52. data/ext/zstdruby/libzstd/compress/zstd_opt.h +1 -1
  53. data/ext/zstdruby/libzstd/compress/zstdmt_compress.c +12 -5
  54. data/ext/zstdruby/libzstd/compress/zstdmt_compress.h +1 -1
  55. data/ext/zstdruby/libzstd/decompress/huf_decompress.c +434 -441
  56. data/ext/zstdruby/libzstd/decompress/huf_decompress_amd64.S +30 -39
  57. data/ext/zstdruby/libzstd/decompress/zstd_ddict.c +3 -4
  58. data/ext/zstdruby/libzstd/decompress/zstd_ddict.h +1 -1
  59. data/ext/zstdruby/libzstd/decompress/zstd_decompress.c +164 -42
  60. data/ext/zstdruby/libzstd/decompress/zstd_decompress_block.c +186 -65
  61. data/ext/zstdruby/libzstd/decompress/zstd_decompress_block.h +1 -1
  62. data/ext/zstdruby/libzstd/decompress/zstd_decompress_internal.h +4 -2
  63. data/ext/zstdruby/libzstd/dictBuilder/cover.c +19 -15
  64. data/ext/zstdruby/libzstd/dictBuilder/cover.h +1 -1
  65. data/ext/zstdruby/libzstd/dictBuilder/fastcover.c +2 -2
  66. data/ext/zstdruby/libzstd/dictBuilder/zdict.c +9 -87
  67. data/ext/zstdruby/libzstd/zdict.h +53 -31
  68. data/ext/zstdruby/libzstd/zstd.h +489 -90
  69. data/ext/zstdruby/libzstd/zstd_errors.h +27 -8
  70. data/ext/zstdruby/main.c +4 -0
  71. data/ext/zstdruby/streaming_compress.c +1 -7
  72. data/ext/zstdruby/zstdruby.c +110 -26
  73. data/lib/zstd-ruby/version.rb +1 -1
  74. data/lib/zstd-ruby.rb +0 -1
  75. metadata +7 -6
@@ -1,5 +1,5 @@
1
1
  /*
2
- * Copyright (c) Yann Collet, Facebook, Inc.
2
+ * Copyright (c) Meta Platforms, Inc. and affiliates.
3
3
  * All rights reserved.
4
4
  *
5
5
  * This source code is licensed under both the BSD-style license (found in the
@@ -16,7 +16,6 @@
16
16
  #include "hist.h" /* HIST_countFast_wksp */
17
17
  #define FSE_STATIC_LINKING_ONLY /* FSE_encodeSymbol */
18
18
  #include "../common/fse.h"
19
- #define HUF_STATIC_LINKING_ONLY
20
19
  #include "../common/huf.h"
21
20
  #include "zstd_compress_internal.h"
22
21
  #include "zstd_compress_sequences.h"
@@ -27,6 +26,7 @@
27
26
  #include "zstd_opt.h"
28
27
  #include "zstd_ldm.h"
29
28
  #include "zstd_compress_superblock.h"
29
+ #include "../common/bits.h" /* ZSTD_highbit32 */
30
30
 
31
31
  /* ***************************************************************
32
32
  * Tuning parameters
@@ -58,14 +58,17 @@
58
58
  * Helper functions
59
59
  ***************************************/
60
60
  /* ZSTD_compressBound()
61
- * Note that the result from this function is only compatible with the "normal"
62
- * full-block strategy.
63
- * When there are a lot of small blocks due to frequent flush in streaming mode
64
- * the overhead of headers can make the compressed data to be larger than the
65
- * return value of ZSTD_compressBound().
61
+ * Note that the result from this function is only valid for
62
+ * the one-pass compression functions.
63
+ * When employing the streaming mode,
64
+ * if flushes are frequently altering the size of blocks,
65
+ * the overhead from block headers can make the compressed data larger
66
+ * than the return value of ZSTD_compressBound().
66
67
  */
67
68
  size_t ZSTD_compressBound(size_t srcSize) {
68
- return ZSTD_COMPRESSBOUND(srcSize);
69
+ size_t const r = ZSTD_COMPRESSBOUND(srcSize);
70
+ if (r==0) return ERROR(srcSize_wrong);
71
+ return r;
69
72
  }
70
73
 
71
74
 
@@ -177,12 +180,9 @@ size_t ZSTD_freeCCtx(ZSTD_CCtx* cctx)
177
180
  if (cctx==NULL) return 0; /* support free on NULL */
178
181
  RETURN_ERROR_IF(cctx->staticSize, memory_allocation,
179
182
  "not compatible with static CCtx");
180
- {
181
- int cctxInWorkspace = ZSTD_cwksp_owns_buffer(&cctx->workspace, cctx);
183
+ { int cctxInWorkspace = ZSTD_cwksp_owns_buffer(&cctx->workspace, cctx);
182
184
  ZSTD_freeCCtxContent(cctx);
183
- if (!cctxInWorkspace) {
184
- ZSTD_customFree(cctx, cctx->customMem);
185
- }
185
+ if (!cctxInWorkspace) ZSTD_customFree(cctx, cctx->customMem);
186
186
  }
187
187
  return 0;
188
188
  }
@@ -267,9 +267,9 @@ static int ZSTD_allocateChainTable(const ZSTD_strategy strategy,
267
267
  return forDDSDict || ((strategy != ZSTD_fast) && !ZSTD_rowMatchFinderUsed(strategy, useRowMatchFinder));
268
268
  }
269
269
 
270
- /* Returns 1 if compression parameters are such that we should
270
+ /* Returns ZSTD_ps_enable if compression parameters are such that we should
271
271
  * enable long distance matching (wlog >= 27, strategy >= btopt).
272
- * Returns 0 otherwise.
272
+ * Returns ZSTD_ps_disable otherwise.
273
273
  */
274
274
  static ZSTD_paramSwitch_e ZSTD_resolveEnableLdm(ZSTD_paramSwitch_e mode,
275
275
  const ZSTD_compressionParameters* const cParams) {
@@ -277,6 +277,34 @@ static ZSTD_paramSwitch_e ZSTD_resolveEnableLdm(ZSTD_paramSwitch_e mode,
277
277
  return (cParams->strategy >= ZSTD_btopt && cParams->windowLog >= 27) ? ZSTD_ps_enable : ZSTD_ps_disable;
278
278
  }
279
279
 
280
+ static int ZSTD_resolveExternalSequenceValidation(int mode) {
281
+ return mode;
282
+ }
283
+
284
+ /* Resolves maxBlockSize to the default if no value is present. */
285
+ static size_t ZSTD_resolveMaxBlockSize(size_t maxBlockSize) {
286
+ if (maxBlockSize == 0) {
287
+ return ZSTD_BLOCKSIZE_MAX;
288
+ } else {
289
+ return maxBlockSize;
290
+ }
291
+ }
292
+
293
+ static ZSTD_paramSwitch_e ZSTD_resolveExternalRepcodeSearch(ZSTD_paramSwitch_e value, int cLevel) {
294
+ if (value != ZSTD_ps_auto) return value;
295
+ if (cLevel < 10) {
296
+ return ZSTD_ps_disable;
297
+ } else {
298
+ return ZSTD_ps_enable;
299
+ }
300
+ }
301
+
302
+ /* Returns 1 if compression parameters are such that CDict hashtable and chaintable indices are tagged.
303
+ * If so, the tags need to be removed in ZSTD_resetCCtx_byCopyingCDict. */
304
+ static int ZSTD_CDictIndicesAreTagged(const ZSTD_compressionParameters* const cParams) {
305
+ return cParams->strategy == ZSTD_fast || cParams->strategy == ZSTD_dfast;
306
+ }
307
+
280
308
  static ZSTD_CCtx_params ZSTD_makeCCtxParamsFromCParams(
281
309
  ZSTD_compressionParameters cParams)
282
310
  {
@@ -294,6 +322,10 @@ static ZSTD_CCtx_params ZSTD_makeCCtxParamsFromCParams(
294
322
  }
295
323
  cctxParams.useBlockSplitter = ZSTD_resolveBlockSplitterMode(cctxParams.useBlockSplitter, &cParams);
296
324
  cctxParams.useRowMatchFinder = ZSTD_resolveRowMatchFinderMode(cctxParams.useRowMatchFinder, &cParams);
325
+ cctxParams.validateSequences = ZSTD_resolveExternalSequenceValidation(cctxParams.validateSequences);
326
+ cctxParams.maxBlockSize = ZSTD_resolveMaxBlockSize(cctxParams.maxBlockSize);
327
+ cctxParams.searchForExternalRepcodes = ZSTD_resolveExternalRepcodeSearch(cctxParams.searchForExternalRepcodes,
328
+ cctxParams.compressionLevel);
297
329
  assert(!ZSTD_checkCParams(cParams));
298
330
  return cctxParams;
299
331
  }
@@ -339,10 +371,13 @@ size_t ZSTD_CCtxParams_init(ZSTD_CCtx_params* cctxParams, int compressionLevel)
339
371
  #define ZSTD_NO_CLEVEL 0
340
372
 
341
373
  /**
342
- * Initializes the cctxParams from params and compressionLevel.
374
+ * Initializes `cctxParams` from `params` and `compressionLevel`.
343
375
  * @param compressionLevel If params are derived from a compression level then that compression level, otherwise ZSTD_NO_CLEVEL.
344
376
  */
345
- static void ZSTD_CCtxParams_init_internal(ZSTD_CCtx_params* cctxParams, ZSTD_parameters const* params, int compressionLevel)
377
+ static void
378
+ ZSTD_CCtxParams_init_internal(ZSTD_CCtx_params* cctxParams,
379
+ const ZSTD_parameters* params,
380
+ int compressionLevel)
346
381
  {
347
382
  assert(!ZSTD_checkCParams(params->cParams));
348
383
  ZSTD_memset(cctxParams, 0, sizeof(*cctxParams));
@@ -355,6 +390,9 @@ static void ZSTD_CCtxParams_init_internal(ZSTD_CCtx_params* cctxParams, ZSTD_par
355
390
  cctxParams->useRowMatchFinder = ZSTD_resolveRowMatchFinderMode(cctxParams->useRowMatchFinder, &params->cParams);
356
391
  cctxParams->useBlockSplitter = ZSTD_resolveBlockSplitterMode(cctxParams->useBlockSplitter, &params->cParams);
357
392
  cctxParams->ldmParams.enableLdm = ZSTD_resolveEnableLdm(cctxParams->ldmParams.enableLdm, &params->cParams);
393
+ cctxParams->validateSequences = ZSTD_resolveExternalSequenceValidation(cctxParams->validateSequences);
394
+ cctxParams->maxBlockSize = ZSTD_resolveMaxBlockSize(cctxParams->maxBlockSize);
395
+ cctxParams->searchForExternalRepcodes = ZSTD_resolveExternalRepcodeSearch(cctxParams->searchForExternalRepcodes, compressionLevel);
358
396
  DEBUGLOG(4, "ZSTD_CCtxParams_init_internal: useRowMatchFinder=%d, useBlockSplitter=%d ldm=%d",
359
397
  cctxParams->useRowMatchFinder, cctxParams->useBlockSplitter, cctxParams->ldmParams.enableLdm);
360
398
  }
@@ -369,7 +407,7 @@ size_t ZSTD_CCtxParams_init_advanced(ZSTD_CCtx_params* cctxParams, ZSTD_paramete
369
407
 
370
408
  /**
371
409
  * Sets cctxParams' cParams and fParams from params, but otherwise leaves them alone.
372
- * @param param Validated zstd parameters.
410
+ * @param params Validated zstd parameters.
373
411
  */
374
412
  static void ZSTD_CCtxParams_setZstdParams(
375
413
  ZSTD_CCtx_params* cctxParams, const ZSTD_parameters* params)
@@ -478,8 +516,8 @@ ZSTD_bounds ZSTD_cParam_getBounds(ZSTD_cParameter param)
478
516
  return bounds;
479
517
 
480
518
  case ZSTD_c_enableLongDistanceMatching:
481
- bounds.lowerBound = 0;
482
- bounds.upperBound = 1;
519
+ bounds.lowerBound = (int)ZSTD_ps_auto;
520
+ bounds.upperBound = (int)ZSTD_ps_disable;
483
521
  return bounds;
484
522
 
485
523
  case ZSTD_c_ldmHashLog:
@@ -572,6 +610,26 @@ ZSTD_bounds ZSTD_cParam_getBounds(ZSTD_cParameter param)
572
610
  bounds.upperBound = 1;
573
611
  return bounds;
574
612
 
613
+ case ZSTD_c_prefetchCDictTables:
614
+ bounds.lowerBound = (int)ZSTD_ps_auto;
615
+ bounds.upperBound = (int)ZSTD_ps_disable;
616
+ return bounds;
617
+
618
+ case ZSTD_c_enableSeqProducerFallback:
619
+ bounds.lowerBound = 0;
620
+ bounds.upperBound = 1;
621
+ return bounds;
622
+
623
+ case ZSTD_c_maxBlockSize:
624
+ bounds.lowerBound = ZSTD_BLOCKSIZE_MAX_MIN;
625
+ bounds.upperBound = ZSTD_BLOCKSIZE_MAX;
626
+ return bounds;
627
+
628
+ case ZSTD_c_searchForExternalRepcodes:
629
+ bounds.lowerBound = (int)ZSTD_ps_auto;
630
+ bounds.upperBound = (int)ZSTD_ps_disable;
631
+ return bounds;
632
+
575
633
  default:
576
634
  bounds.error = ERROR(parameter_unsupported);
577
635
  return bounds;
@@ -636,6 +694,10 @@ static int ZSTD_isUpdateAuthorized(ZSTD_cParameter param)
636
694
  case ZSTD_c_useBlockSplitter:
637
695
  case ZSTD_c_useRowMatchFinder:
638
696
  case ZSTD_c_deterministicRefPrefix:
697
+ case ZSTD_c_prefetchCDictTables:
698
+ case ZSTD_c_enableSeqProducerFallback:
699
+ case ZSTD_c_maxBlockSize:
700
+ case ZSTD_c_searchForExternalRepcodes:
639
701
  default:
640
702
  return 0;
641
703
  }
@@ -648,7 +710,7 @@ size_t ZSTD_CCtx_setParameter(ZSTD_CCtx* cctx, ZSTD_cParameter param, int value)
648
710
  if (ZSTD_isUpdateAuthorized(param)) {
649
711
  cctx->cParamsChanged = 1;
650
712
  } else {
651
- RETURN_ERROR(stage_wrong, "can only set params in ctx init stage");
713
+ RETURN_ERROR(stage_wrong, "can only set params in cctx init stage");
652
714
  } }
653
715
 
654
716
  switch(param)
@@ -691,6 +753,10 @@ size_t ZSTD_CCtx_setParameter(ZSTD_CCtx* cctx, ZSTD_cParameter param, int value)
691
753
  case ZSTD_c_useBlockSplitter:
692
754
  case ZSTD_c_useRowMatchFinder:
693
755
  case ZSTD_c_deterministicRefPrefix:
756
+ case ZSTD_c_prefetchCDictTables:
757
+ case ZSTD_c_enableSeqProducerFallback:
758
+ case ZSTD_c_maxBlockSize:
759
+ case ZSTD_c_searchForExternalRepcodes:
694
760
  break;
695
761
 
696
762
  default: RETURN_ERROR(parameter_unsupported, "unknown parameter");
@@ -746,12 +812,12 @@ size_t ZSTD_CCtxParams_setParameter(ZSTD_CCtx_params* CCtxParams,
746
812
  case ZSTD_c_minMatch :
747
813
  if (value!=0) /* 0 => use default */
748
814
  BOUNDCHECK(ZSTD_c_minMatch, value);
749
- CCtxParams->cParams.minMatch = value;
815
+ CCtxParams->cParams.minMatch = (U32)value;
750
816
  return CCtxParams->cParams.minMatch;
751
817
 
752
818
  case ZSTD_c_targetLength :
753
819
  BOUNDCHECK(ZSTD_c_targetLength, value);
754
- CCtxParams->cParams.targetLength = value;
820
+ CCtxParams->cParams.targetLength = (U32)value;
755
821
  return CCtxParams->cParams.targetLength;
756
822
 
757
823
  case ZSTD_c_strategy :
@@ -764,12 +830,12 @@ size_t ZSTD_CCtxParams_setParameter(ZSTD_CCtx_params* CCtxParams,
764
830
  /* Content size written in frame header _when known_ (default:1) */
765
831
  DEBUGLOG(4, "set content size flag = %u", (value!=0));
766
832
  CCtxParams->fParams.contentSizeFlag = value != 0;
767
- return CCtxParams->fParams.contentSizeFlag;
833
+ return (size_t)CCtxParams->fParams.contentSizeFlag;
768
834
 
769
835
  case ZSTD_c_checksumFlag :
770
836
  /* A 32-bits content checksum will be calculated and written at end of frame (default:0) */
771
837
  CCtxParams->fParams.checksumFlag = value != 0;
772
- return CCtxParams->fParams.checksumFlag;
838
+ return (size_t)CCtxParams->fParams.checksumFlag;
773
839
 
774
840
  case ZSTD_c_dictIDFlag : /* When applicable, dictionary's dictID is provided in frame header (default:1) */
775
841
  DEBUGLOG(4, "set dictIDFlag = %u", (value!=0));
@@ -778,18 +844,18 @@ size_t ZSTD_CCtxParams_setParameter(ZSTD_CCtx_params* CCtxParams,
778
844
 
779
845
  case ZSTD_c_forceMaxWindow :
780
846
  CCtxParams->forceWindow = (value != 0);
781
- return CCtxParams->forceWindow;
847
+ return (size_t)CCtxParams->forceWindow;
782
848
 
783
849
  case ZSTD_c_forceAttachDict : {
784
850
  const ZSTD_dictAttachPref_e pref = (ZSTD_dictAttachPref_e)value;
785
- BOUNDCHECK(ZSTD_c_forceAttachDict, pref);
851
+ BOUNDCHECK(ZSTD_c_forceAttachDict, (int)pref);
786
852
  CCtxParams->attachDictPref = pref;
787
853
  return CCtxParams->attachDictPref;
788
854
  }
789
855
 
790
856
  case ZSTD_c_literalCompressionMode : {
791
857
  const ZSTD_paramSwitch_e lcm = (ZSTD_paramSwitch_e)value;
792
- BOUNDCHECK(ZSTD_c_literalCompressionMode, lcm);
858
+ BOUNDCHECK(ZSTD_c_literalCompressionMode, (int)lcm);
793
859
  CCtxParams->literalCompressionMode = lcm;
794
860
  return CCtxParams->literalCompressionMode;
795
861
  }
@@ -840,47 +906,48 @@ size_t ZSTD_CCtxParams_setParameter(ZSTD_CCtx_params* CCtxParams,
840
906
 
841
907
  case ZSTD_c_enableDedicatedDictSearch :
842
908
  CCtxParams->enableDedicatedDictSearch = (value!=0);
843
- return CCtxParams->enableDedicatedDictSearch;
909
+ return (size_t)CCtxParams->enableDedicatedDictSearch;
844
910
 
845
911
  case ZSTD_c_enableLongDistanceMatching :
912
+ BOUNDCHECK(ZSTD_c_enableLongDistanceMatching, value);
846
913
  CCtxParams->ldmParams.enableLdm = (ZSTD_paramSwitch_e)value;
847
914
  return CCtxParams->ldmParams.enableLdm;
848
915
 
849
916
  case ZSTD_c_ldmHashLog :
850
917
  if (value!=0) /* 0 ==> auto */
851
918
  BOUNDCHECK(ZSTD_c_ldmHashLog, value);
852
- CCtxParams->ldmParams.hashLog = value;
919
+ CCtxParams->ldmParams.hashLog = (U32)value;
853
920
  return CCtxParams->ldmParams.hashLog;
854
921
 
855
922
  case ZSTD_c_ldmMinMatch :
856
923
  if (value!=0) /* 0 ==> default */
857
924
  BOUNDCHECK(ZSTD_c_ldmMinMatch, value);
858
- CCtxParams->ldmParams.minMatchLength = value;
925
+ CCtxParams->ldmParams.minMatchLength = (U32)value;
859
926
  return CCtxParams->ldmParams.minMatchLength;
860
927
 
861
928
  case ZSTD_c_ldmBucketSizeLog :
862
929
  if (value!=0) /* 0 ==> default */
863
930
  BOUNDCHECK(ZSTD_c_ldmBucketSizeLog, value);
864
- CCtxParams->ldmParams.bucketSizeLog = value;
931
+ CCtxParams->ldmParams.bucketSizeLog = (U32)value;
865
932
  return CCtxParams->ldmParams.bucketSizeLog;
866
933
 
867
934
  case ZSTD_c_ldmHashRateLog :
868
935
  if (value!=0) /* 0 ==> default */
869
936
  BOUNDCHECK(ZSTD_c_ldmHashRateLog, value);
870
- CCtxParams->ldmParams.hashRateLog = value;
937
+ CCtxParams->ldmParams.hashRateLog = (U32)value;
871
938
  return CCtxParams->ldmParams.hashRateLog;
872
939
 
873
940
  case ZSTD_c_targetCBlockSize :
874
941
  if (value!=0) /* 0 ==> default */
875
942
  BOUNDCHECK(ZSTD_c_targetCBlockSize, value);
876
- CCtxParams->targetCBlockSize = value;
943
+ CCtxParams->targetCBlockSize = (U32)value;
877
944
  return CCtxParams->targetCBlockSize;
878
945
 
879
946
  case ZSTD_c_srcSizeHint :
880
947
  if (value!=0) /* 0 ==> default */
881
948
  BOUNDCHECK(ZSTD_c_srcSizeHint, value);
882
949
  CCtxParams->srcSizeHint = value;
883
- return CCtxParams->srcSizeHint;
950
+ return (size_t)CCtxParams->srcSizeHint;
884
951
 
885
952
  case ZSTD_c_stableInBuffer:
886
953
  BOUNDCHECK(ZSTD_c_stableInBuffer, value);
@@ -917,6 +984,27 @@ size_t ZSTD_CCtxParams_setParameter(ZSTD_CCtx_params* CCtxParams,
917
984
  CCtxParams->deterministicRefPrefix = !!value;
918
985
  return CCtxParams->deterministicRefPrefix;
919
986
 
987
+ case ZSTD_c_prefetchCDictTables:
988
+ BOUNDCHECK(ZSTD_c_prefetchCDictTables, value);
989
+ CCtxParams->prefetchCDictTables = (ZSTD_paramSwitch_e)value;
990
+ return CCtxParams->prefetchCDictTables;
991
+
992
+ case ZSTD_c_enableSeqProducerFallback:
993
+ BOUNDCHECK(ZSTD_c_enableSeqProducerFallback, value);
994
+ CCtxParams->enableMatchFinderFallback = value;
995
+ return CCtxParams->enableMatchFinderFallback;
996
+
997
+ case ZSTD_c_maxBlockSize:
998
+ if (value!=0) /* 0 ==> default */
999
+ BOUNDCHECK(ZSTD_c_maxBlockSize, value);
1000
+ CCtxParams->maxBlockSize = value;
1001
+ return CCtxParams->maxBlockSize;
1002
+
1003
+ case ZSTD_c_searchForExternalRepcodes:
1004
+ BOUNDCHECK(ZSTD_c_searchForExternalRepcodes, value);
1005
+ CCtxParams->searchForExternalRepcodes = (ZSTD_paramSwitch_e)value;
1006
+ return CCtxParams->searchForExternalRepcodes;
1007
+
920
1008
  default: RETURN_ERROR(parameter_unsupported, "unknown parameter");
921
1009
  }
922
1010
  }
@@ -1049,6 +1137,18 @@ size_t ZSTD_CCtxParams_getParameter(
1049
1137
  case ZSTD_c_deterministicRefPrefix:
1050
1138
  *value = (int)CCtxParams->deterministicRefPrefix;
1051
1139
  break;
1140
+ case ZSTD_c_prefetchCDictTables:
1141
+ *value = (int)CCtxParams->prefetchCDictTables;
1142
+ break;
1143
+ case ZSTD_c_enableSeqProducerFallback:
1144
+ *value = CCtxParams->enableMatchFinderFallback;
1145
+ break;
1146
+ case ZSTD_c_maxBlockSize:
1147
+ *value = (int)CCtxParams->maxBlockSize;
1148
+ break;
1149
+ case ZSTD_c_searchForExternalRepcodes:
1150
+ *value = (int)CCtxParams->searchForExternalRepcodes;
1151
+ break;
1052
1152
  default: RETURN_ERROR(parameter_unsupported, "unknown parameter");
1053
1153
  }
1054
1154
  return 0;
@@ -1075,9 +1175,24 @@ size_t ZSTD_CCtx_setParametersUsingCCtxParams(
1075
1175
  return 0;
1076
1176
  }
1077
1177
 
1178
+ size_t ZSTD_CCtx_setCParams(ZSTD_CCtx* cctx, ZSTD_compressionParameters cparams)
1179
+ {
1180
+ DEBUGLOG(4, "ZSTD_CCtx_setCParams");
1181
+ assert(cctx != NULL);
1182
+ if (cctx->streamStage != zcss_init) {
1183
+ /* All parameters in @cparams are allowed to be updated during MT compression.
1184
+ * This must be signaled, so that MT compression picks up the changes */
1185
+ cctx->cParamsChanged = 1;
1186
+ }
1187
+ /* only update if parameters are valid */
1188
+ FORWARD_IF_ERROR(ZSTD_checkCParams(cparams), "");
1189
+ cctx->requestedParams.cParams = cparams;
1190
+ return 0;
1191
+ }
1192
+
1078
1193
  size_t ZSTD_CCtx_setPledgedSrcSize(ZSTD_CCtx* cctx, unsigned long long pledgedSrcSize)
1079
1194
  {
1080
- DEBUGLOG(4, "ZSTD_CCtx_setPledgedSrcSize to %u bytes", (U32)pledgedSrcSize);
1195
+ DEBUGLOG(4, "ZSTD_CCtx_setPledgedSrcSize to %llu bytes", pledgedSrcSize);
1081
1196
  RETURN_ERROR_IF(cctx->streamStage != zcss_init, stage_wrong,
1082
1197
  "Can't set pledgedSrcSize when not in init stage.");
1083
1198
  cctx->pledgedSrcSizePlusOne = pledgedSrcSize+1;
@@ -1220,6 +1335,7 @@ size_t ZSTD_CCtx_reset(ZSTD_CCtx* cctx, ZSTD_ResetDirective reset)
1220
1335
  RETURN_ERROR_IF(cctx->streamStage != zcss_init, stage_wrong,
1221
1336
  "Can't reset parameters only when not in init stage.");
1222
1337
  ZSTD_clearAllDicts(cctx);
1338
+ ZSTD_memset(&cctx->externalMatchCtx, 0, sizeof(cctx->externalMatchCtx));
1223
1339
  return ZSTD_CCtxParams_reset(&cctx->requestedParams);
1224
1340
  }
1225
1341
  return 0;
@@ -1316,7 +1432,8 @@ static ZSTD_compressionParameters
1316
1432
  ZSTD_adjustCParams_internal(ZSTD_compressionParameters cPar,
1317
1433
  unsigned long long srcSize,
1318
1434
  size_t dictSize,
1319
- ZSTD_cParamMode_e mode)
1435
+ ZSTD_cParamMode_e mode,
1436
+ ZSTD_paramSwitch_e useRowMatchFinder)
1320
1437
  {
1321
1438
  const U64 minSrcSize = 513; /* (1<<9) + 1 */
1322
1439
  const U64 maxWindowResize = 1ULL << (ZSTD_WINDOWLOG_MAX-1);
@@ -1350,8 +1467,8 @@ ZSTD_adjustCParams_internal(ZSTD_compressionParameters cPar,
1350
1467
  }
1351
1468
 
1352
1469
  /* resize windowLog if input is small enough, to use less memory */
1353
- if ( (srcSize < maxWindowResize)
1354
- && (dictSize < maxWindowResize) ) {
1470
+ if ( (srcSize <= maxWindowResize)
1471
+ && (dictSize <= maxWindowResize) ) {
1355
1472
  U32 const tSize = (U32)(srcSize + dictSize);
1356
1473
  static U32 const hashSizeMin = 1 << ZSTD_HASHLOG_MIN;
1357
1474
  U32 const srcLog = (tSize < hashSizeMin) ? ZSTD_HASHLOG_MIN :
@@ -1369,6 +1486,42 @@ ZSTD_adjustCParams_internal(ZSTD_compressionParameters cPar,
1369
1486
  if (cPar.windowLog < ZSTD_WINDOWLOG_ABSOLUTEMIN)
1370
1487
  cPar.windowLog = ZSTD_WINDOWLOG_ABSOLUTEMIN; /* minimum wlog required for valid frame header */
1371
1488
 
1489
+ /* We can't use more than 32 bits of hash in total, so that means that we require:
1490
+ * (hashLog + 8) <= 32 && (chainLog + 8) <= 32
1491
+ */
1492
+ if (mode == ZSTD_cpm_createCDict && ZSTD_CDictIndicesAreTagged(&cPar)) {
1493
+ U32 const maxShortCacheHashLog = 32 - ZSTD_SHORT_CACHE_TAG_BITS;
1494
+ if (cPar.hashLog > maxShortCacheHashLog) {
1495
+ cPar.hashLog = maxShortCacheHashLog;
1496
+ }
1497
+ if (cPar.chainLog > maxShortCacheHashLog) {
1498
+ cPar.chainLog = maxShortCacheHashLog;
1499
+ }
1500
+ }
1501
+
1502
+
1503
+ /* At this point, we aren't 100% sure if we are using the row match finder.
1504
+ * Unless it is explicitly disabled, conservatively assume that it is enabled.
1505
+ * In this case it will only be disabled for small sources, so shrinking the
1506
+ * hash log a little bit shouldn't result in any ratio loss.
1507
+ */
1508
+ if (useRowMatchFinder == ZSTD_ps_auto)
1509
+ useRowMatchFinder = ZSTD_ps_enable;
1510
+
1511
+ /* We can't hash more than 32-bits in total. So that means that we require:
1512
+ * (hashLog - rowLog + 8) <= 32
1513
+ */
1514
+ if (ZSTD_rowMatchFinderUsed(cPar.strategy, useRowMatchFinder)) {
1515
+ /* Switch to 32-entry rows if searchLog is 5 (or more) */
1516
+ U32 const rowLog = BOUNDED(4, cPar.searchLog, 6);
1517
+ U32 const maxRowHashLog = 32 - ZSTD_ROW_HASH_TAG_BITS;
1518
+ U32 const maxHashLog = maxRowHashLog + rowLog;
1519
+ assert(cPar.hashLog >= rowLog);
1520
+ if (cPar.hashLog > maxHashLog) {
1521
+ cPar.hashLog = maxHashLog;
1522
+ }
1523
+ }
1524
+
1372
1525
  return cPar;
1373
1526
  }
1374
1527
 
@@ -1379,7 +1532,7 @@ ZSTD_adjustCParams(ZSTD_compressionParameters cPar,
1379
1532
  {
1380
1533
  cPar = ZSTD_clampCParams(cPar); /* resulting cPar is necessarily valid (all parameters within range) */
1381
1534
  if (srcSize == 0) srcSize = ZSTD_CONTENTSIZE_UNKNOWN;
1382
- return ZSTD_adjustCParams_internal(cPar, srcSize, dictSize, ZSTD_cpm_unknown);
1535
+ return ZSTD_adjustCParams_internal(cPar, srcSize, dictSize, ZSTD_cpm_unknown, ZSTD_ps_auto);
1383
1536
  }
1384
1537
 
1385
1538
  static ZSTD_compressionParameters ZSTD_getCParams_internal(int compressionLevel, unsigned long long srcSizeHint, size_t dictSize, ZSTD_cParamMode_e mode);
@@ -1410,7 +1563,7 @@ ZSTD_compressionParameters ZSTD_getCParamsFromCCtxParams(
1410
1563
  ZSTD_overrideCParams(&cParams, &CCtxParams->cParams);
1411
1564
  assert(!ZSTD_checkCParams(cParams));
1412
1565
  /* srcSizeHint == 0 means 0 */
1413
- return ZSTD_adjustCParams_internal(cParams, srcSizeHint, dictSize, mode);
1566
+ return ZSTD_adjustCParams_internal(cParams, srcSizeHint, dictSize, mode, CCtxParams->useRowMatchFinder);
1414
1567
  }
1415
1568
 
1416
1569
  static size_t
@@ -1455,6 +1608,13 @@ ZSTD_sizeof_matchState(const ZSTD_compressionParameters* const cParams,
1455
1608
  return tableSpace + optSpace + slackSpace + lazyAdditionalSpace;
1456
1609
  }
1457
1610
 
1611
+ /* Helper function for calculating memory requirements.
1612
+ * Gives a tighter bound than ZSTD_sequenceBound() by taking minMatch into account. */
1613
+ static size_t ZSTD_maxNbSeq(size_t blockSize, unsigned minMatch, int useSequenceProducer) {
1614
+ U32 const divider = (minMatch==3 || useSequenceProducer) ? 3 : 4;
1615
+ return blockSize / divider;
1616
+ }
1617
+
1458
1618
  static size_t ZSTD_estimateCCtxSize_usingCCtxParams_internal(
1459
1619
  const ZSTD_compressionParameters* cParams,
1460
1620
  const ldmParams_t* ldmParams,
@@ -1462,12 +1622,13 @@ static size_t ZSTD_estimateCCtxSize_usingCCtxParams_internal(
1462
1622
  const ZSTD_paramSwitch_e useRowMatchFinder,
1463
1623
  const size_t buffInSize,
1464
1624
  const size_t buffOutSize,
1465
- const U64 pledgedSrcSize)
1625
+ const U64 pledgedSrcSize,
1626
+ int useSequenceProducer,
1627
+ size_t maxBlockSize)
1466
1628
  {
1467
1629
  size_t const windowSize = (size_t) BOUNDED(1ULL, 1ULL << cParams->windowLog, pledgedSrcSize);
1468
- size_t const blockSize = MIN(ZSTD_BLOCKSIZE_MAX, windowSize);
1469
- U32 const divider = (cParams->minMatch==3) ? 3 : 4;
1470
- size_t const maxNbSeq = blockSize / divider;
1630
+ size_t const blockSize = MIN(ZSTD_resolveMaxBlockSize(maxBlockSize), windowSize);
1631
+ size_t const maxNbSeq = ZSTD_maxNbSeq(blockSize, cParams->minMatch, useSequenceProducer);
1471
1632
  size_t const tokenSpace = ZSTD_cwksp_alloc_size(WILDCOPY_OVERLENGTH + blockSize)
1472
1633
  + ZSTD_cwksp_aligned_alloc_size(maxNbSeq * sizeof(seqDef))
1473
1634
  + 3 * ZSTD_cwksp_alloc_size(maxNbSeq * sizeof(BYTE));
@@ -1486,6 +1647,11 @@ static size_t ZSTD_estimateCCtxSize_usingCCtxParams_internal(
1486
1647
 
1487
1648
  size_t const cctxSpace = isStatic ? ZSTD_cwksp_alloc_size(sizeof(ZSTD_CCtx)) : 0;
1488
1649
 
1650
+ size_t const maxNbExternalSeq = ZSTD_sequenceBound(blockSize);
1651
+ size_t const externalSeqSpace = useSequenceProducer
1652
+ ? ZSTD_cwksp_aligned_alloc_size(maxNbExternalSeq * sizeof(ZSTD_Sequence))
1653
+ : 0;
1654
+
1489
1655
  size_t const neededSpace =
1490
1656
  cctxSpace +
1491
1657
  entropySpace +
@@ -1494,7 +1660,8 @@ static size_t ZSTD_estimateCCtxSize_usingCCtxParams_internal(
1494
1660
  ldmSeqSpace +
1495
1661
  matchStateSize +
1496
1662
  tokenSpace +
1497
- bufferSpace;
1663
+ bufferSpace +
1664
+ externalSeqSpace;
1498
1665
 
1499
1666
  DEBUGLOG(5, "estimate workspace : %u", (U32)neededSpace);
1500
1667
  return neededSpace;
@@ -1512,7 +1679,7 @@ size_t ZSTD_estimateCCtxSize_usingCCtxParams(const ZSTD_CCtx_params* params)
1512
1679
  * be needed. However, we still allocate two 0-sized buffers, which can
1513
1680
  * take space under ASAN. */
1514
1681
  return ZSTD_estimateCCtxSize_usingCCtxParams_internal(
1515
- &cParams, &params->ldmParams, 1, useRowMatchFinder, 0, 0, ZSTD_CONTENTSIZE_UNKNOWN);
1682
+ &cParams, &params->ldmParams, 1, useRowMatchFinder, 0, 0, ZSTD_CONTENTSIZE_UNKNOWN, params->useSequenceProducer, params->maxBlockSize);
1516
1683
  }
1517
1684
 
1518
1685
  size_t ZSTD_estimateCCtxSize_usingCParams(ZSTD_compressionParameters cParams)
@@ -1562,7 +1729,7 @@ size_t ZSTD_estimateCStreamSize_usingCCtxParams(const ZSTD_CCtx_params* params)
1562
1729
  RETURN_ERROR_IF(params->nbWorkers > 0, GENERIC, "Estimate CCtx size is supported for single-threaded compression only.");
1563
1730
  { ZSTD_compressionParameters const cParams =
1564
1731
  ZSTD_getCParamsFromCCtxParams(params, ZSTD_CONTENTSIZE_UNKNOWN, 0, ZSTD_cpm_noAttachDict);
1565
- size_t const blockSize = MIN(ZSTD_BLOCKSIZE_MAX, (size_t)1 << cParams.windowLog);
1732
+ size_t const blockSize = MIN(ZSTD_resolveMaxBlockSize(params->maxBlockSize), (size_t)1 << cParams.windowLog);
1566
1733
  size_t const inBuffSize = (params->inBufferMode == ZSTD_bm_buffered)
1567
1734
  ? ((size_t)1 << cParams.windowLog) + blockSize
1568
1735
  : 0;
@@ -1573,7 +1740,7 @@ size_t ZSTD_estimateCStreamSize_usingCCtxParams(const ZSTD_CCtx_params* params)
1573
1740
 
1574
1741
  return ZSTD_estimateCCtxSize_usingCCtxParams_internal(
1575
1742
  &cParams, &params->ldmParams, 1, useRowMatchFinder, inBuffSize, outBuffSize,
1576
- ZSTD_CONTENTSIZE_UNKNOWN);
1743
+ ZSTD_CONTENTSIZE_UNKNOWN, params->useSequenceProducer, params->maxBlockSize);
1577
1744
  }
1578
1745
  }
1579
1746
 
@@ -1847,6 +2014,7 @@ static size_t ZSTD_resetCCtx_internal(ZSTD_CCtx* zc,
1847
2014
  assert(params->useRowMatchFinder != ZSTD_ps_auto);
1848
2015
  assert(params->useBlockSplitter != ZSTD_ps_auto);
1849
2016
  assert(params->ldmParams.enableLdm != ZSTD_ps_auto);
2017
+ assert(params->maxBlockSize != 0);
1850
2018
  if (params->ldmParams.enableLdm == ZSTD_ps_enable) {
1851
2019
  /* Adjust long distance matching parameters */
1852
2020
  ZSTD_ldm_adjustParameters(&zc->appliedParams.ldmParams, &params->cParams);
@@ -1855,9 +2023,8 @@ static size_t ZSTD_resetCCtx_internal(ZSTD_CCtx* zc,
1855
2023
  }
1856
2024
 
1857
2025
  { size_t const windowSize = MAX(1, (size_t)MIN(((U64)1 << params->cParams.windowLog), pledgedSrcSize));
1858
- size_t const blockSize = MIN(ZSTD_BLOCKSIZE_MAX, windowSize);
1859
- U32 const divider = (params->cParams.minMatch==3) ? 3 : 4;
1860
- size_t const maxNbSeq = blockSize / divider;
2026
+ size_t const blockSize = MIN(params->maxBlockSize, windowSize);
2027
+ size_t const maxNbSeq = ZSTD_maxNbSeq(blockSize, params->cParams.minMatch, params->useSequenceProducer);
1861
2028
  size_t const buffOutSize = (zbuff == ZSTDb_buffered && params->outBufferMode == ZSTD_bm_buffered)
1862
2029
  ? ZSTD_compressBound(blockSize) + 1
1863
2030
  : 0;
@@ -1874,7 +2041,7 @@ static size_t ZSTD_resetCCtx_internal(ZSTD_CCtx* zc,
1874
2041
  size_t const neededSpace =
1875
2042
  ZSTD_estimateCCtxSize_usingCCtxParams_internal(
1876
2043
  &params->cParams, &params->ldmParams, zc->staticSize != 0, params->useRowMatchFinder,
1877
- buffInSize, buffOutSize, pledgedSrcSize);
2044
+ buffInSize, buffOutSize, pledgedSrcSize, params->useSequenceProducer, params->maxBlockSize);
1878
2045
  int resizeWorkspace;
1879
2046
 
1880
2047
  FORWARD_IF_ERROR(neededSpace, "cctx size estimate failed!");
@@ -1917,6 +2084,7 @@ static size_t ZSTD_resetCCtx_internal(ZSTD_CCtx* zc,
1917
2084
 
1918
2085
  /* init params */
1919
2086
  zc->blockState.matchState.cParams = params->cParams;
2087
+ zc->blockState.matchState.prefetchCDictTables = params->prefetchCDictTables == ZSTD_ps_enable;
1920
2088
  zc->pledgedSrcSizePlusOne = pledgedSrcSize+1;
1921
2089
  zc->consumedSrcSize = 0;
1922
2090
  zc->producedCSize = 0;
@@ -1986,6 +2154,14 @@ static size_t ZSTD_resetCCtx_internal(ZSTD_CCtx* zc,
1986
2154
  zc->ldmState.loadedDictEnd = 0;
1987
2155
  }
1988
2156
 
2157
+ /* reserve space for block-level external sequences */
2158
+ if (params->useSequenceProducer) {
2159
+ size_t const maxNbExternalSeq = ZSTD_sequenceBound(blockSize);
2160
+ zc->externalMatchCtx.seqBufferCapacity = maxNbExternalSeq;
2161
+ zc->externalMatchCtx.seqBuffer =
2162
+ (ZSTD_Sequence*)ZSTD_cwksp_reserve_aligned(ws, maxNbExternalSeq * sizeof(ZSTD_Sequence));
2163
+ }
2164
+
1989
2165
  DEBUGLOG(3, "wksp: finished allocating, %zd bytes remain available", ZSTD_cwksp_available_space(ws));
1990
2166
  assert(ZSTD_cwksp_estimated_space_within_bounds(ws, neededSpace, resizeWorkspace));
1991
2167
 
@@ -2059,7 +2235,8 @@ ZSTD_resetCCtx_byAttachingCDict(ZSTD_CCtx* cctx,
2059
2235
  }
2060
2236
 
2061
2237
  params.cParams = ZSTD_adjustCParams_internal(adjusted_cdict_cParams, pledgedSrcSize,
2062
- cdict->dictContentSize, ZSTD_cpm_attachDict);
2238
+ cdict->dictContentSize, ZSTD_cpm_attachDict,
2239
+ params.useRowMatchFinder);
2063
2240
  params.cParams.windowLog = windowLog;
2064
2241
  params.useRowMatchFinder = cdict->useRowMatchFinder; /* cdict overrides */
2065
2242
  FORWARD_IF_ERROR(ZSTD_resetCCtx_internal(cctx, &params, pledgedSrcSize,
@@ -2098,6 +2275,22 @@ ZSTD_resetCCtx_byAttachingCDict(ZSTD_CCtx* cctx,
2098
2275
  return 0;
2099
2276
  }
2100
2277
 
2278
+ static void ZSTD_copyCDictTableIntoCCtx(U32* dst, U32 const* src, size_t tableSize,
2279
+ ZSTD_compressionParameters const* cParams) {
2280
+ if (ZSTD_CDictIndicesAreTagged(cParams)){
2281
+ /* Remove tags from the CDict table if they are present.
2282
+ * See docs on "short cache" in zstd_compress_internal.h for context. */
2283
+ size_t i;
2284
+ for (i = 0; i < tableSize; i++) {
2285
+ U32 const taggedIndex = src[i];
2286
+ U32 const index = taggedIndex >> ZSTD_SHORT_CACHE_TAG_BITS;
2287
+ dst[i] = index;
2288
+ }
2289
+ } else {
2290
+ ZSTD_memcpy(dst, src, tableSize * sizeof(U32));
2291
+ }
2292
+ }
2293
+
2101
2294
  static size_t ZSTD_resetCCtx_byCopyingCDict(ZSTD_CCtx* cctx,
2102
2295
  const ZSTD_CDict* cdict,
2103
2296
  ZSTD_CCtx_params params,
@@ -2133,14 +2326,15 @@ static size_t ZSTD_resetCCtx_byCopyingCDict(ZSTD_CCtx* cctx,
2133
2326
  : 0;
2134
2327
  size_t const hSize = (size_t)1 << cdict_cParams->hashLog;
2135
2328
 
2136
- ZSTD_memcpy(cctx->blockState.matchState.hashTable,
2137
- cdict->matchState.hashTable,
2138
- hSize * sizeof(U32));
2329
+ ZSTD_copyCDictTableIntoCCtx(cctx->blockState.matchState.hashTable,
2330
+ cdict->matchState.hashTable,
2331
+ hSize, cdict_cParams);
2332
+
2139
2333
  /* Do not copy cdict's chainTable if cctx has parameters such that it would not use chainTable */
2140
2334
  if (ZSTD_allocateChainTable(cctx->appliedParams.cParams.strategy, cctx->appliedParams.useRowMatchFinder, 0 /* forDDSDict */)) {
2141
- ZSTD_memcpy(cctx->blockState.matchState.chainTable,
2142
- cdict->matchState.chainTable,
2143
- chainSize * sizeof(U32));
2335
+ ZSTD_copyCDictTableIntoCCtx(cctx->blockState.matchState.chainTable,
2336
+ cdict->matchState.chainTable,
2337
+ chainSize, cdict_cParams);
2144
2338
  }
2145
2339
  /* copy tag table */
2146
2340
  if (ZSTD_rowMatchFinderUsed(cdict_cParams->strategy, cdict->useRowMatchFinder)) {
@@ -2226,6 +2420,7 @@ static size_t ZSTD_copyCCtx_internal(ZSTD_CCtx* dstCCtx,
2226
2420
  params.useBlockSplitter = srcCCtx->appliedParams.useBlockSplitter;
2227
2421
  params.ldmParams = srcCCtx->appliedParams.ldmParams;
2228
2422
  params.fParams = fParams;
2423
+ params.maxBlockSize = srcCCtx->appliedParams.maxBlockSize;
2229
2424
  ZSTD_resetCCtx_internal(dstCCtx, &params, pledgedSrcSize,
2230
2425
  /* loadedDictSize */ 0,
2231
2426
  ZSTDcrp_leaveDirty, zbuff);
@@ -2385,7 +2580,7 @@ static void ZSTD_reduceIndex (ZSTD_matchState_t* ms, ZSTD_CCtx_params const* par
2385
2580
 
2386
2581
  /* See doc/zstd_compression_format.md for detailed format description */
2387
2582
 
2388
- void ZSTD_seqToCodes(const seqStore_t* seqStorePtr)
2583
+ int ZSTD_seqToCodes(const seqStore_t* seqStorePtr)
2389
2584
  {
2390
2585
  const seqDef* const sequences = seqStorePtr->sequencesStart;
2391
2586
  BYTE* const llCodeTable = seqStorePtr->llCode;
@@ -2393,18 +2588,24 @@ void ZSTD_seqToCodes(const seqStore_t* seqStorePtr)
2393
2588
  BYTE* const mlCodeTable = seqStorePtr->mlCode;
2394
2589
  U32 const nbSeq = (U32)(seqStorePtr->sequences - seqStorePtr->sequencesStart);
2395
2590
  U32 u;
2591
+ int longOffsets = 0;
2396
2592
  assert(nbSeq <= seqStorePtr->maxNbSeq);
2397
2593
  for (u=0; u<nbSeq; u++) {
2398
2594
  U32 const llv = sequences[u].litLength;
2595
+ U32 const ofCode = ZSTD_highbit32(sequences[u].offBase);
2399
2596
  U32 const mlv = sequences[u].mlBase;
2400
2597
  llCodeTable[u] = (BYTE)ZSTD_LLcode(llv);
2401
- ofCodeTable[u] = (BYTE)ZSTD_highbit32(sequences[u].offBase);
2598
+ ofCodeTable[u] = (BYTE)ofCode;
2402
2599
  mlCodeTable[u] = (BYTE)ZSTD_MLcode(mlv);
2600
+ assert(!(MEM_64bits() && ofCode >= STREAM_ACCUMULATOR_MIN));
2601
+ if (MEM_32bits() && ofCode >= STREAM_ACCUMULATOR_MIN)
2602
+ longOffsets = 1;
2403
2603
  }
2404
2604
  if (seqStorePtr->longLengthType==ZSTD_llt_literalLength)
2405
2605
  llCodeTable[seqStorePtr->longLengthPos] = MaxLL;
2406
2606
  if (seqStorePtr->longLengthType==ZSTD_llt_matchLength)
2407
2607
  mlCodeTable[seqStorePtr->longLengthPos] = MaxML;
2608
+ return longOffsets;
2408
2609
  }
2409
2610
 
2410
2611
  /* ZSTD_useTargetCBlockSize():
@@ -2438,6 +2639,7 @@ typedef struct {
2438
2639
  U32 MLtype;
2439
2640
  size_t size;
2440
2641
  size_t lastCountSize; /* Accounts for bug in 1.3.4. More detail in ZSTD_entropyCompressSeqStore_internal() */
2642
+ int longOffsets;
2441
2643
  } ZSTD_symbolEncodingTypeStats_t;
2442
2644
 
2443
2645
  /* ZSTD_buildSequencesStatistics():
@@ -2448,11 +2650,13 @@ typedef struct {
2448
2650
  * entropyWkspSize must be of size at least ENTROPY_WORKSPACE_SIZE - (MaxSeq + 1)*sizeof(U32)
2449
2651
  */
2450
2652
  static ZSTD_symbolEncodingTypeStats_t
2451
- ZSTD_buildSequencesStatistics(seqStore_t* seqStorePtr, size_t nbSeq,
2452
- const ZSTD_fseCTables_t* prevEntropy, ZSTD_fseCTables_t* nextEntropy,
2453
- BYTE* dst, const BYTE* const dstEnd,
2454
- ZSTD_strategy strategy, unsigned* countWorkspace,
2455
- void* entropyWorkspace, size_t entropyWkspSize) {
2653
+ ZSTD_buildSequencesStatistics(
2654
+ const seqStore_t* seqStorePtr, size_t nbSeq,
2655
+ const ZSTD_fseCTables_t* prevEntropy, ZSTD_fseCTables_t* nextEntropy,
2656
+ BYTE* dst, const BYTE* const dstEnd,
2657
+ ZSTD_strategy strategy, unsigned* countWorkspace,
2658
+ void* entropyWorkspace, size_t entropyWkspSize)
2659
+ {
2456
2660
  BYTE* const ostart = dst;
2457
2661
  const BYTE* const oend = dstEnd;
2458
2662
  BYTE* op = ostart;
@@ -2466,7 +2670,7 @@ ZSTD_buildSequencesStatistics(seqStore_t* seqStorePtr, size_t nbSeq,
2466
2670
 
2467
2671
  stats.lastCountSize = 0;
2468
2672
  /* convert length/distances into codes */
2469
- ZSTD_seqToCodes(seqStorePtr);
2673
+ stats.longOffsets = ZSTD_seqToCodes(seqStorePtr);
2470
2674
  assert(op <= oend);
2471
2675
  assert(nbSeq != 0); /* ZSTD_selectEncodingType() divides by nbSeq */
2472
2676
  /* build CTable for Literal Lengths */
@@ -2571,22 +2775,22 @@ ZSTD_buildSequencesStatistics(seqStore_t* seqStorePtr, size_t nbSeq,
2571
2775
  */
2572
2776
  #define SUSPECT_UNCOMPRESSIBLE_LITERAL_RATIO 20
2573
2777
  MEM_STATIC size_t
2574
- ZSTD_entropyCompressSeqStore_internal(seqStore_t* seqStorePtr,
2575
- const ZSTD_entropyCTables_t* prevEntropy,
2576
- ZSTD_entropyCTables_t* nextEntropy,
2577
- const ZSTD_CCtx_params* cctxParams,
2578
- void* dst, size_t dstCapacity,
2579
- void* entropyWorkspace, size_t entropyWkspSize,
2580
- const int bmi2)
2778
+ ZSTD_entropyCompressSeqStore_internal(
2779
+ const seqStore_t* seqStorePtr,
2780
+ const ZSTD_entropyCTables_t* prevEntropy,
2781
+ ZSTD_entropyCTables_t* nextEntropy,
2782
+ const ZSTD_CCtx_params* cctxParams,
2783
+ void* dst, size_t dstCapacity,
2784
+ void* entropyWorkspace, size_t entropyWkspSize,
2785
+ const int bmi2)
2581
2786
  {
2582
- const int longOffsets = cctxParams->cParams.windowLog > STREAM_ACCUMULATOR_MIN;
2583
2787
  ZSTD_strategy const strategy = cctxParams->cParams.strategy;
2584
2788
  unsigned* count = (unsigned*)entropyWorkspace;
2585
2789
  FSE_CTable* CTable_LitLength = nextEntropy->fse.litlengthCTable;
2586
2790
  FSE_CTable* CTable_OffsetBits = nextEntropy->fse.offcodeCTable;
2587
2791
  FSE_CTable* CTable_MatchLength = nextEntropy->fse.matchlengthCTable;
2588
2792
  const seqDef* const sequences = seqStorePtr->sequencesStart;
2589
- const size_t nbSeq = seqStorePtr->sequences - seqStorePtr->sequencesStart;
2793
+ const size_t nbSeq = (size_t)(seqStorePtr->sequences - seqStorePtr->sequencesStart);
2590
2794
  const BYTE* const ofCodeTable = seqStorePtr->ofCode;
2591
2795
  const BYTE* const llCodeTable = seqStorePtr->llCode;
2592
2796
  const BYTE* const mlCodeTable = seqStorePtr->mlCode;
@@ -2594,29 +2798,31 @@ ZSTD_entropyCompressSeqStore_internal(seqStore_t* seqStorePtr,
2594
2798
  BYTE* const oend = ostart + dstCapacity;
2595
2799
  BYTE* op = ostart;
2596
2800
  size_t lastCountSize;
2801
+ int longOffsets = 0;
2597
2802
 
2598
2803
  entropyWorkspace = count + (MaxSeq + 1);
2599
2804
  entropyWkspSize -= (MaxSeq + 1) * sizeof(*count);
2600
2805
 
2601
- DEBUGLOG(4, "ZSTD_entropyCompressSeqStore_internal (nbSeq=%zu)", nbSeq);
2806
+ DEBUGLOG(5, "ZSTD_entropyCompressSeqStore_internal (nbSeq=%zu, dstCapacity=%zu)", nbSeq, dstCapacity);
2602
2807
  ZSTD_STATIC_ASSERT(HUF_WORKSPACE_SIZE >= (1<<MAX(MLFSELog,LLFSELog)));
2603
2808
  assert(entropyWkspSize >= HUF_WORKSPACE_SIZE);
2604
2809
 
2605
2810
  /* Compress literals */
2606
2811
  { const BYTE* const literals = seqStorePtr->litStart;
2607
- size_t const numSequences = seqStorePtr->sequences - seqStorePtr->sequencesStart;
2608
- size_t const numLiterals = seqStorePtr->lit - seqStorePtr->litStart;
2812
+ size_t const numSequences = (size_t)(seqStorePtr->sequences - seqStorePtr->sequencesStart);
2813
+ size_t const numLiterals = (size_t)(seqStorePtr->lit - seqStorePtr->litStart);
2609
2814
  /* Base suspicion of uncompressibility on ratio of literals to sequences */
2610
2815
  unsigned const suspectUncompressible = (numSequences == 0) || (numLiterals / numSequences >= SUSPECT_UNCOMPRESSIBLE_LITERAL_RATIO);
2611
2816
  size_t const litSize = (size_t)(seqStorePtr->lit - literals);
2817
+
2612
2818
  size_t const cSize = ZSTD_compressLiterals(
2613
- &prevEntropy->huf, &nextEntropy->huf,
2614
- cctxParams->cParams.strategy,
2615
- ZSTD_literalsCompressionIsDisabled(cctxParams),
2616
2819
  op, dstCapacity,
2617
2820
  literals, litSize,
2618
2821
  entropyWorkspace, entropyWkspSize,
2619
- bmi2, suspectUncompressible);
2822
+ &prevEntropy->huf, &nextEntropy->huf,
2823
+ cctxParams->cParams.strategy,
2824
+ ZSTD_literalsCompressionIsDisabled(cctxParams),
2825
+ suspectUncompressible, bmi2);
2620
2826
  FORWARD_IF_ERROR(cSize, "ZSTD_compressLiterals failed");
2621
2827
  assert(cSize <= dstCapacity);
2622
2828
  op += cSize;
@@ -2642,11 +2848,10 @@ ZSTD_entropyCompressSeqStore_internal(seqStore_t* seqStorePtr,
2642
2848
  ZSTD_memcpy(&nextEntropy->fse, &prevEntropy->fse, sizeof(prevEntropy->fse));
2643
2849
  return (size_t)(op - ostart);
2644
2850
  }
2645
- {
2646
- ZSTD_symbolEncodingTypeStats_t stats;
2647
- BYTE* seqHead = op++;
2851
+ { BYTE* const seqHead = op++;
2648
2852
  /* build stats for sequences */
2649
- stats = ZSTD_buildSequencesStatistics(seqStorePtr, nbSeq,
2853
+ const ZSTD_symbolEncodingTypeStats_t stats =
2854
+ ZSTD_buildSequencesStatistics(seqStorePtr, nbSeq,
2650
2855
  &prevEntropy->fse, &nextEntropy->fse,
2651
2856
  op, oend,
2652
2857
  strategy, count,
@@ -2655,6 +2860,7 @@ ZSTD_entropyCompressSeqStore_internal(seqStore_t* seqStorePtr,
2655
2860
  *seqHead = (BYTE)((stats.LLtype<<6) + (stats.Offtype<<4) + (stats.MLtype<<2));
2656
2861
  lastCountSize = stats.lastCountSize;
2657
2862
  op += stats.size;
2863
+ longOffsets = stats.longOffsets;
2658
2864
  }
2659
2865
 
2660
2866
  { size_t const bitstreamSize = ZSTD_encodeSequences(
@@ -2689,14 +2895,15 @@ ZSTD_entropyCompressSeqStore_internal(seqStore_t* seqStorePtr,
2689
2895
  }
2690
2896
 
2691
2897
  MEM_STATIC size_t
2692
- ZSTD_entropyCompressSeqStore(seqStore_t* seqStorePtr,
2693
- const ZSTD_entropyCTables_t* prevEntropy,
2694
- ZSTD_entropyCTables_t* nextEntropy,
2695
- const ZSTD_CCtx_params* cctxParams,
2696
- void* dst, size_t dstCapacity,
2697
- size_t srcSize,
2698
- void* entropyWorkspace, size_t entropyWkspSize,
2699
- int bmi2)
2898
+ ZSTD_entropyCompressSeqStore(
2899
+ const seqStore_t* seqStorePtr,
2900
+ const ZSTD_entropyCTables_t* prevEntropy,
2901
+ ZSTD_entropyCTables_t* nextEntropy,
2902
+ const ZSTD_CCtx_params* cctxParams,
2903
+ void* dst, size_t dstCapacity,
2904
+ size_t srcSize,
2905
+ void* entropyWorkspace, size_t entropyWkspSize,
2906
+ int bmi2)
2700
2907
  {
2701
2908
  size_t const cSize = ZSTD_entropyCompressSeqStore_internal(
2702
2909
  seqStorePtr, prevEntropy, nextEntropy, cctxParams,
@@ -2706,15 +2913,21 @@ ZSTD_entropyCompressSeqStore(seqStore_t* seqStorePtr,
2706
2913
  /* When srcSize <= dstCapacity, there is enough space to write a raw uncompressed block.
2707
2914
  * Since we ran out of space, block must be not compressible, so fall back to raw uncompressed block.
2708
2915
  */
2709
- if ((cSize == ERROR(dstSize_tooSmall)) & (srcSize <= dstCapacity))
2916
+ if ((cSize == ERROR(dstSize_tooSmall)) & (srcSize <= dstCapacity)) {
2917
+ DEBUGLOG(4, "not enough dstCapacity (%zu) for ZSTD_entropyCompressSeqStore_internal()=> do not compress block", dstCapacity);
2710
2918
  return 0; /* block not compressed */
2919
+ }
2711
2920
  FORWARD_IF_ERROR(cSize, "ZSTD_entropyCompressSeqStore_internal failed");
2712
2921
 
2713
2922
  /* Check compressibility */
2714
2923
  { size_t const maxCSize = srcSize - ZSTD_minGain(srcSize, cctxParams->cParams.strategy);
2715
2924
  if (cSize >= maxCSize) return 0; /* block not compressed */
2716
2925
  }
2717
- DEBUGLOG(4, "ZSTD_entropyCompressSeqStore() cSize: %zu", cSize);
2926
+ DEBUGLOG(5, "ZSTD_entropyCompressSeqStore() cSize: %zu", cSize);
2927
+ /* libzstd decoder before > v1.5.4 is not compatible with compressed blocks of size ZSTD_BLOCKSIZE_MAX exactly.
2928
+ * This restriction is indirectly already fulfilled by respecting ZSTD_minGain() condition above.
2929
+ */
2930
+ assert(cSize < ZSTD_BLOCKSIZE_MAX);
2718
2931
  return cSize;
2719
2932
  }
2720
2933
 
@@ -2809,6 +3022,72 @@ void ZSTD_resetSeqStore(seqStore_t* ssPtr)
2809
3022
  ssPtr->longLengthType = ZSTD_llt_none;
2810
3023
  }
2811
3024
 
3025
+ /* ZSTD_postProcessSequenceProducerResult() :
3026
+ * Validates and post-processes sequences obtained through the external matchfinder API:
3027
+ * - Checks whether nbExternalSeqs represents an error condition.
3028
+ * - Appends a block delimiter to outSeqs if one is not already present.
3029
+ * See zstd.h for context regarding block delimiters.
3030
+ * Returns the number of sequences after post-processing, or an error code. */
3031
+ static size_t ZSTD_postProcessSequenceProducerResult(
3032
+ ZSTD_Sequence* outSeqs, size_t nbExternalSeqs, size_t outSeqsCapacity, size_t srcSize
3033
+ ) {
3034
+ RETURN_ERROR_IF(
3035
+ nbExternalSeqs > outSeqsCapacity,
3036
+ sequenceProducer_failed,
3037
+ "External sequence producer returned error code %lu",
3038
+ (unsigned long)nbExternalSeqs
3039
+ );
3040
+
3041
+ RETURN_ERROR_IF(
3042
+ nbExternalSeqs == 0 && srcSize > 0,
3043
+ sequenceProducer_failed,
3044
+ "Got zero sequences from external sequence producer for a non-empty src buffer!"
3045
+ );
3046
+
3047
+ if (srcSize == 0) {
3048
+ ZSTD_memset(&outSeqs[0], 0, sizeof(ZSTD_Sequence));
3049
+ return 1;
3050
+ }
3051
+
3052
+ {
3053
+ ZSTD_Sequence const lastSeq = outSeqs[nbExternalSeqs - 1];
3054
+
3055
+ /* We can return early if lastSeq is already a block delimiter. */
3056
+ if (lastSeq.offset == 0 && lastSeq.matchLength == 0) {
3057
+ return nbExternalSeqs;
3058
+ }
3059
+
3060
+ /* This error condition is only possible if the external matchfinder
3061
+ * produced an invalid parse, by definition of ZSTD_sequenceBound(). */
3062
+ RETURN_ERROR_IF(
3063
+ nbExternalSeqs == outSeqsCapacity,
3064
+ sequenceProducer_failed,
3065
+ "nbExternalSeqs == outSeqsCapacity but lastSeq is not a block delimiter!"
3066
+ );
3067
+
3068
+ /* lastSeq is not a block delimiter, so we need to append one. */
3069
+ ZSTD_memset(&outSeqs[nbExternalSeqs], 0, sizeof(ZSTD_Sequence));
3070
+ return nbExternalSeqs + 1;
3071
+ }
3072
+ }
3073
+
3074
+ /* ZSTD_fastSequenceLengthSum() :
3075
+ * Returns sum(litLen) + sum(matchLen) + lastLits for *seqBuf*.
3076
+ * Similar to another function in zstd_compress.c (determine_blockSize),
3077
+ * except it doesn't check for a block delimiter to end summation.
3078
+ * Removing the early exit allows the compiler to auto-vectorize (https://godbolt.org/z/cY1cajz9P).
3079
+ * This function can be deleted and replaced by determine_blockSize after we resolve issue #3456. */
3080
+ static size_t ZSTD_fastSequenceLengthSum(ZSTD_Sequence const* seqBuf, size_t seqBufSize) {
3081
+ size_t matchLenSum, litLenSum, i;
3082
+ matchLenSum = 0;
3083
+ litLenSum = 0;
3084
+ for (i = 0; i < seqBufSize; i++) {
3085
+ litLenSum += seqBuf[i].litLength;
3086
+ matchLenSum += seqBuf[i].matchLength;
3087
+ }
3088
+ return litLenSum + matchLenSum;
3089
+ }
3090
+
2812
3091
  typedef enum { ZSTDbss_compress, ZSTDbss_noCompress } ZSTD_buildSeqStore_e;
2813
3092
 
2814
3093
  static size_t ZSTD_buildSeqStore(ZSTD_CCtx* zc, const void* src, size_t srcSize)
@@ -2818,7 +3097,9 @@ static size_t ZSTD_buildSeqStore(ZSTD_CCtx* zc, const void* src, size_t srcSize)
2818
3097
  assert(srcSize <= ZSTD_BLOCKSIZE_MAX);
2819
3098
  /* Assert that we have correctly flushed the ctx params into the ms's copy */
2820
3099
  ZSTD_assertEqualCParams(zc->appliedParams.cParams, ms->cParams);
2821
- if (srcSize < MIN_CBLOCK_SIZE+ZSTD_blockHeaderSize+1) {
3100
+ /* TODO: See 3090. We reduced MIN_CBLOCK_SIZE from 3 to 2 so to compensate we are adding
3101
+ * additional 1. We need to revisit and change this logic to be more consistent */
3102
+ if (srcSize < MIN_CBLOCK_SIZE+ZSTD_blockHeaderSize+1+1) {
2822
3103
  if (zc->appliedParams.cParams.strategy >= ZSTD_btopt) {
2823
3104
  ZSTD_ldm_skipRawSeqStoreBytes(&zc->externSeqStore, srcSize);
2824
3105
  } else {
@@ -2854,6 +3135,15 @@ static size_t ZSTD_buildSeqStore(ZSTD_CCtx* zc, const void* src, size_t srcSize)
2854
3135
  }
2855
3136
  if (zc->externSeqStore.pos < zc->externSeqStore.size) {
2856
3137
  assert(zc->appliedParams.ldmParams.enableLdm == ZSTD_ps_disable);
3138
+
3139
+ /* External matchfinder + LDM is technically possible, just not implemented yet.
3140
+ * We need to revisit soon and implement it. */
3141
+ RETURN_ERROR_IF(
3142
+ zc->appliedParams.useSequenceProducer,
3143
+ parameter_combination_unsupported,
3144
+ "Long-distance matching with external sequence producer enabled is not currently supported."
3145
+ );
3146
+
2857
3147
  /* Updates ldmSeqStore.pos */
2858
3148
  lastLLSize =
2859
3149
  ZSTD_ldm_blockCompress(&zc->externSeqStore,
@@ -2865,6 +3155,14 @@ static size_t ZSTD_buildSeqStore(ZSTD_CCtx* zc, const void* src, size_t srcSize)
2865
3155
  } else if (zc->appliedParams.ldmParams.enableLdm == ZSTD_ps_enable) {
2866
3156
  rawSeqStore_t ldmSeqStore = kNullRawSeqStore;
2867
3157
 
3158
+ /* External matchfinder + LDM is technically possible, just not implemented yet.
3159
+ * We need to revisit soon and implement it. */
3160
+ RETURN_ERROR_IF(
3161
+ zc->appliedParams.useSequenceProducer,
3162
+ parameter_combination_unsupported,
3163
+ "Long-distance matching with external sequence producer enabled is not currently supported."
3164
+ );
3165
+
2868
3166
  ldmSeqStore.seq = zc->ldmSequences;
2869
3167
  ldmSeqStore.capacity = zc->maxNbLdmSequences;
2870
3168
  /* Updates ldmSeqStore.size */
@@ -2879,7 +3177,68 @@ static size_t ZSTD_buildSeqStore(ZSTD_CCtx* zc, const void* src, size_t srcSize)
2879
3177
  zc->appliedParams.useRowMatchFinder,
2880
3178
  src, srcSize);
2881
3179
  assert(ldmSeqStore.pos == ldmSeqStore.size);
2882
- } else { /* not long range mode */
3180
+ } else if (zc->appliedParams.useSequenceProducer) {
3181
+ assert(
3182
+ zc->externalMatchCtx.seqBufferCapacity >= ZSTD_sequenceBound(srcSize)
3183
+ );
3184
+ assert(zc->externalMatchCtx.mFinder != NULL);
3185
+
3186
+ { U32 const windowSize = (U32)1 << zc->appliedParams.cParams.windowLog;
3187
+
3188
+ size_t const nbExternalSeqs = (zc->externalMatchCtx.mFinder)(
3189
+ zc->externalMatchCtx.mState,
3190
+ zc->externalMatchCtx.seqBuffer,
3191
+ zc->externalMatchCtx.seqBufferCapacity,
3192
+ src, srcSize,
3193
+ NULL, 0, /* dict and dictSize, currently not supported */
3194
+ zc->appliedParams.compressionLevel,
3195
+ windowSize
3196
+ );
3197
+
3198
+ size_t const nbPostProcessedSeqs = ZSTD_postProcessSequenceProducerResult(
3199
+ zc->externalMatchCtx.seqBuffer,
3200
+ nbExternalSeqs,
3201
+ zc->externalMatchCtx.seqBufferCapacity,
3202
+ srcSize
3203
+ );
3204
+
3205
+ /* Return early if there is no error, since we don't need to worry about last literals */
3206
+ if (!ZSTD_isError(nbPostProcessedSeqs)) {
3207
+ ZSTD_sequencePosition seqPos = {0,0,0};
3208
+ size_t const seqLenSum = ZSTD_fastSequenceLengthSum(zc->externalMatchCtx.seqBuffer, nbPostProcessedSeqs);
3209
+ RETURN_ERROR_IF(seqLenSum > srcSize, externalSequences_invalid, "External sequences imply too large a block!");
3210
+ FORWARD_IF_ERROR(
3211
+ ZSTD_copySequencesToSeqStoreExplicitBlockDelim(
3212
+ zc, &seqPos,
3213
+ zc->externalMatchCtx.seqBuffer, nbPostProcessedSeqs,
3214
+ src, srcSize,
3215
+ zc->appliedParams.searchForExternalRepcodes
3216
+ ),
3217
+ "Failed to copy external sequences to seqStore!"
3218
+ );
3219
+ ms->ldmSeqStore = NULL;
3220
+ DEBUGLOG(5, "Copied %lu sequences from external sequence producer to internal seqStore.", (unsigned long)nbExternalSeqs);
3221
+ return ZSTDbss_compress;
3222
+ }
3223
+
3224
+ /* Propagate the error if fallback is disabled */
3225
+ if (!zc->appliedParams.enableMatchFinderFallback) {
3226
+ return nbPostProcessedSeqs;
3227
+ }
3228
+
3229
+ /* Fallback to software matchfinder */
3230
+ { ZSTD_blockCompressor const blockCompressor = ZSTD_selectBlockCompressor(zc->appliedParams.cParams.strategy,
3231
+ zc->appliedParams.useRowMatchFinder,
3232
+ dictMode);
3233
+ ms->ldmSeqStore = NULL;
3234
+ DEBUGLOG(
3235
+ 5,
3236
+ "External sequence producer returned error code %lu. Falling back to internal parser.",
3237
+ (unsigned long)nbExternalSeqs
3238
+ );
3239
+ lastLLSize = blockCompressor(ms, &zc->seqStore, zc->blockState.nextCBlock->rep, src, srcSize);
3240
+ } }
3241
+ } else { /* not long range mode and no external matchfinder */
2883
3242
  ZSTD_blockCompressor const blockCompressor = ZSTD_selectBlockCompressor(zc->appliedParams.cParams.strategy,
2884
3243
  zc->appliedParams.useRowMatchFinder,
2885
3244
  dictMode);
@@ -2940,7 +3299,7 @@ static void ZSTD_copyBlockSequences(ZSTD_CCtx* zc)
2940
3299
  /* seqStoreSeqs[i].offset == offCode+1, and ZSTD_updateRep() expects offCode
2941
3300
  so we provide seqStoreSeqs[i].offset - 1 */
2942
3301
  ZSTD_updateRep(updatedRepcodes.rep,
2943
- seqStoreSeqs[i].offBase - 1,
3302
+ seqStoreSeqs[i].offBase,
2944
3303
  seqStoreSeqs[i].litLength == 0);
2945
3304
  literalsRead += outSeqs[i].litLength;
2946
3305
  }
@@ -2956,6 +3315,10 @@ static void ZSTD_copyBlockSequences(ZSTD_CCtx* zc)
2956
3315
  zc->seqCollector.seqIndex += seqStoreSeqSize;
2957
3316
  }
2958
3317
 
3318
+ size_t ZSTD_sequenceBound(size_t srcSize) {
3319
+ return (srcSize / ZSTD_MINMATCH_MIN) + 1;
3320
+ }
3321
+
2959
3322
  size_t ZSTD_generateSequences(ZSTD_CCtx* zc, ZSTD_Sequence* outSeqs,
2960
3323
  size_t outSeqsSize, const void* src, size_t srcSize)
2961
3324
  {
@@ -3001,19 +3364,17 @@ static int ZSTD_isRLE(const BYTE* src, size_t length) {
3001
3364
  const size_t unrollMask = unrollSize - 1;
3002
3365
  const size_t prefixLength = length & unrollMask;
3003
3366
  size_t i;
3004
- size_t u;
3005
3367
  if (length == 1) return 1;
3006
3368
  /* Check if prefix is RLE first before using unrolled loop */
3007
3369
  if (prefixLength && ZSTD_count(ip+1, ip, ip+prefixLength) != prefixLength-1) {
3008
3370
  return 0;
3009
3371
  }
3010
3372
  for (i = prefixLength; i != length; i += unrollSize) {
3373
+ size_t u;
3011
3374
  for (u = 0; u < unrollSize; u += sizeof(size_t)) {
3012
3375
  if (MEM_readST(ip + i + u) != valueST) {
3013
3376
  return 0;
3014
- }
3015
- }
3016
- }
3377
+ } } }
3017
3378
  return 1;
3018
3379
  }
3019
3380
 
@@ -3029,7 +3390,8 @@ static int ZSTD_maybeRLE(seqStore_t const* seqStore)
3029
3390
  return nbSeqs < 4 && nbLits < 10;
3030
3391
  }
3031
3392
 
3032
- static void ZSTD_blockState_confirmRepcodesAndEntropyTables(ZSTD_blockState_t* const bs)
3393
+ static void
3394
+ ZSTD_blockState_confirmRepcodesAndEntropyTables(ZSTD_blockState_t* const bs)
3033
3395
  {
3034
3396
  ZSTD_compressedBlockState_t* const tmp = bs->prevCBlock;
3035
3397
  bs->prevCBlock = bs->nextCBlock;
@@ -3037,7 +3399,9 @@ static void ZSTD_blockState_confirmRepcodesAndEntropyTables(ZSTD_blockState_t* c
3037
3399
  }
3038
3400
 
3039
3401
  /* Writes the block header */
3040
- static void writeBlockHeader(void* op, size_t cSize, size_t blockSize, U32 lastBlock) {
3402
+ static void
3403
+ writeBlockHeader(void* op, size_t cSize, size_t blockSize, U32 lastBlock)
3404
+ {
3041
3405
  U32 const cBlockHeader = cSize == 1 ?
3042
3406
  lastBlock + (((U32)bt_rle)<<1) + (U32)(blockSize << 3) :
3043
3407
  lastBlock + (((U32)bt_compressed)<<1) + (U32)(cSize << 3);
@@ -3050,13 +3414,16 @@ static void writeBlockHeader(void* op, size_t cSize, size_t blockSize, U32 lastB
3050
3414
  * Stores literals block type (raw, rle, compressed, repeat) and
3051
3415
  * huffman description table to hufMetadata.
3052
3416
  * Requires ENTROPY_WORKSPACE_SIZE workspace
3053
- * @return : size of huffman description table or error code */
3054
- static size_t ZSTD_buildBlockEntropyStats_literals(void* const src, size_t srcSize,
3055
- const ZSTD_hufCTables_t* prevHuf,
3056
- ZSTD_hufCTables_t* nextHuf,
3057
- ZSTD_hufCTablesMetadata_t* hufMetadata,
3058
- const int literalsCompressionIsDisabled,
3059
- void* workspace, size_t wkspSize)
3417
+ * @return : size of huffman description table, or an error code
3418
+ */
3419
+ static size_t
3420
+ ZSTD_buildBlockEntropyStats_literals(void* const src, size_t srcSize,
3421
+ const ZSTD_hufCTables_t* prevHuf,
3422
+ ZSTD_hufCTables_t* nextHuf,
3423
+ ZSTD_hufCTablesMetadata_t* hufMetadata,
3424
+ const int literalsCompressionIsDisabled,
3425
+ void* workspace, size_t wkspSize,
3426
+ int hufFlags)
3060
3427
  {
3061
3428
  BYTE* const wkspStart = (BYTE*)workspace;
3062
3429
  BYTE* const wkspEnd = wkspStart + wkspSize;
@@ -3064,9 +3431,9 @@ static size_t ZSTD_buildBlockEntropyStats_literals(void* const src, size_t srcSi
3064
3431
  unsigned* const countWksp = (unsigned*)workspace;
3065
3432
  const size_t countWkspSize = (HUF_SYMBOLVALUE_MAX + 1) * sizeof(unsigned);
3066
3433
  BYTE* const nodeWksp = countWkspStart + countWkspSize;
3067
- const size_t nodeWkspSize = wkspEnd-nodeWksp;
3434
+ const size_t nodeWkspSize = (size_t)(wkspEnd - nodeWksp);
3068
3435
  unsigned maxSymbolValue = HUF_SYMBOLVALUE_MAX;
3069
- unsigned huffLog = HUF_TABLELOG_DEFAULT;
3436
+ unsigned huffLog = LitHufLog;
3070
3437
  HUF_repeat repeat = prevHuf->repeatMode;
3071
3438
  DEBUGLOG(5, "ZSTD_buildBlockEntropyStats_literals (srcSize=%zu)", srcSize);
3072
3439
 
@@ -3081,73 +3448,77 @@ static size_t ZSTD_buildBlockEntropyStats_literals(void* const src, size_t srcSi
3081
3448
 
3082
3449
  /* small ? don't even attempt compression (speed opt) */
3083
3450
  #ifndef COMPRESS_LITERALS_SIZE_MIN
3084
- #define COMPRESS_LITERALS_SIZE_MIN 63
3451
+ # define COMPRESS_LITERALS_SIZE_MIN 63 /* heuristic */
3085
3452
  #endif
3086
3453
  { size_t const minLitSize = (prevHuf->repeatMode == HUF_repeat_valid) ? 6 : COMPRESS_LITERALS_SIZE_MIN;
3087
3454
  if (srcSize <= minLitSize) {
3088
3455
  DEBUGLOG(5, "set_basic - too small");
3089
3456
  hufMetadata->hType = set_basic;
3090
3457
  return 0;
3091
- }
3092
- }
3458
+ } }
3093
3459
 
3094
3460
  /* Scan input and build symbol stats */
3095
- { size_t const largest = HIST_count_wksp (countWksp, &maxSymbolValue, (const BYTE*)src, srcSize, workspace, wkspSize);
3461
+ { size_t const largest =
3462
+ HIST_count_wksp (countWksp, &maxSymbolValue,
3463
+ (const BYTE*)src, srcSize,
3464
+ workspace, wkspSize);
3096
3465
  FORWARD_IF_ERROR(largest, "HIST_count_wksp failed");
3097
3466
  if (largest == srcSize) {
3467
+ /* only one literal symbol */
3098
3468
  DEBUGLOG(5, "set_rle");
3099
3469
  hufMetadata->hType = set_rle;
3100
3470
  return 0;
3101
3471
  }
3102
3472
  if (largest <= (srcSize >> 7)+4) {
3473
+ /* heuristic: likely not compressible */
3103
3474
  DEBUGLOG(5, "set_basic - no gain");
3104
3475
  hufMetadata->hType = set_basic;
3105
3476
  return 0;
3106
- }
3107
- }
3477
+ } }
3108
3478
 
3109
3479
  /* Validate the previous Huffman table */
3110
- if (repeat == HUF_repeat_check && !HUF_validateCTable((HUF_CElt const*)prevHuf->CTable, countWksp, maxSymbolValue)) {
3480
+ if (repeat == HUF_repeat_check
3481
+ && !HUF_validateCTable((HUF_CElt const*)prevHuf->CTable, countWksp, maxSymbolValue)) {
3111
3482
  repeat = HUF_repeat_none;
3112
3483
  }
3113
3484
 
3114
3485
  /* Build Huffman Tree */
3115
3486
  ZSTD_memset(nextHuf->CTable, 0, sizeof(nextHuf->CTable));
3116
- huffLog = HUF_optimalTableLog(huffLog, srcSize, maxSymbolValue);
3487
+ huffLog = HUF_optimalTableLog(huffLog, srcSize, maxSymbolValue, nodeWksp, nodeWkspSize, nextHuf->CTable, countWksp, hufFlags);
3488
+ assert(huffLog <= LitHufLog);
3117
3489
  { size_t const maxBits = HUF_buildCTable_wksp((HUF_CElt*)nextHuf->CTable, countWksp,
3118
3490
  maxSymbolValue, huffLog,
3119
3491
  nodeWksp, nodeWkspSize);
3120
3492
  FORWARD_IF_ERROR(maxBits, "HUF_buildCTable_wksp");
3121
3493
  huffLog = (U32)maxBits;
3122
- { /* Build and write the CTable */
3123
- size_t const newCSize = HUF_estimateCompressedSize(
3124
- (HUF_CElt*)nextHuf->CTable, countWksp, maxSymbolValue);
3125
- size_t const hSize = HUF_writeCTable_wksp(
3126
- hufMetadata->hufDesBuffer, sizeof(hufMetadata->hufDesBuffer),
3127
- (HUF_CElt*)nextHuf->CTable, maxSymbolValue, huffLog,
3128
- nodeWksp, nodeWkspSize);
3129
- /* Check against repeating the previous CTable */
3130
- if (repeat != HUF_repeat_none) {
3131
- size_t const oldCSize = HUF_estimateCompressedSize(
3132
- (HUF_CElt const*)prevHuf->CTable, countWksp, maxSymbolValue);
3133
- if (oldCSize < srcSize && (oldCSize <= hSize + newCSize || hSize + 12 >= srcSize)) {
3134
- DEBUGLOG(5, "set_repeat - smaller");
3135
- ZSTD_memcpy(nextHuf, prevHuf, sizeof(*prevHuf));
3136
- hufMetadata->hType = set_repeat;
3137
- return 0;
3138
- }
3139
- }
3140
- if (newCSize + hSize >= srcSize) {
3141
- DEBUGLOG(5, "set_basic - no gains");
3494
+ }
3495
+ { /* Build and write the CTable */
3496
+ size_t const newCSize = HUF_estimateCompressedSize(
3497
+ (HUF_CElt*)nextHuf->CTable, countWksp, maxSymbolValue);
3498
+ size_t const hSize = HUF_writeCTable_wksp(
3499
+ hufMetadata->hufDesBuffer, sizeof(hufMetadata->hufDesBuffer),
3500
+ (HUF_CElt*)nextHuf->CTable, maxSymbolValue, huffLog,
3501
+ nodeWksp, nodeWkspSize);
3502
+ /* Check against repeating the previous CTable */
3503
+ if (repeat != HUF_repeat_none) {
3504
+ size_t const oldCSize = HUF_estimateCompressedSize(
3505
+ (HUF_CElt const*)prevHuf->CTable, countWksp, maxSymbolValue);
3506
+ if (oldCSize < srcSize && (oldCSize <= hSize + newCSize || hSize + 12 >= srcSize)) {
3507
+ DEBUGLOG(5, "set_repeat - smaller");
3142
3508
  ZSTD_memcpy(nextHuf, prevHuf, sizeof(*prevHuf));
3143
- hufMetadata->hType = set_basic;
3509
+ hufMetadata->hType = set_repeat;
3144
3510
  return 0;
3145
- }
3146
- DEBUGLOG(5, "set_compressed (hSize=%u)", (U32)hSize);
3147
- hufMetadata->hType = set_compressed;
3148
- nextHuf->repeatMode = HUF_repeat_check;
3149
- return hSize;
3511
+ } }
3512
+ if (newCSize + hSize >= srcSize) {
3513
+ DEBUGLOG(5, "set_basic - no gains");
3514
+ ZSTD_memcpy(nextHuf, prevHuf, sizeof(*prevHuf));
3515
+ hufMetadata->hType = set_basic;
3516
+ return 0;
3150
3517
  }
3518
+ DEBUGLOG(5, "set_compressed (hSize=%u)", (U32)hSize);
3519
+ hufMetadata->hType = set_compressed;
3520
+ nextHuf->repeatMode = HUF_repeat_check;
3521
+ return hSize;
3151
3522
  }
3152
3523
  }
3153
3524
 
@@ -3157,8 +3528,9 @@ static size_t ZSTD_buildBlockEntropyStats_literals(void* const src, size_t srcSi
3157
3528
  * and updates nextEntropy to the appropriate repeatMode.
3158
3529
  */
3159
3530
  static ZSTD_symbolEncodingTypeStats_t
3160
- ZSTD_buildDummySequencesStatistics(ZSTD_fseCTables_t* nextEntropy) {
3161
- ZSTD_symbolEncodingTypeStats_t stats = {set_basic, set_basic, set_basic, 0, 0};
3531
+ ZSTD_buildDummySequencesStatistics(ZSTD_fseCTables_t* nextEntropy)
3532
+ {
3533
+ ZSTD_symbolEncodingTypeStats_t stats = {set_basic, set_basic, set_basic, 0, 0, 0};
3162
3534
  nextEntropy->litlength_repeatMode = FSE_repeat_none;
3163
3535
  nextEntropy->offcode_repeatMode = FSE_repeat_none;
3164
3536
  nextEntropy->matchlength_repeatMode = FSE_repeat_none;
@@ -3169,16 +3541,18 @@ ZSTD_buildDummySequencesStatistics(ZSTD_fseCTables_t* nextEntropy) {
3169
3541
  * Builds entropy for the sequences.
3170
3542
  * Stores symbol compression modes and fse table to fseMetadata.
3171
3543
  * Requires ENTROPY_WORKSPACE_SIZE wksp.
3172
- * @return : size of fse tables or error code */
3173
- static size_t ZSTD_buildBlockEntropyStats_sequences(seqStore_t* seqStorePtr,
3174
- const ZSTD_fseCTables_t* prevEntropy,
3175
- ZSTD_fseCTables_t* nextEntropy,
3176
- const ZSTD_CCtx_params* cctxParams,
3177
- ZSTD_fseCTablesMetadata_t* fseMetadata,
3178
- void* workspace, size_t wkspSize)
3544
+ * @return : size of fse tables or error code */
3545
+ static size_t
3546
+ ZSTD_buildBlockEntropyStats_sequences(
3547
+ const seqStore_t* seqStorePtr,
3548
+ const ZSTD_fseCTables_t* prevEntropy,
3549
+ ZSTD_fseCTables_t* nextEntropy,
3550
+ const ZSTD_CCtx_params* cctxParams,
3551
+ ZSTD_fseCTablesMetadata_t* fseMetadata,
3552
+ void* workspace, size_t wkspSize)
3179
3553
  {
3180
3554
  ZSTD_strategy const strategy = cctxParams->cParams.strategy;
3181
- size_t const nbSeq = seqStorePtr->sequences - seqStorePtr->sequencesStart;
3555
+ size_t const nbSeq = (size_t)(seqStorePtr->sequences - seqStorePtr->sequencesStart);
3182
3556
  BYTE* const ostart = fseMetadata->fseTablesBuffer;
3183
3557
  BYTE* const oend = ostart + sizeof(fseMetadata->fseTablesBuffer);
3184
3558
  BYTE* op = ostart;
@@ -3205,23 +3579,28 @@ static size_t ZSTD_buildBlockEntropyStats_sequences(seqStore_t* seqStorePtr,
3205
3579
  /** ZSTD_buildBlockEntropyStats() :
3206
3580
  * Builds entropy for the block.
3207
3581
  * Requires workspace size ENTROPY_WORKSPACE_SIZE
3208
- *
3209
- * @return : 0 on success or error code
3582
+ * @return : 0 on success, or an error code
3583
+ * Note : also employed in superblock
3210
3584
  */
3211
- size_t ZSTD_buildBlockEntropyStats(seqStore_t* seqStorePtr,
3212
- const ZSTD_entropyCTables_t* prevEntropy,
3213
- ZSTD_entropyCTables_t* nextEntropy,
3214
- const ZSTD_CCtx_params* cctxParams,
3215
- ZSTD_entropyCTablesMetadata_t* entropyMetadata,
3216
- void* workspace, size_t wkspSize)
3217
- {
3218
- size_t const litSize = seqStorePtr->lit - seqStorePtr->litStart;
3585
+ size_t ZSTD_buildBlockEntropyStats(
3586
+ const seqStore_t* seqStorePtr,
3587
+ const ZSTD_entropyCTables_t* prevEntropy,
3588
+ ZSTD_entropyCTables_t* nextEntropy,
3589
+ const ZSTD_CCtx_params* cctxParams,
3590
+ ZSTD_entropyCTablesMetadata_t* entropyMetadata,
3591
+ void* workspace, size_t wkspSize)
3592
+ {
3593
+ size_t const litSize = (size_t)(seqStorePtr->lit - seqStorePtr->litStart);
3594
+ int const huf_useOptDepth = (cctxParams->cParams.strategy >= HUF_OPTIMAL_DEPTH_THRESHOLD);
3595
+ int const hufFlags = huf_useOptDepth ? HUF_flags_optimalDepth : 0;
3596
+
3219
3597
  entropyMetadata->hufMetadata.hufDesSize =
3220
3598
  ZSTD_buildBlockEntropyStats_literals(seqStorePtr->litStart, litSize,
3221
3599
  &prevEntropy->huf, &nextEntropy->huf,
3222
3600
  &entropyMetadata->hufMetadata,
3223
3601
  ZSTD_literalsCompressionIsDisabled(cctxParams),
3224
- workspace, wkspSize);
3602
+ workspace, wkspSize, hufFlags);
3603
+
3225
3604
  FORWARD_IF_ERROR(entropyMetadata->hufMetadata.hufDesSize, "ZSTD_buildBlockEntropyStats_literals failed");
3226
3605
  entropyMetadata->fseMetadata.fseTablesSize =
3227
3606
  ZSTD_buildBlockEntropyStats_sequences(seqStorePtr,
@@ -3234,11 +3613,12 @@ size_t ZSTD_buildBlockEntropyStats(seqStore_t* seqStorePtr,
3234
3613
  }
3235
3614
 
3236
3615
  /* Returns the size estimate for the literals section (header + content) of a block */
3237
- static size_t ZSTD_estimateBlockSize_literal(const BYTE* literals, size_t litSize,
3238
- const ZSTD_hufCTables_t* huf,
3239
- const ZSTD_hufCTablesMetadata_t* hufMetadata,
3240
- void* workspace, size_t wkspSize,
3241
- int writeEntropy)
3616
+ static size_t
3617
+ ZSTD_estimateBlockSize_literal(const BYTE* literals, size_t litSize,
3618
+ const ZSTD_hufCTables_t* huf,
3619
+ const ZSTD_hufCTablesMetadata_t* hufMetadata,
3620
+ void* workspace, size_t wkspSize,
3621
+ int writeEntropy)
3242
3622
  {
3243
3623
  unsigned* const countWksp = (unsigned*)workspace;
3244
3624
  unsigned maxSymbolValue = HUF_SYMBOLVALUE_MAX;
@@ -3260,12 +3640,13 @@ static size_t ZSTD_estimateBlockSize_literal(const BYTE* literals, size_t litSiz
3260
3640
  }
3261
3641
 
3262
3642
  /* Returns the size estimate for the FSE-compressed symbols (of, ml, ll) of a block */
3263
- static size_t ZSTD_estimateBlockSize_symbolType(symbolEncodingType_e type,
3264
- const BYTE* codeTable, size_t nbSeq, unsigned maxCode,
3265
- const FSE_CTable* fseCTable,
3266
- const U8* additionalBits,
3267
- short const* defaultNorm, U32 defaultNormLog, U32 defaultMax,
3268
- void* workspace, size_t wkspSize)
3643
+ static size_t
3644
+ ZSTD_estimateBlockSize_symbolType(symbolEncodingType_e type,
3645
+ const BYTE* codeTable, size_t nbSeq, unsigned maxCode,
3646
+ const FSE_CTable* fseCTable,
3647
+ const U8* additionalBits,
3648
+ short const* defaultNorm, U32 defaultNormLog, U32 defaultMax,
3649
+ void* workspace, size_t wkspSize)
3269
3650
  {
3270
3651
  unsigned* const countWksp = (unsigned*)workspace;
3271
3652
  const BYTE* ctp = codeTable;
@@ -3297,99 +3678,107 @@ static size_t ZSTD_estimateBlockSize_symbolType(symbolEncodingType_e type,
3297
3678
  }
3298
3679
 
3299
3680
  /* Returns the size estimate for the sequences section (header + content) of a block */
3300
- static size_t ZSTD_estimateBlockSize_sequences(const BYTE* ofCodeTable,
3301
- const BYTE* llCodeTable,
3302
- const BYTE* mlCodeTable,
3303
- size_t nbSeq,
3304
- const ZSTD_fseCTables_t* fseTables,
3305
- const ZSTD_fseCTablesMetadata_t* fseMetadata,
3306
- void* workspace, size_t wkspSize,
3307
- int writeEntropy)
3681
+ static size_t
3682
+ ZSTD_estimateBlockSize_sequences(const BYTE* ofCodeTable,
3683
+ const BYTE* llCodeTable,
3684
+ const BYTE* mlCodeTable,
3685
+ size_t nbSeq,
3686
+ const ZSTD_fseCTables_t* fseTables,
3687
+ const ZSTD_fseCTablesMetadata_t* fseMetadata,
3688
+ void* workspace, size_t wkspSize,
3689
+ int writeEntropy)
3308
3690
  {
3309
3691
  size_t sequencesSectionHeaderSize = 1 /* seqHead */ + 1 /* min seqSize size */ + (nbSeq >= 128) + (nbSeq >= LONGNBSEQ);
3310
3692
  size_t cSeqSizeEstimate = 0;
3311
3693
  cSeqSizeEstimate += ZSTD_estimateBlockSize_symbolType(fseMetadata->ofType, ofCodeTable, nbSeq, MaxOff,
3312
- fseTables->offcodeCTable, NULL,
3313
- OF_defaultNorm, OF_defaultNormLog, DefaultMaxOff,
3314
- workspace, wkspSize);
3694
+ fseTables->offcodeCTable, NULL,
3695
+ OF_defaultNorm, OF_defaultNormLog, DefaultMaxOff,
3696
+ workspace, wkspSize);
3315
3697
  cSeqSizeEstimate += ZSTD_estimateBlockSize_symbolType(fseMetadata->llType, llCodeTable, nbSeq, MaxLL,
3316
- fseTables->litlengthCTable, LL_bits,
3317
- LL_defaultNorm, LL_defaultNormLog, MaxLL,
3318
- workspace, wkspSize);
3698
+ fseTables->litlengthCTable, LL_bits,
3699
+ LL_defaultNorm, LL_defaultNormLog, MaxLL,
3700
+ workspace, wkspSize);
3319
3701
  cSeqSizeEstimate += ZSTD_estimateBlockSize_symbolType(fseMetadata->mlType, mlCodeTable, nbSeq, MaxML,
3320
- fseTables->matchlengthCTable, ML_bits,
3321
- ML_defaultNorm, ML_defaultNormLog, MaxML,
3322
- workspace, wkspSize);
3702
+ fseTables->matchlengthCTable, ML_bits,
3703
+ ML_defaultNorm, ML_defaultNormLog, MaxML,
3704
+ workspace, wkspSize);
3323
3705
  if (writeEntropy) cSeqSizeEstimate += fseMetadata->fseTablesSize;
3324
3706
  return cSeqSizeEstimate + sequencesSectionHeaderSize;
3325
3707
  }
3326
3708
 
3327
3709
  /* Returns the size estimate for a given stream of literals, of, ll, ml */
3328
- static size_t ZSTD_estimateBlockSize(const BYTE* literals, size_t litSize,
3329
- const BYTE* ofCodeTable,
3330
- const BYTE* llCodeTable,
3331
- const BYTE* mlCodeTable,
3332
- size_t nbSeq,
3333
- const ZSTD_entropyCTables_t* entropy,
3334
- const ZSTD_entropyCTablesMetadata_t* entropyMetadata,
3335
- void* workspace, size_t wkspSize,
3336
- int writeLitEntropy, int writeSeqEntropy) {
3710
+ static size_t
3711
+ ZSTD_estimateBlockSize(const BYTE* literals, size_t litSize,
3712
+ const BYTE* ofCodeTable,
3713
+ const BYTE* llCodeTable,
3714
+ const BYTE* mlCodeTable,
3715
+ size_t nbSeq,
3716
+ const ZSTD_entropyCTables_t* entropy,
3717
+ const ZSTD_entropyCTablesMetadata_t* entropyMetadata,
3718
+ void* workspace, size_t wkspSize,
3719
+ int writeLitEntropy, int writeSeqEntropy)
3720
+ {
3337
3721
  size_t const literalsSize = ZSTD_estimateBlockSize_literal(literals, litSize,
3338
- &entropy->huf, &entropyMetadata->hufMetadata,
3339
- workspace, wkspSize, writeLitEntropy);
3722
+ &entropy->huf, &entropyMetadata->hufMetadata,
3723
+ workspace, wkspSize, writeLitEntropy);
3340
3724
  size_t const seqSize = ZSTD_estimateBlockSize_sequences(ofCodeTable, llCodeTable, mlCodeTable,
3341
- nbSeq, &entropy->fse, &entropyMetadata->fseMetadata,
3342
- workspace, wkspSize, writeSeqEntropy);
3725
+ nbSeq, &entropy->fse, &entropyMetadata->fseMetadata,
3726
+ workspace, wkspSize, writeSeqEntropy);
3343
3727
  return seqSize + literalsSize + ZSTD_blockHeaderSize;
3344
3728
  }
3345
3729
 
3346
3730
  /* Builds entropy statistics and uses them for blocksize estimation.
3347
3731
  *
3348
- * Returns the estimated compressed size of the seqStore, or a zstd error.
3732
+ * @return: estimated compressed size of the seqStore, or a zstd error.
3349
3733
  */
3350
- static size_t ZSTD_buildEntropyStatisticsAndEstimateSubBlockSize(seqStore_t* seqStore, ZSTD_CCtx* zc) {
3351
- ZSTD_entropyCTablesMetadata_t* entropyMetadata = &zc->blockSplitCtx.entropyMetadata;
3734
+ static size_t
3735
+ ZSTD_buildEntropyStatisticsAndEstimateSubBlockSize(seqStore_t* seqStore, ZSTD_CCtx* zc)
3736
+ {
3737
+ ZSTD_entropyCTablesMetadata_t* const entropyMetadata = &zc->blockSplitCtx.entropyMetadata;
3352
3738
  DEBUGLOG(6, "ZSTD_buildEntropyStatisticsAndEstimateSubBlockSize()");
3353
3739
  FORWARD_IF_ERROR(ZSTD_buildBlockEntropyStats(seqStore,
3354
3740
  &zc->blockState.prevCBlock->entropy,
3355
3741
  &zc->blockState.nextCBlock->entropy,
3356
3742
  &zc->appliedParams,
3357
3743
  entropyMetadata,
3358
- zc->entropyWorkspace, ENTROPY_WORKSPACE_SIZE /* statically allocated in resetCCtx */), "");
3359
- return ZSTD_estimateBlockSize(seqStore->litStart, (size_t)(seqStore->lit - seqStore->litStart),
3744
+ zc->entropyWorkspace, ENTROPY_WORKSPACE_SIZE), "");
3745
+ return ZSTD_estimateBlockSize(
3746
+ seqStore->litStart, (size_t)(seqStore->lit - seqStore->litStart),
3360
3747
  seqStore->ofCode, seqStore->llCode, seqStore->mlCode,
3361
3748
  (size_t)(seqStore->sequences - seqStore->sequencesStart),
3362
- &zc->blockState.nextCBlock->entropy, entropyMetadata, zc->entropyWorkspace, ENTROPY_WORKSPACE_SIZE,
3749
+ &zc->blockState.nextCBlock->entropy,
3750
+ entropyMetadata,
3751
+ zc->entropyWorkspace, ENTROPY_WORKSPACE_SIZE,
3363
3752
  (int)(entropyMetadata->hufMetadata.hType == set_compressed), 1);
3364
3753
  }
3365
3754
 
3366
3755
  /* Returns literals bytes represented in a seqStore */
3367
- static size_t ZSTD_countSeqStoreLiteralsBytes(const seqStore_t* const seqStore) {
3756
+ static size_t ZSTD_countSeqStoreLiteralsBytes(const seqStore_t* const seqStore)
3757
+ {
3368
3758
  size_t literalsBytes = 0;
3369
- size_t const nbSeqs = seqStore->sequences - seqStore->sequencesStart;
3759
+ size_t const nbSeqs = (size_t)(seqStore->sequences - seqStore->sequencesStart);
3370
3760
  size_t i;
3371
3761
  for (i = 0; i < nbSeqs; ++i) {
3372
- seqDef seq = seqStore->sequencesStart[i];
3762
+ seqDef const seq = seqStore->sequencesStart[i];
3373
3763
  literalsBytes += seq.litLength;
3374
3764
  if (i == seqStore->longLengthPos && seqStore->longLengthType == ZSTD_llt_literalLength) {
3375
3765
  literalsBytes += 0x10000;
3376
- }
3377
- }
3766
+ } }
3378
3767
  return literalsBytes;
3379
3768
  }
3380
3769
 
3381
3770
  /* Returns match bytes represented in a seqStore */
3382
- static size_t ZSTD_countSeqStoreMatchBytes(const seqStore_t* const seqStore) {
3771
+ static size_t ZSTD_countSeqStoreMatchBytes(const seqStore_t* const seqStore)
3772
+ {
3383
3773
  size_t matchBytes = 0;
3384
- size_t const nbSeqs = seqStore->sequences - seqStore->sequencesStart;
3774
+ size_t const nbSeqs = (size_t)(seqStore->sequences - seqStore->sequencesStart);
3385
3775
  size_t i;
3386
3776
  for (i = 0; i < nbSeqs; ++i) {
3387
3777
  seqDef seq = seqStore->sequencesStart[i];
3388
3778
  matchBytes += seq.mlBase + MINMATCH;
3389
3779
  if (i == seqStore->longLengthPos && seqStore->longLengthType == ZSTD_llt_matchLength) {
3390
3780
  matchBytes += 0x10000;
3391
- }
3392
- }
3781
+ } }
3393
3782
  return matchBytes;
3394
3783
  }
3395
3784
 
@@ -3398,15 +3787,12 @@ static size_t ZSTD_countSeqStoreMatchBytes(const seqStore_t* const seqStore) {
3398
3787
  */
3399
3788
  static void ZSTD_deriveSeqStoreChunk(seqStore_t* resultSeqStore,
3400
3789
  const seqStore_t* originalSeqStore,
3401
- size_t startIdx, size_t endIdx) {
3402
- BYTE* const litEnd = originalSeqStore->lit;
3403
- size_t literalsBytes;
3404
- size_t literalsBytesPreceding = 0;
3405
-
3790
+ size_t startIdx, size_t endIdx)
3791
+ {
3406
3792
  *resultSeqStore = *originalSeqStore;
3407
3793
  if (startIdx > 0) {
3408
3794
  resultSeqStore->sequences = originalSeqStore->sequencesStart + startIdx;
3409
- literalsBytesPreceding = ZSTD_countSeqStoreLiteralsBytes(resultSeqStore);
3795
+ resultSeqStore->litStart += ZSTD_countSeqStoreLiteralsBytes(resultSeqStore);
3410
3796
  }
3411
3797
 
3412
3798
  /* Move longLengthPos into the correct position if necessary */
@@ -3419,13 +3805,12 @@ static void ZSTD_deriveSeqStoreChunk(seqStore_t* resultSeqStore,
3419
3805
  }
3420
3806
  resultSeqStore->sequencesStart = originalSeqStore->sequencesStart + startIdx;
3421
3807
  resultSeqStore->sequences = originalSeqStore->sequencesStart + endIdx;
3422
- literalsBytes = ZSTD_countSeqStoreLiteralsBytes(resultSeqStore);
3423
- resultSeqStore->litStart += literalsBytesPreceding;
3424
3808
  if (endIdx == (size_t)(originalSeqStore->sequences - originalSeqStore->sequencesStart)) {
3425
3809
  /* This accounts for possible last literals if the derived chunk reaches the end of the block */
3426
- resultSeqStore->lit = litEnd;
3810
+ assert(resultSeqStore->lit == originalSeqStore->lit);
3427
3811
  } else {
3428
- resultSeqStore->lit = resultSeqStore->litStart+literalsBytes;
3812
+ size_t const literalsBytes = ZSTD_countSeqStoreLiteralsBytes(resultSeqStore);
3813
+ resultSeqStore->lit = resultSeqStore->litStart + literalsBytes;
3429
3814
  }
3430
3815
  resultSeqStore->llCode += startIdx;
3431
3816
  resultSeqStore->mlCode += startIdx;
@@ -3433,20 +3818,26 @@ static void ZSTD_deriveSeqStoreChunk(seqStore_t* resultSeqStore,
3433
3818
  }
3434
3819
 
3435
3820
  /**
3436
- * Returns the raw offset represented by the combination of offCode, ll0, and repcode history.
3437
- * offCode must represent a repcode in the numeric representation of ZSTD_storeSeq().
3821
+ * Returns the raw offset represented by the combination of offBase, ll0, and repcode history.
3822
+ * offBase must represent a repcode in the numeric representation of ZSTD_storeSeq().
3438
3823
  */
3439
3824
  static U32
3440
- ZSTD_resolveRepcodeToRawOffset(const U32 rep[ZSTD_REP_NUM], const U32 offCode, const U32 ll0)
3441
- {
3442
- U32 const adjustedOffCode = STORED_REPCODE(offCode) - 1 + ll0; /* [ 0 - 3 ] */
3443
- assert(STORED_IS_REPCODE(offCode));
3444
- if (adjustedOffCode == ZSTD_REP_NUM) {
3445
- /* litlength == 0 and offCode == 2 implies selection of first repcode - 1 */
3446
- assert(rep[0] > 0);
3825
+ ZSTD_resolveRepcodeToRawOffset(const U32 rep[ZSTD_REP_NUM], const U32 offBase, const U32 ll0)
3826
+ {
3827
+ U32 const adjustedRepCode = OFFBASE_TO_REPCODE(offBase) - 1 + ll0; /* [ 0 - 3 ] */
3828
+ assert(OFFBASE_IS_REPCODE(offBase));
3829
+ if (adjustedRepCode == ZSTD_REP_NUM) {
3830
+ assert(ll0);
3831
+ /* litlength == 0 and offCode == 2 implies selection of first repcode - 1
3832
+ * This is only valid if it results in a valid offset value, aka > 0.
3833
+ * Note : it may happen that `rep[0]==1` in exceptional circumstances.
3834
+ * In which case this function will return 0, which is an invalid offset.
3835
+ * It's not an issue though, since this value will be
3836
+ * compared and discarded within ZSTD_seqStore_resolveOffCodes().
3837
+ */
3447
3838
  return rep[0] - 1;
3448
3839
  }
3449
- return rep[adjustedOffCode];
3840
+ return rep[adjustedRepCode];
3450
3841
  }
3451
3842
 
3452
3843
  /**
@@ -3462,30 +3853,32 @@ ZSTD_resolveRepcodeToRawOffset(const U32 rep[ZSTD_REP_NUM], const U32 offCode, c
3462
3853
  * 1-3 : repcode 1-3
3463
3854
  * 4+ : real_offset+3
3464
3855
  */
3465
- static void ZSTD_seqStore_resolveOffCodes(repcodes_t* const dRepcodes, repcodes_t* const cRepcodes,
3466
- seqStore_t* const seqStore, U32 const nbSeq) {
3856
+ static void
3857
+ ZSTD_seqStore_resolveOffCodes(repcodes_t* const dRepcodes, repcodes_t* const cRepcodes,
3858
+ const seqStore_t* const seqStore, U32 const nbSeq)
3859
+ {
3467
3860
  U32 idx = 0;
3468
3861
  for (; idx < nbSeq; ++idx) {
3469
3862
  seqDef* const seq = seqStore->sequencesStart + idx;
3470
3863
  U32 const ll0 = (seq->litLength == 0);
3471
- U32 const offCode = OFFBASE_TO_STORED(seq->offBase);
3472
- assert(seq->offBase > 0);
3473
- if (STORED_IS_REPCODE(offCode)) {
3474
- U32 const dRawOffset = ZSTD_resolveRepcodeToRawOffset(dRepcodes->rep, offCode, ll0);
3475
- U32 const cRawOffset = ZSTD_resolveRepcodeToRawOffset(cRepcodes->rep, offCode, ll0);
3864
+ U32 const offBase = seq->offBase;
3865
+ assert(offBase > 0);
3866
+ if (OFFBASE_IS_REPCODE(offBase)) {
3867
+ U32 const dRawOffset = ZSTD_resolveRepcodeToRawOffset(dRepcodes->rep, offBase, ll0);
3868
+ U32 const cRawOffset = ZSTD_resolveRepcodeToRawOffset(cRepcodes->rep, offBase, ll0);
3476
3869
  /* Adjust simulated decompression repcode history if we come across a mismatch. Replace
3477
3870
  * the repcode with the offset it actually references, determined by the compression
3478
3871
  * repcode history.
3479
3872
  */
3480
3873
  if (dRawOffset != cRawOffset) {
3481
- seq->offBase = cRawOffset + ZSTD_REP_NUM;
3874
+ seq->offBase = OFFSET_TO_OFFBASE(cRawOffset);
3482
3875
  }
3483
3876
  }
3484
3877
  /* Compression repcode history is always updated with values directly from the unmodified seqStore.
3485
3878
  * Decompression repcode history may use modified seq->offset value taken from compression repcode history.
3486
3879
  */
3487
- ZSTD_updateRep(dRepcodes->rep, OFFBASE_TO_STORED(seq->offBase), ll0);
3488
- ZSTD_updateRep(cRepcodes->rep, offCode, ll0);
3880
+ ZSTD_updateRep(dRepcodes->rep, seq->offBase, ll0);
3881
+ ZSTD_updateRep(cRepcodes->rep, offBase, ll0);
3489
3882
  }
3490
3883
  }
3491
3884
 
@@ -3495,10 +3888,11 @@ static void ZSTD_seqStore_resolveOffCodes(repcodes_t* const dRepcodes, repcodes_
3495
3888
  * Returns the total size of that block (including header) or a ZSTD error code.
3496
3889
  */
3497
3890
  static size_t
3498
- ZSTD_compressSeqStore_singleBlock(ZSTD_CCtx* zc, seqStore_t* const seqStore,
3891
+ ZSTD_compressSeqStore_singleBlock(ZSTD_CCtx* zc,
3892
+ const seqStore_t* const seqStore,
3499
3893
  repcodes_t* const dRep, repcodes_t* const cRep,
3500
3894
  void* dst, size_t dstCapacity,
3501
- const void* src, size_t srcSize,
3895
+ const void* src, size_t srcSize,
3502
3896
  U32 lastBlock, U32 isPartition)
3503
3897
  {
3504
3898
  const U32 rleMaxLength = 25;
@@ -3572,45 +3966,49 @@ typedef struct {
3572
3966
 
3573
3967
  /* Helper function to perform the recursive search for block splits.
3574
3968
  * Estimates the cost of seqStore prior to split, and estimates the cost of splitting the sequences in half.
3575
- * If advantageous to split, then we recurse down the two sub-blocks. If not, or if an error occurred in estimation, then
3576
- * we do not recurse.
3969
+ * If advantageous to split, then we recurse down the two sub-blocks.
3970
+ * If not, or if an error occurred in estimation, then we do not recurse.
3577
3971
  *
3578
- * Note: The recursion depth is capped by a heuristic minimum number of sequences, defined by MIN_SEQUENCES_BLOCK_SPLITTING.
3972
+ * Note: The recursion depth is capped by a heuristic minimum number of sequences,
3973
+ * defined by MIN_SEQUENCES_BLOCK_SPLITTING.
3579
3974
  * In theory, this means the absolute largest recursion depth is 10 == log2(maxNbSeqInBlock/MIN_SEQUENCES_BLOCK_SPLITTING).
3580
3975
  * In practice, recursion depth usually doesn't go beyond 4.
3581
3976
  *
3582
- * Furthermore, the number of splits is capped by ZSTD_MAX_NB_BLOCK_SPLITS. At ZSTD_MAX_NB_BLOCK_SPLITS == 196 with the current existing blockSize
3977
+ * Furthermore, the number of splits is capped by ZSTD_MAX_NB_BLOCK_SPLITS.
3978
+ * At ZSTD_MAX_NB_BLOCK_SPLITS == 196 with the current existing blockSize
3583
3979
  * maximum of 128 KB, this value is actually impossible to reach.
3584
3980
  */
3585
3981
  static void
3586
3982
  ZSTD_deriveBlockSplitsHelper(seqStoreSplits* splits, size_t startIdx, size_t endIdx,
3587
3983
  ZSTD_CCtx* zc, const seqStore_t* origSeqStore)
3588
3984
  {
3589
- seqStore_t* fullSeqStoreChunk = &zc->blockSplitCtx.fullSeqStoreChunk;
3590
- seqStore_t* firstHalfSeqStore = &zc->blockSplitCtx.firstHalfSeqStore;
3591
- seqStore_t* secondHalfSeqStore = &zc->blockSplitCtx.secondHalfSeqStore;
3985
+ seqStore_t* const fullSeqStoreChunk = &zc->blockSplitCtx.fullSeqStoreChunk;
3986
+ seqStore_t* const firstHalfSeqStore = &zc->blockSplitCtx.firstHalfSeqStore;
3987
+ seqStore_t* const secondHalfSeqStore = &zc->blockSplitCtx.secondHalfSeqStore;
3592
3988
  size_t estimatedOriginalSize;
3593
3989
  size_t estimatedFirstHalfSize;
3594
3990
  size_t estimatedSecondHalfSize;
3595
3991
  size_t midIdx = (startIdx + endIdx)/2;
3596
3992
 
3993
+ DEBUGLOG(5, "ZSTD_deriveBlockSplitsHelper: startIdx=%zu endIdx=%zu", startIdx, endIdx);
3994
+ assert(endIdx >= startIdx);
3597
3995
  if (endIdx - startIdx < MIN_SEQUENCES_BLOCK_SPLITTING || splits->idx >= ZSTD_MAX_NB_BLOCK_SPLITS) {
3598
- DEBUGLOG(6, "ZSTD_deriveBlockSplitsHelper: Too few sequences");
3996
+ DEBUGLOG(6, "ZSTD_deriveBlockSplitsHelper: Too few sequences (%zu)", endIdx - startIdx);
3599
3997
  return;
3600
3998
  }
3601
- DEBUGLOG(4, "ZSTD_deriveBlockSplitsHelper: startIdx=%zu endIdx=%zu", startIdx, endIdx);
3602
3999
  ZSTD_deriveSeqStoreChunk(fullSeqStoreChunk, origSeqStore, startIdx, endIdx);
3603
4000
  ZSTD_deriveSeqStoreChunk(firstHalfSeqStore, origSeqStore, startIdx, midIdx);
3604
4001
  ZSTD_deriveSeqStoreChunk(secondHalfSeqStore, origSeqStore, midIdx, endIdx);
3605
4002
  estimatedOriginalSize = ZSTD_buildEntropyStatisticsAndEstimateSubBlockSize(fullSeqStoreChunk, zc);
3606
4003
  estimatedFirstHalfSize = ZSTD_buildEntropyStatisticsAndEstimateSubBlockSize(firstHalfSeqStore, zc);
3607
4004
  estimatedSecondHalfSize = ZSTD_buildEntropyStatisticsAndEstimateSubBlockSize(secondHalfSeqStore, zc);
3608
- DEBUGLOG(4, "Estimated original block size: %zu -- First half split: %zu -- Second half split: %zu",
4005
+ DEBUGLOG(5, "Estimated original block size: %zu -- First half split: %zu -- Second half split: %zu",
3609
4006
  estimatedOriginalSize, estimatedFirstHalfSize, estimatedSecondHalfSize);
3610
4007
  if (ZSTD_isError(estimatedOriginalSize) || ZSTD_isError(estimatedFirstHalfSize) || ZSTD_isError(estimatedSecondHalfSize)) {
3611
4008
  return;
3612
4009
  }
3613
4010
  if (estimatedFirstHalfSize + estimatedSecondHalfSize < estimatedOriginalSize) {
4011
+ DEBUGLOG(5, "split decided at seqNb:%zu", midIdx);
3614
4012
  ZSTD_deriveBlockSplitsHelper(splits, startIdx, midIdx, zc, origSeqStore);
3615
4013
  splits->splitLocations[splits->idx] = (U32)midIdx;
3616
4014
  splits->idx++;
@@ -3618,14 +4016,18 @@ ZSTD_deriveBlockSplitsHelper(seqStoreSplits* splits, size_t startIdx, size_t end
3618
4016
  }
3619
4017
  }
3620
4018
 
3621
- /* Base recursive function. Populates a table with intra-block partition indices that can improve compression ratio.
4019
+ /* Base recursive function.
4020
+ * Populates a table with intra-block partition indices that can improve compression ratio.
3622
4021
  *
3623
- * Returns the number of splits made (which equals the size of the partition table - 1).
4022
+ * @return: number of splits made (which equals the size of the partition table - 1).
3624
4023
  */
3625
- static size_t ZSTD_deriveBlockSplits(ZSTD_CCtx* zc, U32 partitions[], U32 nbSeq) {
3626
- seqStoreSplits splits = {partitions, 0};
4024
+ static size_t ZSTD_deriveBlockSplits(ZSTD_CCtx* zc, U32 partitions[], U32 nbSeq)
4025
+ {
4026
+ seqStoreSplits splits;
4027
+ splits.splitLocations = partitions;
4028
+ splits.idx = 0;
3627
4029
  if (nbSeq <= 4) {
3628
- DEBUGLOG(4, "ZSTD_deriveBlockSplits: Too few sequences to split");
4030
+ DEBUGLOG(5, "ZSTD_deriveBlockSplits: Too few sequences to split (%u <= 4)", nbSeq);
3629
4031
  /* Refuse to try and split anything with less than 4 sequences */
3630
4032
  return 0;
3631
4033
  }
@@ -3641,18 +4043,20 @@ static size_t ZSTD_deriveBlockSplits(ZSTD_CCtx* zc, U32 partitions[], U32 nbSeq)
3641
4043
  * Returns combined size of all blocks (which includes headers), or a ZSTD error code.
3642
4044
  */
3643
4045
  static size_t
3644
- ZSTD_compressBlock_splitBlock_internal(ZSTD_CCtx* zc, void* dst, size_t dstCapacity,
3645
- const void* src, size_t blockSize, U32 lastBlock, U32 nbSeq)
4046
+ ZSTD_compressBlock_splitBlock_internal(ZSTD_CCtx* zc,
4047
+ void* dst, size_t dstCapacity,
4048
+ const void* src, size_t blockSize,
4049
+ U32 lastBlock, U32 nbSeq)
3646
4050
  {
3647
4051
  size_t cSize = 0;
3648
4052
  const BYTE* ip = (const BYTE*)src;
3649
4053
  BYTE* op = (BYTE*)dst;
3650
4054
  size_t i = 0;
3651
4055
  size_t srcBytesTotal = 0;
3652
- U32* partitions = zc->blockSplitCtx.partitions; /* size == ZSTD_MAX_NB_BLOCK_SPLITS */
3653
- seqStore_t* nextSeqStore = &zc->blockSplitCtx.nextSeqStore;
3654
- seqStore_t* currSeqStore = &zc->blockSplitCtx.currSeqStore;
3655
- size_t numSplits = ZSTD_deriveBlockSplits(zc, partitions, nbSeq);
4056
+ U32* const partitions = zc->blockSplitCtx.partitions; /* size == ZSTD_MAX_NB_BLOCK_SPLITS */
4057
+ seqStore_t* const nextSeqStore = &zc->blockSplitCtx.nextSeqStore;
4058
+ seqStore_t* const currSeqStore = &zc->blockSplitCtx.currSeqStore;
4059
+ size_t const numSplits = ZSTD_deriveBlockSplits(zc, partitions, nbSeq);
3656
4060
 
3657
4061
  /* If a block is split and some partitions are emitted as RLE/uncompressed, then repcode history
3658
4062
  * may become invalid. In order to reconcile potentially invalid repcodes, we keep track of two
@@ -3674,30 +4078,31 @@ ZSTD_compressBlock_splitBlock_internal(ZSTD_CCtx* zc, void* dst, size_t dstCapac
3674
4078
  ZSTD_memcpy(cRep.rep, zc->blockState.prevCBlock->rep, sizeof(repcodes_t));
3675
4079
  ZSTD_memset(nextSeqStore, 0, sizeof(seqStore_t));
3676
4080
 
3677
- DEBUGLOG(4, "ZSTD_compressBlock_splitBlock_internal (dstCapacity=%u, dictLimit=%u, nextToUpdate=%u)",
4081
+ DEBUGLOG(5, "ZSTD_compressBlock_splitBlock_internal (dstCapacity=%u, dictLimit=%u, nextToUpdate=%u)",
3678
4082
  (unsigned)dstCapacity, (unsigned)zc->blockState.matchState.window.dictLimit,
3679
4083
  (unsigned)zc->blockState.matchState.nextToUpdate);
3680
4084
 
3681
4085
  if (numSplits == 0) {
3682
- size_t cSizeSingleBlock = ZSTD_compressSeqStore_singleBlock(zc, &zc->seqStore,
3683
- &dRep, &cRep,
3684
- op, dstCapacity,
3685
- ip, blockSize,
3686
- lastBlock, 0 /* isPartition */);
4086
+ size_t cSizeSingleBlock =
4087
+ ZSTD_compressSeqStore_singleBlock(zc, &zc->seqStore,
4088
+ &dRep, &cRep,
4089
+ op, dstCapacity,
4090
+ ip, blockSize,
4091
+ lastBlock, 0 /* isPartition */);
3687
4092
  FORWARD_IF_ERROR(cSizeSingleBlock, "Compressing single block from splitBlock_internal() failed!");
3688
4093
  DEBUGLOG(5, "ZSTD_compressBlock_splitBlock_internal: No splits");
3689
- assert(cSizeSingleBlock <= ZSTD_BLOCKSIZE_MAX + ZSTD_blockHeaderSize);
4094
+ assert(zc->blockSize <= ZSTD_BLOCKSIZE_MAX);
4095
+ assert(cSizeSingleBlock <= zc->blockSize + ZSTD_blockHeaderSize);
3690
4096
  return cSizeSingleBlock;
3691
4097
  }
3692
4098
 
3693
4099
  ZSTD_deriveSeqStoreChunk(currSeqStore, &zc->seqStore, 0, partitions[0]);
3694
4100
  for (i = 0; i <= numSplits; ++i) {
3695
- size_t srcBytes;
3696
4101
  size_t cSizeChunk;
3697
4102
  U32 const lastPartition = (i == numSplits);
3698
4103
  U32 lastBlockEntireSrc = 0;
3699
4104
 
3700
- srcBytes = ZSTD_countSeqStoreLiteralsBytes(currSeqStore) + ZSTD_countSeqStoreMatchBytes(currSeqStore);
4105
+ size_t srcBytes = ZSTD_countSeqStoreLiteralsBytes(currSeqStore) + ZSTD_countSeqStoreMatchBytes(currSeqStore);
3701
4106
  srcBytesTotal += srcBytes;
3702
4107
  if (lastPartition) {
3703
4108
  /* This is the final partition, need to account for possible last literals */
@@ -3712,7 +4117,8 @@ ZSTD_compressBlock_splitBlock_internal(ZSTD_CCtx* zc, void* dst, size_t dstCapac
3712
4117
  op, dstCapacity,
3713
4118
  ip, srcBytes,
3714
4119
  lastBlockEntireSrc, 1 /* isPartition */);
3715
- DEBUGLOG(5, "Estimated size: %zu actual size: %zu", ZSTD_buildEntropyStatisticsAndEstimateSubBlockSize(currSeqStore, zc), cSizeChunk);
4120
+ DEBUGLOG(5, "Estimated size: %zu vs %zu : actual size",
4121
+ ZSTD_buildEntropyStatisticsAndEstimateSubBlockSize(currSeqStore, zc), cSizeChunk);
3716
4122
  FORWARD_IF_ERROR(cSizeChunk, "Compressing chunk failed!");
3717
4123
 
3718
4124
  ip += srcBytes;
@@ -3720,10 +4126,10 @@ ZSTD_compressBlock_splitBlock_internal(ZSTD_CCtx* zc, void* dst, size_t dstCapac
3720
4126
  dstCapacity -= cSizeChunk;
3721
4127
  cSize += cSizeChunk;
3722
4128
  *currSeqStore = *nextSeqStore;
3723
- assert(cSizeChunk <= ZSTD_BLOCKSIZE_MAX + ZSTD_blockHeaderSize);
4129
+ assert(cSizeChunk <= zc->blockSize + ZSTD_blockHeaderSize);
3724
4130
  }
3725
- /* cRep and dRep may have diverged during the compression. If so, we use the dRep repcodes
3726
- * for the next block.
4131
+ /* cRep and dRep may have diverged during the compression.
4132
+ * If so, we use the dRep repcodes for the next block.
3727
4133
  */
3728
4134
  ZSTD_memcpy(zc->blockState.prevCBlock->rep, dRep.rep, sizeof(repcodes_t));
3729
4135
  return cSize;
@@ -3734,8 +4140,6 @@ ZSTD_compressBlock_splitBlock(ZSTD_CCtx* zc,
3734
4140
  void* dst, size_t dstCapacity,
3735
4141
  const void* src, size_t srcSize, U32 lastBlock)
3736
4142
  {
3737
- const BYTE* ip = (const BYTE*)src;
3738
- BYTE* op = (BYTE*)dst;
3739
4143
  U32 nbSeq;
3740
4144
  size_t cSize;
3741
4145
  DEBUGLOG(4, "ZSTD_compressBlock_splitBlock");
@@ -3746,7 +4150,7 @@ ZSTD_compressBlock_splitBlock(ZSTD_CCtx* zc,
3746
4150
  if (bss == ZSTDbss_noCompress) {
3747
4151
  if (zc->blockState.prevCBlock->entropy.fse.offcode_repeatMode == FSE_repeat_valid)
3748
4152
  zc->blockState.prevCBlock->entropy.fse.offcode_repeatMode = FSE_repeat_check;
3749
- cSize = ZSTD_noCompressBlock(op, dstCapacity, ip, srcSize, lastBlock);
4153
+ cSize = ZSTD_noCompressBlock(dst, dstCapacity, src, srcSize, lastBlock);
3750
4154
  FORWARD_IF_ERROR(cSize, "ZSTD_noCompressBlock failed");
3751
4155
  DEBUGLOG(4, "ZSTD_compressBlock_splitBlock: Nocompress block");
3752
4156
  return cSize;
@@ -3764,9 +4168,9 @@ ZSTD_compressBlock_internal(ZSTD_CCtx* zc,
3764
4168
  void* dst, size_t dstCapacity,
3765
4169
  const void* src, size_t srcSize, U32 frame)
3766
4170
  {
3767
- /* This the upper bound for the length of an rle block.
3768
- * This isn't the actual upper bound. Finding the real threshold
3769
- * needs further investigation.
4171
+ /* This is an estimated upper bound for the length of an rle block.
4172
+ * This isn't the actual upper bound.
4173
+ * Finding the real threshold needs further investigation.
3770
4174
  */
3771
4175
  const U32 rleMaxLength = 25;
3772
4176
  size_t cSize;
@@ -3858,10 +4262,11 @@ static size_t ZSTD_compressBlock_targetCBlockSize_body(ZSTD_CCtx* zc,
3858
4262
  * * cSize >= blockBound(srcSize): We have expanded the block too much so
3859
4263
  * emit an uncompressed block.
3860
4264
  */
3861
- {
3862
- size_t const cSize = ZSTD_compressSuperBlock(zc, dst, dstCapacity, src, srcSize, lastBlock);
4265
+ { size_t const cSize =
4266
+ ZSTD_compressSuperBlock(zc, dst, dstCapacity, src, srcSize, lastBlock);
3863
4267
  if (cSize != ERROR(dstSize_tooSmall)) {
3864
- size_t const maxCSize = srcSize - ZSTD_minGain(srcSize, zc->appliedParams.cParams.strategy);
4268
+ size_t const maxCSize =
4269
+ srcSize - ZSTD_minGain(srcSize, zc->appliedParams.cParams.strategy);
3865
4270
  FORWARD_IF_ERROR(cSize, "ZSTD_compressSuperBlock failed");
3866
4271
  if (cSize != 0 && cSize < maxCSize + ZSTD_blockHeaderSize) {
3867
4272
  ZSTD_blockState_confirmRepcodesAndEntropyTables(&zc->blockState);
@@ -3869,7 +4274,7 @@ static size_t ZSTD_compressBlock_targetCBlockSize_body(ZSTD_CCtx* zc,
3869
4274
  }
3870
4275
  }
3871
4276
  }
3872
- }
4277
+ } /* if (bss == ZSTDbss_compress)*/
3873
4278
 
3874
4279
  DEBUGLOG(6, "Resorting to ZSTD_noCompressBlock()");
3875
4280
  /* Superblock compression failed, attempt to emit a single no compress block.
@@ -3927,7 +4332,7 @@ static void ZSTD_overflowCorrectIfNeeded(ZSTD_matchState_t* ms,
3927
4332
  * All blocks will be terminated, all input will be consumed.
3928
4333
  * Function will issue an error if there is not enough `dstCapacity` to hold the compressed content.
3929
4334
  * Frame is supposed already started (header already produced)
3930
- * @return : compressed size, or an error code
4335
+ * @return : compressed size, or an error code
3931
4336
  */
3932
4337
  static size_t ZSTD_compress_frameChunk(ZSTD_CCtx* cctx,
3933
4338
  void* dst, size_t dstCapacity,
@@ -3951,7 +4356,9 @@ static size_t ZSTD_compress_frameChunk(ZSTD_CCtx* cctx,
3951
4356
  ZSTD_matchState_t* const ms = &cctx->blockState.matchState;
3952
4357
  U32 const lastBlock = lastFrameChunk & (blockSize >= remaining);
3953
4358
 
3954
- RETURN_ERROR_IF(dstCapacity < ZSTD_blockHeaderSize + MIN_CBLOCK_SIZE,
4359
+ /* TODO: See 3090. We reduced MIN_CBLOCK_SIZE from 3 to 2 so to compensate we are adding
4360
+ * additional 1. We need to revisit and change this logic to be more consistent */
4361
+ RETURN_ERROR_IF(dstCapacity < ZSTD_blockHeaderSize + MIN_CBLOCK_SIZE + 1,
3955
4362
  dstSize_tooSmall,
3956
4363
  "not enough space to store compressed block");
3957
4364
  if (remaining < blockSize) blockSize = remaining;
@@ -3990,7 +4397,7 @@ static size_t ZSTD_compress_frameChunk(ZSTD_CCtx* cctx,
3990
4397
  MEM_writeLE24(op, cBlockHeader);
3991
4398
  cSize += ZSTD_blockHeaderSize;
3992
4399
  }
3993
- }
4400
+ } /* if (ZSTD_useTargetCBlockSize(&cctx->appliedParams))*/
3994
4401
 
3995
4402
 
3996
4403
  ip += blockSize;
@@ -4182,7 +4589,7 @@ size_t ZSTD_getBlockSize(const ZSTD_CCtx* cctx)
4182
4589
  {
4183
4590
  ZSTD_compressionParameters const cParams = cctx->appliedParams.cParams;
4184
4591
  assert(!ZSTD_checkCParams(cParams));
4185
- return MIN (ZSTD_BLOCKSIZE_MAX, (U32)1 << cParams.windowLog);
4592
+ return MIN(cctx->appliedParams.maxBlockSize, (size_t)1 << cParams.windowLog);
4186
4593
  }
4187
4594
 
4188
4595
  size_t ZSTD_compressBlock(ZSTD_CCtx* cctx, void* dst, size_t dstCapacity, const void* src, size_t srcSize)
@@ -4202,31 +4609,47 @@ static size_t ZSTD_loadDictionaryContent(ZSTD_matchState_t* ms,
4202
4609
  ZSTD_cwksp* ws,
4203
4610
  ZSTD_CCtx_params const* params,
4204
4611
  const void* src, size_t srcSize,
4205
- ZSTD_dictTableLoadMethod_e dtlm)
4612
+ ZSTD_dictTableLoadMethod_e dtlm,
4613
+ ZSTD_tableFillPurpose_e tfp)
4206
4614
  {
4207
4615
  const BYTE* ip = (const BYTE*) src;
4208
4616
  const BYTE* const iend = ip + srcSize;
4209
4617
  int const loadLdmDict = params->ldmParams.enableLdm == ZSTD_ps_enable && ls != NULL;
4210
4618
 
4211
- /* Assert that we the ms params match the params we're being given */
4619
+ /* Assert that the ms params match the params we're being given */
4212
4620
  ZSTD_assertEqualCParams(params->cParams, ms->cParams);
4213
4621
 
4214
- if (srcSize > ZSTD_CHUNKSIZE_MAX) {
4622
+ { /* Ensure large dictionaries can't cause index overflow */
4623
+
4215
4624
  /* Allow the dictionary to set indices up to exactly ZSTD_CURRENT_MAX.
4216
4625
  * Dictionaries right at the edge will immediately trigger overflow
4217
4626
  * correction, but I don't want to insert extra constraints here.
4218
4627
  */
4219
- U32 const maxDictSize = ZSTD_CURRENT_MAX - 1;
4220
- /* We must have cleared our windows when our source is this large. */
4221
- assert(ZSTD_window_isEmpty(ms->window));
4222
- if (loadLdmDict)
4223
- assert(ZSTD_window_isEmpty(ls->window));
4628
+ U32 maxDictSize = ZSTD_CURRENT_MAX - ZSTD_WINDOW_START_INDEX;
4629
+
4630
+ int const CDictTaggedIndices = ZSTD_CDictIndicesAreTagged(&params->cParams);
4631
+ if (CDictTaggedIndices && tfp == ZSTD_tfp_forCDict) {
4632
+ /* Some dictionary matchfinders in zstd use "short cache",
4633
+ * which treats the lower ZSTD_SHORT_CACHE_TAG_BITS of each
4634
+ * CDict hashtable entry as a tag rather than as part of an index.
4635
+ * When short cache is used, we need to truncate the dictionary
4636
+ * so that its indices don't overlap with the tag. */
4637
+ U32 const shortCacheMaxDictSize = (1u << (32 - ZSTD_SHORT_CACHE_TAG_BITS)) - ZSTD_WINDOW_START_INDEX;
4638
+ maxDictSize = MIN(maxDictSize, shortCacheMaxDictSize);
4639
+ assert(!loadLdmDict);
4640
+ }
4641
+
4224
4642
  /* If the dictionary is too large, only load the suffix of the dictionary. */
4225
4643
  if (srcSize > maxDictSize) {
4226
4644
  ip = iend - maxDictSize;
4227
4645
  src = ip;
4228
4646
  srcSize = maxDictSize;
4229
- }
4647
+ } }
4648
+
4649
+ if (srcSize > ZSTD_CHUNKSIZE_MAX) {
4650
+ /* We must have cleared our windows when our source is this large. */
4651
+ assert(ZSTD_window_isEmpty(ms->window));
4652
+ if (loadLdmDict) assert(ZSTD_window_isEmpty(ls->window));
4230
4653
  }
4231
4654
 
4232
4655
  DEBUGLOG(4, "ZSTD_loadDictionaryContent(): useRowMatchFinder=%d", (int)params->useRowMatchFinder);
@@ -4249,10 +4672,10 @@ static size_t ZSTD_loadDictionaryContent(ZSTD_matchState_t* ms,
4249
4672
  switch(params->cParams.strategy)
4250
4673
  {
4251
4674
  case ZSTD_fast:
4252
- ZSTD_fillHashTable(ms, iend, dtlm);
4675
+ ZSTD_fillHashTable(ms, iend, dtlm, tfp);
4253
4676
  break;
4254
4677
  case ZSTD_dfast:
4255
- ZSTD_fillDoubleHashTable(ms, iend, dtlm);
4678
+ ZSTD_fillDoubleHashTable(ms, iend, dtlm, tfp);
4256
4679
  break;
4257
4680
 
4258
4681
  case ZSTD_greedy:
@@ -4418,6 +4841,7 @@ static size_t ZSTD_loadZstdDictionary(ZSTD_compressedBlockState_t* bs,
4418
4841
  ZSTD_CCtx_params const* params,
4419
4842
  const void* dict, size_t dictSize,
4420
4843
  ZSTD_dictTableLoadMethod_e dtlm,
4844
+ ZSTD_tableFillPurpose_e tfp,
4421
4845
  void* workspace)
4422
4846
  {
4423
4847
  const BYTE* dictPtr = (const BYTE*)dict;
@@ -4436,7 +4860,7 @@ static size_t ZSTD_loadZstdDictionary(ZSTD_compressedBlockState_t* bs,
4436
4860
  {
4437
4861
  size_t const dictContentSize = (size_t)(dictEnd - dictPtr);
4438
4862
  FORWARD_IF_ERROR(ZSTD_loadDictionaryContent(
4439
- ms, NULL, ws, params, dictPtr, dictContentSize, dtlm), "");
4863
+ ms, NULL, ws, params, dictPtr, dictContentSize, dtlm, tfp), "");
4440
4864
  }
4441
4865
  return dictID;
4442
4866
  }
@@ -4452,6 +4876,7 @@ ZSTD_compress_insertDictionary(ZSTD_compressedBlockState_t* bs,
4452
4876
  const void* dict, size_t dictSize,
4453
4877
  ZSTD_dictContentType_e dictContentType,
4454
4878
  ZSTD_dictTableLoadMethod_e dtlm,
4879
+ ZSTD_tableFillPurpose_e tfp,
4455
4880
  void* workspace)
4456
4881
  {
4457
4882
  DEBUGLOG(4, "ZSTD_compress_insertDictionary (dictSize=%u)", (U32)dictSize);
@@ -4464,13 +4889,13 @@ ZSTD_compress_insertDictionary(ZSTD_compressedBlockState_t* bs,
4464
4889
 
4465
4890
  /* dict restricted modes */
4466
4891
  if (dictContentType == ZSTD_dct_rawContent)
4467
- return ZSTD_loadDictionaryContent(ms, ls, ws, params, dict, dictSize, dtlm);
4892
+ return ZSTD_loadDictionaryContent(ms, ls, ws, params, dict, dictSize, dtlm, tfp);
4468
4893
 
4469
4894
  if (MEM_readLE32(dict) != ZSTD_MAGIC_DICTIONARY) {
4470
4895
  if (dictContentType == ZSTD_dct_auto) {
4471
4896
  DEBUGLOG(4, "raw content dictionary detected");
4472
4897
  return ZSTD_loadDictionaryContent(
4473
- ms, ls, ws, params, dict, dictSize, dtlm);
4898
+ ms, ls, ws, params, dict, dictSize, dtlm, tfp);
4474
4899
  }
4475
4900
  RETURN_ERROR_IF(dictContentType == ZSTD_dct_fullDict, dictionary_wrong, "");
4476
4901
  assert(0); /* impossible */
@@ -4478,13 +4903,14 @@ ZSTD_compress_insertDictionary(ZSTD_compressedBlockState_t* bs,
4478
4903
 
4479
4904
  /* dict as full zstd dictionary */
4480
4905
  return ZSTD_loadZstdDictionary(
4481
- bs, ms, ws, params, dict, dictSize, dtlm, workspace);
4906
+ bs, ms, ws, params, dict, dictSize, dtlm, tfp, workspace);
4482
4907
  }
4483
4908
 
4484
4909
  #define ZSTD_USE_CDICT_PARAMS_SRCSIZE_CUTOFF (128 KB)
4485
4910
  #define ZSTD_USE_CDICT_PARAMS_DICTSIZE_MULTIPLIER (6ULL)
4486
4911
 
4487
4912
  /*! ZSTD_compressBegin_internal() :
4913
+ * Assumption : either @dict OR @cdict (or none) is non-NULL, never both
4488
4914
  * @return : 0, or an error code */
4489
4915
  static size_t ZSTD_compressBegin_internal(ZSTD_CCtx* cctx,
4490
4916
  const void* dict, size_t dictSize,
@@ -4520,11 +4946,11 @@ static size_t ZSTD_compressBegin_internal(ZSTD_CCtx* cctx,
4520
4946
  cctx->blockState.prevCBlock, &cctx->blockState.matchState,
4521
4947
  &cctx->ldmState, &cctx->workspace, &cctx->appliedParams, cdict->dictContent,
4522
4948
  cdict->dictContentSize, cdict->dictContentType, dtlm,
4523
- cctx->entropyWorkspace)
4949
+ ZSTD_tfp_forCCtx, cctx->entropyWorkspace)
4524
4950
  : ZSTD_compress_insertDictionary(
4525
4951
  cctx->blockState.prevCBlock, &cctx->blockState.matchState,
4526
4952
  &cctx->ldmState, &cctx->workspace, &cctx->appliedParams, dict, dictSize,
4527
- dictContentType, dtlm, cctx->entropyWorkspace);
4953
+ dictContentType, dtlm, ZSTD_tfp_forCCtx, cctx->entropyWorkspace);
4528
4954
  FORWARD_IF_ERROR(dictID, "ZSTD_compress_insertDictionary failed");
4529
4955
  assert(dictID <= UINT_MAX);
4530
4956
  cctx->dictID = (U32)dictID;
@@ -4565,11 +4991,11 @@ size_t ZSTD_compressBegin_advanced(ZSTD_CCtx* cctx,
4565
4991
  &cctxParams, pledgedSrcSize);
4566
4992
  }
4567
4993
 
4568
- size_t ZSTD_compressBegin_usingDict(ZSTD_CCtx* cctx, const void* dict, size_t dictSize, int compressionLevel)
4994
+ size_t
4995
+ ZSTD_compressBegin_usingDict(ZSTD_CCtx* cctx, const void* dict, size_t dictSize, int compressionLevel)
4569
4996
  {
4570
4997
  ZSTD_CCtx_params cctxParams;
4571
- {
4572
- ZSTD_parameters const params = ZSTD_getParams_internal(compressionLevel, ZSTD_CONTENTSIZE_UNKNOWN, dictSize, ZSTD_cpm_noAttachDict);
4998
+ { ZSTD_parameters const params = ZSTD_getParams_internal(compressionLevel, ZSTD_CONTENTSIZE_UNKNOWN, dictSize, ZSTD_cpm_noAttachDict);
4573
4999
  ZSTD_CCtxParams_init_internal(&cctxParams, &params, (compressionLevel == 0) ? ZSTD_CLEVEL_DEFAULT : compressionLevel);
4574
5000
  }
4575
5001
  DEBUGLOG(4, "ZSTD_compressBegin_usingDict (dictSize=%u)", (unsigned)dictSize);
@@ -4828,7 +5254,7 @@ static size_t ZSTD_initCDict_internal(
4828
5254
  { size_t const dictID = ZSTD_compress_insertDictionary(
4829
5255
  &cdict->cBlockState, &cdict->matchState, NULL, &cdict->workspace,
4830
5256
  &params, cdict->dictContent, cdict->dictContentSize,
4831
- dictContentType, ZSTD_dtlm_full, cdict->entropyWorkspace);
5257
+ dictContentType, ZSTD_dtlm_full, ZSTD_tfp_forCDict, cdict->entropyWorkspace);
4832
5258
  FORWARD_IF_ERROR(dictID, "ZSTD_compress_insertDictionary failed");
4833
5259
  assert(dictID <= (size_t)(U32)-1);
4834
5260
  cdict->dictID = (U32)dictID;
@@ -5316,30 +5742,41 @@ size_t ZSTD_initCStream(ZSTD_CStream* zcs, int compressionLevel)
5316
5742
 
5317
5743
  static size_t ZSTD_nextInputSizeHint(const ZSTD_CCtx* cctx)
5318
5744
  {
5319
- size_t hintInSize = cctx->inBuffTarget - cctx->inBuffPos;
5320
- if (hintInSize==0) hintInSize = cctx->blockSize;
5321
- return hintInSize;
5745
+ if (cctx->appliedParams.inBufferMode == ZSTD_bm_stable) {
5746
+ return cctx->blockSize - cctx->stableIn_notConsumed;
5747
+ }
5748
+ assert(cctx->appliedParams.inBufferMode == ZSTD_bm_buffered);
5749
+ { size_t hintInSize = cctx->inBuffTarget - cctx->inBuffPos;
5750
+ if (hintInSize==0) hintInSize = cctx->blockSize;
5751
+ return hintInSize;
5752
+ }
5322
5753
  }
5323
5754
 
5324
5755
  /** ZSTD_compressStream_generic():
5325
5756
  * internal function for all *compressStream*() variants
5326
- * non-static, because can be called from zstdmt_compress.c
5327
- * @return : hint size for next input */
5757
+ * @return : hint size for next input to complete ongoing block */
5328
5758
  static size_t ZSTD_compressStream_generic(ZSTD_CStream* zcs,
5329
5759
  ZSTD_outBuffer* output,
5330
5760
  ZSTD_inBuffer* input,
5331
5761
  ZSTD_EndDirective const flushMode)
5332
5762
  {
5333
- const char* const istart = (const char*)input->src;
5334
- const char* const iend = input->size != 0 ? istart + input->size : istart;
5335
- const char* ip = input->pos != 0 ? istart + input->pos : istart;
5336
- char* const ostart = (char*)output->dst;
5337
- char* const oend = output->size != 0 ? ostart + output->size : ostart;
5338
- char* op = output->pos != 0 ? ostart + output->pos : ostart;
5763
+ const char* const istart = (assert(input != NULL), (const char*)input->src);
5764
+ const char* const iend = (istart != NULL) ? istart + input->size : istart;
5765
+ const char* ip = (istart != NULL) ? istart + input->pos : istart;
5766
+ char* const ostart = (assert(output != NULL), (char*)output->dst);
5767
+ char* const oend = (ostart != NULL) ? ostart + output->size : ostart;
5768
+ char* op = (ostart != NULL) ? ostart + output->pos : ostart;
5339
5769
  U32 someMoreWork = 1;
5340
5770
 
5341
5771
  /* check expectations */
5342
- DEBUGLOG(5, "ZSTD_compressStream_generic, flush=%u", (unsigned)flushMode);
5772
+ DEBUGLOG(5, "ZSTD_compressStream_generic, flush=%i, srcSize = %zu", (int)flushMode, input->size - input->pos);
5773
+ assert(zcs != NULL);
5774
+ if (zcs->appliedParams.inBufferMode == ZSTD_bm_stable) {
5775
+ assert(input->pos >= zcs->stableIn_notConsumed);
5776
+ input->pos -= zcs->stableIn_notConsumed;
5777
+ ip -= zcs->stableIn_notConsumed;
5778
+ zcs->stableIn_notConsumed = 0;
5779
+ }
5343
5780
  if (zcs->appliedParams.inBufferMode == ZSTD_bm_buffered) {
5344
5781
  assert(zcs->inBuff != NULL);
5345
5782
  assert(zcs->inBuffSize > 0);
@@ -5348,8 +5785,10 @@ static size_t ZSTD_compressStream_generic(ZSTD_CStream* zcs,
5348
5785
  assert(zcs->outBuff != NULL);
5349
5786
  assert(zcs->outBuffSize > 0);
5350
5787
  }
5351
- assert(output->pos <= output->size);
5788
+ if (input->src == NULL) assert(input->size == 0);
5352
5789
  assert(input->pos <= input->size);
5790
+ if (output->dst == NULL) assert(output->size == 0);
5791
+ assert(output->pos <= output->size);
5353
5792
  assert((U32)flushMode <= (U32)ZSTD_e_end);
5354
5793
 
5355
5794
  while (someMoreWork) {
@@ -5381,8 +5820,7 @@ static size_t ZSTD_compressStream_generic(ZSTD_CStream* zcs,
5381
5820
  zcs->inBuff + zcs->inBuffPos, toLoad,
5382
5821
  ip, iend-ip);
5383
5822
  zcs->inBuffPos += loaded;
5384
- if (loaded != 0)
5385
- ip += loaded;
5823
+ if (ip) ip += loaded;
5386
5824
  if ( (flushMode == ZSTD_e_continue)
5387
5825
  && (zcs->inBuffPos < zcs->inBuffTarget) ) {
5388
5826
  /* not enough input to fill full block : stop here */
@@ -5393,6 +5831,20 @@ static size_t ZSTD_compressStream_generic(ZSTD_CStream* zcs,
5393
5831
  /* empty */
5394
5832
  someMoreWork = 0; break;
5395
5833
  }
5834
+ } else {
5835
+ assert(zcs->appliedParams.inBufferMode == ZSTD_bm_stable);
5836
+ if ( (flushMode == ZSTD_e_continue)
5837
+ && ( (size_t)(iend - ip) < zcs->blockSize) ) {
5838
+ /* can't compress a full block : stop here */
5839
+ zcs->stableIn_notConsumed = (size_t)(iend - ip);
5840
+ ip = iend; /* pretend to have consumed input */
5841
+ someMoreWork = 0; break;
5842
+ }
5843
+ if ( (flushMode == ZSTD_e_flush)
5844
+ && (ip == iend) ) {
5845
+ /* empty */
5846
+ someMoreWork = 0; break;
5847
+ }
5396
5848
  }
5397
5849
  /* compress current block (note : this stage cannot be stopped in the middle) */
5398
5850
  DEBUGLOG(5, "stream compression stage (flushMode==%u)", flushMode);
@@ -5400,9 +5852,8 @@ static size_t ZSTD_compressStream_generic(ZSTD_CStream* zcs,
5400
5852
  void* cDst;
5401
5853
  size_t cSize;
5402
5854
  size_t oSize = oend-op;
5403
- size_t const iSize = inputBuffered
5404
- ? zcs->inBuffPos - zcs->inToCompress
5405
- : MIN((size_t)(iend - ip), zcs->blockSize);
5855
+ size_t const iSize = inputBuffered ? zcs->inBuffPos - zcs->inToCompress
5856
+ : MIN((size_t)(iend - ip), zcs->blockSize);
5406
5857
  if (oSize >= ZSTD_compressBound(iSize) || zcs->appliedParams.outBufferMode == ZSTD_bm_stable)
5407
5858
  cDst = op; /* compress into output buffer, to skip flush stage */
5408
5859
  else
@@ -5425,19 +5876,16 @@ static size_t ZSTD_compressStream_generic(ZSTD_CStream* zcs,
5425
5876
  if (!lastBlock)
5426
5877
  assert(zcs->inBuffTarget <= zcs->inBuffSize);
5427
5878
  zcs->inToCompress = zcs->inBuffPos;
5428
- } else {
5429
- unsigned const lastBlock = (ip + iSize == iend);
5430
- assert(flushMode == ZSTD_e_end /* Already validated */);
5879
+ } else { /* !inputBuffered, hence ZSTD_bm_stable */
5880
+ unsigned const lastBlock = (flushMode == ZSTD_e_end) && (ip + iSize == iend);
5431
5881
  cSize = lastBlock ?
5432
5882
  ZSTD_compressEnd(zcs, cDst, oSize, ip, iSize) :
5433
5883
  ZSTD_compressContinue(zcs, cDst, oSize, ip, iSize);
5434
5884
  /* Consume the input prior to error checking to mirror buffered mode. */
5435
- if (iSize > 0)
5436
- ip += iSize;
5885
+ if (ip) ip += iSize;
5437
5886
  FORWARD_IF_ERROR(cSize, "%s", lastBlock ? "ZSTD_compressEnd failed" : "ZSTD_compressContinue failed");
5438
5887
  zcs->frameEnded = lastBlock;
5439
- if (lastBlock)
5440
- assert(ip == iend);
5888
+ if (lastBlock) assert(ip == iend);
5441
5889
  }
5442
5890
  if (cDst == op) { /* no need to flush */
5443
5891
  op += cSize;
@@ -5513,8 +5961,10 @@ size_t ZSTD_compressStream(ZSTD_CStream* zcs, ZSTD_outBuffer* output, ZSTD_inBuf
5513
5961
  /* After a compression call set the expected input/output buffer.
5514
5962
  * This is validated at the start of the next compression call.
5515
5963
  */
5516
- static void ZSTD_setBufferExpectations(ZSTD_CCtx* cctx, ZSTD_outBuffer const* output, ZSTD_inBuffer const* input)
5964
+ static void
5965
+ ZSTD_setBufferExpectations(ZSTD_CCtx* cctx, const ZSTD_outBuffer* output, const ZSTD_inBuffer* input)
5517
5966
  {
5967
+ DEBUGLOG(5, "ZSTD_setBufferExpectations (for advanced stable in/out modes)");
5518
5968
  if (cctx->appliedParams.inBufferMode == ZSTD_bm_stable) {
5519
5969
  cctx->expectedInBuffer = *input;
5520
5970
  }
@@ -5533,22 +5983,22 @@ static size_t ZSTD_checkBufferStability(ZSTD_CCtx const* cctx,
5533
5983
  {
5534
5984
  if (cctx->appliedParams.inBufferMode == ZSTD_bm_stable) {
5535
5985
  ZSTD_inBuffer const expect = cctx->expectedInBuffer;
5536
- if (expect.src != input->src || expect.pos != input->pos || expect.size != input->size)
5537
- RETURN_ERROR(srcBuffer_wrong, "ZSTD_c_stableInBuffer enabled but input differs!");
5538
- if (endOp != ZSTD_e_end)
5539
- RETURN_ERROR(srcBuffer_wrong, "ZSTD_c_stableInBuffer can only be used with ZSTD_e_end!");
5986
+ if (expect.src != input->src || expect.pos != input->pos)
5987
+ RETURN_ERROR(stabilityCondition_notRespected, "ZSTD_c_stableInBuffer enabled but input differs!");
5540
5988
  }
5989
+ (void)endOp;
5541
5990
  if (cctx->appliedParams.outBufferMode == ZSTD_bm_stable) {
5542
5991
  size_t const outBufferSize = output->size - output->pos;
5543
5992
  if (cctx->expectedOutBufferSize != outBufferSize)
5544
- RETURN_ERROR(dstBuffer_wrong, "ZSTD_c_stableOutBuffer enabled but output size differs!");
5993
+ RETURN_ERROR(stabilityCondition_notRespected, "ZSTD_c_stableOutBuffer enabled but output size differs!");
5545
5994
  }
5546
5995
  return 0;
5547
5996
  }
5548
5997
 
5549
5998
  static size_t ZSTD_CCtx_init_compressStream2(ZSTD_CCtx* cctx,
5550
5999
  ZSTD_EndDirective endOp,
5551
- size_t inSize) {
6000
+ size_t inSize)
6001
+ {
5552
6002
  ZSTD_CCtx_params params = cctx->requestedParams;
5553
6003
  ZSTD_prefixDict const prefixDict = cctx->prefixDict;
5554
6004
  FORWARD_IF_ERROR( ZSTD_initLocalDict(cctx) , ""); /* Init the local dict if present. */
@@ -5562,9 +6012,9 @@ static size_t ZSTD_CCtx_init_compressStream2(ZSTD_CCtx* cctx,
5562
6012
  params.compressionLevel = cctx->cdict->compressionLevel;
5563
6013
  }
5564
6014
  DEBUGLOG(4, "ZSTD_compressStream2 : transparent init stage");
5565
- if (endOp == ZSTD_e_end) cctx->pledgedSrcSizePlusOne = inSize + 1; /* auto-fix pledgedSrcSize */
5566
- {
5567
- size_t const dictSize = prefixDict.dict
6015
+ if (endOp == ZSTD_e_end) cctx->pledgedSrcSizePlusOne = inSize + 1; /* auto-determine pledgedSrcSize */
6016
+
6017
+ { size_t const dictSize = prefixDict.dict
5568
6018
  ? prefixDict.dictSize
5569
6019
  : (cctx->cdict ? cctx->cdict->dictContentSize : 0);
5570
6020
  ZSTD_cParamMode_e const mode = ZSTD_getCParamMode(cctx->cdict, &params, cctx->pledgedSrcSizePlusOne - 1);
@@ -5576,8 +6026,18 @@ static size_t ZSTD_CCtx_init_compressStream2(ZSTD_CCtx* cctx,
5576
6026
  params.useBlockSplitter = ZSTD_resolveBlockSplitterMode(params.useBlockSplitter, &params.cParams);
5577
6027
  params.ldmParams.enableLdm = ZSTD_resolveEnableLdm(params.ldmParams.enableLdm, &params.cParams);
5578
6028
  params.useRowMatchFinder = ZSTD_resolveRowMatchFinderMode(params.useRowMatchFinder, &params.cParams);
6029
+ params.validateSequences = ZSTD_resolveExternalSequenceValidation(params.validateSequences);
6030
+ params.maxBlockSize = ZSTD_resolveMaxBlockSize(params.maxBlockSize);
6031
+ params.searchForExternalRepcodes = ZSTD_resolveExternalRepcodeSearch(params.searchForExternalRepcodes, params.compressionLevel);
5579
6032
 
5580
6033
  #ifdef ZSTD_MULTITHREAD
6034
+ /* If external matchfinder is enabled, make sure to fail before checking job size (for consistency) */
6035
+ RETURN_ERROR_IF(
6036
+ params.useSequenceProducer == 1 && params.nbWorkers >= 1,
6037
+ parameter_combination_unsupported,
6038
+ "External sequence producer isn't supported with nbWorkers >= 1"
6039
+ );
6040
+
5581
6041
  if ((cctx->pledgedSrcSizePlusOne-1) <= ZSTDMT_JOBSIZE_MIN) {
5582
6042
  params.nbWorkers = 0; /* do not invoke multi-threading when src size is too small */
5583
6043
  }
@@ -5605,7 +6065,7 @@ static size_t ZSTD_CCtx_init_compressStream2(ZSTD_CCtx* cctx,
5605
6065
  cctx->streamStage = zcss_load;
5606
6066
  cctx->appliedParams = params;
5607
6067
  } else
5608
- #endif
6068
+ #endif /* ZSTD_MULTITHREAD */
5609
6069
  { U64 const pledgedSrcSize = cctx->pledgedSrcSizePlusOne - 1;
5610
6070
  assert(!ZSTD_isError(ZSTD_checkCParams(params.cParams)));
5611
6071
  FORWARD_IF_ERROR( ZSTD_compressBegin_internal(cctx,
@@ -5631,6 +6091,8 @@ static size_t ZSTD_CCtx_init_compressStream2(ZSTD_CCtx* cctx,
5631
6091
  return 0;
5632
6092
  }
5633
6093
 
6094
+ /* @return provides a minimum amount of data remaining to be flushed from internal buffers
6095
+ */
5634
6096
  size_t ZSTD_compressStream2( ZSTD_CCtx* cctx,
5635
6097
  ZSTD_outBuffer* output,
5636
6098
  ZSTD_inBuffer* input,
@@ -5645,8 +6107,27 @@ size_t ZSTD_compressStream2( ZSTD_CCtx* cctx,
5645
6107
 
5646
6108
  /* transparent initialization stage */
5647
6109
  if (cctx->streamStage == zcss_init) {
5648
- FORWARD_IF_ERROR(ZSTD_CCtx_init_compressStream2(cctx, endOp, input->size), "CompressStream2 initialization failed");
5649
- ZSTD_setBufferExpectations(cctx, output, input); /* Set initial buffer expectations now that we've initialized */
6110
+ size_t const inputSize = input->size - input->pos; /* no obligation to start from pos==0 */
6111
+ size_t const totalInputSize = inputSize + cctx->stableIn_notConsumed;
6112
+ if ( (cctx->requestedParams.inBufferMode == ZSTD_bm_stable) /* input is presumed stable, across invocations */
6113
+ && (endOp == ZSTD_e_continue) /* no flush requested, more input to come */
6114
+ && (totalInputSize < ZSTD_BLOCKSIZE_MAX) ) { /* not even reached one block yet */
6115
+ if (cctx->stableIn_notConsumed) { /* not the first time */
6116
+ /* check stable source guarantees */
6117
+ RETURN_ERROR_IF(input->src != cctx->expectedInBuffer.src, stabilityCondition_notRespected, "stableInBuffer condition not respected: wrong src pointer");
6118
+ RETURN_ERROR_IF(input->pos != cctx->expectedInBuffer.size, stabilityCondition_notRespected, "stableInBuffer condition not respected: externally modified pos");
6119
+ }
6120
+ /* pretend input was consumed, to give a sense forward progress */
6121
+ input->pos = input->size;
6122
+ /* save stable inBuffer, for later control, and flush/end */
6123
+ cctx->expectedInBuffer = *input;
6124
+ /* but actually input wasn't consumed, so keep track of position from where compression shall resume */
6125
+ cctx->stableIn_notConsumed += inputSize;
6126
+ /* don't initialize yet, wait for the first block of flush() order, for better parameters adaptation */
6127
+ return ZSTD_FRAMEHEADERSIZE_MIN(cctx->requestedParams.format); /* at least some header to produce */
6128
+ }
6129
+ FORWARD_IF_ERROR(ZSTD_CCtx_init_compressStream2(cctx, endOp, totalInputSize), "compressStream2 initialization failed");
6130
+ ZSTD_setBufferExpectations(cctx, output, input); /* Set initial buffer expectations now that we've initialized */
5650
6131
  }
5651
6132
  /* end of transparent initialization stage */
5652
6133
 
@@ -5659,6 +6140,13 @@ size_t ZSTD_compressStream2( ZSTD_CCtx* cctx,
5659
6140
  ZSTDMT_updateCParams_whileCompressing(cctx->mtctx, &cctx->requestedParams);
5660
6141
  cctx->cParamsChanged = 0;
5661
6142
  }
6143
+ if (cctx->stableIn_notConsumed) {
6144
+ assert(cctx->appliedParams.inBufferMode == ZSTD_bm_stable);
6145
+ /* some early data was skipped - make it available for consumption */
6146
+ assert(input->pos >= cctx->stableIn_notConsumed);
6147
+ input->pos -= cctx->stableIn_notConsumed;
6148
+ cctx->stableIn_notConsumed = 0;
6149
+ }
5662
6150
  for (;;) {
5663
6151
  size_t const ipos = input->pos;
5664
6152
  size_t const opos = output->pos;
@@ -5697,7 +6185,7 @@ size_t ZSTD_compressStream2( ZSTD_CCtx* cctx,
5697
6185
  ZSTD_setBufferExpectations(cctx, output, input);
5698
6186
  return flushMin;
5699
6187
  }
5700
- #endif
6188
+ #endif /* ZSTD_MULTITHREAD */
5701
6189
  FORWARD_IF_ERROR( ZSTD_compressStream_generic(cctx, output, input, endOp) , "");
5702
6190
  DEBUGLOG(5, "completed ZSTD_compressStream2");
5703
6191
  ZSTD_setBufferExpectations(cctx, output, input);
@@ -5710,13 +6198,20 @@ size_t ZSTD_compressStream2_simpleArgs (
5710
6198
  const void* src, size_t srcSize, size_t* srcPos,
5711
6199
  ZSTD_EndDirective endOp)
5712
6200
  {
5713
- ZSTD_outBuffer output = { dst, dstCapacity, *dstPos };
5714
- ZSTD_inBuffer input = { src, srcSize, *srcPos };
6201
+ ZSTD_outBuffer output;
6202
+ ZSTD_inBuffer input;
6203
+ output.dst = dst;
6204
+ output.size = dstCapacity;
6205
+ output.pos = *dstPos;
6206
+ input.src = src;
6207
+ input.size = srcSize;
6208
+ input.pos = *srcPos;
5715
6209
  /* ZSTD_compressStream2() will check validity of dstPos and srcPos */
5716
- size_t const cErr = ZSTD_compressStream2(cctx, &output, &input, endOp);
5717
- *dstPos = output.pos;
5718
- *srcPos = input.pos;
5719
- return cErr;
6210
+ { size_t const cErr = ZSTD_compressStream2(cctx, &output, &input, endOp);
6211
+ *dstPos = output.pos;
6212
+ *srcPos = input.pos;
6213
+ return cErr;
6214
+ }
5720
6215
  }
5721
6216
 
5722
6217
  size_t ZSTD_compress2(ZSTD_CCtx* cctx,
@@ -5739,6 +6234,7 @@ size_t ZSTD_compress2(ZSTD_CCtx* cctx,
5739
6234
  /* Reset to the original values. */
5740
6235
  cctx->requestedParams.inBufferMode = originalInBufferMode;
5741
6236
  cctx->requestedParams.outBufferMode = originalOutBufferMode;
6237
+
5742
6238
  FORWARD_IF_ERROR(result, "ZSTD_compressStream2_simpleArgs failed");
5743
6239
  if (result != 0) { /* compression not completed, due to lack of output space */
5744
6240
  assert(oPos == dstCapacity);
@@ -5749,64 +6245,61 @@ size_t ZSTD_compress2(ZSTD_CCtx* cctx,
5749
6245
  }
5750
6246
  }
5751
6247
 
5752
- typedef struct {
5753
- U32 idx; /* Index in array of ZSTD_Sequence */
5754
- U32 posInSequence; /* Position within sequence at idx */
5755
- size_t posInSrc; /* Number of bytes given by sequences provided so far */
5756
- } ZSTD_sequencePosition;
5757
-
5758
6248
  /* ZSTD_validateSequence() :
5759
6249
  * @offCode : is presumed to follow format required by ZSTD_storeSeq()
5760
6250
  * @returns a ZSTD error code if sequence is not valid
5761
6251
  */
5762
6252
  static size_t
5763
- ZSTD_validateSequence(U32 offCode, U32 matchLength,
5764
- size_t posInSrc, U32 windowLog, size_t dictSize)
6253
+ ZSTD_validateSequence(U32 offCode, U32 matchLength, U32 minMatch,
6254
+ size_t posInSrc, U32 windowLog, size_t dictSize, int useSequenceProducer)
5765
6255
  {
5766
- U32 const windowSize = 1 << windowLog;
5767
- /* posInSrc represents the amount of data the the decoder would decode up to this point.
6256
+ U32 const windowSize = 1u << windowLog;
6257
+ /* posInSrc represents the amount of data the decoder would decode up to this point.
5768
6258
  * As long as the amount of data decoded is less than or equal to window size, offsets may be
5769
6259
  * larger than the total length of output decoded in order to reference the dict, even larger than
5770
6260
  * window size. After output surpasses windowSize, we're limited to windowSize offsets again.
5771
6261
  */
5772
6262
  size_t const offsetBound = posInSrc > windowSize ? (size_t)windowSize : posInSrc + (size_t)dictSize;
5773
- RETURN_ERROR_IF(offCode > STORE_OFFSET(offsetBound), corruption_detected, "Offset too large!");
5774
- RETURN_ERROR_IF(matchLength < MINMATCH, corruption_detected, "Matchlength too small");
6263
+ size_t const matchLenLowerBound = (minMatch == 3 || useSequenceProducer) ? 3 : 4;
6264
+ RETURN_ERROR_IF(offCode > OFFSET_TO_OFFBASE(offsetBound), externalSequences_invalid, "Offset too large!");
6265
+ /* Validate maxNbSeq is large enough for the given matchLength and minMatch */
6266
+ RETURN_ERROR_IF(matchLength < matchLenLowerBound, externalSequences_invalid, "Matchlength too small for the minMatch");
5775
6267
  return 0;
5776
6268
  }
5777
6269
 
5778
6270
  /* Returns an offset code, given a sequence's raw offset, the ongoing repcode array, and whether litLength == 0 */
5779
- static U32 ZSTD_finalizeOffCode(U32 rawOffset, const U32 rep[ZSTD_REP_NUM], U32 ll0)
6271
+ static U32 ZSTD_finalizeOffBase(U32 rawOffset, const U32 rep[ZSTD_REP_NUM], U32 ll0)
5780
6272
  {
5781
- U32 offCode = STORE_OFFSET(rawOffset);
6273
+ U32 offBase = OFFSET_TO_OFFBASE(rawOffset);
5782
6274
 
5783
6275
  if (!ll0 && rawOffset == rep[0]) {
5784
- offCode = STORE_REPCODE_1;
6276
+ offBase = REPCODE1_TO_OFFBASE;
5785
6277
  } else if (rawOffset == rep[1]) {
5786
- offCode = STORE_REPCODE(2 - ll0);
6278
+ offBase = REPCODE_TO_OFFBASE(2 - ll0);
5787
6279
  } else if (rawOffset == rep[2]) {
5788
- offCode = STORE_REPCODE(3 - ll0);
6280
+ offBase = REPCODE_TO_OFFBASE(3 - ll0);
5789
6281
  } else if (ll0 && rawOffset == rep[0] - 1) {
5790
- offCode = STORE_REPCODE_3;
6282
+ offBase = REPCODE3_TO_OFFBASE;
5791
6283
  }
5792
- return offCode;
6284
+ return offBase;
5793
6285
  }
5794
6286
 
5795
- /* Returns 0 on success, and a ZSTD_error otherwise. This function scans through an array of
5796
- * ZSTD_Sequence, storing the sequences it finds, until it reaches a block delimiter.
5797
- */
5798
- static size_t
6287
+ size_t
5799
6288
  ZSTD_copySequencesToSeqStoreExplicitBlockDelim(ZSTD_CCtx* cctx,
5800
6289
  ZSTD_sequencePosition* seqPos,
5801
6290
  const ZSTD_Sequence* const inSeqs, size_t inSeqsSize,
5802
- const void* src, size_t blockSize)
6291
+ const void* src, size_t blockSize,
6292
+ ZSTD_paramSwitch_e externalRepSearch)
5803
6293
  {
5804
6294
  U32 idx = seqPos->idx;
6295
+ U32 const startIdx = idx;
5805
6296
  BYTE const* ip = (BYTE const*)(src);
5806
6297
  const BYTE* const iend = ip + blockSize;
5807
6298
  repcodes_t updatedRepcodes;
5808
6299
  U32 dictSize;
5809
6300
 
6301
+ DEBUGLOG(5, "ZSTD_copySequencesToSeqStoreExplicitBlockDelim (blockSize = %zu)", blockSize);
6302
+
5810
6303
  if (cctx->cdict) {
5811
6304
  dictSize = (U32)cctx->cdict->dictContentSize;
5812
6305
  } else if (cctx->prefixDict.dict) {
@@ -5815,25 +6308,55 @@ ZSTD_copySequencesToSeqStoreExplicitBlockDelim(ZSTD_CCtx* cctx,
5815
6308
  dictSize = 0;
5816
6309
  }
5817
6310
  ZSTD_memcpy(updatedRepcodes.rep, cctx->blockState.prevCBlock->rep, sizeof(repcodes_t));
5818
- for (; (inSeqs[idx].matchLength != 0 || inSeqs[idx].offset != 0) && idx < inSeqsSize; ++idx) {
6311
+ for (; idx < inSeqsSize && (inSeqs[idx].matchLength != 0 || inSeqs[idx].offset != 0); ++idx) {
5819
6312
  U32 const litLength = inSeqs[idx].litLength;
5820
- U32 const ll0 = (litLength == 0);
5821
6313
  U32 const matchLength = inSeqs[idx].matchLength;
5822
- U32 const offCode = ZSTD_finalizeOffCode(inSeqs[idx].offset, updatedRepcodes.rep, ll0);
5823
- ZSTD_updateRep(updatedRepcodes.rep, offCode, ll0);
6314
+ U32 offBase;
6315
+
6316
+ if (externalRepSearch == ZSTD_ps_disable) {
6317
+ offBase = OFFSET_TO_OFFBASE(inSeqs[idx].offset);
6318
+ } else {
6319
+ U32 const ll0 = (litLength == 0);
6320
+ offBase = ZSTD_finalizeOffBase(inSeqs[idx].offset, updatedRepcodes.rep, ll0);
6321
+ ZSTD_updateRep(updatedRepcodes.rep, offBase, ll0);
6322
+ }
5824
6323
 
5825
- DEBUGLOG(6, "Storing sequence: (of: %u, ml: %u, ll: %u)", offCode, matchLength, litLength);
6324
+ DEBUGLOG(6, "Storing sequence: (of: %u, ml: %u, ll: %u)", offBase, matchLength, litLength);
5826
6325
  if (cctx->appliedParams.validateSequences) {
5827
6326
  seqPos->posInSrc += litLength + matchLength;
5828
- FORWARD_IF_ERROR(ZSTD_validateSequence(offCode, matchLength, seqPos->posInSrc,
5829
- cctx->appliedParams.cParams.windowLog, dictSize),
6327
+ FORWARD_IF_ERROR(ZSTD_validateSequence(offBase, matchLength, cctx->appliedParams.cParams.minMatch, seqPos->posInSrc,
6328
+ cctx->appliedParams.cParams.windowLog, dictSize, cctx->appliedParams.useSequenceProducer),
5830
6329
  "Sequence validation failed");
5831
6330
  }
5832
- RETURN_ERROR_IF(idx - seqPos->idx > cctx->seqStore.maxNbSeq, memory_allocation,
6331
+ RETURN_ERROR_IF(idx - seqPos->idx >= cctx->seqStore.maxNbSeq, externalSequences_invalid,
5833
6332
  "Not enough memory allocated. Try adjusting ZSTD_c_minMatch.");
5834
- ZSTD_storeSeq(&cctx->seqStore, litLength, ip, iend, offCode, matchLength);
6333
+ ZSTD_storeSeq(&cctx->seqStore, litLength, ip, iend, offBase, matchLength);
5835
6334
  ip += matchLength + litLength;
5836
6335
  }
6336
+
6337
+ /* If we skipped repcode search while parsing, we need to update repcodes now */
6338
+ assert(externalRepSearch != ZSTD_ps_auto);
6339
+ assert(idx >= startIdx);
6340
+ if (externalRepSearch == ZSTD_ps_disable && idx != startIdx) {
6341
+ U32* const rep = updatedRepcodes.rep;
6342
+ U32 lastSeqIdx = idx - 1; /* index of last non-block-delimiter sequence */
6343
+
6344
+ if (lastSeqIdx >= startIdx + 2) {
6345
+ rep[2] = inSeqs[lastSeqIdx - 2].offset;
6346
+ rep[1] = inSeqs[lastSeqIdx - 1].offset;
6347
+ rep[0] = inSeqs[lastSeqIdx].offset;
6348
+ } else if (lastSeqIdx == startIdx + 1) {
6349
+ rep[2] = rep[0];
6350
+ rep[1] = inSeqs[lastSeqIdx - 1].offset;
6351
+ rep[0] = inSeqs[lastSeqIdx].offset;
6352
+ } else {
6353
+ assert(lastSeqIdx == startIdx);
6354
+ rep[2] = rep[1];
6355
+ rep[1] = rep[0];
6356
+ rep[0] = inSeqs[lastSeqIdx].offset;
6357
+ }
6358
+ }
6359
+
5837
6360
  ZSTD_memcpy(cctx->blockState.nextCBlock->rep, updatedRepcodes.rep, sizeof(repcodes_t));
5838
6361
 
5839
6362
  if (inSeqs[idx].litLength) {
@@ -5842,26 +6365,15 @@ ZSTD_copySequencesToSeqStoreExplicitBlockDelim(ZSTD_CCtx* cctx,
5842
6365
  ip += inSeqs[idx].litLength;
5843
6366
  seqPos->posInSrc += inSeqs[idx].litLength;
5844
6367
  }
5845
- RETURN_ERROR_IF(ip != iend, corruption_detected, "Blocksize doesn't agree with block delimiter!");
6368
+ RETURN_ERROR_IF(ip != iend, externalSequences_invalid, "Blocksize doesn't agree with block delimiter!");
5846
6369
  seqPos->idx = idx+1;
5847
6370
  return 0;
5848
6371
  }
5849
6372
 
5850
- /* Returns the number of bytes to move the current read position back by. Only non-zero
5851
- * if we ended up splitting a sequence. Otherwise, it may return a ZSTD error if something
5852
- * went wrong.
5853
- *
5854
- * This function will attempt to scan through blockSize bytes represented by the sequences
5855
- * in inSeqs, storing any (partial) sequences.
5856
- *
5857
- * Occasionally, we may want to change the actual number of bytes we consumed from inSeqs to
5858
- * avoid splitting a match, or to avoid splitting a match such that it would produce a match
5859
- * smaller than MINMATCH. In this case, we return the number of bytes that we didn't read from this block.
5860
- */
5861
- static size_t
6373
+ size_t
5862
6374
  ZSTD_copySequencesToSeqStoreNoBlockDelim(ZSTD_CCtx* cctx, ZSTD_sequencePosition* seqPos,
5863
6375
  const ZSTD_Sequence* const inSeqs, size_t inSeqsSize,
5864
- const void* src, size_t blockSize)
6376
+ const void* src, size_t blockSize, ZSTD_paramSwitch_e externalRepSearch)
5865
6377
  {
5866
6378
  U32 idx = seqPos->idx;
5867
6379
  U32 startPosInSequence = seqPos->posInSequence;
@@ -5873,6 +6385,9 @@ ZSTD_copySequencesToSeqStoreNoBlockDelim(ZSTD_CCtx* cctx, ZSTD_sequencePosition*
5873
6385
  U32 bytesAdjustment = 0;
5874
6386
  U32 finalMatchSplit = 0;
5875
6387
 
6388
+ /* TODO(embg) support fast parsing mode in noBlockDelim mode */
6389
+ (void)externalRepSearch;
6390
+
5876
6391
  if (cctx->cdict) {
5877
6392
  dictSize = cctx->cdict->dictContentSize;
5878
6393
  } else if (cctx->prefixDict.dict) {
@@ -5880,7 +6395,7 @@ ZSTD_copySequencesToSeqStoreNoBlockDelim(ZSTD_CCtx* cctx, ZSTD_sequencePosition*
5880
6395
  } else {
5881
6396
  dictSize = 0;
5882
6397
  }
5883
- DEBUGLOG(5, "ZSTD_copySequencesToSeqStore: idx: %u PIS: %u blockSize: %zu", idx, startPosInSequence, blockSize);
6398
+ DEBUGLOG(5, "ZSTD_copySequencesToSeqStoreNoBlockDelim: idx: %u PIS: %u blockSize: %zu", idx, startPosInSequence, blockSize);
5884
6399
  DEBUGLOG(5, "Start seq: idx: %u (of: %u ml: %u ll: %u)", idx, inSeqs[idx].offset, inSeqs[idx].matchLength, inSeqs[idx].litLength);
5885
6400
  ZSTD_memcpy(updatedRepcodes.rep, cctx->blockState.prevCBlock->rep, sizeof(repcodes_t));
5886
6401
  while (endPosInSequence && idx < inSeqsSize && !finalMatchSplit) {
@@ -5888,7 +6403,7 @@ ZSTD_copySequencesToSeqStoreNoBlockDelim(ZSTD_CCtx* cctx, ZSTD_sequencePosition*
5888
6403
  U32 litLength = currSeq.litLength;
5889
6404
  U32 matchLength = currSeq.matchLength;
5890
6405
  U32 const rawOffset = currSeq.offset;
5891
- U32 offCode;
6406
+ U32 offBase;
5892
6407
 
5893
6408
  /* Modify the sequence depending on where endPosInSequence lies */
5894
6409
  if (endPosInSequence >= currSeq.litLength + currSeq.matchLength) {
@@ -5902,7 +6417,6 @@ ZSTD_copySequencesToSeqStoreNoBlockDelim(ZSTD_CCtx* cctx, ZSTD_sequencePosition*
5902
6417
  /* Move to the next sequence */
5903
6418
  endPosInSequence -= currSeq.litLength + currSeq.matchLength;
5904
6419
  startPosInSequence = 0;
5905
- idx++;
5906
6420
  } else {
5907
6421
  /* This is the final (partial) sequence we're adding from inSeqs, and endPosInSequence
5908
6422
  does not reach the end of the match. So, we have to split the sequence */
@@ -5942,21 +6456,23 @@ ZSTD_copySequencesToSeqStoreNoBlockDelim(ZSTD_CCtx* cctx, ZSTD_sequencePosition*
5942
6456
  }
5943
6457
  /* Check if this offset can be represented with a repcode */
5944
6458
  { U32 const ll0 = (litLength == 0);
5945
- offCode = ZSTD_finalizeOffCode(rawOffset, updatedRepcodes.rep, ll0);
5946
- ZSTD_updateRep(updatedRepcodes.rep, offCode, ll0);
6459
+ offBase = ZSTD_finalizeOffBase(rawOffset, updatedRepcodes.rep, ll0);
6460
+ ZSTD_updateRep(updatedRepcodes.rep, offBase, ll0);
5947
6461
  }
5948
6462
 
5949
6463
  if (cctx->appliedParams.validateSequences) {
5950
6464
  seqPos->posInSrc += litLength + matchLength;
5951
- FORWARD_IF_ERROR(ZSTD_validateSequence(offCode, matchLength, seqPos->posInSrc,
5952
- cctx->appliedParams.cParams.windowLog, dictSize),
6465
+ FORWARD_IF_ERROR(ZSTD_validateSequence(offBase, matchLength, cctx->appliedParams.cParams.minMatch, seqPos->posInSrc,
6466
+ cctx->appliedParams.cParams.windowLog, dictSize, cctx->appliedParams.useSequenceProducer),
5953
6467
  "Sequence validation failed");
5954
6468
  }
5955
- DEBUGLOG(6, "Storing sequence: (of: %u, ml: %u, ll: %u)", offCode, matchLength, litLength);
5956
- RETURN_ERROR_IF(idx - seqPos->idx > cctx->seqStore.maxNbSeq, memory_allocation,
6469
+ DEBUGLOG(6, "Storing sequence: (of: %u, ml: %u, ll: %u)", offBase, matchLength, litLength);
6470
+ RETURN_ERROR_IF(idx - seqPos->idx >= cctx->seqStore.maxNbSeq, externalSequences_invalid,
5957
6471
  "Not enough memory allocated. Try adjusting ZSTD_c_minMatch.");
5958
- ZSTD_storeSeq(&cctx->seqStore, litLength, ip, iend, offCode, matchLength);
6472
+ ZSTD_storeSeq(&cctx->seqStore, litLength, ip, iend, offBase, matchLength);
5959
6473
  ip += matchLength + litLength;
6474
+ if (!finalMatchSplit)
6475
+ idx++; /* Next Sequence */
5960
6476
  }
5961
6477
  DEBUGLOG(5, "Ending seq: idx: %u (of: %u ml: %u ll: %u)", idx, inSeqs[idx].offset, inSeqs[idx].matchLength, inSeqs[idx].litLength);
5962
6478
  assert(idx == inSeqsSize || endPosInSequence <= inSeqs[idx].litLength + inSeqs[idx].matchLength);
@@ -5979,7 +6495,7 @@ ZSTD_copySequencesToSeqStoreNoBlockDelim(ZSTD_CCtx* cctx, ZSTD_sequencePosition*
5979
6495
 
5980
6496
  typedef size_t (*ZSTD_sequenceCopier) (ZSTD_CCtx* cctx, ZSTD_sequencePosition* seqPos,
5981
6497
  const ZSTD_Sequence* const inSeqs, size_t inSeqsSize,
5982
- const void* src, size_t blockSize);
6498
+ const void* src, size_t blockSize, ZSTD_paramSwitch_e externalRepSearch);
5983
6499
  static ZSTD_sequenceCopier ZSTD_selectSequenceCopier(ZSTD_sequenceFormat_e mode)
5984
6500
  {
5985
6501
  ZSTD_sequenceCopier sequenceCopier = NULL;
@@ -5993,6 +6509,57 @@ static ZSTD_sequenceCopier ZSTD_selectSequenceCopier(ZSTD_sequenceFormat_e mode)
5993
6509
  return sequenceCopier;
5994
6510
  }
5995
6511
 
6512
+ /* Discover the size of next block by searching for the delimiter.
6513
+ * Note that a block delimiter **must** exist in this mode,
6514
+ * otherwise it's an input error.
6515
+ * The block size retrieved will be later compared to ensure it remains within bounds */
6516
+ static size_t
6517
+ blockSize_explicitDelimiter(const ZSTD_Sequence* inSeqs, size_t inSeqsSize, ZSTD_sequencePosition seqPos)
6518
+ {
6519
+ int end = 0;
6520
+ size_t blockSize = 0;
6521
+ size_t spos = seqPos.idx;
6522
+ DEBUGLOG(6, "blockSize_explicitDelimiter : seq %zu / %zu", spos, inSeqsSize);
6523
+ assert(spos <= inSeqsSize);
6524
+ while (spos < inSeqsSize) {
6525
+ end = (inSeqs[spos].offset == 0);
6526
+ blockSize += inSeqs[spos].litLength + inSeqs[spos].matchLength;
6527
+ if (end) {
6528
+ if (inSeqs[spos].matchLength != 0)
6529
+ RETURN_ERROR(externalSequences_invalid, "delimiter format error : both matchlength and offset must be == 0");
6530
+ break;
6531
+ }
6532
+ spos++;
6533
+ }
6534
+ if (!end)
6535
+ RETURN_ERROR(externalSequences_invalid, "Reached end of sequences without finding a block delimiter");
6536
+ return blockSize;
6537
+ }
6538
+
6539
+ /* More a "target" block size */
6540
+ static size_t blockSize_noDelimiter(size_t blockSize, size_t remaining)
6541
+ {
6542
+ int const lastBlock = (remaining <= blockSize);
6543
+ return lastBlock ? remaining : blockSize;
6544
+ }
6545
+
6546
+ static size_t determine_blockSize(ZSTD_sequenceFormat_e mode,
6547
+ size_t blockSize, size_t remaining,
6548
+ const ZSTD_Sequence* inSeqs, size_t inSeqsSize, ZSTD_sequencePosition seqPos)
6549
+ {
6550
+ DEBUGLOG(6, "determine_blockSize : remainingSize = %zu", remaining);
6551
+ if (mode == ZSTD_sf_noBlockDelimiters)
6552
+ return blockSize_noDelimiter(blockSize, remaining);
6553
+ { size_t const explicitBlockSize = blockSize_explicitDelimiter(inSeqs, inSeqsSize, seqPos);
6554
+ FORWARD_IF_ERROR(explicitBlockSize, "Error while determining block size with explicit delimiters");
6555
+ if (explicitBlockSize > blockSize)
6556
+ RETURN_ERROR(externalSequences_invalid, "sequences incorrectly define a too large block");
6557
+ if (explicitBlockSize > remaining)
6558
+ RETURN_ERROR(externalSequences_invalid, "sequences define a frame longer than source");
6559
+ return explicitBlockSize;
6560
+ }
6561
+ }
6562
+
5996
6563
  /* Compress, block-by-block, all of the sequences given.
5997
6564
  *
5998
6565
  * Returns the cumulative size of all compressed blocks (including their headers),
@@ -6005,9 +6572,6 @@ ZSTD_compressSequences_internal(ZSTD_CCtx* cctx,
6005
6572
  const void* src, size_t srcSize)
6006
6573
  {
6007
6574
  size_t cSize = 0;
6008
- U32 lastBlock;
6009
- size_t blockSize;
6010
- size_t compressedSeqsSize;
6011
6575
  size_t remaining = srcSize;
6012
6576
  ZSTD_sequencePosition seqPos = {0, 0, 0};
6013
6577
 
@@ -6027,22 +6591,29 @@ ZSTD_compressSequences_internal(ZSTD_CCtx* cctx,
6027
6591
  }
6028
6592
 
6029
6593
  while (remaining) {
6594
+ size_t compressedSeqsSize;
6030
6595
  size_t cBlockSize;
6031
6596
  size_t additionalByteAdjustment;
6032
- lastBlock = remaining <= cctx->blockSize;
6033
- blockSize = lastBlock ? (U32)remaining : (U32)cctx->blockSize;
6597
+ size_t blockSize = determine_blockSize(cctx->appliedParams.blockDelimiters,
6598
+ cctx->blockSize, remaining,
6599
+ inSeqs, inSeqsSize, seqPos);
6600
+ U32 const lastBlock = (blockSize == remaining);
6601
+ FORWARD_IF_ERROR(blockSize, "Error while trying to determine block size");
6602
+ assert(blockSize <= remaining);
6034
6603
  ZSTD_resetSeqStore(&cctx->seqStore);
6035
- DEBUGLOG(4, "Working on new block. Blocksize: %zu", blockSize);
6604
+ DEBUGLOG(5, "Working on new block. Blocksize: %zu (total:%zu)", blockSize, (ip - (const BYTE*)src) + blockSize);
6036
6605
 
6037
- additionalByteAdjustment = sequenceCopier(cctx, &seqPos, inSeqs, inSeqsSize, ip, blockSize);
6606
+ additionalByteAdjustment = sequenceCopier(cctx, &seqPos, inSeqs, inSeqsSize, ip, blockSize, cctx->appliedParams.searchForExternalRepcodes);
6038
6607
  FORWARD_IF_ERROR(additionalByteAdjustment, "Bad sequence copy");
6039
6608
  blockSize -= additionalByteAdjustment;
6040
6609
 
6041
6610
  /* If blocks are too small, emit as a nocompress block */
6042
- if (blockSize < MIN_CBLOCK_SIZE+ZSTD_blockHeaderSize+1) {
6611
+ /* TODO: See 3090. We reduced MIN_CBLOCK_SIZE from 3 to 2 so to compensate we are adding
6612
+ * additional 1. We need to revisit and change this logic to be more consistent */
6613
+ if (blockSize < MIN_CBLOCK_SIZE+ZSTD_blockHeaderSize+1+1) {
6043
6614
  cBlockSize = ZSTD_noCompressBlock(op, dstCapacity, ip, blockSize, lastBlock);
6044
6615
  FORWARD_IF_ERROR(cBlockSize, "Nocompress block failed");
6045
- DEBUGLOG(4, "Block too small, writing out nocompress block: cSize: %zu", cBlockSize);
6616
+ DEBUGLOG(5, "Block too small, writing out nocompress block: cSize: %zu", cBlockSize);
6046
6617
  cSize += cBlockSize;
6047
6618
  ip += blockSize;
6048
6619
  op += cBlockSize;
@@ -6051,6 +6622,7 @@ ZSTD_compressSequences_internal(ZSTD_CCtx* cctx,
6051
6622
  continue;
6052
6623
  }
6053
6624
 
6625
+ RETURN_ERROR_IF(dstCapacity < ZSTD_blockHeaderSize, dstSize_tooSmall, "not enough dstCapacity to write a new compressed block");
6054
6626
  compressedSeqsSize = ZSTD_entropyCompressSeqStore(&cctx->seqStore,
6055
6627
  &cctx->blockState.prevCBlock->entropy, &cctx->blockState.nextCBlock->entropy,
6056
6628
  &cctx->appliedParams,
@@ -6059,11 +6631,11 @@ ZSTD_compressSequences_internal(ZSTD_CCtx* cctx,
6059
6631
  cctx->entropyWorkspace, ENTROPY_WORKSPACE_SIZE /* statically allocated in resetCCtx */,
6060
6632
  cctx->bmi2);
6061
6633
  FORWARD_IF_ERROR(compressedSeqsSize, "Compressing sequences of block failed");
6062
- DEBUGLOG(4, "Compressed sequences size: %zu", compressedSeqsSize);
6634
+ DEBUGLOG(5, "Compressed sequences size: %zu", compressedSeqsSize);
6063
6635
 
6064
6636
  if (!cctx->isFirstBlock &&
6065
6637
  ZSTD_maybeRLE(&cctx->seqStore) &&
6066
- ZSTD_isRLE((BYTE const*)src, srcSize)) {
6638
+ ZSTD_isRLE(ip, blockSize)) {
6067
6639
  /* We don't want to emit our first block as a RLE even if it qualifies because
6068
6640
  * doing so will cause the decoder (cli only) to throw a "should consume all input error."
6069
6641
  * This is only an issue for zstd <= v1.4.3
@@ -6074,12 +6646,12 @@ ZSTD_compressSequences_internal(ZSTD_CCtx* cctx,
6074
6646
  if (compressedSeqsSize == 0) {
6075
6647
  /* ZSTD_noCompressBlock writes the block header as well */
6076
6648
  cBlockSize = ZSTD_noCompressBlock(op, dstCapacity, ip, blockSize, lastBlock);
6077
- FORWARD_IF_ERROR(cBlockSize, "Nocompress block failed");
6078
- DEBUGLOG(4, "Writing out nocompress block, size: %zu", cBlockSize);
6649
+ FORWARD_IF_ERROR(cBlockSize, "ZSTD_noCompressBlock failed");
6650
+ DEBUGLOG(5, "Writing out nocompress block, size: %zu", cBlockSize);
6079
6651
  } else if (compressedSeqsSize == 1) {
6080
6652
  cBlockSize = ZSTD_rleCompressBlock(op, dstCapacity, *ip, blockSize, lastBlock);
6081
- FORWARD_IF_ERROR(cBlockSize, "RLE compress block failed");
6082
- DEBUGLOG(4, "Writing out RLE block, size: %zu", cBlockSize);
6653
+ FORWARD_IF_ERROR(cBlockSize, "ZSTD_rleCompressBlock failed");
6654
+ DEBUGLOG(5, "Writing out RLE block, size: %zu", cBlockSize);
6083
6655
  } else {
6084
6656
  U32 cBlockHeader;
6085
6657
  /* Error checking and repcodes update */
@@ -6091,11 +6663,10 @@ ZSTD_compressSequences_internal(ZSTD_CCtx* cctx,
6091
6663
  cBlockHeader = lastBlock + (((U32)bt_compressed)<<1) + (U32)(compressedSeqsSize << 3);
6092
6664
  MEM_writeLE24(op, cBlockHeader);
6093
6665
  cBlockSize = ZSTD_blockHeaderSize + compressedSeqsSize;
6094
- DEBUGLOG(4, "Writing out compressed block, size: %zu", cBlockSize);
6666
+ DEBUGLOG(5, "Writing out compressed block, size: %zu", cBlockSize);
6095
6667
  }
6096
6668
 
6097
6669
  cSize += cBlockSize;
6098
- DEBUGLOG(4, "cSize running total: %zu", cSize);
6099
6670
 
6100
6671
  if (lastBlock) {
6101
6672
  break;
@@ -6106,12 +6677,15 @@ ZSTD_compressSequences_internal(ZSTD_CCtx* cctx,
6106
6677
  dstCapacity -= cBlockSize;
6107
6678
  cctx->isFirstBlock = 0;
6108
6679
  }
6680
+ DEBUGLOG(5, "cSize running total: %zu (remaining dstCapacity=%zu)", cSize, dstCapacity);
6109
6681
  }
6110
6682
 
6683
+ DEBUGLOG(4, "cSize final total: %zu", cSize);
6111
6684
  return cSize;
6112
6685
  }
6113
6686
 
6114
- size_t ZSTD_compressSequences(ZSTD_CCtx* const cctx, void* dst, size_t dstCapacity,
6687
+ size_t ZSTD_compressSequences(ZSTD_CCtx* cctx,
6688
+ void* dst, size_t dstCapacity,
6115
6689
  const ZSTD_Sequence* inSeqs, size_t inSeqsSize,
6116
6690
  const void* src, size_t srcSize)
6117
6691
  {
@@ -6121,7 +6695,7 @@ size_t ZSTD_compressSequences(ZSTD_CCtx* const cctx, void* dst, size_t dstCapaci
6121
6695
  size_t frameHeaderSize = 0;
6122
6696
 
6123
6697
  /* Transparent initialization stage, same as compressStream2() */
6124
- DEBUGLOG(3, "ZSTD_compressSequences()");
6698
+ DEBUGLOG(4, "ZSTD_compressSequences (dstCapacity=%zu)", dstCapacity);
6125
6699
  assert(cctx != NULL);
6126
6700
  FORWARD_IF_ERROR(ZSTD_CCtx_init_compressStream2(cctx, ZSTD_e_end, srcSize), "CCtx initialization failed");
6127
6701
  /* Begin writing output, starting with frame header */
@@ -6149,26 +6723,34 @@ size_t ZSTD_compressSequences(ZSTD_CCtx* const cctx, void* dst, size_t dstCapaci
6149
6723
  cSize += 4;
6150
6724
  }
6151
6725
 
6152
- DEBUGLOG(3, "Final compressed size: %zu", cSize);
6726
+ DEBUGLOG(4, "Final compressed size: %zu", cSize);
6153
6727
  return cSize;
6154
6728
  }
6155
6729
 
6156
6730
  /*====== Finalize ======*/
6157
6731
 
6732
+ static ZSTD_inBuffer inBuffer_forEndFlush(const ZSTD_CStream* zcs)
6733
+ {
6734
+ const ZSTD_inBuffer nullInput = { NULL, 0, 0 };
6735
+ const int stableInput = (zcs->appliedParams.inBufferMode == ZSTD_bm_stable);
6736
+ return stableInput ? zcs->expectedInBuffer : nullInput;
6737
+ }
6738
+
6158
6739
  /*! ZSTD_flushStream() :
6159
6740
  * @return : amount of data remaining to flush */
6160
6741
  size_t ZSTD_flushStream(ZSTD_CStream* zcs, ZSTD_outBuffer* output)
6161
6742
  {
6162
- ZSTD_inBuffer input = { NULL, 0, 0 };
6743
+ ZSTD_inBuffer input = inBuffer_forEndFlush(zcs);
6744
+ input.size = input.pos; /* do not ingest more input during flush */
6163
6745
  return ZSTD_compressStream2(zcs, output, &input, ZSTD_e_flush);
6164
6746
  }
6165
6747
 
6166
6748
 
6167
6749
  size_t ZSTD_endStream(ZSTD_CStream* zcs, ZSTD_outBuffer* output)
6168
6750
  {
6169
- ZSTD_inBuffer input = { NULL, 0, 0 };
6751
+ ZSTD_inBuffer input = inBuffer_forEndFlush(zcs);
6170
6752
  size_t const remainingToFlush = ZSTD_compressStream2(zcs, output, &input, ZSTD_e_end);
6171
- FORWARD_IF_ERROR( remainingToFlush , "ZSTD_compressStream2 failed");
6753
+ FORWARD_IF_ERROR(remainingToFlush , "ZSTD_compressStream2(,,ZSTD_e_end) failed");
6172
6754
  if (zcs->appliedParams.nbWorkers > 0) return remainingToFlush; /* minimal estimation */
6173
6755
  /* single thread mode : attempt to calculate remaining to flush more precisely */
6174
6756
  { size_t const lastBlockSize = zcs->frameEnded ? 0 : ZSTD_BLOCKHEADERSIZE;
@@ -6290,7 +6872,7 @@ static ZSTD_compressionParameters ZSTD_getCParams_internal(int compressionLevel,
6290
6872
  cp.targetLength = (unsigned)(-clampedCompressionLevel);
6291
6873
  }
6292
6874
  /* refine parameters based on srcSize & dictSize */
6293
- return ZSTD_adjustCParams_internal(cp, srcSizeHint, dictSize, mode);
6875
+ return ZSTD_adjustCParams_internal(cp, srcSizeHint, dictSize, mode, ZSTD_ps_auto);
6294
6876
  }
6295
6877
  }
6296
6878
 
@@ -6325,3 +6907,21 @@ ZSTD_parameters ZSTD_getParams(int compressionLevel, unsigned long long srcSizeH
6325
6907
  if (srcSizeHint == 0) srcSizeHint = ZSTD_CONTENTSIZE_UNKNOWN;
6326
6908
  return ZSTD_getParams_internal(compressionLevel, srcSizeHint, dictSize, ZSTD_cpm_unknown);
6327
6909
  }
6910
+
6911
+ void ZSTD_registerSequenceProducer(
6912
+ ZSTD_CCtx* zc, void* mState,
6913
+ ZSTD_sequenceProducer_F* mFinder
6914
+ ) {
6915
+ if (mFinder != NULL) {
6916
+ ZSTD_externalMatchCtx emctx;
6917
+ emctx.mState = mState;
6918
+ emctx.mFinder = mFinder;
6919
+ emctx.seqBuffer = NULL;
6920
+ emctx.seqBufferCapacity = 0;
6921
+ zc->externalMatchCtx = emctx;
6922
+ zc->requestedParams.useSequenceProducer = 1;
6923
+ } else {
6924
+ ZSTD_memset(&zc->externalMatchCtx, 0, sizeof(zc->externalMatchCtx));
6925
+ zc->requestedParams.useSequenceProducer = 0;
6926
+ }
6927
+ }