zstdlib 0.10.0-x86_64-darwin → 0.11.0-x86_64-darwin

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (85) hide show
  1. checksums.yaml +4 -4
  2. data/CHANGES.md +8 -0
  3. data/ext/zstdlib_c/extconf.rb +2 -2
  4. data/ext/zstdlib_c/ruby/zlib-3.2/zstdlib.c +5090 -0
  5. data/ext/zstdlib_c/zstd-1.5.5/lib/common/allocations.h +55 -0
  6. data/ext/zstdlib_c/zstd-1.5.5/lib/common/bits.h +200 -0
  7. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/common/bitstream.h +19 -60
  8. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/common/compiler.h +26 -3
  9. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/common/cpu.h +1 -1
  10. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/common/debug.c +1 -1
  11. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/common/debug.h +1 -1
  12. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/common/entropy_common.c +12 -40
  13. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/common/error_private.c +9 -2
  14. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/common/error_private.h +1 -1
  15. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/common/fse.h +5 -83
  16. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/common/fse_decompress.c +7 -99
  17. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/common/huf.h +65 -156
  18. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/common/mem.h +39 -46
  19. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/common/pool.c +26 -10
  20. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/common/pool.h +7 -1
  21. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/common/portability_macros.h +22 -3
  22. data/ext/zstdlib_c/zstd-1.5.5/lib/common/threading.c +176 -0
  23. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/common/threading.h +5 -10
  24. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/common/xxhash.c +2 -2
  25. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/common/xxhash.h +8 -8
  26. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/common/zstd_common.c +1 -36
  27. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/common/zstd_deps.h +1 -1
  28. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/common/zstd_internal.h +17 -118
  29. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/common/zstd_trace.h +3 -3
  30. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/compress/clevels.h +1 -1
  31. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/compress/fse_compress.c +7 -124
  32. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/compress/hist.c +1 -1
  33. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/compress/hist.h +1 -1
  34. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/compress/huf_compress.c +234 -169
  35. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/compress/zstd_compress.c +1243 -538
  36. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/compress/zstd_compress_internal.h +225 -151
  37. data/ext/zstdlib_c/zstd-1.5.5/lib/compress/zstd_compress_literals.c +235 -0
  38. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/compress/zstd_compress_literals.h +16 -8
  39. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/compress/zstd_compress_sequences.c +3 -3
  40. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/compress/zstd_compress_sequences.h +1 -1
  41. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/compress/zstd_compress_superblock.c +25 -21
  42. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/compress/zstd_compress_superblock.h +1 -1
  43. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/compress/zstd_cwksp.h +128 -62
  44. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/compress/zstd_double_fast.c +95 -33
  45. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/compress/zstd_double_fast.h +3 -2
  46. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/compress/zstd_fast.c +433 -148
  47. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/compress/zstd_fast.h +3 -2
  48. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/compress/zstd_lazy.c +398 -345
  49. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/compress/zstd_lazy.h +4 -2
  50. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/compress/zstd_ldm.c +5 -5
  51. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/compress/zstd_ldm.h +1 -1
  52. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/compress/zstd_ldm_geartab.h +1 -1
  53. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/compress/zstd_opt.c +106 -80
  54. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/compress/zstd_opt.h +1 -1
  55. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/compress/zstdmt_compress.c +17 -9
  56. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/compress/zstdmt_compress.h +1 -1
  57. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/decompress/huf_decompress.c +434 -441
  58. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/decompress/huf_decompress_amd64.S +30 -39
  59. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/decompress/zstd_ddict.c +4 -4
  60. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/decompress/zstd_ddict.h +1 -1
  61. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/decompress/zstd_decompress.c +205 -80
  62. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/decompress/zstd_decompress_block.c +201 -81
  63. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/decompress/zstd_decompress_block.h +6 -1
  64. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/decompress/zstd_decompress_internal.h +4 -2
  65. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/zdict.h +53 -31
  66. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/zstd.h +580 -135
  67. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/lib/zstd_errors.h +27 -8
  68. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/zlibWrapper/gzclose.c +1 -1
  69. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/zlibWrapper/gzcompatibility.h +8 -8
  70. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/zlibWrapper/gzguts.h +10 -10
  71. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/zlibWrapper/gzlib.c +3 -3
  72. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/zlibWrapper/gzread.c +10 -10
  73. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/zlibWrapper/gzwrite.c +5 -5
  74. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/zlibWrapper/zstd_zlibwrapper.c +46 -44
  75. data/ext/zstdlib_c/{zstd-1.5.2 → zstd-1.5.5}/zlibWrapper/zstd_zlibwrapper.h +4 -1
  76. data/lib/2.4/zstdlib_c.bundle +0 -0
  77. data/lib/2.5/zstdlib_c.bundle +0 -0
  78. data/lib/2.6/zstdlib_c.bundle +0 -0
  79. data/lib/2.7/zstdlib_c.bundle +0 -0
  80. data/lib/3.0/zstdlib_c.bundle +0 -0
  81. data/lib/3.1/zstdlib_c.bundle +0 -0
  82. data/lib/3.2/zstdlib_c.bundle +0 -0
  83. metadata +82 -78
  84. data/ext/zstdlib_c/zstd-1.5.2/lib/common/threading.c +0 -122
  85. data/ext/zstdlib_c/zstd-1.5.2/lib/compress/zstd_compress_literals.c +0 -159
@@ -1,5 +1,5 @@
1
1
  /*
2
- * Copyright (c) Yann Collet, Facebook, Inc.
2
+ * Copyright (c) Meta Platforms, Inc. and affiliates.
3
3
  * All rights reserved.
4
4
  *
5
5
  * This source code is licensed under both the BSD-style license (found in the
@@ -11,12 +11,12 @@
11
11
  /*-*************************************
12
12
  * Dependencies
13
13
  ***************************************/
14
+ #include "../common/allocations.h" /* ZSTD_customMalloc, ZSTD_customCalloc, ZSTD_customFree */
14
15
  #include "../common/zstd_deps.h" /* INT_MAX, ZSTD_memset, ZSTD_memcpy */
15
16
  #include "../common/mem.h"
16
17
  #include "hist.h" /* HIST_countFast_wksp */
17
18
  #define FSE_STATIC_LINKING_ONLY /* FSE_encodeSymbol */
18
19
  #include "../common/fse.h"
19
- #define HUF_STATIC_LINKING_ONLY
20
20
  #include "../common/huf.h"
21
21
  #include "zstd_compress_internal.h"
22
22
  #include "zstd_compress_sequences.h"
@@ -27,6 +27,7 @@
27
27
  #include "zstd_opt.h"
28
28
  #include "zstd_ldm.h"
29
29
  #include "zstd_compress_superblock.h"
30
+ #include "../common/bits.h" /* ZSTD_highbit32, ZSTD_rotateRight_U64 */
30
31
 
31
32
  /* ***************************************************************
32
33
  * Tuning parameters
@@ -58,14 +59,17 @@
58
59
  * Helper functions
59
60
  ***************************************/
60
61
  /* ZSTD_compressBound()
61
- * Note that the result from this function is only compatible with the "normal"
62
- * full-block strategy.
63
- * When there are a lot of small blocks due to frequent flush in streaming mode
64
- * the overhead of headers can make the compressed data to be larger than the
65
- * return value of ZSTD_compressBound().
62
+ * Note that the result from this function is only valid for
63
+ * the one-pass compression functions.
64
+ * When employing the streaming mode,
65
+ * if flushes are frequently altering the size of blocks,
66
+ * the overhead from block headers can make the compressed data larger
67
+ * than the return value of ZSTD_compressBound().
66
68
  */
67
69
  size_t ZSTD_compressBound(size_t srcSize) {
68
- return ZSTD_COMPRESSBOUND(srcSize);
70
+ size_t const r = ZSTD_COMPRESSBOUND(srcSize);
71
+ if (r==0) return ERROR(srcSize_wrong);
72
+ return r;
69
73
  }
70
74
 
71
75
 
@@ -177,12 +181,9 @@ size_t ZSTD_freeCCtx(ZSTD_CCtx* cctx)
177
181
  if (cctx==NULL) return 0; /* support free on NULL */
178
182
  RETURN_ERROR_IF(cctx->staticSize, memory_allocation,
179
183
  "not compatible with static CCtx");
180
- {
181
- int cctxInWorkspace = ZSTD_cwksp_owns_buffer(&cctx->workspace, cctx);
184
+ { int cctxInWorkspace = ZSTD_cwksp_owns_buffer(&cctx->workspace, cctx);
182
185
  ZSTD_freeCCtxContent(cctx);
183
- if (!cctxInWorkspace) {
184
- ZSTD_customFree(cctx, cctx->customMem);
185
- }
186
+ if (!cctxInWorkspace) ZSTD_customFree(cctx, cctx->customMem);
186
187
  }
187
188
  return 0;
188
189
  }
@@ -267,9 +268,9 @@ static int ZSTD_allocateChainTable(const ZSTD_strategy strategy,
267
268
  return forDDSDict || ((strategy != ZSTD_fast) && !ZSTD_rowMatchFinderUsed(strategy, useRowMatchFinder));
268
269
  }
269
270
 
270
- /* Returns 1 if compression parameters are such that we should
271
+ /* Returns ZSTD_ps_enable if compression parameters are such that we should
271
272
  * enable long distance matching (wlog >= 27, strategy >= btopt).
272
- * Returns 0 otherwise.
273
+ * Returns ZSTD_ps_disable otherwise.
273
274
  */
274
275
  static ZSTD_paramSwitch_e ZSTD_resolveEnableLdm(ZSTD_paramSwitch_e mode,
275
276
  const ZSTD_compressionParameters* const cParams) {
@@ -277,6 +278,34 @@ static ZSTD_paramSwitch_e ZSTD_resolveEnableLdm(ZSTD_paramSwitch_e mode,
277
278
  return (cParams->strategy >= ZSTD_btopt && cParams->windowLog >= 27) ? ZSTD_ps_enable : ZSTD_ps_disable;
278
279
  }
279
280
 
281
+ static int ZSTD_resolveExternalSequenceValidation(int mode) {
282
+ return mode;
283
+ }
284
+
285
+ /* Resolves maxBlockSize to the default if no value is present. */
286
+ static size_t ZSTD_resolveMaxBlockSize(size_t maxBlockSize) {
287
+ if (maxBlockSize == 0) {
288
+ return ZSTD_BLOCKSIZE_MAX;
289
+ } else {
290
+ return maxBlockSize;
291
+ }
292
+ }
293
+
294
+ static ZSTD_paramSwitch_e ZSTD_resolveExternalRepcodeSearch(ZSTD_paramSwitch_e value, int cLevel) {
295
+ if (value != ZSTD_ps_auto) return value;
296
+ if (cLevel < 10) {
297
+ return ZSTD_ps_disable;
298
+ } else {
299
+ return ZSTD_ps_enable;
300
+ }
301
+ }
302
+
303
+ /* Returns 1 if compression parameters are such that CDict hashtable and chaintable indices are tagged.
304
+ * If so, the tags need to be removed in ZSTD_resetCCtx_byCopyingCDict. */
305
+ static int ZSTD_CDictIndicesAreTagged(const ZSTD_compressionParameters* const cParams) {
306
+ return cParams->strategy == ZSTD_fast || cParams->strategy == ZSTD_dfast;
307
+ }
308
+
280
309
  static ZSTD_CCtx_params ZSTD_makeCCtxParamsFromCParams(
281
310
  ZSTD_compressionParameters cParams)
282
311
  {
@@ -294,6 +323,10 @@ static ZSTD_CCtx_params ZSTD_makeCCtxParamsFromCParams(
294
323
  }
295
324
  cctxParams.useBlockSplitter = ZSTD_resolveBlockSplitterMode(cctxParams.useBlockSplitter, &cParams);
296
325
  cctxParams.useRowMatchFinder = ZSTD_resolveRowMatchFinderMode(cctxParams.useRowMatchFinder, &cParams);
326
+ cctxParams.validateSequences = ZSTD_resolveExternalSequenceValidation(cctxParams.validateSequences);
327
+ cctxParams.maxBlockSize = ZSTD_resolveMaxBlockSize(cctxParams.maxBlockSize);
328
+ cctxParams.searchForExternalRepcodes = ZSTD_resolveExternalRepcodeSearch(cctxParams.searchForExternalRepcodes,
329
+ cctxParams.compressionLevel);
297
330
  assert(!ZSTD_checkCParams(cParams));
298
331
  return cctxParams;
299
332
  }
@@ -339,10 +372,13 @@ size_t ZSTD_CCtxParams_init(ZSTD_CCtx_params* cctxParams, int compressionLevel)
339
372
  #define ZSTD_NO_CLEVEL 0
340
373
 
341
374
  /**
342
- * Initializes the cctxParams from params and compressionLevel.
375
+ * Initializes `cctxParams` from `params` and `compressionLevel`.
343
376
  * @param compressionLevel If params are derived from a compression level then that compression level, otherwise ZSTD_NO_CLEVEL.
344
377
  */
345
- static void ZSTD_CCtxParams_init_internal(ZSTD_CCtx_params* cctxParams, ZSTD_parameters const* params, int compressionLevel)
378
+ static void
379
+ ZSTD_CCtxParams_init_internal(ZSTD_CCtx_params* cctxParams,
380
+ const ZSTD_parameters* params,
381
+ int compressionLevel)
346
382
  {
347
383
  assert(!ZSTD_checkCParams(params->cParams));
348
384
  ZSTD_memset(cctxParams, 0, sizeof(*cctxParams));
@@ -355,6 +391,9 @@ static void ZSTD_CCtxParams_init_internal(ZSTD_CCtx_params* cctxParams, ZSTD_par
355
391
  cctxParams->useRowMatchFinder = ZSTD_resolveRowMatchFinderMode(cctxParams->useRowMatchFinder, &params->cParams);
356
392
  cctxParams->useBlockSplitter = ZSTD_resolveBlockSplitterMode(cctxParams->useBlockSplitter, &params->cParams);
357
393
  cctxParams->ldmParams.enableLdm = ZSTD_resolveEnableLdm(cctxParams->ldmParams.enableLdm, &params->cParams);
394
+ cctxParams->validateSequences = ZSTD_resolveExternalSequenceValidation(cctxParams->validateSequences);
395
+ cctxParams->maxBlockSize = ZSTD_resolveMaxBlockSize(cctxParams->maxBlockSize);
396
+ cctxParams->searchForExternalRepcodes = ZSTD_resolveExternalRepcodeSearch(cctxParams->searchForExternalRepcodes, compressionLevel);
358
397
  DEBUGLOG(4, "ZSTD_CCtxParams_init_internal: useRowMatchFinder=%d, useBlockSplitter=%d ldm=%d",
359
398
  cctxParams->useRowMatchFinder, cctxParams->useBlockSplitter, cctxParams->ldmParams.enableLdm);
360
399
  }
@@ -369,7 +408,7 @@ size_t ZSTD_CCtxParams_init_advanced(ZSTD_CCtx_params* cctxParams, ZSTD_paramete
369
408
 
370
409
  /**
371
410
  * Sets cctxParams' cParams and fParams from params, but otherwise leaves them alone.
372
- * @param param Validated zstd parameters.
411
+ * @param params Validated zstd parameters.
373
412
  */
374
413
  static void ZSTD_CCtxParams_setZstdParams(
375
414
  ZSTD_CCtx_params* cctxParams, const ZSTD_parameters* params)
@@ -478,8 +517,8 @@ ZSTD_bounds ZSTD_cParam_getBounds(ZSTD_cParameter param)
478
517
  return bounds;
479
518
 
480
519
  case ZSTD_c_enableLongDistanceMatching:
481
- bounds.lowerBound = 0;
482
- bounds.upperBound = 1;
520
+ bounds.lowerBound = (int)ZSTD_ps_auto;
521
+ bounds.upperBound = (int)ZSTD_ps_disable;
483
522
  return bounds;
484
523
 
485
524
  case ZSTD_c_ldmHashLog:
@@ -572,6 +611,26 @@ ZSTD_bounds ZSTD_cParam_getBounds(ZSTD_cParameter param)
572
611
  bounds.upperBound = 1;
573
612
  return bounds;
574
613
 
614
+ case ZSTD_c_prefetchCDictTables:
615
+ bounds.lowerBound = (int)ZSTD_ps_auto;
616
+ bounds.upperBound = (int)ZSTD_ps_disable;
617
+ return bounds;
618
+
619
+ case ZSTD_c_enableSeqProducerFallback:
620
+ bounds.lowerBound = 0;
621
+ bounds.upperBound = 1;
622
+ return bounds;
623
+
624
+ case ZSTD_c_maxBlockSize:
625
+ bounds.lowerBound = ZSTD_BLOCKSIZE_MAX_MIN;
626
+ bounds.upperBound = ZSTD_BLOCKSIZE_MAX;
627
+ return bounds;
628
+
629
+ case ZSTD_c_searchForExternalRepcodes:
630
+ bounds.lowerBound = (int)ZSTD_ps_auto;
631
+ bounds.upperBound = (int)ZSTD_ps_disable;
632
+ return bounds;
633
+
575
634
  default:
576
635
  bounds.error = ERROR(parameter_unsupported);
577
636
  return bounds;
@@ -636,6 +695,10 @@ static int ZSTD_isUpdateAuthorized(ZSTD_cParameter param)
636
695
  case ZSTD_c_useBlockSplitter:
637
696
  case ZSTD_c_useRowMatchFinder:
638
697
  case ZSTD_c_deterministicRefPrefix:
698
+ case ZSTD_c_prefetchCDictTables:
699
+ case ZSTD_c_enableSeqProducerFallback:
700
+ case ZSTD_c_maxBlockSize:
701
+ case ZSTD_c_searchForExternalRepcodes:
639
702
  default:
640
703
  return 0;
641
704
  }
@@ -648,7 +711,7 @@ size_t ZSTD_CCtx_setParameter(ZSTD_CCtx* cctx, ZSTD_cParameter param, int value)
648
711
  if (ZSTD_isUpdateAuthorized(param)) {
649
712
  cctx->cParamsChanged = 1;
650
713
  } else {
651
- RETURN_ERROR(stage_wrong, "can only set params in ctx init stage");
714
+ RETURN_ERROR(stage_wrong, "can only set params in cctx init stage");
652
715
  } }
653
716
 
654
717
  switch(param)
@@ -691,6 +754,10 @@ size_t ZSTD_CCtx_setParameter(ZSTD_CCtx* cctx, ZSTD_cParameter param, int value)
691
754
  case ZSTD_c_useBlockSplitter:
692
755
  case ZSTD_c_useRowMatchFinder:
693
756
  case ZSTD_c_deterministicRefPrefix:
757
+ case ZSTD_c_prefetchCDictTables:
758
+ case ZSTD_c_enableSeqProducerFallback:
759
+ case ZSTD_c_maxBlockSize:
760
+ case ZSTD_c_searchForExternalRepcodes:
694
761
  break;
695
762
 
696
763
  default: RETURN_ERROR(parameter_unsupported, "unknown parameter");
@@ -746,12 +813,12 @@ size_t ZSTD_CCtxParams_setParameter(ZSTD_CCtx_params* CCtxParams,
746
813
  case ZSTD_c_minMatch :
747
814
  if (value!=0) /* 0 => use default */
748
815
  BOUNDCHECK(ZSTD_c_minMatch, value);
749
- CCtxParams->cParams.minMatch = value;
816
+ CCtxParams->cParams.minMatch = (U32)value;
750
817
  return CCtxParams->cParams.minMatch;
751
818
 
752
819
  case ZSTD_c_targetLength :
753
820
  BOUNDCHECK(ZSTD_c_targetLength, value);
754
- CCtxParams->cParams.targetLength = value;
821
+ CCtxParams->cParams.targetLength = (U32)value;
755
822
  return CCtxParams->cParams.targetLength;
756
823
 
757
824
  case ZSTD_c_strategy :
@@ -764,12 +831,12 @@ size_t ZSTD_CCtxParams_setParameter(ZSTD_CCtx_params* CCtxParams,
764
831
  /* Content size written in frame header _when known_ (default:1) */
765
832
  DEBUGLOG(4, "set content size flag = %u", (value!=0));
766
833
  CCtxParams->fParams.contentSizeFlag = value != 0;
767
- return CCtxParams->fParams.contentSizeFlag;
834
+ return (size_t)CCtxParams->fParams.contentSizeFlag;
768
835
 
769
836
  case ZSTD_c_checksumFlag :
770
837
  /* A 32-bits content checksum will be calculated and written at end of frame (default:0) */
771
838
  CCtxParams->fParams.checksumFlag = value != 0;
772
- return CCtxParams->fParams.checksumFlag;
839
+ return (size_t)CCtxParams->fParams.checksumFlag;
773
840
 
774
841
  case ZSTD_c_dictIDFlag : /* When applicable, dictionary's dictID is provided in frame header (default:1) */
775
842
  DEBUGLOG(4, "set dictIDFlag = %u", (value!=0));
@@ -778,18 +845,18 @@ size_t ZSTD_CCtxParams_setParameter(ZSTD_CCtx_params* CCtxParams,
778
845
 
779
846
  case ZSTD_c_forceMaxWindow :
780
847
  CCtxParams->forceWindow = (value != 0);
781
- return CCtxParams->forceWindow;
848
+ return (size_t)CCtxParams->forceWindow;
782
849
 
783
850
  case ZSTD_c_forceAttachDict : {
784
851
  const ZSTD_dictAttachPref_e pref = (ZSTD_dictAttachPref_e)value;
785
- BOUNDCHECK(ZSTD_c_forceAttachDict, pref);
852
+ BOUNDCHECK(ZSTD_c_forceAttachDict, (int)pref);
786
853
  CCtxParams->attachDictPref = pref;
787
854
  return CCtxParams->attachDictPref;
788
855
  }
789
856
 
790
857
  case ZSTD_c_literalCompressionMode : {
791
858
  const ZSTD_paramSwitch_e lcm = (ZSTD_paramSwitch_e)value;
792
- BOUNDCHECK(ZSTD_c_literalCompressionMode, lcm);
859
+ BOUNDCHECK(ZSTD_c_literalCompressionMode, (int)lcm);
793
860
  CCtxParams->literalCompressionMode = lcm;
794
861
  return CCtxParams->literalCompressionMode;
795
862
  }
@@ -840,47 +907,48 @@ size_t ZSTD_CCtxParams_setParameter(ZSTD_CCtx_params* CCtxParams,
840
907
 
841
908
  case ZSTD_c_enableDedicatedDictSearch :
842
909
  CCtxParams->enableDedicatedDictSearch = (value!=0);
843
- return CCtxParams->enableDedicatedDictSearch;
910
+ return (size_t)CCtxParams->enableDedicatedDictSearch;
844
911
 
845
912
  case ZSTD_c_enableLongDistanceMatching :
913
+ BOUNDCHECK(ZSTD_c_enableLongDistanceMatching, value);
846
914
  CCtxParams->ldmParams.enableLdm = (ZSTD_paramSwitch_e)value;
847
915
  return CCtxParams->ldmParams.enableLdm;
848
916
 
849
917
  case ZSTD_c_ldmHashLog :
850
918
  if (value!=0) /* 0 ==> auto */
851
919
  BOUNDCHECK(ZSTD_c_ldmHashLog, value);
852
- CCtxParams->ldmParams.hashLog = value;
920
+ CCtxParams->ldmParams.hashLog = (U32)value;
853
921
  return CCtxParams->ldmParams.hashLog;
854
922
 
855
923
  case ZSTD_c_ldmMinMatch :
856
924
  if (value!=0) /* 0 ==> default */
857
925
  BOUNDCHECK(ZSTD_c_ldmMinMatch, value);
858
- CCtxParams->ldmParams.minMatchLength = value;
926
+ CCtxParams->ldmParams.minMatchLength = (U32)value;
859
927
  return CCtxParams->ldmParams.minMatchLength;
860
928
 
861
929
  case ZSTD_c_ldmBucketSizeLog :
862
930
  if (value!=0) /* 0 ==> default */
863
931
  BOUNDCHECK(ZSTD_c_ldmBucketSizeLog, value);
864
- CCtxParams->ldmParams.bucketSizeLog = value;
932
+ CCtxParams->ldmParams.bucketSizeLog = (U32)value;
865
933
  return CCtxParams->ldmParams.bucketSizeLog;
866
934
 
867
935
  case ZSTD_c_ldmHashRateLog :
868
936
  if (value!=0) /* 0 ==> default */
869
937
  BOUNDCHECK(ZSTD_c_ldmHashRateLog, value);
870
- CCtxParams->ldmParams.hashRateLog = value;
938
+ CCtxParams->ldmParams.hashRateLog = (U32)value;
871
939
  return CCtxParams->ldmParams.hashRateLog;
872
940
 
873
941
  case ZSTD_c_targetCBlockSize :
874
942
  if (value!=0) /* 0 ==> default */
875
943
  BOUNDCHECK(ZSTD_c_targetCBlockSize, value);
876
- CCtxParams->targetCBlockSize = value;
944
+ CCtxParams->targetCBlockSize = (U32)value;
877
945
  return CCtxParams->targetCBlockSize;
878
946
 
879
947
  case ZSTD_c_srcSizeHint :
880
948
  if (value!=0) /* 0 ==> default */
881
949
  BOUNDCHECK(ZSTD_c_srcSizeHint, value);
882
950
  CCtxParams->srcSizeHint = value;
883
- return CCtxParams->srcSizeHint;
951
+ return (size_t)CCtxParams->srcSizeHint;
884
952
 
885
953
  case ZSTD_c_stableInBuffer:
886
954
  BOUNDCHECK(ZSTD_c_stableInBuffer, value);
@@ -917,6 +985,27 @@ size_t ZSTD_CCtxParams_setParameter(ZSTD_CCtx_params* CCtxParams,
917
985
  CCtxParams->deterministicRefPrefix = !!value;
918
986
  return CCtxParams->deterministicRefPrefix;
919
987
 
988
+ case ZSTD_c_prefetchCDictTables:
989
+ BOUNDCHECK(ZSTD_c_prefetchCDictTables, value);
990
+ CCtxParams->prefetchCDictTables = (ZSTD_paramSwitch_e)value;
991
+ return CCtxParams->prefetchCDictTables;
992
+
993
+ case ZSTD_c_enableSeqProducerFallback:
994
+ BOUNDCHECK(ZSTD_c_enableSeqProducerFallback, value);
995
+ CCtxParams->enableMatchFinderFallback = value;
996
+ return CCtxParams->enableMatchFinderFallback;
997
+
998
+ case ZSTD_c_maxBlockSize:
999
+ if (value!=0) /* 0 ==> default */
1000
+ BOUNDCHECK(ZSTD_c_maxBlockSize, value);
1001
+ CCtxParams->maxBlockSize = value;
1002
+ return CCtxParams->maxBlockSize;
1003
+
1004
+ case ZSTD_c_searchForExternalRepcodes:
1005
+ BOUNDCHECK(ZSTD_c_searchForExternalRepcodes, value);
1006
+ CCtxParams->searchForExternalRepcodes = (ZSTD_paramSwitch_e)value;
1007
+ return CCtxParams->searchForExternalRepcodes;
1008
+
920
1009
  default: RETURN_ERROR(parameter_unsupported, "unknown parameter");
921
1010
  }
922
1011
  }
@@ -1049,6 +1138,18 @@ size_t ZSTD_CCtxParams_getParameter(
1049
1138
  case ZSTD_c_deterministicRefPrefix:
1050
1139
  *value = (int)CCtxParams->deterministicRefPrefix;
1051
1140
  break;
1141
+ case ZSTD_c_prefetchCDictTables:
1142
+ *value = (int)CCtxParams->prefetchCDictTables;
1143
+ break;
1144
+ case ZSTD_c_enableSeqProducerFallback:
1145
+ *value = CCtxParams->enableMatchFinderFallback;
1146
+ break;
1147
+ case ZSTD_c_maxBlockSize:
1148
+ *value = (int)CCtxParams->maxBlockSize;
1149
+ break;
1150
+ case ZSTD_c_searchForExternalRepcodes:
1151
+ *value = (int)CCtxParams->searchForExternalRepcodes;
1152
+ break;
1052
1153
  default: RETURN_ERROR(parameter_unsupported, "unknown parameter");
1053
1154
  }
1054
1155
  return 0;
@@ -1075,9 +1176,47 @@ size_t ZSTD_CCtx_setParametersUsingCCtxParams(
1075
1176
  return 0;
1076
1177
  }
1077
1178
 
1179
+ size_t ZSTD_CCtx_setCParams(ZSTD_CCtx* cctx, ZSTD_compressionParameters cparams)
1180
+ {
1181
+ ZSTD_STATIC_ASSERT(sizeof(cparams) == 7 * 4 /* all params are listed below */);
1182
+ DEBUGLOG(4, "ZSTD_CCtx_setCParams");
1183
+ /* only update if all parameters are valid */
1184
+ FORWARD_IF_ERROR(ZSTD_checkCParams(cparams), "");
1185
+ FORWARD_IF_ERROR(ZSTD_CCtx_setParameter(cctx, ZSTD_c_windowLog, cparams.windowLog), "");
1186
+ FORWARD_IF_ERROR(ZSTD_CCtx_setParameter(cctx, ZSTD_c_chainLog, cparams.chainLog), "");
1187
+ FORWARD_IF_ERROR(ZSTD_CCtx_setParameter(cctx, ZSTD_c_hashLog, cparams.hashLog), "");
1188
+ FORWARD_IF_ERROR(ZSTD_CCtx_setParameter(cctx, ZSTD_c_searchLog, cparams.searchLog), "");
1189
+ FORWARD_IF_ERROR(ZSTD_CCtx_setParameter(cctx, ZSTD_c_minMatch, cparams.minMatch), "");
1190
+ FORWARD_IF_ERROR(ZSTD_CCtx_setParameter(cctx, ZSTD_c_targetLength, cparams.targetLength), "");
1191
+ FORWARD_IF_ERROR(ZSTD_CCtx_setParameter(cctx, ZSTD_c_strategy, cparams.strategy), "");
1192
+ return 0;
1193
+ }
1194
+
1195
+ size_t ZSTD_CCtx_setFParams(ZSTD_CCtx* cctx, ZSTD_frameParameters fparams)
1196
+ {
1197
+ ZSTD_STATIC_ASSERT(sizeof(fparams) == 3 * 4 /* all params are listed below */);
1198
+ DEBUGLOG(4, "ZSTD_CCtx_setFParams");
1199
+ FORWARD_IF_ERROR(ZSTD_CCtx_setParameter(cctx, ZSTD_c_contentSizeFlag, fparams.contentSizeFlag != 0), "");
1200
+ FORWARD_IF_ERROR(ZSTD_CCtx_setParameter(cctx, ZSTD_c_checksumFlag, fparams.checksumFlag != 0), "");
1201
+ FORWARD_IF_ERROR(ZSTD_CCtx_setParameter(cctx, ZSTD_c_dictIDFlag, fparams.noDictIDFlag == 0), "");
1202
+ return 0;
1203
+ }
1204
+
1205
+ size_t ZSTD_CCtx_setParams(ZSTD_CCtx* cctx, ZSTD_parameters params)
1206
+ {
1207
+ DEBUGLOG(4, "ZSTD_CCtx_setParams");
1208
+ /* First check cParams, because we want to update all or none. */
1209
+ FORWARD_IF_ERROR(ZSTD_checkCParams(params.cParams), "");
1210
+ /* Next set fParams, because this could fail if the cctx isn't in init stage. */
1211
+ FORWARD_IF_ERROR(ZSTD_CCtx_setFParams(cctx, params.fParams), "");
1212
+ /* Finally set cParams, which should succeed. */
1213
+ FORWARD_IF_ERROR(ZSTD_CCtx_setCParams(cctx, params.cParams), "");
1214
+ return 0;
1215
+ }
1216
+
1078
1217
  size_t ZSTD_CCtx_setPledgedSrcSize(ZSTD_CCtx* cctx, unsigned long long pledgedSrcSize)
1079
1218
  {
1080
- DEBUGLOG(4, "ZSTD_CCtx_setPledgedSrcSize to %u bytes", (U32)pledgedSrcSize);
1219
+ DEBUGLOG(4, "ZSTD_CCtx_setPledgedSrcSize to %llu bytes", pledgedSrcSize);
1081
1220
  RETURN_ERROR_IF(cctx->streamStage != zcss_init, stage_wrong,
1082
1221
  "Can't set pledgedSrcSize when not in init stage.");
1083
1222
  cctx->pledgedSrcSizePlusOne = pledgedSrcSize+1;
@@ -1093,9 +1232,9 @@ static void ZSTD_dedicatedDictSearch_revertCParams(
1093
1232
  ZSTD_compressionParameters* cParams);
1094
1233
 
1095
1234
  /**
1096
- * Initializes the local dict using the requested parameters.
1097
- * NOTE: This does not use the pledged src size, because it may be used for more
1098
- * than one compression.
1235
+ * Initializes the local dictionary using requested parameters.
1236
+ * NOTE: Initialization does not employ the pledged src size,
1237
+ * because the dictionary may be used for multiple compressions.
1099
1238
  */
1100
1239
  static size_t ZSTD_initLocalDict(ZSTD_CCtx* cctx)
1101
1240
  {
@@ -1108,8 +1247,8 @@ static size_t ZSTD_initLocalDict(ZSTD_CCtx* cctx)
1108
1247
  return 0;
1109
1248
  }
1110
1249
  if (dl->cdict != NULL) {
1111
- assert(cctx->cdict == dl->cdict);
1112
1250
  /* Local dictionary already initialized. */
1251
+ assert(cctx->cdict == dl->cdict);
1113
1252
  return 0;
1114
1253
  }
1115
1254
  assert(dl->dictSize > 0);
@@ -1129,26 +1268,30 @@ static size_t ZSTD_initLocalDict(ZSTD_CCtx* cctx)
1129
1268
  }
1130
1269
 
1131
1270
  size_t ZSTD_CCtx_loadDictionary_advanced(
1132
- ZSTD_CCtx* cctx, const void* dict, size_t dictSize,
1133
- ZSTD_dictLoadMethod_e dictLoadMethod, ZSTD_dictContentType_e dictContentType)
1271
+ ZSTD_CCtx* cctx,
1272
+ const void* dict, size_t dictSize,
1273
+ ZSTD_dictLoadMethod_e dictLoadMethod,
1274
+ ZSTD_dictContentType_e dictContentType)
1134
1275
  {
1135
- RETURN_ERROR_IF(cctx->streamStage != zcss_init, stage_wrong,
1136
- "Can't load a dictionary when ctx is not in init stage.");
1137
1276
  DEBUGLOG(4, "ZSTD_CCtx_loadDictionary_advanced (size: %u)", (U32)dictSize);
1138
- ZSTD_clearAllDicts(cctx); /* in case one already exists */
1139
- if (dict == NULL || dictSize == 0) /* no dictionary mode */
1277
+ RETURN_ERROR_IF(cctx->streamStage != zcss_init, stage_wrong,
1278
+ "Can't load a dictionary when cctx is not in init stage.");
1279
+ ZSTD_clearAllDicts(cctx); /* erase any previously set dictionary */
1280
+ if (dict == NULL || dictSize == 0) /* no dictionary */
1140
1281
  return 0;
1141
1282
  if (dictLoadMethod == ZSTD_dlm_byRef) {
1142
1283
  cctx->localDict.dict = dict;
1143
1284
  } else {
1285
+ /* copy dictionary content inside CCtx to own its lifetime */
1144
1286
  void* dictBuffer;
1145
1287
  RETURN_ERROR_IF(cctx->staticSize, memory_allocation,
1146
- "no malloc for static CCtx");
1288
+ "static CCtx can't allocate for an internal copy of dictionary");
1147
1289
  dictBuffer = ZSTD_customMalloc(dictSize, cctx->customMem);
1148
- RETURN_ERROR_IF(!dictBuffer, memory_allocation, "NULL pointer!");
1290
+ RETURN_ERROR_IF(dictBuffer==NULL, memory_allocation,
1291
+ "allocation failed for dictionary content");
1149
1292
  ZSTD_memcpy(dictBuffer, dict, dictSize);
1150
- cctx->localDict.dictBuffer = dictBuffer;
1151
- cctx->localDict.dict = dictBuffer;
1293
+ cctx->localDict.dictBuffer = dictBuffer; /* owned ptr to free */
1294
+ cctx->localDict.dict = dictBuffer; /* read-only reference */
1152
1295
  }
1153
1296
  cctx->localDict.dictSize = dictSize;
1154
1297
  cctx->localDict.dictContentType = dictContentType;
@@ -1218,8 +1361,9 @@ size_t ZSTD_CCtx_reset(ZSTD_CCtx* cctx, ZSTD_ResetDirective reset)
1218
1361
  if ( (reset == ZSTD_reset_parameters)
1219
1362
  || (reset == ZSTD_reset_session_and_parameters) ) {
1220
1363
  RETURN_ERROR_IF(cctx->streamStage != zcss_init, stage_wrong,
1221
- "Can't reset parameters only when not in init stage.");
1364
+ "Reset parameters is only possible during init stage.");
1222
1365
  ZSTD_clearAllDicts(cctx);
1366
+ ZSTD_memset(&cctx->externalMatchCtx, 0, sizeof(cctx->externalMatchCtx));
1223
1367
  return ZSTD_CCtxParams_reset(&cctx->requestedParams);
1224
1368
  }
1225
1369
  return 0;
@@ -1316,7 +1460,8 @@ static ZSTD_compressionParameters
1316
1460
  ZSTD_adjustCParams_internal(ZSTD_compressionParameters cPar,
1317
1461
  unsigned long long srcSize,
1318
1462
  size_t dictSize,
1319
- ZSTD_cParamMode_e mode)
1463
+ ZSTD_cParamMode_e mode,
1464
+ ZSTD_paramSwitch_e useRowMatchFinder)
1320
1465
  {
1321
1466
  const U64 minSrcSize = 513; /* (1<<9) + 1 */
1322
1467
  const U64 maxWindowResize = 1ULL << (ZSTD_WINDOWLOG_MAX-1);
@@ -1350,8 +1495,8 @@ ZSTD_adjustCParams_internal(ZSTD_compressionParameters cPar,
1350
1495
  }
1351
1496
 
1352
1497
  /* resize windowLog if input is small enough, to use less memory */
1353
- if ( (srcSize < maxWindowResize)
1354
- && (dictSize < maxWindowResize) ) {
1498
+ if ( (srcSize <= maxWindowResize)
1499
+ && (dictSize <= maxWindowResize) ) {
1355
1500
  U32 const tSize = (U32)(srcSize + dictSize);
1356
1501
  static U32 const hashSizeMin = 1 << ZSTD_HASHLOG_MIN;
1357
1502
  U32 const srcLog = (tSize < hashSizeMin) ? ZSTD_HASHLOG_MIN :
@@ -1369,6 +1514,42 @@ ZSTD_adjustCParams_internal(ZSTD_compressionParameters cPar,
1369
1514
  if (cPar.windowLog < ZSTD_WINDOWLOG_ABSOLUTEMIN)
1370
1515
  cPar.windowLog = ZSTD_WINDOWLOG_ABSOLUTEMIN; /* minimum wlog required for valid frame header */
1371
1516
 
1517
+ /* We can't use more than 32 bits of hash in total, so that means that we require:
1518
+ * (hashLog + 8) <= 32 && (chainLog + 8) <= 32
1519
+ */
1520
+ if (mode == ZSTD_cpm_createCDict && ZSTD_CDictIndicesAreTagged(&cPar)) {
1521
+ U32 const maxShortCacheHashLog = 32 - ZSTD_SHORT_CACHE_TAG_BITS;
1522
+ if (cPar.hashLog > maxShortCacheHashLog) {
1523
+ cPar.hashLog = maxShortCacheHashLog;
1524
+ }
1525
+ if (cPar.chainLog > maxShortCacheHashLog) {
1526
+ cPar.chainLog = maxShortCacheHashLog;
1527
+ }
1528
+ }
1529
+
1530
+
1531
+ /* At this point, we aren't 100% sure if we are using the row match finder.
1532
+ * Unless it is explicitly disabled, conservatively assume that it is enabled.
1533
+ * In this case it will only be disabled for small sources, so shrinking the
1534
+ * hash log a little bit shouldn't result in any ratio loss.
1535
+ */
1536
+ if (useRowMatchFinder == ZSTD_ps_auto)
1537
+ useRowMatchFinder = ZSTD_ps_enable;
1538
+
1539
+ /* We can't hash more than 32-bits in total. So that means that we require:
1540
+ * (hashLog - rowLog + 8) <= 32
1541
+ */
1542
+ if (ZSTD_rowMatchFinderUsed(cPar.strategy, useRowMatchFinder)) {
1543
+ /* Switch to 32-entry rows if searchLog is 5 (or more) */
1544
+ U32 const rowLog = BOUNDED(4, cPar.searchLog, 6);
1545
+ U32 const maxRowHashLog = 32 - ZSTD_ROW_HASH_TAG_BITS;
1546
+ U32 const maxHashLog = maxRowHashLog + rowLog;
1547
+ assert(cPar.hashLog >= rowLog);
1548
+ if (cPar.hashLog > maxHashLog) {
1549
+ cPar.hashLog = maxHashLog;
1550
+ }
1551
+ }
1552
+
1372
1553
  return cPar;
1373
1554
  }
1374
1555
 
@@ -1379,7 +1560,7 @@ ZSTD_adjustCParams(ZSTD_compressionParameters cPar,
1379
1560
  {
1380
1561
  cPar = ZSTD_clampCParams(cPar); /* resulting cPar is necessarily valid (all parameters within range) */
1381
1562
  if (srcSize == 0) srcSize = ZSTD_CONTENTSIZE_UNKNOWN;
1382
- return ZSTD_adjustCParams_internal(cPar, srcSize, dictSize, ZSTD_cpm_unknown);
1563
+ return ZSTD_adjustCParams_internal(cPar, srcSize, dictSize, ZSTD_cpm_unknown, ZSTD_ps_auto);
1383
1564
  }
1384
1565
 
1385
1566
  static ZSTD_compressionParameters ZSTD_getCParams_internal(int compressionLevel, unsigned long long srcSizeHint, size_t dictSize, ZSTD_cParamMode_e mode);
@@ -1410,7 +1591,7 @@ ZSTD_compressionParameters ZSTD_getCParamsFromCCtxParams(
1410
1591
  ZSTD_overrideCParams(&cParams, &CCtxParams->cParams);
1411
1592
  assert(!ZSTD_checkCParams(cParams));
1412
1593
  /* srcSizeHint == 0 means 0 */
1413
- return ZSTD_adjustCParams_internal(cParams, srcSizeHint, dictSize, mode);
1594
+ return ZSTD_adjustCParams_internal(cParams, srcSizeHint, dictSize, mode, CCtxParams->useRowMatchFinder);
1414
1595
  }
1415
1596
 
1416
1597
  static size_t
@@ -1439,7 +1620,7 @@ ZSTD_sizeof_matchState(const ZSTD_compressionParameters* const cParams,
1439
1620
  + ZSTD_cwksp_aligned_alloc_size((ZSTD_OPT_NUM+1) * sizeof(ZSTD_match_t))
1440
1621
  + ZSTD_cwksp_aligned_alloc_size((ZSTD_OPT_NUM+1) * sizeof(ZSTD_optimal_t));
1441
1622
  size_t const lazyAdditionalSpace = ZSTD_rowMatchFinderUsed(cParams->strategy, useRowMatchFinder)
1442
- ? ZSTD_cwksp_aligned_alloc_size(hSize*sizeof(U16))
1623
+ ? ZSTD_cwksp_aligned_alloc_size(hSize)
1443
1624
  : 0;
1444
1625
  size_t const optSpace = (forCCtx && (cParams->strategy >= ZSTD_btopt))
1445
1626
  ? optPotentialSpace
@@ -1455,6 +1636,13 @@ ZSTD_sizeof_matchState(const ZSTD_compressionParameters* const cParams,
1455
1636
  return tableSpace + optSpace + slackSpace + lazyAdditionalSpace;
1456
1637
  }
1457
1638
 
1639
+ /* Helper function for calculating memory requirements.
1640
+ * Gives a tighter bound than ZSTD_sequenceBound() by taking minMatch into account. */
1641
+ static size_t ZSTD_maxNbSeq(size_t blockSize, unsigned minMatch, int useSequenceProducer) {
1642
+ U32 const divider = (minMatch==3 || useSequenceProducer) ? 3 : 4;
1643
+ return blockSize / divider;
1644
+ }
1645
+
1458
1646
  static size_t ZSTD_estimateCCtxSize_usingCCtxParams_internal(
1459
1647
  const ZSTD_compressionParameters* cParams,
1460
1648
  const ldmParams_t* ldmParams,
@@ -1462,12 +1650,13 @@ static size_t ZSTD_estimateCCtxSize_usingCCtxParams_internal(
1462
1650
  const ZSTD_paramSwitch_e useRowMatchFinder,
1463
1651
  const size_t buffInSize,
1464
1652
  const size_t buffOutSize,
1465
- const U64 pledgedSrcSize)
1653
+ const U64 pledgedSrcSize,
1654
+ int useSequenceProducer,
1655
+ size_t maxBlockSize)
1466
1656
  {
1467
1657
  size_t const windowSize = (size_t) BOUNDED(1ULL, 1ULL << cParams->windowLog, pledgedSrcSize);
1468
- size_t const blockSize = MIN(ZSTD_BLOCKSIZE_MAX, windowSize);
1469
- U32 const divider = (cParams->minMatch==3) ? 3 : 4;
1470
- size_t const maxNbSeq = blockSize / divider;
1658
+ size_t const blockSize = MIN(ZSTD_resolveMaxBlockSize(maxBlockSize), windowSize);
1659
+ size_t const maxNbSeq = ZSTD_maxNbSeq(blockSize, cParams->minMatch, useSequenceProducer);
1471
1660
  size_t const tokenSpace = ZSTD_cwksp_alloc_size(WILDCOPY_OVERLENGTH + blockSize)
1472
1661
  + ZSTD_cwksp_aligned_alloc_size(maxNbSeq * sizeof(seqDef))
1473
1662
  + 3 * ZSTD_cwksp_alloc_size(maxNbSeq * sizeof(BYTE));
@@ -1486,6 +1675,11 @@ static size_t ZSTD_estimateCCtxSize_usingCCtxParams_internal(
1486
1675
 
1487
1676
  size_t const cctxSpace = isStatic ? ZSTD_cwksp_alloc_size(sizeof(ZSTD_CCtx)) : 0;
1488
1677
 
1678
+ size_t const maxNbExternalSeq = ZSTD_sequenceBound(blockSize);
1679
+ size_t const externalSeqSpace = useSequenceProducer
1680
+ ? ZSTD_cwksp_aligned_alloc_size(maxNbExternalSeq * sizeof(ZSTD_Sequence))
1681
+ : 0;
1682
+
1489
1683
  size_t const neededSpace =
1490
1684
  cctxSpace +
1491
1685
  entropySpace +
@@ -1494,7 +1688,8 @@ static size_t ZSTD_estimateCCtxSize_usingCCtxParams_internal(
1494
1688
  ldmSeqSpace +
1495
1689
  matchStateSize +
1496
1690
  tokenSpace +
1497
- bufferSpace;
1691
+ bufferSpace +
1692
+ externalSeqSpace;
1498
1693
 
1499
1694
  DEBUGLOG(5, "estimate workspace : %u", (U32)neededSpace);
1500
1695
  return neededSpace;
@@ -1512,7 +1707,7 @@ size_t ZSTD_estimateCCtxSize_usingCCtxParams(const ZSTD_CCtx_params* params)
1512
1707
  * be needed. However, we still allocate two 0-sized buffers, which can
1513
1708
  * take space under ASAN. */
1514
1709
  return ZSTD_estimateCCtxSize_usingCCtxParams_internal(
1515
- &cParams, &params->ldmParams, 1, useRowMatchFinder, 0, 0, ZSTD_CONTENTSIZE_UNKNOWN);
1710
+ &cParams, &params->ldmParams, 1, useRowMatchFinder, 0, 0, ZSTD_CONTENTSIZE_UNKNOWN, params->useSequenceProducer, params->maxBlockSize);
1516
1711
  }
1517
1712
 
1518
1713
  size_t ZSTD_estimateCCtxSize_usingCParams(ZSTD_compressionParameters cParams)
@@ -1562,7 +1757,7 @@ size_t ZSTD_estimateCStreamSize_usingCCtxParams(const ZSTD_CCtx_params* params)
1562
1757
  RETURN_ERROR_IF(params->nbWorkers > 0, GENERIC, "Estimate CCtx size is supported for single-threaded compression only.");
1563
1758
  { ZSTD_compressionParameters const cParams =
1564
1759
  ZSTD_getCParamsFromCCtxParams(params, ZSTD_CONTENTSIZE_UNKNOWN, 0, ZSTD_cpm_noAttachDict);
1565
- size_t const blockSize = MIN(ZSTD_BLOCKSIZE_MAX, (size_t)1 << cParams.windowLog);
1760
+ size_t const blockSize = MIN(ZSTD_resolveMaxBlockSize(params->maxBlockSize), (size_t)1 << cParams.windowLog);
1566
1761
  size_t const inBuffSize = (params->inBufferMode == ZSTD_bm_buffered)
1567
1762
  ? ((size_t)1 << cParams.windowLog) + blockSize
1568
1763
  : 0;
@@ -1573,7 +1768,7 @@ size_t ZSTD_estimateCStreamSize_usingCCtxParams(const ZSTD_CCtx_params* params)
1573
1768
 
1574
1769
  return ZSTD_estimateCCtxSize_usingCCtxParams_internal(
1575
1770
  &cParams, &params->ldmParams, 1, useRowMatchFinder, inBuffSize, outBuffSize,
1576
- ZSTD_CONTENTSIZE_UNKNOWN);
1771
+ ZSTD_CONTENTSIZE_UNKNOWN, params->useSequenceProducer, params->maxBlockSize);
1577
1772
  }
1578
1773
  }
1579
1774
 
@@ -1716,6 +1911,19 @@ typedef enum {
1716
1911
  ZSTD_resetTarget_CCtx
1717
1912
  } ZSTD_resetTarget_e;
1718
1913
 
1914
+ /* Mixes bits in a 64 bits in a value, based on XXH3_rrmxmx */
1915
+ static U64 ZSTD_bitmix(U64 val, U64 len) {
1916
+ val ^= ZSTD_rotateRight_U64(val, 49) ^ ZSTD_rotateRight_U64(val, 24);
1917
+ val *= 0x9FB21C651E98DF25ULL;
1918
+ val ^= (val >> 35) + len ;
1919
+ val *= 0x9FB21C651E98DF25ULL;
1920
+ return val ^ (val >> 28);
1921
+ }
1922
+
1923
+ /* Mixes in the hashSalt and hashSaltEntropy to create a new hashSalt */
1924
+ static void ZSTD_advanceHashSalt(ZSTD_matchState_t* ms) {
1925
+ ms->hashSalt = ZSTD_bitmix(ms->hashSalt, 8) ^ ZSTD_bitmix((U64) ms->hashSaltEntropy, 4);
1926
+ }
1719
1927
 
1720
1928
  static size_t
1721
1929
  ZSTD_reset_matchState(ZSTD_matchState_t* ms,
@@ -1743,6 +1951,7 @@ ZSTD_reset_matchState(ZSTD_matchState_t* ms,
1743
1951
  }
1744
1952
 
1745
1953
  ms->hashLog3 = hashLog3;
1954
+ ms->lazySkipping = 0;
1746
1955
 
1747
1956
  ZSTD_invalidateMatchState(ms);
1748
1957
 
@@ -1764,6 +1973,27 @@ ZSTD_reset_matchState(ZSTD_matchState_t* ms,
1764
1973
  ZSTD_cwksp_clean_tables(ws);
1765
1974
  }
1766
1975
 
1976
+ if (ZSTD_rowMatchFinderUsed(cParams->strategy, useRowMatchFinder)) {
1977
+ /* Row match finder needs an additional table of hashes ("tags") */
1978
+ size_t const tagTableSize = hSize;
1979
+ /* We want to generate a new salt in case we reset a Cctx, but we always want to use
1980
+ * 0 when we reset a Cdict */
1981
+ if(forWho == ZSTD_resetTarget_CCtx) {
1982
+ ms->tagTable = (BYTE*) ZSTD_cwksp_reserve_aligned_init_once(ws, tagTableSize);
1983
+ ZSTD_advanceHashSalt(ms);
1984
+ } else {
1985
+ /* When we are not salting we want to always memset the memory */
1986
+ ms->tagTable = (BYTE*) ZSTD_cwksp_reserve_aligned(ws, tagTableSize);
1987
+ ZSTD_memset(ms->tagTable, 0, tagTableSize);
1988
+ ms->hashSalt = 0;
1989
+ }
1990
+ { /* Switch to 32-entry rows if searchLog is 5 (or more) */
1991
+ U32 const rowLog = BOUNDED(4, cParams->searchLog, 6);
1992
+ assert(cParams->hashLog >= rowLog);
1993
+ ms->rowHashLog = cParams->hashLog - rowLog;
1994
+ }
1995
+ }
1996
+
1767
1997
  /* opt parser space */
1768
1998
  if ((forWho == ZSTD_resetTarget_CCtx) && (cParams->strategy >= ZSTD_btopt)) {
1769
1999
  DEBUGLOG(4, "reserving optimal parser space");
@@ -1775,19 +2005,6 @@ ZSTD_reset_matchState(ZSTD_matchState_t* ms,
1775
2005
  ms->opt.priceTable = (ZSTD_optimal_t*)ZSTD_cwksp_reserve_aligned(ws, (ZSTD_OPT_NUM+1) * sizeof(ZSTD_optimal_t));
1776
2006
  }
1777
2007
 
1778
- if (ZSTD_rowMatchFinderUsed(cParams->strategy, useRowMatchFinder)) {
1779
- { /* Row match finder needs an additional table of hashes ("tags") */
1780
- size_t const tagTableSize = hSize*sizeof(U16);
1781
- ms->tagTable = (U16*)ZSTD_cwksp_reserve_aligned(ws, tagTableSize);
1782
- if (ms->tagTable) ZSTD_memset(ms->tagTable, 0, tagTableSize);
1783
- }
1784
- { /* Switch to 32-entry rows if searchLog is 5 (or more) */
1785
- U32 const rowLog = BOUNDED(4, cParams->searchLog, 6);
1786
- assert(cParams->hashLog >= rowLog);
1787
- ms->rowHashLog = cParams->hashLog - rowLog;
1788
- }
1789
- }
1790
-
1791
2008
  ms->cParams = *cParams;
1792
2009
 
1793
2010
  RETURN_ERROR_IF(ZSTD_cwksp_reserve_failed(ws), memory_allocation,
@@ -1847,6 +2064,7 @@ static size_t ZSTD_resetCCtx_internal(ZSTD_CCtx* zc,
1847
2064
  assert(params->useRowMatchFinder != ZSTD_ps_auto);
1848
2065
  assert(params->useBlockSplitter != ZSTD_ps_auto);
1849
2066
  assert(params->ldmParams.enableLdm != ZSTD_ps_auto);
2067
+ assert(params->maxBlockSize != 0);
1850
2068
  if (params->ldmParams.enableLdm == ZSTD_ps_enable) {
1851
2069
  /* Adjust long distance matching parameters */
1852
2070
  ZSTD_ldm_adjustParameters(&zc->appliedParams.ldmParams, &params->cParams);
@@ -1855,9 +2073,8 @@ static size_t ZSTD_resetCCtx_internal(ZSTD_CCtx* zc,
1855
2073
  }
1856
2074
 
1857
2075
  { size_t const windowSize = MAX(1, (size_t)MIN(((U64)1 << params->cParams.windowLog), pledgedSrcSize));
1858
- size_t const blockSize = MIN(ZSTD_BLOCKSIZE_MAX, windowSize);
1859
- U32 const divider = (params->cParams.minMatch==3) ? 3 : 4;
1860
- size_t const maxNbSeq = blockSize / divider;
2076
+ size_t const blockSize = MIN(params->maxBlockSize, windowSize);
2077
+ size_t const maxNbSeq = ZSTD_maxNbSeq(blockSize, params->cParams.minMatch, params->useSequenceProducer);
1861
2078
  size_t const buffOutSize = (zbuff == ZSTDb_buffered && params->outBufferMode == ZSTD_bm_buffered)
1862
2079
  ? ZSTD_compressBound(blockSize) + 1
1863
2080
  : 0;
@@ -1874,7 +2091,7 @@ static size_t ZSTD_resetCCtx_internal(ZSTD_CCtx* zc,
1874
2091
  size_t const neededSpace =
1875
2092
  ZSTD_estimateCCtxSize_usingCCtxParams_internal(
1876
2093
  &params->cParams, &params->ldmParams, zc->staticSize != 0, params->useRowMatchFinder,
1877
- buffInSize, buffOutSize, pledgedSrcSize);
2094
+ buffInSize, buffOutSize, pledgedSrcSize, params->useSequenceProducer, params->maxBlockSize);
1878
2095
  int resizeWorkspace;
1879
2096
 
1880
2097
  FORWARD_IF_ERROR(neededSpace, "cctx size estimate failed!");
@@ -1917,6 +2134,7 @@ static size_t ZSTD_resetCCtx_internal(ZSTD_CCtx* zc,
1917
2134
 
1918
2135
  /* init params */
1919
2136
  zc->blockState.matchState.cParams = params->cParams;
2137
+ zc->blockState.matchState.prefetchCDictTables = params->prefetchCDictTables == ZSTD_ps_enable;
1920
2138
  zc->pledgedSrcSizePlusOne = pledgedSrcSize+1;
1921
2139
  zc->consumedSrcSize = 0;
1922
2140
  zc->producedCSize = 0;
@@ -1933,13 +2151,46 @@ static size_t ZSTD_resetCCtx_internal(ZSTD_CCtx* zc,
1933
2151
 
1934
2152
  ZSTD_reset_compressedBlockState(zc->blockState.prevCBlock);
1935
2153
 
2154
+ FORWARD_IF_ERROR(ZSTD_reset_matchState(
2155
+ &zc->blockState.matchState,
2156
+ ws,
2157
+ &params->cParams,
2158
+ params->useRowMatchFinder,
2159
+ crp,
2160
+ needsIndexReset,
2161
+ ZSTD_resetTarget_CCtx), "");
2162
+
2163
+ zc->seqStore.sequencesStart = (seqDef*)ZSTD_cwksp_reserve_aligned(ws, maxNbSeq * sizeof(seqDef));
2164
+
2165
+ /* ldm hash table */
2166
+ if (params->ldmParams.enableLdm == ZSTD_ps_enable) {
2167
+ /* TODO: avoid memset? */
2168
+ size_t const ldmHSize = ((size_t)1) << params->ldmParams.hashLog;
2169
+ zc->ldmState.hashTable = (ldmEntry_t*)ZSTD_cwksp_reserve_aligned(ws, ldmHSize * sizeof(ldmEntry_t));
2170
+ ZSTD_memset(zc->ldmState.hashTable, 0, ldmHSize * sizeof(ldmEntry_t));
2171
+ zc->ldmSequences = (rawSeq*)ZSTD_cwksp_reserve_aligned(ws, maxNbLdmSeq * sizeof(rawSeq));
2172
+ zc->maxNbLdmSequences = maxNbLdmSeq;
2173
+
2174
+ ZSTD_window_init(&zc->ldmState.window);
2175
+ zc->ldmState.loadedDictEnd = 0;
2176
+ }
2177
+
2178
+ /* reserve space for block-level external sequences */
2179
+ if (params->useSequenceProducer) {
2180
+ size_t const maxNbExternalSeq = ZSTD_sequenceBound(blockSize);
2181
+ zc->externalMatchCtx.seqBufferCapacity = maxNbExternalSeq;
2182
+ zc->externalMatchCtx.seqBuffer =
2183
+ (ZSTD_Sequence*)ZSTD_cwksp_reserve_aligned(ws, maxNbExternalSeq * sizeof(ZSTD_Sequence));
2184
+ }
2185
+
2186
+ /* buffers */
2187
+
1936
2188
  /* ZSTD_wildcopy() is used to copy into the literals buffer,
1937
2189
  * so we have to oversize the buffer by WILDCOPY_OVERLENGTH bytes.
1938
2190
  */
1939
2191
  zc->seqStore.litStart = ZSTD_cwksp_reserve_buffer(ws, blockSize + WILDCOPY_OVERLENGTH);
1940
2192
  zc->seqStore.maxNbLit = blockSize;
1941
2193
 
1942
- /* buffers */
1943
2194
  zc->bufferedPolicy = zbuff;
1944
2195
  zc->inBuffSize = buffInSize;
1945
2196
  zc->inBuff = (char*)ZSTD_cwksp_reserve_buffer(ws, buffInSize);
@@ -1962,32 +2213,9 @@ static size_t ZSTD_resetCCtx_internal(ZSTD_CCtx* zc,
1962
2213
  zc->seqStore.llCode = ZSTD_cwksp_reserve_buffer(ws, maxNbSeq * sizeof(BYTE));
1963
2214
  zc->seqStore.mlCode = ZSTD_cwksp_reserve_buffer(ws, maxNbSeq * sizeof(BYTE));
1964
2215
  zc->seqStore.ofCode = ZSTD_cwksp_reserve_buffer(ws, maxNbSeq * sizeof(BYTE));
1965
- zc->seqStore.sequencesStart = (seqDef*)ZSTD_cwksp_reserve_aligned(ws, maxNbSeq * sizeof(seqDef));
1966
-
1967
- FORWARD_IF_ERROR(ZSTD_reset_matchState(
1968
- &zc->blockState.matchState,
1969
- ws,
1970
- &params->cParams,
1971
- params->useRowMatchFinder,
1972
- crp,
1973
- needsIndexReset,
1974
- ZSTD_resetTarget_CCtx), "");
1975
-
1976
- /* ldm hash table */
1977
- if (params->ldmParams.enableLdm == ZSTD_ps_enable) {
1978
- /* TODO: avoid memset? */
1979
- size_t const ldmHSize = ((size_t)1) << params->ldmParams.hashLog;
1980
- zc->ldmState.hashTable = (ldmEntry_t*)ZSTD_cwksp_reserve_aligned(ws, ldmHSize * sizeof(ldmEntry_t));
1981
- ZSTD_memset(zc->ldmState.hashTable, 0, ldmHSize * sizeof(ldmEntry_t));
1982
- zc->ldmSequences = (rawSeq*)ZSTD_cwksp_reserve_aligned(ws, maxNbLdmSeq * sizeof(rawSeq));
1983
- zc->maxNbLdmSequences = maxNbLdmSeq;
1984
-
1985
- ZSTD_window_init(&zc->ldmState.window);
1986
- zc->ldmState.loadedDictEnd = 0;
1987
- }
1988
2216
 
1989
2217
  DEBUGLOG(3, "wksp: finished allocating, %zd bytes remain available", ZSTD_cwksp_available_space(ws));
1990
- assert(ZSTD_cwksp_estimated_space_within_bounds(ws, neededSpace, resizeWorkspace));
2218
+ assert(ZSTD_cwksp_estimated_space_within_bounds(ws, neededSpace));
1991
2219
 
1992
2220
  zc->initialized = 1;
1993
2221
 
@@ -2059,7 +2287,8 @@ ZSTD_resetCCtx_byAttachingCDict(ZSTD_CCtx* cctx,
2059
2287
  }
2060
2288
 
2061
2289
  params.cParams = ZSTD_adjustCParams_internal(adjusted_cdict_cParams, pledgedSrcSize,
2062
- cdict->dictContentSize, ZSTD_cpm_attachDict);
2290
+ cdict->dictContentSize, ZSTD_cpm_attachDict,
2291
+ params.useRowMatchFinder);
2063
2292
  params.cParams.windowLog = windowLog;
2064
2293
  params.useRowMatchFinder = cdict->useRowMatchFinder; /* cdict overrides */
2065
2294
  FORWARD_IF_ERROR(ZSTD_resetCCtx_internal(cctx, &params, pledgedSrcSize,
@@ -2098,6 +2327,22 @@ ZSTD_resetCCtx_byAttachingCDict(ZSTD_CCtx* cctx,
2098
2327
  return 0;
2099
2328
  }
2100
2329
 
2330
+ static void ZSTD_copyCDictTableIntoCCtx(U32* dst, U32 const* src, size_t tableSize,
2331
+ ZSTD_compressionParameters const* cParams) {
2332
+ if (ZSTD_CDictIndicesAreTagged(cParams)){
2333
+ /* Remove tags from the CDict table if they are present.
2334
+ * See docs on "short cache" in zstd_compress_internal.h for context. */
2335
+ size_t i;
2336
+ for (i = 0; i < tableSize; i++) {
2337
+ U32 const taggedIndex = src[i];
2338
+ U32 const index = taggedIndex >> ZSTD_SHORT_CACHE_TAG_BITS;
2339
+ dst[i] = index;
2340
+ }
2341
+ } else {
2342
+ ZSTD_memcpy(dst, src, tableSize * sizeof(U32));
2343
+ }
2344
+ }
2345
+
2101
2346
  static size_t ZSTD_resetCCtx_byCopyingCDict(ZSTD_CCtx* cctx,
2102
2347
  const ZSTD_CDict* cdict,
2103
2348
  ZSTD_CCtx_params params,
@@ -2133,21 +2378,23 @@ static size_t ZSTD_resetCCtx_byCopyingCDict(ZSTD_CCtx* cctx,
2133
2378
  : 0;
2134
2379
  size_t const hSize = (size_t)1 << cdict_cParams->hashLog;
2135
2380
 
2136
- ZSTD_memcpy(cctx->blockState.matchState.hashTable,
2137
- cdict->matchState.hashTable,
2138
- hSize * sizeof(U32));
2381
+ ZSTD_copyCDictTableIntoCCtx(cctx->blockState.matchState.hashTable,
2382
+ cdict->matchState.hashTable,
2383
+ hSize, cdict_cParams);
2384
+
2139
2385
  /* Do not copy cdict's chainTable if cctx has parameters such that it would not use chainTable */
2140
2386
  if (ZSTD_allocateChainTable(cctx->appliedParams.cParams.strategy, cctx->appliedParams.useRowMatchFinder, 0 /* forDDSDict */)) {
2141
- ZSTD_memcpy(cctx->blockState.matchState.chainTable,
2142
- cdict->matchState.chainTable,
2143
- chainSize * sizeof(U32));
2387
+ ZSTD_copyCDictTableIntoCCtx(cctx->blockState.matchState.chainTable,
2388
+ cdict->matchState.chainTable,
2389
+ chainSize, cdict_cParams);
2144
2390
  }
2145
2391
  /* copy tag table */
2146
2392
  if (ZSTD_rowMatchFinderUsed(cdict_cParams->strategy, cdict->useRowMatchFinder)) {
2147
- size_t const tagTableSize = hSize*sizeof(U16);
2393
+ size_t const tagTableSize = hSize;
2148
2394
  ZSTD_memcpy(cctx->blockState.matchState.tagTable,
2149
- cdict->matchState.tagTable,
2150
- tagTableSize);
2395
+ cdict->matchState.tagTable,
2396
+ tagTableSize);
2397
+ cctx->blockState.matchState.hashSalt = cdict->matchState.hashSalt;
2151
2398
  }
2152
2399
  }
2153
2400
 
@@ -2226,6 +2473,7 @@ static size_t ZSTD_copyCCtx_internal(ZSTD_CCtx* dstCCtx,
2226
2473
  params.useBlockSplitter = srcCCtx->appliedParams.useBlockSplitter;
2227
2474
  params.ldmParams = srcCCtx->appliedParams.ldmParams;
2228
2475
  params.fParams = fParams;
2476
+ params.maxBlockSize = srcCCtx->appliedParams.maxBlockSize;
2229
2477
  ZSTD_resetCCtx_internal(dstCCtx, &params, pledgedSrcSize,
2230
2478
  /* loadedDictSize */ 0,
2231
2479
  ZSTDcrp_leaveDirty, zbuff);
@@ -2385,7 +2633,7 @@ static void ZSTD_reduceIndex (ZSTD_matchState_t* ms, ZSTD_CCtx_params const* par
2385
2633
 
2386
2634
  /* See doc/zstd_compression_format.md for detailed format description */
2387
2635
 
2388
- void ZSTD_seqToCodes(const seqStore_t* seqStorePtr)
2636
+ int ZSTD_seqToCodes(const seqStore_t* seqStorePtr)
2389
2637
  {
2390
2638
  const seqDef* const sequences = seqStorePtr->sequencesStart;
2391
2639
  BYTE* const llCodeTable = seqStorePtr->llCode;
@@ -2393,18 +2641,24 @@ void ZSTD_seqToCodes(const seqStore_t* seqStorePtr)
2393
2641
  BYTE* const mlCodeTable = seqStorePtr->mlCode;
2394
2642
  U32 const nbSeq = (U32)(seqStorePtr->sequences - seqStorePtr->sequencesStart);
2395
2643
  U32 u;
2644
+ int longOffsets = 0;
2396
2645
  assert(nbSeq <= seqStorePtr->maxNbSeq);
2397
2646
  for (u=0; u<nbSeq; u++) {
2398
2647
  U32 const llv = sequences[u].litLength;
2648
+ U32 const ofCode = ZSTD_highbit32(sequences[u].offBase);
2399
2649
  U32 const mlv = sequences[u].mlBase;
2400
2650
  llCodeTable[u] = (BYTE)ZSTD_LLcode(llv);
2401
- ofCodeTable[u] = (BYTE)ZSTD_highbit32(sequences[u].offBase);
2651
+ ofCodeTable[u] = (BYTE)ofCode;
2402
2652
  mlCodeTable[u] = (BYTE)ZSTD_MLcode(mlv);
2653
+ assert(!(MEM_64bits() && ofCode >= STREAM_ACCUMULATOR_MIN));
2654
+ if (MEM_32bits() && ofCode >= STREAM_ACCUMULATOR_MIN)
2655
+ longOffsets = 1;
2403
2656
  }
2404
2657
  if (seqStorePtr->longLengthType==ZSTD_llt_literalLength)
2405
2658
  llCodeTable[seqStorePtr->longLengthPos] = MaxLL;
2406
2659
  if (seqStorePtr->longLengthType==ZSTD_llt_matchLength)
2407
2660
  mlCodeTable[seqStorePtr->longLengthPos] = MaxML;
2661
+ return longOffsets;
2408
2662
  }
2409
2663
 
2410
2664
  /* ZSTD_useTargetCBlockSize():
@@ -2438,6 +2692,7 @@ typedef struct {
2438
2692
  U32 MLtype;
2439
2693
  size_t size;
2440
2694
  size_t lastCountSize; /* Accounts for bug in 1.3.4. More detail in ZSTD_entropyCompressSeqStore_internal() */
2695
+ int longOffsets;
2441
2696
  } ZSTD_symbolEncodingTypeStats_t;
2442
2697
 
2443
2698
  /* ZSTD_buildSequencesStatistics():
@@ -2448,11 +2703,13 @@ typedef struct {
2448
2703
  * entropyWkspSize must be of size at least ENTROPY_WORKSPACE_SIZE - (MaxSeq + 1)*sizeof(U32)
2449
2704
  */
2450
2705
  static ZSTD_symbolEncodingTypeStats_t
2451
- ZSTD_buildSequencesStatistics(seqStore_t* seqStorePtr, size_t nbSeq,
2452
- const ZSTD_fseCTables_t* prevEntropy, ZSTD_fseCTables_t* nextEntropy,
2453
- BYTE* dst, const BYTE* const dstEnd,
2454
- ZSTD_strategy strategy, unsigned* countWorkspace,
2455
- void* entropyWorkspace, size_t entropyWkspSize) {
2706
+ ZSTD_buildSequencesStatistics(
2707
+ const seqStore_t* seqStorePtr, size_t nbSeq,
2708
+ const ZSTD_fseCTables_t* prevEntropy, ZSTD_fseCTables_t* nextEntropy,
2709
+ BYTE* dst, const BYTE* const dstEnd,
2710
+ ZSTD_strategy strategy, unsigned* countWorkspace,
2711
+ void* entropyWorkspace, size_t entropyWkspSize)
2712
+ {
2456
2713
  BYTE* const ostart = dst;
2457
2714
  const BYTE* const oend = dstEnd;
2458
2715
  BYTE* op = ostart;
@@ -2466,7 +2723,7 @@ ZSTD_buildSequencesStatistics(seqStore_t* seqStorePtr, size_t nbSeq,
2466
2723
 
2467
2724
  stats.lastCountSize = 0;
2468
2725
  /* convert length/distances into codes */
2469
- ZSTD_seqToCodes(seqStorePtr);
2726
+ stats.longOffsets = ZSTD_seqToCodes(seqStorePtr);
2470
2727
  assert(op <= oend);
2471
2728
  assert(nbSeq != 0); /* ZSTD_selectEncodingType() divides by nbSeq */
2472
2729
  /* build CTable for Literal Lengths */
@@ -2571,22 +2828,22 @@ ZSTD_buildSequencesStatistics(seqStore_t* seqStorePtr, size_t nbSeq,
2571
2828
  */
2572
2829
  #define SUSPECT_UNCOMPRESSIBLE_LITERAL_RATIO 20
2573
2830
  MEM_STATIC size_t
2574
- ZSTD_entropyCompressSeqStore_internal(seqStore_t* seqStorePtr,
2575
- const ZSTD_entropyCTables_t* prevEntropy,
2576
- ZSTD_entropyCTables_t* nextEntropy,
2577
- const ZSTD_CCtx_params* cctxParams,
2578
- void* dst, size_t dstCapacity,
2579
- void* entropyWorkspace, size_t entropyWkspSize,
2580
- const int bmi2)
2831
+ ZSTD_entropyCompressSeqStore_internal(
2832
+ const seqStore_t* seqStorePtr,
2833
+ const ZSTD_entropyCTables_t* prevEntropy,
2834
+ ZSTD_entropyCTables_t* nextEntropy,
2835
+ const ZSTD_CCtx_params* cctxParams,
2836
+ void* dst, size_t dstCapacity,
2837
+ void* entropyWorkspace, size_t entropyWkspSize,
2838
+ const int bmi2)
2581
2839
  {
2582
- const int longOffsets = cctxParams->cParams.windowLog > STREAM_ACCUMULATOR_MIN;
2583
2840
  ZSTD_strategy const strategy = cctxParams->cParams.strategy;
2584
2841
  unsigned* count = (unsigned*)entropyWorkspace;
2585
2842
  FSE_CTable* CTable_LitLength = nextEntropy->fse.litlengthCTable;
2586
2843
  FSE_CTable* CTable_OffsetBits = nextEntropy->fse.offcodeCTable;
2587
2844
  FSE_CTable* CTable_MatchLength = nextEntropy->fse.matchlengthCTable;
2588
2845
  const seqDef* const sequences = seqStorePtr->sequencesStart;
2589
- const size_t nbSeq = seqStorePtr->sequences - seqStorePtr->sequencesStart;
2846
+ const size_t nbSeq = (size_t)(seqStorePtr->sequences - seqStorePtr->sequencesStart);
2590
2847
  const BYTE* const ofCodeTable = seqStorePtr->ofCode;
2591
2848
  const BYTE* const llCodeTable = seqStorePtr->llCode;
2592
2849
  const BYTE* const mlCodeTable = seqStorePtr->mlCode;
@@ -2594,29 +2851,31 @@ ZSTD_entropyCompressSeqStore_internal(seqStore_t* seqStorePtr,
2594
2851
  BYTE* const oend = ostart + dstCapacity;
2595
2852
  BYTE* op = ostart;
2596
2853
  size_t lastCountSize;
2854
+ int longOffsets = 0;
2597
2855
 
2598
2856
  entropyWorkspace = count + (MaxSeq + 1);
2599
2857
  entropyWkspSize -= (MaxSeq + 1) * sizeof(*count);
2600
2858
 
2601
- DEBUGLOG(4, "ZSTD_entropyCompressSeqStore_internal (nbSeq=%zu)", nbSeq);
2859
+ DEBUGLOG(5, "ZSTD_entropyCompressSeqStore_internal (nbSeq=%zu, dstCapacity=%zu)", nbSeq, dstCapacity);
2602
2860
  ZSTD_STATIC_ASSERT(HUF_WORKSPACE_SIZE >= (1<<MAX(MLFSELog,LLFSELog)));
2603
2861
  assert(entropyWkspSize >= HUF_WORKSPACE_SIZE);
2604
2862
 
2605
2863
  /* Compress literals */
2606
2864
  { const BYTE* const literals = seqStorePtr->litStart;
2607
- size_t const numSequences = seqStorePtr->sequences - seqStorePtr->sequencesStart;
2608
- size_t const numLiterals = seqStorePtr->lit - seqStorePtr->litStart;
2865
+ size_t const numSequences = (size_t)(seqStorePtr->sequences - seqStorePtr->sequencesStart);
2866
+ size_t const numLiterals = (size_t)(seqStorePtr->lit - seqStorePtr->litStart);
2609
2867
  /* Base suspicion of uncompressibility on ratio of literals to sequences */
2610
2868
  unsigned const suspectUncompressible = (numSequences == 0) || (numLiterals / numSequences >= SUSPECT_UNCOMPRESSIBLE_LITERAL_RATIO);
2611
2869
  size_t const litSize = (size_t)(seqStorePtr->lit - literals);
2870
+
2612
2871
  size_t const cSize = ZSTD_compressLiterals(
2613
- &prevEntropy->huf, &nextEntropy->huf,
2614
- cctxParams->cParams.strategy,
2615
- ZSTD_literalsCompressionIsDisabled(cctxParams),
2616
2872
  op, dstCapacity,
2617
2873
  literals, litSize,
2618
2874
  entropyWorkspace, entropyWkspSize,
2619
- bmi2, suspectUncompressible);
2875
+ &prevEntropy->huf, &nextEntropy->huf,
2876
+ cctxParams->cParams.strategy,
2877
+ ZSTD_literalsCompressionIsDisabled(cctxParams),
2878
+ suspectUncompressible, bmi2);
2620
2879
  FORWARD_IF_ERROR(cSize, "ZSTD_compressLiterals failed");
2621
2880
  assert(cSize <= dstCapacity);
2622
2881
  op += cSize;
@@ -2642,11 +2901,10 @@ ZSTD_entropyCompressSeqStore_internal(seqStore_t* seqStorePtr,
2642
2901
  ZSTD_memcpy(&nextEntropy->fse, &prevEntropy->fse, sizeof(prevEntropy->fse));
2643
2902
  return (size_t)(op - ostart);
2644
2903
  }
2645
- {
2646
- ZSTD_symbolEncodingTypeStats_t stats;
2647
- BYTE* seqHead = op++;
2904
+ { BYTE* const seqHead = op++;
2648
2905
  /* build stats for sequences */
2649
- stats = ZSTD_buildSequencesStatistics(seqStorePtr, nbSeq,
2906
+ const ZSTD_symbolEncodingTypeStats_t stats =
2907
+ ZSTD_buildSequencesStatistics(seqStorePtr, nbSeq,
2650
2908
  &prevEntropy->fse, &nextEntropy->fse,
2651
2909
  op, oend,
2652
2910
  strategy, count,
@@ -2655,6 +2913,7 @@ ZSTD_entropyCompressSeqStore_internal(seqStore_t* seqStorePtr,
2655
2913
  *seqHead = (BYTE)((stats.LLtype<<6) + (stats.Offtype<<4) + (stats.MLtype<<2));
2656
2914
  lastCountSize = stats.lastCountSize;
2657
2915
  op += stats.size;
2916
+ longOffsets = stats.longOffsets;
2658
2917
  }
2659
2918
 
2660
2919
  { size_t const bitstreamSize = ZSTD_encodeSequences(
@@ -2689,14 +2948,15 @@ ZSTD_entropyCompressSeqStore_internal(seqStore_t* seqStorePtr,
2689
2948
  }
2690
2949
 
2691
2950
  MEM_STATIC size_t
2692
- ZSTD_entropyCompressSeqStore(seqStore_t* seqStorePtr,
2693
- const ZSTD_entropyCTables_t* prevEntropy,
2694
- ZSTD_entropyCTables_t* nextEntropy,
2695
- const ZSTD_CCtx_params* cctxParams,
2696
- void* dst, size_t dstCapacity,
2697
- size_t srcSize,
2698
- void* entropyWorkspace, size_t entropyWkspSize,
2699
- int bmi2)
2951
+ ZSTD_entropyCompressSeqStore(
2952
+ const seqStore_t* seqStorePtr,
2953
+ const ZSTD_entropyCTables_t* prevEntropy,
2954
+ ZSTD_entropyCTables_t* nextEntropy,
2955
+ const ZSTD_CCtx_params* cctxParams,
2956
+ void* dst, size_t dstCapacity,
2957
+ size_t srcSize,
2958
+ void* entropyWorkspace, size_t entropyWkspSize,
2959
+ int bmi2)
2700
2960
  {
2701
2961
  size_t const cSize = ZSTD_entropyCompressSeqStore_internal(
2702
2962
  seqStorePtr, prevEntropy, nextEntropy, cctxParams,
@@ -2706,15 +2966,21 @@ ZSTD_entropyCompressSeqStore(seqStore_t* seqStorePtr,
2706
2966
  /* When srcSize <= dstCapacity, there is enough space to write a raw uncompressed block.
2707
2967
  * Since we ran out of space, block must be not compressible, so fall back to raw uncompressed block.
2708
2968
  */
2709
- if ((cSize == ERROR(dstSize_tooSmall)) & (srcSize <= dstCapacity))
2969
+ if ((cSize == ERROR(dstSize_tooSmall)) & (srcSize <= dstCapacity)) {
2970
+ DEBUGLOG(4, "not enough dstCapacity (%zu) for ZSTD_entropyCompressSeqStore_internal()=> do not compress block", dstCapacity);
2710
2971
  return 0; /* block not compressed */
2972
+ }
2711
2973
  FORWARD_IF_ERROR(cSize, "ZSTD_entropyCompressSeqStore_internal failed");
2712
2974
 
2713
2975
  /* Check compressibility */
2714
2976
  { size_t const maxCSize = srcSize - ZSTD_minGain(srcSize, cctxParams->cParams.strategy);
2715
2977
  if (cSize >= maxCSize) return 0; /* block not compressed */
2716
2978
  }
2717
- DEBUGLOG(4, "ZSTD_entropyCompressSeqStore() cSize: %zu", cSize);
2979
+ DEBUGLOG(5, "ZSTD_entropyCompressSeqStore() cSize: %zu", cSize);
2980
+ /* libzstd decoder before > v1.5.4 is not compatible with compressed blocks of size ZSTD_BLOCKSIZE_MAX exactly.
2981
+ * This restriction is indirectly already fulfilled by respecting ZSTD_minGain() condition above.
2982
+ */
2983
+ assert(cSize < ZSTD_BLOCKSIZE_MAX);
2718
2984
  return cSize;
2719
2985
  }
2720
2986
 
@@ -2809,6 +3075,72 @@ void ZSTD_resetSeqStore(seqStore_t* ssPtr)
2809
3075
  ssPtr->longLengthType = ZSTD_llt_none;
2810
3076
  }
2811
3077
 
3078
+ /* ZSTD_postProcessSequenceProducerResult() :
3079
+ * Validates and post-processes sequences obtained through the external matchfinder API:
3080
+ * - Checks whether nbExternalSeqs represents an error condition.
3081
+ * - Appends a block delimiter to outSeqs if one is not already present.
3082
+ * See zstd.h for context regarding block delimiters.
3083
+ * Returns the number of sequences after post-processing, or an error code. */
3084
+ static size_t ZSTD_postProcessSequenceProducerResult(
3085
+ ZSTD_Sequence* outSeqs, size_t nbExternalSeqs, size_t outSeqsCapacity, size_t srcSize
3086
+ ) {
3087
+ RETURN_ERROR_IF(
3088
+ nbExternalSeqs > outSeqsCapacity,
3089
+ sequenceProducer_failed,
3090
+ "External sequence producer returned error code %lu",
3091
+ (unsigned long)nbExternalSeqs
3092
+ );
3093
+
3094
+ RETURN_ERROR_IF(
3095
+ nbExternalSeqs == 0 && srcSize > 0,
3096
+ sequenceProducer_failed,
3097
+ "Got zero sequences from external sequence producer for a non-empty src buffer!"
3098
+ );
3099
+
3100
+ if (srcSize == 0) {
3101
+ ZSTD_memset(&outSeqs[0], 0, sizeof(ZSTD_Sequence));
3102
+ return 1;
3103
+ }
3104
+
3105
+ {
3106
+ ZSTD_Sequence const lastSeq = outSeqs[nbExternalSeqs - 1];
3107
+
3108
+ /* We can return early if lastSeq is already a block delimiter. */
3109
+ if (lastSeq.offset == 0 && lastSeq.matchLength == 0) {
3110
+ return nbExternalSeqs;
3111
+ }
3112
+
3113
+ /* This error condition is only possible if the external matchfinder
3114
+ * produced an invalid parse, by definition of ZSTD_sequenceBound(). */
3115
+ RETURN_ERROR_IF(
3116
+ nbExternalSeqs == outSeqsCapacity,
3117
+ sequenceProducer_failed,
3118
+ "nbExternalSeqs == outSeqsCapacity but lastSeq is not a block delimiter!"
3119
+ );
3120
+
3121
+ /* lastSeq is not a block delimiter, so we need to append one. */
3122
+ ZSTD_memset(&outSeqs[nbExternalSeqs], 0, sizeof(ZSTD_Sequence));
3123
+ return nbExternalSeqs + 1;
3124
+ }
3125
+ }
3126
+
3127
+ /* ZSTD_fastSequenceLengthSum() :
3128
+ * Returns sum(litLen) + sum(matchLen) + lastLits for *seqBuf*.
3129
+ * Similar to another function in zstd_compress.c (determine_blockSize),
3130
+ * except it doesn't check for a block delimiter to end summation.
3131
+ * Removing the early exit allows the compiler to auto-vectorize (https://godbolt.org/z/cY1cajz9P).
3132
+ * This function can be deleted and replaced by determine_blockSize after we resolve issue #3456. */
3133
+ static size_t ZSTD_fastSequenceLengthSum(ZSTD_Sequence const* seqBuf, size_t seqBufSize) {
3134
+ size_t matchLenSum, litLenSum, i;
3135
+ matchLenSum = 0;
3136
+ litLenSum = 0;
3137
+ for (i = 0; i < seqBufSize; i++) {
3138
+ litLenSum += seqBuf[i].litLength;
3139
+ matchLenSum += seqBuf[i].matchLength;
3140
+ }
3141
+ return litLenSum + matchLenSum;
3142
+ }
3143
+
2812
3144
  typedef enum { ZSTDbss_compress, ZSTDbss_noCompress } ZSTD_buildSeqStore_e;
2813
3145
 
2814
3146
  static size_t ZSTD_buildSeqStore(ZSTD_CCtx* zc, const void* src, size_t srcSize)
@@ -2818,7 +3150,9 @@ static size_t ZSTD_buildSeqStore(ZSTD_CCtx* zc, const void* src, size_t srcSize)
2818
3150
  assert(srcSize <= ZSTD_BLOCKSIZE_MAX);
2819
3151
  /* Assert that we have correctly flushed the ctx params into the ms's copy */
2820
3152
  ZSTD_assertEqualCParams(zc->appliedParams.cParams, ms->cParams);
2821
- if (srcSize < MIN_CBLOCK_SIZE+ZSTD_blockHeaderSize+1) {
3153
+ /* TODO: See 3090. We reduced MIN_CBLOCK_SIZE from 3 to 2 so to compensate we are adding
3154
+ * additional 1. We need to revisit and change this logic to be more consistent */
3155
+ if (srcSize < MIN_CBLOCK_SIZE+ZSTD_blockHeaderSize+1+1) {
2822
3156
  if (zc->appliedParams.cParams.strategy >= ZSTD_btopt) {
2823
3157
  ZSTD_ldm_skipRawSeqStoreBytes(&zc->externSeqStore, srcSize);
2824
3158
  } else {
@@ -2854,6 +3188,15 @@ static size_t ZSTD_buildSeqStore(ZSTD_CCtx* zc, const void* src, size_t srcSize)
2854
3188
  }
2855
3189
  if (zc->externSeqStore.pos < zc->externSeqStore.size) {
2856
3190
  assert(zc->appliedParams.ldmParams.enableLdm == ZSTD_ps_disable);
3191
+
3192
+ /* External matchfinder + LDM is technically possible, just not implemented yet.
3193
+ * We need to revisit soon and implement it. */
3194
+ RETURN_ERROR_IF(
3195
+ zc->appliedParams.useSequenceProducer,
3196
+ parameter_combination_unsupported,
3197
+ "Long-distance matching with external sequence producer enabled is not currently supported."
3198
+ );
3199
+
2857
3200
  /* Updates ldmSeqStore.pos */
2858
3201
  lastLLSize =
2859
3202
  ZSTD_ldm_blockCompress(&zc->externSeqStore,
@@ -2865,6 +3208,14 @@ static size_t ZSTD_buildSeqStore(ZSTD_CCtx* zc, const void* src, size_t srcSize)
2865
3208
  } else if (zc->appliedParams.ldmParams.enableLdm == ZSTD_ps_enable) {
2866
3209
  rawSeqStore_t ldmSeqStore = kNullRawSeqStore;
2867
3210
 
3211
+ /* External matchfinder + LDM is technically possible, just not implemented yet.
3212
+ * We need to revisit soon and implement it. */
3213
+ RETURN_ERROR_IF(
3214
+ zc->appliedParams.useSequenceProducer,
3215
+ parameter_combination_unsupported,
3216
+ "Long-distance matching with external sequence producer enabled is not currently supported."
3217
+ );
3218
+
2868
3219
  ldmSeqStore.seq = zc->ldmSequences;
2869
3220
  ldmSeqStore.capacity = zc->maxNbLdmSequences;
2870
3221
  /* Updates ldmSeqStore.size */
@@ -2879,7 +3230,68 @@ static size_t ZSTD_buildSeqStore(ZSTD_CCtx* zc, const void* src, size_t srcSize)
2879
3230
  zc->appliedParams.useRowMatchFinder,
2880
3231
  src, srcSize);
2881
3232
  assert(ldmSeqStore.pos == ldmSeqStore.size);
2882
- } else { /* not long range mode */
3233
+ } else if (zc->appliedParams.useSequenceProducer) {
3234
+ assert(
3235
+ zc->externalMatchCtx.seqBufferCapacity >= ZSTD_sequenceBound(srcSize)
3236
+ );
3237
+ assert(zc->externalMatchCtx.mFinder != NULL);
3238
+
3239
+ { U32 const windowSize = (U32)1 << zc->appliedParams.cParams.windowLog;
3240
+
3241
+ size_t const nbExternalSeqs = (zc->externalMatchCtx.mFinder)(
3242
+ zc->externalMatchCtx.mState,
3243
+ zc->externalMatchCtx.seqBuffer,
3244
+ zc->externalMatchCtx.seqBufferCapacity,
3245
+ src, srcSize,
3246
+ NULL, 0, /* dict and dictSize, currently not supported */
3247
+ zc->appliedParams.compressionLevel,
3248
+ windowSize
3249
+ );
3250
+
3251
+ size_t const nbPostProcessedSeqs = ZSTD_postProcessSequenceProducerResult(
3252
+ zc->externalMatchCtx.seqBuffer,
3253
+ nbExternalSeqs,
3254
+ zc->externalMatchCtx.seqBufferCapacity,
3255
+ srcSize
3256
+ );
3257
+
3258
+ /* Return early if there is no error, since we don't need to worry about last literals */
3259
+ if (!ZSTD_isError(nbPostProcessedSeqs)) {
3260
+ ZSTD_sequencePosition seqPos = {0,0,0};
3261
+ size_t const seqLenSum = ZSTD_fastSequenceLengthSum(zc->externalMatchCtx.seqBuffer, nbPostProcessedSeqs);
3262
+ RETURN_ERROR_IF(seqLenSum > srcSize, externalSequences_invalid, "External sequences imply too large a block!");
3263
+ FORWARD_IF_ERROR(
3264
+ ZSTD_copySequencesToSeqStoreExplicitBlockDelim(
3265
+ zc, &seqPos,
3266
+ zc->externalMatchCtx.seqBuffer, nbPostProcessedSeqs,
3267
+ src, srcSize,
3268
+ zc->appliedParams.searchForExternalRepcodes
3269
+ ),
3270
+ "Failed to copy external sequences to seqStore!"
3271
+ );
3272
+ ms->ldmSeqStore = NULL;
3273
+ DEBUGLOG(5, "Copied %lu sequences from external sequence producer to internal seqStore.", (unsigned long)nbExternalSeqs);
3274
+ return ZSTDbss_compress;
3275
+ }
3276
+
3277
+ /* Propagate the error if fallback is disabled */
3278
+ if (!zc->appliedParams.enableMatchFinderFallback) {
3279
+ return nbPostProcessedSeqs;
3280
+ }
3281
+
3282
+ /* Fallback to software matchfinder */
3283
+ { ZSTD_blockCompressor const blockCompressor = ZSTD_selectBlockCompressor(zc->appliedParams.cParams.strategy,
3284
+ zc->appliedParams.useRowMatchFinder,
3285
+ dictMode);
3286
+ ms->ldmSeqStore = NULL;
3287
+ DEBUGLOG(
3288
+ 5,
3289
+ "External sequence producer returned error code %lu. Falling back to internal parser.",
3290
+ (unsigned long)nbExternalSeqs
3291
+ );
3292
+ lastLLSize = blockCompressor(ms, &zc->seqStore, zc->blockState.nextCBlock->rep, src, srcSize);
3293
+ } }
3294
+ } else { /* not long range mode and no external matchfinder */
2883
3295
  ZSTD_blockCompressor const blockCompressor = ZSTD_selectBlockCompressor(zc->appliedParams.cParams.strategy,
2884
3296
  zc->appliedParams.useRowMatchFinder,
2885
3297
  dictMode);
@@ -2940,7 +3352,7 @@ static void ZSTD_copyBlockSequences(ZSTD_CCtx* zc)
2940
3352
  /* seqStoreSeqs[i].offset == offCode+1, and ZSTD_updateRep() expects offCode
2941
3353
  so we provide seqStoreSeqs[i].offset - 1 */
2942
3354
  ZSTD_updateRep(updatedRepcodes.rep,
2943
- seqStoreSeqs[i].offBase - 1,
3355
+ seqStoreSeqs[i].offBase,
2944
3356
  seqStoreSeqs[i].litLength == 0);
2945
3357
  literalsRead += outSeqs[i].litLength;
2946
3358
  }
@@ -2956,6 +3368,10 @@ static void ZSTD_copyBlockSequences(ZSTD_CCtx* zc)
2956
3368
  zc->seqCollector.seqIndex += seqStoreSeqSize;
2957
3369
  }
2958
3370
 
3371
+ size_t ZSTD_sequenceBound(size_t srcSize) {
3372
+ return (srcSize / ZSTD_MINMATCH_MIN) + 1;
3373
+ }
3374
+
2959
3375
  size_t ZSTD_generateSequences(ZSTD_CCtx* zc, ZSTD_Sequence* outSeqs,
2960
3376
  size_t outSeqsSize, const void* src, size_t srcSize)
2961
3377
  {
@@ -3001,19 +3417,17 @@ static int ZSTD_isRLE(const BYTE* src, size_t length) {
3001
3417
  const size_t unrollMask = unrollSize - 1;
3002
3418
  const size_t prefixLength = length & unrollMask;
3003
3419
  size_t i;
3004
- size_t u;
3005
3420
  if (length == 1) return 1;
3006
3421
  /* Check if prefix is RLE first before using unrolled loop */
3007
3422
  if (prefixLength && ZSTD_count(ip+1, ip, ip+prefixLength) != prefixLength-1) {
3008
3423
  return 0;
3009
3424
  }
3010
3425
  for (i = prefixLength; i != length; i += unrollSize) {
3426
+ size_t u;
3011
3427
  for (u = 0; u < unrollSize; u += sizeof(size_t)) {
3012
3428
  if (MEM_readST(ip + i + u) != valueST) {
3013
3429
  return 0;
3014
- }
3015
- }
3016
- }
3430
+ } } }
3017
3431
  return 1;
3018
3432
  }
3019
3433
 
@@ -3029,7 +3443,8 @@ static int ZSTD_maybeRLE(seqStore_t const* seqStore)
3029
3443
  return nbSeqs < 4 && nbLits < 10;
3030
3444
  }
3031
3445
 
3032
- static void ZSTD_blockState_confirmRepcodesAndEntropyTables(ZSTD_blockState_t* const bs)
3446
+ static void
3447
+ ZSTD_blockState_confirmRepcodesAndEntropyTables(ZSTD_blockState_t* const bs)
3033
3448
  {
3034
3449
  ZSTD_compressedBlockState_t* const tmp = bs->prevCBlock;
3035
3450
  bs->prevCBlock = bs->nextCBlock;
@@ -3037,7 +3452,9 @@ static void ZSTD_blockState_confirmRepcodesAndEntropyTables(ZSTD_blockState_t* c
3037
3452
  }
3038
3453
 
3039
3454
  /* Writes the block header */
3040
- static void writeBlockHeader(void* op, size_t cSize, size_t blockSize, U32 lastBlock) {
3455
+ static void
3456
+ writeBlockHeader(void* op, size_t cSize, size_t blockSize, U32 lastBlock)
3457
+ {
3041
3458
  U32 const cBlockHeader = cSize == 1 ?
3042
3459
  lastBlock + (((U32)bt_rle)<<1) + (U32)(blockSize << 3) :
3043
3460
  lastBlock + (((U32)bt_compressed)<<1) + (U32)(cSize << 3);
@@ -3050,13 +3467,16 @@ static void writeBlockHeader(void* op, size_t cSize, size_t blockSize, U32 lastB
3050
3467
  * Stores literals block type (raw, rle, compressed, repeat) and
3051
3468
  * huffman description table to hufMetadata.
3052
3469
  * Requires ENTROPY_WORKSPACE_SIZE workspace
3053
- * @return : size of huffman description table or error code */
3054
- static size_t ZSTD_buildBlockEntropyStats_literals(void* const src, size_t srcSize,
3055
- const ZSTD_hufCTables_t* prevHuf,
3056
- ZSTD_hufCTables_t* nextHuf,
3057
- ZSTD_hufCTablesMetadata_t* hufMetadata,
3058
- const int literalsCompressionIsDisabled,
3059
- void* workspace, size_t wkspSize)
3470
+ * @return : size of huffman description table, or an error code
3471
+ */
3472
+ static size_t
3473
+ ZSTD_buildBlockEntropyStats_literals(void* const src, size_t srcSize,
3474
+ const ZSTD_hufCTables_t* prevHuf,
3475
+ ZSTD_hufCTables_t* nextHuf,
3476
+ ZSTD_hufCTablesMetadata_t* hufMetadata,
3477
+ const int literalsCompressionIsDisabled,
3478
+ void* workspace, size_t wkspSize,
3479
+ int hufFlags)
3060
3480
  {
3061
3481
  BYTE* const wkspStart = (BYTE*)workspace;
3062
3482
  BYTE* const wkspEnd = wkspStart + wkspSize;
@@ -3064,9 +3484,9 @@ static size_t ZSTD_buildBlockEntropyStats_literals(void* const src, size_t srcSi
3064
3484
  unsigned* const countWksp = (unsigned*)workspace;
3065
3485
  const size_t countWkspSize = (HUF_SYMBOLVALUE_MAX + 1) * sizeof(unsigned);
3066
3486
  BYTE* const nodeWksp = countWkspStart + countWkspSize;
3067
- const size_t nodeWkspSize = wkspEnd-nodeWksp;
3487
+ const size_t nodeWkspSize = (size_t)(wkspEnd - nodeWksp);
3068
3488
  unsigned maxSymbolValue = HUF_SYMBOLVALUE_MAX;
3069
- unsigned huffLog = HUF_TABLELOG_DEFAULT;
3489
+ unsigned huffLog = LitHufLog;
3070
3490
  HUF_repeat repeat = prevHuf->repeatMode;
3071
3491
  DEBUGLOG(5, "ZSTD_buildBlockEntropyStats_literals (srcSize=%zu)", srcSize);
3072
3492
 
@@ -3081,73 +3501,77 @@ static size_t ZSTD_buildBlockEntropyStats_literals(void* const src, size_t srcSi
3081
3501
 
3082
3502
  /* small ? don't even attempt compression (speed opt) */
3083
3503
  #ifndef COMPRESS_LITERALS_SIZE_MIN
3084
- #define COMPRESS_LITERALS_SIZE_MIN 63
3504
+ # define COMPRESS_LITERALS_SIZE_MIN 63 /* heuristic */
3085
3505
  #endif
3086
3506
  { size_t const minLitSize = (prevHuf->repeatMode == HUF_repeat_valid) ? 6 : COMPRESS_LITERALS_SIZE_MIN;
3087
3507
  if (srcSize <= minLitSize) {
3088
3508
  DEBUGLOG(5, "set_basic - too small");
3089
3509
  hufMetadata->hType = set_basic;
3090
3510
  return 0;
3091
- }
3092
- }
3511
+ } }
3093
3512
 
3094
3513
  /* Scan input and build symbol stats */
3095
- { size_t const largest = HIST_count_wksp (countWksp, &maxSymbolValue, (const BYTE*)src, srcSize, workspace, wkspSize);
3514
+ { size_t const largest =
3515
+ HIST_count_wksp (countWksp, &maxSymbolValue,
3516
+ (const BYTE*)src, srcSize,
3517
+ workspace, wkspSize);
3096
3518
  FORWARD_IF_ERROR(largest, "HIST_count_wksp failed");
3097
3519
  if (largest == srcSize) {
3520
+ /* only one literal symbol */
3098
3521
  DEBUGLOG(5, "set_rle");
3099
3522
  hufMetadata->hType = set_rle;
3100
3523
  return 0;
3101
3524
  }
3102
3525
  if (largest <= (srcSize >> 7)+4) {
3526
+ /* heuristic: likely not compressible */
3103
3527
  DEBUGLOG(5, "set_basic - no gain");
3104
3528
  hufMetadata->hType = set_basic;
3105
3529
  return 0;
3106
- }
3107
- }
3530
+ } }
3108
3531
 
3109
3532
  /* Validate the previous Huffman table */
3110
- if (repeat == HUF_repeat_check && !HUF_validateCTable((HUF_CElt const*)prevHuf->CTable, countWksp, maxSymbolValue)) {
3533
+ if (repeat == HUF_repeat_check
3534
+ && !HUF_validateCTable((HUF_CElt const*)prevHuf->CTable, countWksp, maxSymbolValue)) {
3111
3535
  repeat = HUF_repeat_none;
3112
3536
  }
3113
3537
 
3114
3538
  /* Build Huffman Tree */
3115
3539
  ZSTD_memset(nextHuf->CTable, 0, sizeof(nextHuf->CTable));
3116
- huffLog = HUF_optimalTableLog(huffLog, srcSize, maxSymbolValue);
3540
+ huffLog = HUF_optimalTableLog(huffLog, srcSize, maxSymbolValue, nodeWksp, nodeWkspSize, nextHuf->CTable, countWksp, hufFlags);
3541
+ assert(huffLog <= LitHufLog);
3117
3542
  { size_t const maxBits = HUF_buildCTable_wksp((HUF_CElt*)nextHuf->CTable, countWksp,
3118
3543
  maxSymbolValue, huffLog,
3119
3544
  nodeWksp, nodeWkspSize);
3120
3545
  FORWARD_IF_ERROR(maxBits, "HUF_buildCTable_wksp");
3121
3546
  huffLog = (U32)maxBits;
3122
- { /* Build and write the CTable */
3123
- size_t const newCSize = HUF_estimateCompressedSize(
3124
- (HUF_CElt*)nextHuf->CTable, countWksp, maxSymbolValue);
3125
- size_t const hSize = HUF_writeCTable_wksp(
3126
- hufMetadata->hufDesBuffer, sizeof(hufMetadata->hufDesBuffer),
3127
- (HUF_CElt*)nextHuf->CTable, maxSymbolValue, huffLog,
3128
- nodeWksp, nodeWkspSize);
3129
- /* Check against repeating the previous CTable */
3130
- if (repeat != HUF_repeat_none) {
3131
- size_t const oldCSize = HUF_estimateCompressedSize(
3132
- (HUF_CElt const*)prevHuf->CTable, countWksp, maxSymbolValue);
3133
- if (oldCSize < srcSize && (oldCSize <= hSize + newCSize || hSize + 12 >= srcSize)) {
3134
- DEBUGLOG(5, "set_repeat - smaller");
3135
- ZSTD_memcpy(nextHuf, prevHuf, sizeof(*prevHuf));
3136
- hufMetadata->hType = set_repeat;
3137
- return 0;
3138
- }
3139
- }
3140
- if (newCSize + hSize >= srcSize) {
3141
- DEBUGLOG(5, "set_basic - no gains");
3547
+ }
3548
+ { /* Build and write the CTable */
3549
+ size_t const newCSize = HUF_estimateCompressedSize(
3550
+ (HUF_CElt*)nextHuf->CTable, countWksp, maxSymbolValue);
3551
+ size_t const hSize = HUF_writeCTable_wksp(
3552
+ hufMetadata->hufDesBuffer, sizeof(hufMetadata->hufDesBuffer),
3553
+ (HUF_CElt*)nextHuf->CTable, maxSymbolValue, huffLog,
3554
+ nodeWksp, nodeWkspSize);
3555
+ /* Check against repeating the previous CTable */
3556
+ if (repeat != HUF_repeat_none) {
3557
+ size_t const oldCSize = HUF_estimateCompressedSize(
3558
+ (HUF_CElt const*)prevHuf->CTable, countWksp, maxSymbolValue);
3559
+ if (oldCSize < srcSize && (oldCSize <= hSize + newCSize || hSize + 12 >= srcSize)) {
3560
+ DEBUGLOG(5, "set_repeat - smaller");
3142
3561
  ZSTD_memcpy(nextHuf, prevHuf, sizeof(*prevHuf));
3143
- hufMetadata->hType = set_basic;
3562
+ hufMetadata->hType = set_repeat;
3144
3563
  return 0;
3145
- }
3146
- DEBUGLOG(5, "set_compressed (hSize=%u)", (U32)hSize);
3147
- hufMetadata->hType = set_compressed;
3148
- nextHuf->repeatMode = HUF_repeat_check;
3149
- return hSize;
3564
+ } }
3565
+ if (newCSize + hSize >= srcSize) {
3566
+ DEBUGLOG(5, "set_basic - no gains");
3567
+ ZSTD_memcpy(nextHuf, prevHuf, sizeof(*prevHuf));
3568
+ hufMetadata->hType = set_basic;
3569
+ return 0;
3150
3570
  }
3571
+ DEBUGLOG(5, "set_compressed (hSize=%u)", (U32)hSize);
3572
+ hufMetadata->hType = set_compressed;
3573
+ nextHuf->repeatMode = HUF_repeat_check;
3574
+ return hSize;
3151
3575
  }
3152
3576
  }
3153
3577
 
@@ -3157,8 +3581,9 @@ static size_t ZSTD_buildBlockEntropyStats_literals(void* const src, size_t srcSi
3157
3581
  * and updates nextEntropy to the appropriate repeatMode.
3158
3582
  */
3159
3583
  static ZSTD_symbolEncodingTypeStats_t
3160
- ZSTD_buildDummySequencesStatistics(ZSTD_fseCTables_t* nextEntropy) {
3161
- ZSTD_symbolEncodingTypeStats_t stats = {set_basic, set_basic, set_basic, 0, 0};
3584
+ ZSTD_buildDummySequencesStatistics(ZSTD_fseCTables_t* nextEntropy)
3585
+ {
3586
+ ZSTD_symbolEncodingTypeStats_t stats = {set_basic, set_basic, set_basic, 0, 0, 0};
3162
3587
  nextEntropy->litlength_repeatMode = FSE_repeat_none;
3163
3588
  nextEntropy->offcode_repeatMode = FSE_repeat_none;
3164
3589
  nextEntropy->matchlength_repeatMode = FSE_repeat_none;
@@ -3169,16 +3594,18 @@ ZSTD_buildDummySequencesStatistics(ZSTD_fseCTables_t* nextEntropy) {
3169
3594
  * Builds entropy for the sequences.
3170
3595
  * Stores symbol compression modes and fse table to fseMetadata.
3171
3596
  * Requires ENTROPY_WORKSPACE_SIZE wksp.
3172
- * @return : size of fse tables or error code */
3173
- static size_t ZSTD_buildBlockEntropyStats_sequences(seqStore_t* seqStorePtr,
3174
- const ZSTD_fseCTables_t* prevEntropy,
3175
- ZSTD_fseCTables_t* nextEntropy,
3176
- const ZSTD_CCtx_params* cctxParams,
3177
- ZSTD_fseCTablesMetadata_t* fseMetadata,
3178
- void* workspace, size_t wkspSize)
3597
+ * @return : size of fse tables or error code */
3598
+ static size_t
3599
+ ZSTD_buildBlockEntropyStats_sequences(
3600
+ const seqStore_t* seqStorePtr,
3601
+ const ZSTD_fseCTables_t* prevEntropy,
3602
+ ZSTD_fseCTables_t* nextEntropy,
3603
+ const ZSTD_CCtx_params* cctxParams,
3604
+ ZSTD_fseCTablesMetadata_t* fseMetadata,
3605
+ void* workspace, size_t wkspSize)
3179
3606
  {
3180
3607
  ZSTD_strategy const strategy = cctxParams->cParams.strategy;
3181
- size_t const nbSeq = seqStorePtr->sequences - seqStorePtr->sequencesStart;
3608
+ size_t const nbSeq = (size_t)(seqStorePtr->sequences - seqStorePtr->sequencesStart);
3182
3609
  BYTE* const ostart = fseMetadata->fseTablesBuffer;
3183
3610
  BYTE* const oend = ostart + sizeof(fseMetadata->fseTablesBuffer);
3184
3611
  BYTE* op = ostart;
@@ -3205,23 +3632,28 @@ static size_t ZSTD_buildBlockEntropyStats_sequences(seqStore_t* seqStorePtr,
3205
3632
  /** ZSTD_buildBlockEntropyStats() :
3206
3633
  * Builds entropy for the block.
3207
3634
  * Requires workspace size ENTROPY_WORKSPACE_SIZE
3208
- *
3209
- * @return : 0 on success or error code
3635
+ * @return : 0 on success, or an error code
3636
+ * Note : also employed in superblock
3210
3637
  */
3211
- size_t ZSTD_buildBlockEntropyStats(seqStore_t* seqStorePtr,
3212
- const ZSTD_entropyCTables_t* prevEntropy,
3213
- ZSTD_entropyCTables_t* nextEntropy,
3214
- const ZSTD_CCtx_params* cctxParams,
3215
- ZSTD_entropyCTablesMetadata_t* entropyMetadata,
3216
- void* workspace, size_t wkspSize)
3217
- {
3218
- size_t const litSize = seqStorePtr->lit - seqStorePtr->litStart;
3638
+ size_t ZSTD_buildBlockEntropyStats(
3639
+ const seqStore_t* seqStorePtr,
3640
+ const ZSTD_entropyCTables_t* prevEntropy,
3641
+ ZSTD_entropyCTables_t* nextEntropy,
3642
+ const ZSTD_CCtx_params* cctxParams,
3643
+ ZSTD_entropyCTablesMetadata_t* entropyMetadata,
3644
+ void* workspace, size_t wkspSize)
3645
+ {
3646
+ size_t const litSize = (size_t)(seqStorePtr->lit - seqStorePtr->litStart);
3647
+ int const huf_useOptDepth = (cctxParams->cParams.strategy >= HUF_OPTIMAL_DEPTH_THRESHOLD);
3648
+ int const hufFlags = huf_useOptDepth ? HUF_flags_optimalDepth : 0;
3649
+
3219
3650
  entropyMetadata->hufMetadata.hufDesSize =
3220
3651
  ZSTD_buildBlockEntropyStats_literals(seqStorePtr->litStart, litSize,
3221
3652
  &prevEntropy->huf, &nextEntropy->huf,
3222
3653
  &entropyMetadata->hufMetadata,
3223
3654
  ZSTD_literalsCompressionIsDisabled(cctxParams),
3224
- workspace, wkspSize);
3655
+ workspace, wkspSize, hufFlags);
3656
+
3225
3657
  FORWARD_IF_ERROR(entropyMetadata->hufMetadata.hufDesSize, "ZSTD_buildBlockEntropyStats_literals failed");
3226
3658
  entropyMetadata->fseMetadata.fseTablesSize =
3227
3659
  ZSTD_buildBlockEntropyStats_sequences(seqStorePtr,
@@ -3234,11 +3666,12 @@ size_t ZSTD_buildBlockEntropyStats(seqStore_t* seqStorePtr,
3234
3666
  }
3235
3667
 
3236
3668
  /* Returns the size estimate for the literals section (header + content) of a block */
3237
- static size_t ZSTD_estimateBlockSize_literal(const BYTE* literals, size_t litSize,
3238
- const ZSTD_hufCTables_t* huf,
3239
- const ZSTD_hufCTablesMetadata_t* hufMetadata,
3240
- void* workspace, size_t wkspSize,
3241
- int writeEntropy)
3669
+ static size_t
3670
+ ZSTD_estimateBlockSize_literal(const BYTE* literals, size_t litSize,
3671
+ const ZSTD_hufCTables_t* huf,
3672
+ const ZSTD_hufCTablesMetadata_t* hufMetadata,
3673
+ void* workspace, size_t wkspSize,
3674
+ int writeEntropy)
3242
3675
  {
3243
3676
  unsigned* const countWksp = (unsigned*)workspace;
3244
3677
  unsigned maxSymbolValue = HUF_SYMBOLVALUE_MAX;
@@ -3260,12 +3693,13 @@ static size_t ZSTD_estimateBlockSize_literal(const BYTE* literals, size_t litSiz
3260
3693
  }
3261
3694
 
3262
3695
  /* Returns the size estimate for the FSE-compressed symbols (of, ml, ll) of a block */
3263
- static size_t ZSTD_estimateBlockSize_symbolType(symbolEncodingType_e type,
3264
- const BYTE* codeTable, size_t nbSeq, unsigned maxCode,
3265
- const FSE_CTable* fseCTable,
3266
- const U8* additionalBits,
3267
- short const* defaultNorm, U32 defaultNormLog, U32 defaultMax,
3268
- void* workspace, size_t wkspSize)
3696
+ static size_t
3697
+ ZSTD_estimateBlockSize_symbolType(symbolEncodingType_e type,
3698
+ const BYTE* codeTable, size_t nbSeq, unsigned maxCode,
3699
+ const FSE_CTable* fseCTable,
3700
+ const U8* additionalBits,
3701
+ short const* defaultNorm, U32 defaultNormLog, U32 defaultMax,
3702
+ void* workspace, size_t wkspSize)
3269
3703
  {
3270
3704
  unsigned* const countWksp = (unsigned*)workspace;
3271
3705
  const BYTE* ctp = codeTable;
@@ -3297,99 +3731,107 @@ static size_t ZSTD_estimateBlockSize_symbolType(symbolEncodingType_e type,
3297
3731
  }
3298
3732
 
3299
3733
  /* Returns the size estimate for the sequences section (header + content) of a block */
3300
- static size_t ZSTD_estimateBlockSize_sequences(const BYTE* ofCodeTable,
3301
- const BYTE* llCodeTable,
3302
- const BYTE* mlCodeTable,
3303
- size_t nbSeq,
3304
- const ZSTD_fseCTables_t* fseTables,
3305
- const ZSTD_fseCTablesMetadata_t* fseMetadata,
3306
- void* workspace, size_t wkspSize,
3307
- int writeEntropy)
3734
+ static size_t
3735
+ ZSTD_estimateBlockSize_sequences(const BYTE* ofCodeTable,
3736
+ const BYTE* llCodeTable,
3737
+ const BYTE* mlCodeTable,
3738
+ size_t nbSeq,
3739
+ const ZSTD_fseCTables_t* fseTables,
3740
+ const ZSTD_fseCTablesMetadata_t* fseMetadata,
3741
+ void* workspace, size_t wkspSize,
3742
+ int writeEntropy)
3308
3743
  {
3309
3744
  size_t sequencesSectionHeaderSize = 1 /* seqHead */ + 1 /* min seqSize size */ + (nbSeq >= 128) + (nbSeq >= LONGNBSEQ);
3310
3745
  size_t cSeqSizeEstimate = 0;
3311
3746
  cSeqSizeEstimate += ZSTD_estimateBlockSize_symbolType(fseMetadata->ofType, ofCodeTable, nbSeq, MaxOff,
3312
- fseTables->offcodeCTable, NULL,
3313
- OF_defaultNorm, OF_defaultNormLog, DefaultMaxOff,
3314
- workspace, wkspSize);
3747
+ fseTables->offcodeCTable, NULL,
3748
+ OF_defaultNorm, OF_defaultNormLog, DefaultMaxOff,
3749
+ workspace, wkspSize);
3315
3750
  cSeqSizeEstimate += ZSTD_estimateBlockSize_symbolType(fseMetadata->llType, llCodeTable, nbSeq, MaxLL,
3316
- fseTables->litlengthCTable, LL_bits,
3317
- LL_defaultNorm, LL_defaultNormLog, MaxLL,
3318
- workspace, wkspSize);
3751
+ fseTables->litlengthCTable, LL_bits,
3752
+ LL_defaultNorm, LL_defaultNormLog, MaxLL,
3753
+ workspace, wkspSize);
3319
3754
  cSeqSizeEstimate += ZSTD_estimateBlockSize_symbolType(fseMetadata->mlType, mlCodeTable, nbSeq, MaxML,
3320
- fseTables->matchlengthCTable, ML_bits,
3321
- ML_defaultNorm, ML_defaultNormLog, MaxML,
3322
- workspace, wkspSize);
3755
+ fseTables->matchlengthCTable, ML_bits,
3756
+ ML_defaultNorm, ML_defaultNormLog, MaxML,
3757
+ workspace, wkspSize);
3323
3758
  if (writeEntropy) cSeqSizeEstimate += fseMetadata->fseTablesSize;
3324
3759
  return cSeqSizeEstimate + sequencesSectionHeaderSize;
3325
3760
  }
3326
3761
 
3327
3762
  /* Returns the size estimate for a given stream of literals, of, ll, ml */
3328
- static size_t ZSTD_estimateBlockSize(const BYTE* literals, size_t litSize,
3329
- const BYTE* ofCodeTable,
3330
- const BYTE* llCodeTable,
3331
- const BYTE* mlCodeTable,
3332
- size_t nbSeq,
3333
- const ZSTD_entropyCTables_t* entropy,
3334
- const ZSTD_entropyCTablesMetadata_t* entropyMetadata,
3335
- void* workspace, size_t wkspSize,
3336
- int writeLitEntropy, int writeSeqEntropy) {
3763
+ static size_t
3764
+ ZSTD_estimateBlockSize(const BYTE* literals, size_t litSize,
3765
+ const BYTE* ofCodeTable,
3766
+ const BYTE* llCodeTable,
3767
+ const BYTE* mlCodeTable,
3768
+ size_t nbSeq,
3769
+ const ZSTD_entropyCTables_t* entropy,
3770
+ const ZSTD_entropyCTablesMetadata_t* entropyMetadata,
3771
+ void* workspace, size_t wkspSize,
3772
+ int writeLitEntropy, int writeSeqEntropy)
3773
+ {
3337
3774
  size_t const literalsSize = ZSTD_estimateBlockSize_literal(literals, litSize,
3338
- &entropy->huf, &entropyMetadata->hufMetadata,
3339
- workspace, wkspSize, writeLitEntropy);
3775
+ &entropy->huf, &entropyMetadata->hufMetadata,
3776
+ workspace, wkspSize, writeLitEntropy);
3340
3777
  size_t const seqSize = ZSTD_estimateBlockSize_sequences(ofCodeTable, llCodeTable, mlCodeTable,
3341
- nbSeq, &entropy->fse, &entropyMetadata->fseMetadata,
3342
- workspace, wkspSize, writeSeqEntropy);
3778
+ nbSeq, &entropy->fse, &entropyMetadata->fseMetadata,
3779
+ workspace, wkspSize, writeSeqEntropy);
3343
3780
  return seqSize + literalsSize + ZSTD_blockHeaderSize;
3344
3781
  }
3345
3782
 
3346
3783
  /* Builds entropy statistics and uses them for blocksize estimation.
3347
3784
  *
3348
- * Returns the estimated compressed size of the seqStore, or a zstd error.
3785
+ * @return: estimated compressed size of the seqStore, or a zstd error.
3349
3786
  */
3350
- static size_t ZSTD_buildEntropyStatisticsAndEstimateSubBlockSize(seqStore_t* seqStore, ZSTD_CCtx* zc) {
3351
- ZSTD_entropyCTablesMetadata_t* entropyMetadata = &zc->blockSplitCtx.entropyMetadata;
3787
+ static size_t
3788
+ ZSTD_buildEntropyStatisticsAndEstimateSubBlockSize(seqStore_t* seqStore, ZSTD_CCtx* zc)
3789
+ {
3790
+ ZSTD_entropyCTablesMetadata_t* const entropyMetadata = &zc->blockSplitCtx.entropyMetadata;
3352
3791
  DEBUGLOG(6, "ZSTD_buildEntropyStatisticsAndEstimateSubBlockSize()");
3353
3792
  FORWARD_IF_ERROR(ZSTD_buildBlockEntropyStats(seqStore,
3354
3793
  &zc->blockState.prevCBlock->entropy,
3355
3794
  &zc->blockState.nextCBlock->entropy,
3356
3795
  &zc->appliedParams,
3357
3796
  entropyMetadata,
3358
- zc->entropyWorkspace, ENTROPY_WORKSPACE_SIZE /* statically allocated in resetCCtx */), "");
3359
- return ZSTD_estimateBlockSize(seqStore->litStart, (size_t)(seqStore->lit - seqStore->litStart),
3797
+ zc->entropyWorkspace, ENTROPY_WORKSPACE_SIZE), "");
3798
+ return ZSTD_estimateBlockSize(
3799
+ seqStore->litStart, (size_t)(seqStore->lit - seqStore->litStart),
3360
3800
  seqStore->ofCode, seqStore->llCode, seqStore->mlCode,
3361
3801
  (size_t)(seqStore->sequences - seqStore->sequencesStart),
3362
- &zc->blockState.nextCBlock->entropy, entropyMetadata, zc->entropyWorkspace, ENTROPY_WORKSPACE_SIZE,
3802
+ &zc->blockState.nextCBlock->entropy,
3803
+ entropyMetadata,
3804
+ zc->entropyWorkspace, ENTROPY_WORKSPACE_SIZE,
3363
3805
  (int)(entropyMetadata->hufMetadata.hType == set_compressed), 1);
3364
3806
  }
3365
3807
 
3366
3808
  /* Returns literals bytes represented in a seqStore */
3367
- static size_t ZSTD_countSeqStoreLiteralsBytes(const seqStore_t* const seqStore) {
3809
+ static size_t ZSTD_countSeqStoreLiteralsBytes(const seqStore_t* const seqStore)
3810
+ {
3368
3811
  size_t literalsBytes = 0;
3369
- size_t const nbSeqs = seqStore->sequences - seqStore->sequencesStart;
3812
+ size_t const nbSeqs = (size_t)(seqStore->sequences - seqStore->sequencesStart);
3370
3813
  size_t i;
3371
3814
  for (i = 0; i < nbSeqs; ++i) {
3372
- seqDef seq = seqStore->sequencesStart[i];
3815
+ seqDef const seq = seqStore->sequencesStart[i];
3373
3816
  literalsBytes += seq.litLength;
3374
3817
  if (i == seqStore->longLengthPos && seqStore->longLengthType == ZSTD_llt_literalLength) {
3375
3818
  literalsBytes += 0x10000;
3376
- }
3377
- }
3819
+ } }
3378
3820
  return literalsBytes;
3379
3821
  }
3380
3822
 
3381
3823
  /* Returns match bytes represented in a seqStore */
3382
- static size_t ZSTD_countSeqStoreMatchBytes(const seqStore_t* const seqStore) {
3824
+ static size_t ZSTD_countSeqStoreMatchBytes(const seqStore_t* const seqStore)
3825
+ {
3383
3826
  size_t matchBytes = 0;
3384
- size_t const nbSeqs = seqStore->sequences - seqStore->sequencesStart;
3827
+ size_t const nbSeqs = (size_t)(seqStore->sequences - seqStore->sequencesStart);
3385
3828
  size_t i;
3386
3829
  for (i = 0; i < nbSeqs; ++i) {
3387
3830
  seqDef seq = seqStore->sequencesStart[i];
3388
3831
  matchBytes += seq.mlBase + MINMATCH;
3389
3832
  if (i == seqStore->longLengthPos && seqStore->longLengthType == ZSTD_llt_matchLength) {
3390
3833
  matchBytes += 0x10000;
3391
- }
3392
- }
3834
+ } }
3393
3835
  return matchBytes;
3394
3836
  }
3395
3837
 
@@ -3398,15 +3840,12 @@ static size_t ZSTD_countSeqStoreMatchBytes(const seqStore_t* const seqStore) {
3398
3840
  */
3399
3841
  static void ZSTD_deriveSeqStoreChunk(seqStore_t* resultSeqStore,
3400
3842
  const seqStore_t* originalSeqStore,
3401
- size_t startIdx, size_t endIdx) {
3402
- BYTE* const litEnd = originalSeqStore->lit;
3403
- size_t literalsBytes;
3404
- size_t literalsBytesPreceding = 0;
3405
-
3843
+ size_t startIdx, size_t endIdx)
3844
+ {
3406
3845
  *resultSeqStore = *originalSeqStore;
3407
3846
  if (startIdx > 0) {
3408
3847
  resultSeqStore->sequences = originalSeqStore->sequencesStart + startIdx;
3409
- literalsBytesPreceding = ZSTD_countSeqStoreLiteralsBytes(resultSeqStore);
3848
+ resultSeqStore->litStart += ZSTD_countSeqStoreLiteralsBytes(resultSeqStore);
3410
3849
  }
3411
3850
 
3412
3851
  /* Move longLengthPos into the correct position if necessary */
@@ -3419,13 +3858,12 @@ static void ZSTD_deriveSeqStoreChunk(seqStore_t* resultSeqStore,
3419
3858
  }
3420
3859
  resultSeqStore->sequencesStart = originalSeqStore->sequencesStart + startIdx;
3421
3860
  resultSeqStore->sequences = originalSeqStore->sequencesStart + endIdx;
3422
- literalsBytes = ZSTD_countSeqStoreLiteralsBytes(resultSeqStore);
3423
- resultSeqStore->litStart += literalsBytesPreceding;
3424
3861
  if (endIdx == (size_t)(originalSeqStore->sequences - originalSeqStore->sequencesStart)) {
3425
3862
  /* This accounts for possible last literals if the derived chunk reaches the end of the block */
3426
- resultSeqStore->lit = litEnd;
3863
+ assert(resultSeqStore->lit == originalSeqStore->lit);
3427
3864
  } else {
3428
- resultSeqStore->lit = resultSeqStore->litStart+literalsBytes;
3865
+ size_t const literalsBytes = ZSTD_countSeqStoreLiteralsBytes(resultSeqStore);
3866
+ resultSeqStore->lit = resultSeqStore->litStart + literalsBytes;
3429
3867
  }
3430
3868
  resultSeqStore->llCode += startIdx;
3431
3869
  resultSeqStore->mlCode += startIdx;
@@ -3433,20 +3871,26 @@ static void ZSTD_deriveSeqStoreChunk(seqStore_t* resultSeqStore,
3433
3871
  }
3434
3872
 
3435
3873
  /**
3436
- * Returns the raw offset represented by the combination of offCode, ll0, and repcode history.
3437
- * offCode must represent a repcode in the numeric representation of ZSTD_storeSeq().
3874
+ * Returns the raw offset represented by the combination of offBase, ll0, and repcode history.
3875
+ * offBase must represent a repcode in the numeric representation of ZSTD_storeSeq().
3438
3876
  */
3439
3877
  static U32
3440
- ZSTD_resolveRepcodeToRawOffset(const U32 rep[ZSTD_REP_NUM], const U32 offCode, const U32 ll0)
3441
- {
3442
- U32 const adjustedOffCode = STORED_REPCODE(offCode) - 1 + ll0; /* [ 0 - 3 ] */
3443
- assert(STORED_IS_REPCODE(offCode));
3444
- if (adjustedOffCode == ZSTD_REP_NUM) {
3445
- /* litlength == 0 and offCode == 2 implies selection of first repcode - 1 */
3446
- assert(rep[0] > 0);
3878
+ ZSTD_resolveRepcodeToRawOffset(const U32 rep[ZSTD_REP_NUM], const U32 offBase, const U32 ll0)
3879
+ {
3880
+ U32 const adjustedRepCode = OFFBASE_TO_REPCODE(offBase) - 1 + ll0; /* [ 0 - 3 ] */
3881
+ assert(OFFBASE_IS_REPCODE(offBase));
3882
+ if (adjustedRepCode == ZSTD_REP_NUM) {
3883
+ assert(ll0);
3884
+ /* litlength == 0 and offCode == 2 implies selection of first repcode - 1
3885
+ * This is only valid if it results in a valid offset value, aka > 0.
3886
+ * Note : it may happen that `rep[0]==1` in exceptional circumstances.
3887
+ * In which case this function will return 0, which is an invalid offset.
3888
+ * It's not an issue though, since this value will be
3889
+ * compared and discarded within ZSTD_seqStore_resolveOffCodes().
3890
+ */
3447
3891
  return rep[0] - 1;
3448
3892
  }
3449
- return rep[adjustedOffCode];
3893
+ return rep[adjustedRepCode];
3450
3894
  }
3451
3895
 
3452
3896
  /**
@@ -3462,30 +3906,33 @@ ZSTD_resolveRepcodeToRawOffset(const U32 rep[ZSTD_REP_NUM], const U32 offCode, c
3462
3906
  * 1-3 : repcode 1-3
3463
3907
  * 4+ : real_offset+3
3464
3908
  */
3465
- static void ZSTD_seqStore_resolveOffCodes(repcodes_t* const dRepcodes, repcodes_t* const cRepcodes,
3466
- seqStore_t* const seqStore, U32 const nbSeq) {
3909
+ static void
3910
+ ZSTD_seqStore_resolveOffCodes(repcodes_t* const dRepcodes, repcodes_t* const cRepcodes,
3911
+ const seqStore_t* const seqStore, U32 const nbSeq)
3912
+ {
3467
3913
  U32 idx = 0;
3914
+ U32 const longLitLenIdx = seqStore->longLengthType == ZSTD_llt_literalLength ? seqStore->longLengthPos : nbSeq;
3468
3915
  for (; idx < nbSeq; ++idx) {
3469
3916
  seqDef* const seq = seqStore->sequencesStart + idx;
3470
- U32 const ll0 = (seq->litLength == 0);
3471
- U32 const offCode = OFFBASE_TO_STORED(seq->offBase);
3472
- assert(seq->offBase > 0);
3473
- if (STORED_IS_REPCODE(offCode)) {
3474
- U32 const dRawOffset = ZSTD_resolveRepcodeToRawOffset(dRepcodes->rep, offCode, ll0);
3475
- U32 const cRawOffset = ZSTD_resolveRepcodeToRawOffset(cRepcodes->rep, offCode, ll0);
3917
+ U32 const ll0 = (seq->litLength == 0) && (idx != longLitLenIdx);
3918
+ U32 const offBase = seq->offBase;
3919
+ assert(offBase > 0);
3920
+ if (OFFBASE_IS_REPCODE(offBase)) {
3921
+ U32 const dRawOffset = ZSTD_resolveRepcodeToRawOffset(dRepcodes->rep, offBase, ll0);
3922
+ U32 const cRawOffset = ZSTD_resolveRepcodeToRawOffset(cRepcodes->rep, offBase, ll0);
3476
3923
  /* Adjust simulated decompression repcode history if we come across a mismatch. Replace
3477
3924
  * the repcode with the offset it actually references, determined by the compression
3478
3925
  * repcode history.
3479
3926
  */
3480
3927
  if (dRawOffset != cRawOffset) {
3481
- seq->offBase = cRawOffset + ZSTD_REP_NUM;
3928
+ seq->offBase = OFFSET_TO_OFFBASE(cRawOffset);
3482
3929
  }
3483
3930
  }
3484
3931
  /* Compression repcode history is always updated with values directly from the unmodified seqStore.
3485
3932
  * Decompression repcode history may use modified seq->offset value taken from compression repcode history.
3486
3933
  */
3487
- ZSTD_updateRep(dRepcodes->rep, OFFBASE_TO_STORED(seq->offBase), ll0);
3488
- ZSTD_updateRep(cRepcodes->rep, offCode, ll0);
3934
+ ZSTD_updateRep(dRepcodes->rep, seq->offBase, ll0);
3935
+ ZSTD_updateRep(cRepcodes->rep, offBase, ll0);
3489
3936
  }
3490
3937
  }
3491
3938
 
@@ -3495,10 +3942,11 @@ static void ZSTD_seqStore_resolveOffCodes(repcodes_t* const dRepcodes, repcodes_
3495
3942
  * Returns the total size of that block (including header) or a ZSTD error code.
3496
3943
  */
3497
3944
  static size_t
3498
- ZSTD_compressSeqStore_singleBlock(ZSTD_CCtx* zc, seqStore_t* const seqStore,
3945
+ ZSTD_compressSeqStore_singleBlock(ZSTD_CCtx* zc,
3946
+ const seqStore_t* const seqStore,
3499
3947
  repcodes_t* const dRep, repcodes_t* const cRep,
3500
3948
  void* dst, size_t dstCapacity,
3501
- const void* src, size_t srcSize,
3949
+ const void* src, size_t srcSize,
3502
3950
  U32 lastBlock, U32 isPartition)
3503
3951
  {
3504
3952
  const U32 rleMaxLength = 25;
@@ -3572,45 +4020,49 @@ typedef struct {
3572
4020
 
3573
4021
  /* Helper function to perform the recursive search for block splits.
3574
4022
  * Estimates the cost of seqStore prior to split, and estimates the cost of splitting the sequences in half.
3575
- * If advantageous to split, then we recurse down the two sub-blocks. If not, or if an error occurred in estimation, then
3576
- * we do not recurse.
4023
+ * If advantageous to split, then we recurse down the two sub-blocks.
4024
+ * If not, or if an error occurred in estimation, then we do not recurse.
3577
4025
  *
3578
- * Note: The recursion depth is capped by a heuristic minimum number of sequences, defined by MIN_SEQUENCES_BLOCK_SPLITTING.
4026
+ * Note: The recursion depth is capped by a heuristic minimum number of sequences,
4027
+ * defined by MIN_SEQUENCES_BLOCK_SPLITTING.
3579
4028
  * In theory, this means the absolute largest recursion depth is 10 == log2(maxNbSeqInBlock/MIN_SEQUENCES_BLOCK_SPLITTING).
3580
4029
  * In practice, recursion depth usually doesn't go beyond 4.
3581
4030
  *
3582
- * Furthermore, the number of splits is capped by ZSTD_MAX_NB_BLOCK_SPLITS. At ZSTD_MAX_NB_BLOCK_SPLITS == 196 with the current existing blockSize
4031
+ * Furthermore, the number of splits is capped by ZSTD_MAX_NB_BLOCK_SPLITS.
4032
+ * At ZSTD_MAX_NB_BLOCK_SPLITS == 196 with the current existing blockSize
3583
4033
  * maximum of 128 KB, this value is actually impossible to reach.
3584
4034
  */
3585
4035
  static void
3586
4036
  ZSTD_deriveBlockSplitsHelper(seqStoreSplits* splits, size_t startIdx, size_t endIdx,
3587
4037
  ZSTD_CCtx* zc, const seqStore_t* origSeqStore)
3588
4038
  {
3589
- seqStore_t* fullSeqStoreChunk = &zc->blockSplitCtx.fullSeqStoreChunk;
3590
- seqStore_t* firstHalfSeqStore = &zc->blockSplitCtx.firstHalfSeqStore;
3591
- seqStore_t* secondHalfSeqStore = &zc->blockSplitCtx.secondHalfSeqStore;
4039
+ seqStore_t* const fullSeqStoreChunk = &zc->blockSplitCtx.fullSeqStoreChunk;
4040
+ seqStore_t* const firstHalfSeqStore = &zc->blockSplitCtx.firstHalfSeqStore;
4041
+ seqStore_t* const secondHalfSeqStore = &zc->blockSplitCtx.secondHalfSeqStore;
3592
4042
  size_t estimatedOriginalSize;
3593
4043
  size_t estimatedFirstHalfSize;
3594
4044
  size_t estimatedSecondHalfSize;
3595
4045
  size_t midIdx = (startIdx + endIdx)/2;
3596
4046
 
4047
+ DEBUGLOG(5, "ZSTD_deriveBlockSplitsHelper: startIdx=%zu endIdx=%zu", startIdx, endIdx);
4048
+ assert(endIdx >= startIdx);
3597
4049
  if (endIdx - startIdx < MIN_SEQUENCES_BLOCK_SPLITTING || splits->idx >= ZSTD_MAX_NB_BLOCK_SPLITS) {
3598
- DEBUGLOG(6, "ZSTD_deriveBlockSplitsHelper: Too few sequences");
4050
+ DEBUGLOG(6, "ZSTD_deriveBlockSplitsHelper: Too few sequences (%zu)", endIdx - startIdx);
3599
4051
  return;
3600
4052
  }
3601
- DEBUGLOG(4, "ZSTD_deriveBlockSplitsHelper: startIdx=%zu endIdx=%zu", startIdx, endIdx);
3602
4053
  ZSTD_deriveSeqStoreChunk(fullSeqStoreChunk, origSeqStore, startIdx, endIdx);
3603
4054
  ZSTD_deriveSeqStoreChunk(firstHalfSeqStore, origSeqStore, startIdx, midIdx);
3604
4055
  ZSTD_deriveSeqStoreChunk(secondHalfSeqStore, origSeqStore, midIdx, endIdx);
3605
4056
  estimatedOriginalSize = ZSTD_buildEntropyStatisticsAndEstimateSubBlockSize(fullSeqStoreChunk, zc);
3606
4057
  estimatedFirstHalfSize = ZSTD_buildEntropyStatisticsAndEstimateSubBlockSize(firstHalfSeqStore, zc);
3607
4058
  estimatedSecondHalfSize = ZSTD_buildEntropyStatisticsAndEstimateSubBlockSize(secondHalfSeqStore, zc);
3608
- DEBUGLOG(4, "Estimated original block size: %zu -- First half split: %zu -- Second half split: %zu",
4059
+ DEBUGLOG(5, "Estimated original block size: %zu -- First half split: %zu -- Second half split: %zu",
3609
4060
  estimatedOriginalSize, estimatedFirstHalfSize, estimatedSecondHalfSize);
3610
4061
  if (ZSTD_isError(estimatedOriginalSize) || ZSTD_isError(estimatedFirstHalfSize) || ZSTD_isError(estimatedSecondHalfSize)) {
3611
4062
  return;
3612
4063
  }
3613
4064
  if (estimatedFirstHalfSize + estimatedSecondHalfSize < estimatedOriginalSize) {
4065
+ DEBUGLOG(5, "split decided at seqNb:%zu", midIdx);
3614
4066
  ZSTD_deriveBlockSplitsHelper(splits, startIdx, midIdx, zc, origSeqStore);
3615
4067
  splits->splitLocations[splits->idx] = (U32)midIdx;
3616
4068
  splits->idx++;
@@ -3618,14 +4070,18 @@ ZSTD_deriveBlockSplitsHelper(seqStoreSplits* splits, size_t startIdx, size_t end
3618
4070
  }
3619
4071
  }
3620
4072
 
3621
- /* Base recursive function. Populates a table with intra-block partition indices that can improve compression ratio.
4073
+ /* Base recursive function.
4074
+ * Populates a table with intra-block partition indices that can improve compression ratio.
3622
4075
  *
3623
- * Returns the number of splits made (which equals the size of the partition table - 1).
4076
+ * @return: number of splits made (which equals the size of the partition table - 1).
3624
4077
  */
3625
- static size_t ZSTD_deriveBlockSplits(ZSTD_CCtx* zc, U32 partitions[], U32 nbSeq) {
3626
- seqStoreSplits splits = {partitions, 0};
4078
+ static size_t ZSTD_deriveBlockSplits(ZSTD_CCtx* zc, U32 partitions[], U32 nbSeq)
4079
+ {
4080
+ seqStoreSplits splits;
4081
+ splits.splitLocations = partitions;
4082
+ splits.idx = 0;
3627
4083
  if (nbSeq <= 4) {
3628
- DEBUGLOG(4, "ZSTD_deriveBlockSplits: Too few sequences to split");
4084
+ DEBUGLOG(5, "ZSTD_deriveBlockSplits: Too few sequences to split (%u <= 4)", nbSeq);
3629
4085
  /* Refuse to try and split anything with less than 4 sequences */
3630
4086
  return 0;
3631
4087
  }
@@ -3641,18 +4097,20 @@ static size_t ZSTD_deriveBlockSplits(ZSTD_CCtx* zc, U32 partitions[], U32 nbSeq)
3641
4097
  * Returns combined size of all blocks (which includes headers), or a ZSTD error code.
3642
4098
  */
3643
4099
  static size_t
3644
- ZSTD_compressBlock_splitBlock_internal(ZSTD_CCtx* zc, void* dst, size_t dstCapacity,
3645
- const void* src, size_t blockSize, U32 lastBlock, U32 nbSeq)
4100
+ ZSTD_compressBlock_splitBlock_internal(ZSTD_CCtx* zc,
4101
+ void* dst, size_t dstCapacity,
4102
+ const void* src, size_t blockSize,
4103
+ U32 lastBlock, U32 nbSeq)
3646
4104
  {
3647
4105
  size_t cSize = 0;
3648
4106
  const BYTE* ip = (const BYTE*)src;
3649
4107
  BYTE* op = (BYTE*)dst;
3650
4108
  size_t i = 0;
3651
4109
  size_t srcBytesTotal = 0;
3652
- U32* partitions = zc->blockSplitCtx.partitions; /* size == ZSTD_MAX_NB_BLOCK_SPLITS */
3653
- seqStore_t* nextSeqStore = &zc->blockSplitCtx.nextSeqStore;
3654
- seqStore_t* currSeqStore = &zc->blockSplitCtx.currSeqStore;
3655
- size_t numSplits = ZSTD_deriveBlockSplits(zc, partitions, nbSeq);
4110
+ U32* const partitions = zc->blockSplitCtx.partitions; /* size == ZSTD_MAX_NB_BLOCK_SPLITS */
4111
+ seqStore_t* const nextSeqStore = &zc->blockSplitCtx.nextSeqStore;
4112
+ seqStore_t* const currSeqStore = &zc->blockSplitCtx.currSeqStore;
4113
+ size_t const numSplits = ZSTD_deriveBlockSplits(zc, partitions, nbSeq);
3656
4114
 
3657
4115
  /* If a block is split and some partitions are emitted as RLE/uncompressed, then repcode history
3658
4116
  * may become invalid. In order to reconcile potentially invalid repcodes, we keep track of two
@@ -3674,30 +4132,31 @@ ZSTD_compressBlock_splitBlock_internal(ZSTD_CCtx* zc, void* dst, size_t dstCapac
3674
4132
  ZSTD_memcpy(cRep.rep, zc->blockState.prevCBlock->rep, sizeof(repcodes_t));
3675
4133
  ZSTD_memset(nextSeqStore, 0, sizeof(seqStore_t));
3676
4134
 
3677
- DEBUGLOG(4, "ZSTD_compressBlock_splitBlock_internal (dstCapacity=%u, dictLimit=%u, nextToUpdate=%u)",
4135
+ DEBUGLOG(5, "ZSTD_compressBlock_splitBlock_internal (dstCapacity=%u, dictLimit=%u, nextToUpdate=%u)",
3678
4136
  (unsigned)dstCapacity, (unsigned)zc->blockState.matchState.window.dictLimit,
3679
4137
  (unsigned)zc->blockState.matchState.nextToUpdate);
3680
4138
 
3681
4139
  if (numSplits == 0) {
3682
- size_t cSizeSingleBlock = ZSTD_compressSeqStore_singleBlock(zc, &zc->seqStore,
3683
- &dRep, &cRep,
3684
- op, dstCapacity,
3685
- ip, blockSize,
3686
- lastBlock, 0 /* isPartition */);
4140
+ size_t cSizeSingleBlock =
4141
+ ZSTD_compressSeqStore_singleBlock(zc, &zc->seqStore,
4142
+ &dRep, &cRep,
4143
+ op, dstCapacity,
4144
+ ip, blockSize,
4145
+ lastBlock, 0 /* isPartition */);
3687
4146
  FORWARD_IF_ERROR(cSizeSingleBlock, "Compressing single block from splitBlock_internal() failed!");
3688
4147
  DEBUGLOG(5, "ZSTD_compressBlock_splitBlock_internal: No splits");
3689
- assert(cSizeSingleBlock <= ZSTD_BLOCKSIZE_MAX + ZSTD_blockHeaderSize);
4148
+ assert(zc->blockSize <= ZSTD_BLOCKSIZE_MAX);
4149
+ assert(cSizeSingleBlock <= zc->blockSize + ZSTD_blockHeaderSize);
3690
4150
  return cSizeSingleBlock;
3691
4151
  }
3692
4152
 
3693
4153
  ZSTD_deriveSeqStoreChunk(currSeqStore, &zc->seqStore, 0, partitions[0]);
3694
4154
  for (i = 0; i <= numSplits; ++i) {
3695
- size_t srcBytes;
3696
4155
  size_t cSizeChunk;
3697
4156
  U32 const lastPartition = (i == numSplits);
3698
4157
  U32 lastBlockEntireSrc = 0;
3699
4158
 
3700
- srcBytes = ZSTD_countSeqStoreLiteralsBytes(currSeqStore) + ZSTD_countSeqStoreMatchBytes(currSeqStore);
4159
+ size_t srcBytes = ZSTD_countSeqStoreLiteralsBytes(currSeqStore) + ZSTD_countSeqStoreMatchBytes(currSeqStore);
3701
4160
  srcBytesTotal += srcBytes;
3702
4161
  if (lastPartition) {
3703
4162
  /* This is the final partition, need to account for possible last literals */
@@ -3712,7 +4171,8 @@ ZSTD_compressBlock_splitBlock_internal(ZSTD_CCtx* zc, void* dst, size_t dstCapac
3712
4171
  op, dstCapacity,
3713
4172
  ip, srcBytes,
3714
4173
  lastBlockEntireSrc, 1 /* isPartition */);
3715
- DEBUGLOG(5, "Estimated size: %zu actual size: %zu", ZSTD_buildEntropyStatisticsAndEstimateSubBlockSize(currSeqStore, zc), cSizeChunk);
4174
+ DEBUGLOG(5, "Estimated size: %zu vs %zu : actual size",
4175
+ ZSTD_buildEntropyStatisticsAndEstimateSubBlockSize(currSeqStore, zc), cSizeChunk);
3716
4176
  FORWARD_IF_ERROR(cSizeChunk, "Compressing chunk failed!");
3717
4177
 
3718
4178
  ip += srcBytes;
@@ -3720,10 +4180,10 @@ ZSTD_compressBlock_splitBlock_internal(ZSTD_CCtx* zc, void* dst, size_t dstCapac
3720
4180
  dstCapacity -= cSizeChunk;
3721
4181
  cSize += cSizeChunk;
3722
4182
  *currSeqStore = *nextSeqStore;
3723
- assert(cSizeChunk <= ZSTD_BLOCKSIZE_MAX + ZSTD_blockHeaderSize);
4183
+ assert(cSizeChunk <= zc->blockSize + ZSTD_blockHeaderSize);
3724
4184
  }
3725
- /* cRep and dRep may have diverged during the compression. If so, we use the dRep repcodes
3726
- * for the next block.
4185
+ /* cRep and dRep may have diverged during the compression.
4186
+ * If so, we use the dRep repcodes for the next block.
3727
4187
  */
3728
4188
  ZSTD_memcpy(zc->blockState.prevCBlock->rep, dRep.rep, sizeof(repcodes_t));
3729
4189
  return cSize;
@@ -3734,8 +4194,6 @@ ZSTD_compressBlock_splitBlock(ZSTD_CCtx* zc,
3734
4194
  void* dst, size_t dstCapacity,
3735
4195
  const void* src, size_t srcSize, U32 lastBlock)
3736
4196
  {
3737
- const BYTE* ip = (const BYTE*)src;
3738
- BYTE* op = (BYTE*)dst;
3739
4197
  U32 nbSeq;
3740
4198
  size_t cSize;
3741
4199
  DEBUGLOG(4, "ZSTD_compressBlock_splitBlock");
@@ -3746,7 +4204,7 @@ ZSTD_compressBlock_splitBlock(ZSTD_CCtx* zc,
3746
4204
  if (bss == ZSTDbss_noCompress) {
3747
4205
  if (zc->blockState.prevCBlock->entropy.fse.offcode_repeatMode == FSE_repeat_valid)
3748
4206
  zc->blockState.prevCBlock->entropy.fse.offcode_repeatMode = FSE_repeat_check;
3749
- cSize = ZSTD_noCompressBlock(op, dstCapacity, ip, srcSize, lastBlock);
4207
+ cSize = ZSTD_noCompressBlock(dst, dstCapacity, src, srcSize, lastBlock);
3750
4208
  FORWARD_IF_ERROR(cSize, "ZSTD_noCompressBlock failed");
3751
4209
  DEBUGLOG(4, "ZSTD_compressBlock_splitBlock: Nocompress block");
3752
4210
  return cSize;
@@ -3764,9 +4222,9 @@ ZSTD_compressBlock_internal(ZSTD_CCtx* zc,
3764
4222
  void* dst, size_t dstCapacity,
3765
4223
  const void* src, size_t srcSize, U32 frame)
3766
4224
  {
3767
- /* This the upper bound for the length of an rle block.
3768
- * This isn't the actual upper bound. Finding the real threshold
3769
- * needs further investigation.
4225
+ /* This is an estimated upper bound for the length of an rle block.
4226
+ * This isn't the actual upper bound.
4227
+ * Finding the real threshold needs further investigation.
3770
4228
  */
3771
4229
  const U32 rleMaxLength = 25;
3772
4230
  size_t cSize;
@@ -3858,10 +4316,11 @@ static size_t ZSTD_compressBlock_targetCBlockSize_body(ZSTD_CCtx* zc,
3858
4316
  * * cSize >= blockBound(srcSize): We have expanded the block too much so
3859
4317
  * emit an uncompressed block.
3860
4318
  */
3861
- {
3862
- size_t const cSize = ZSTD_compressSuperBlock(zc, dst, dstCapacity, src, srcSize, lastBlock);
4319
+ { size_t const cSize =
4320
+ ZSTD_compressSuperBlock(zc, dst, dstCapacity, src, srcSize, lastBlock);
3863
4321
  if (cSize != ERROR(dstSize_tooSmall)) {
3864
- size_t const maxCSize = srcSize - ZSTD_minGain(srcSize, zc->appliedParams.cParams.strategy);
4322
+ size_t const maxCSize =
4323
+ srcSize - ZSTD_minGain(srcSize, zc->appliedParams.cParams.strategy);
3865
4324
  FORWARD_IF_ERROR(cSize, "ZSTD_compressSuperBlock failed");
3866
4325
  if (cSize != 0 && cSize < maxCSize + ZSTD_blockHeaderSize) {
3867
4326
  ZSTD_blockState_confirmRepcodesAndEntropyTables(&zc->blockState);
@@ -3869,7 +4328,7 @@ static size_t ZSTD_compressBlock_targetCBlockSize_body(ZSTD_CCtx* zc,
3869
4328
  }
3870
4329
  }
3871
4330
  }
3872
- }
4331
+ } /* if (bss == ZSTDbss_compress)*/
3873
4332
 
3874
4333
  DEBUGLOG(6, "Resorting to ZSTD_noCompressBlock()");
3875
4334
  /* Superblock compression failed, attempt to emit a single no compress block.
@@ -3927,7 +4386,7 @@ static void ZSTD_overflowCorrectIfNeeded(ZSTD_matchState_t* ms,
3927
4386
  * All blocks will be terminated, all input will be consumed.
3928
4387
  * Function will issue an error if there is not enough `dstCapacity` to hold the compressed content.
3929
4388
  * Frame is supposed already started (header already produced)
3930
- * @return : compressed size, or an error code
4389
+ * @return : compressed size, or an error code
3931
4390
  */
3932
4391
  static size_t ZSTD_compress_frameChunk(ZSTD_CCtx* cctx,
3933
4392
  void* dst, size_t dstCapacity,
@@ -3951,7 +4410,9 @@ static size_t ZSTD_compress_frameChunk(ZSTD_CCtx* cctx,
3951
4410
  ZSTD_matchState_t* const ms = &cctx->blockState.matchState;
3952
4411
  U32 const lastBlock = lastFrameChunk & (blockSize >= remaining);
3953
4412
 
3954
- RETURN_ERROR_IF(dstCapacity < ZSTD_blockHeaderSize + MIN_CBLOCK_SIZE,
4413
+ /* TODO: See 3090. We reduced MIN_CBLOCK_SIZE from 3 to 2 so to compensate we are adding
4414
+ * additional 1. We need to revisit and change this logic to be more consistent */
4415
+ RETURN_ERROR_IF(dstCapacity < ZSTD_blockHeaderSize + MIN_CBLOCK_SIZE + 1,
3955
4416
  dstSize_tooSmall,
3956
4417
  "not enough space to store compressed block");
3957
4418
  if (remaining < blockSize) blockSize = remaining;
@@ -3990,7 +4451,7 @@ static size_t ZSTD_compress_frameChunk(ZSTD_CCtx* cctx,
3990
4451
  MEM_writeLE24(op, cBlockHeader);
3991
4452
  cSize += ZSTD_blockHeaderSize;
3992
4453
  }
3993
- }
4454
+ } /* if (ZSTD_useTargetCBlockSize(&cctx->appliedParams))*/
3994
4455
 
3995
4456
 
3996
4457
  ip += blockSize;
@@ -4169,31 +4630,51 @@ static size_t ZSTD_compressContinue_internal (ZSTD_CCtx* cctx,
4169
4630
  }
4170
4631
  }
4171
4632
 
4172
- size_t ZSTD_compressContinue (ZSTD_CCtx* cctx,
4173
- void* dst, size_t dstCapacity,
4174
- const void* src, size_t srcSize)
4633
+ size_t ZSTD_compressContinue_public(ZSTD_CCtx* cctx,
4634
+ void* dst, size_t dstCapacity,
4635
+ const void* src, size_t srcSize)
4175
4636
  {
4176
4637
  DEBUGLOG(5, "ZSTD_compressContinue (srcSize=%u)", (unsigned)srcSize);
4177
4638
  return ZSTD_compressContinue_internal(cctx, dst, dstCapacity, src, srcSize, 1 /* frame mode */, 0 /* last chunk */);
4178
4639
  }
4179
4640
 
4641
+ /* NOTE: Must just wrap ZSTD_compressContinue_public() */
4642
+ size_t ZSTD_compressContinue(ZSTD_CCtx* cctx,
4643
+ void* dst, size_t dstCapacity,
4644
+ const void* src, size_t srcSize)
4645
+ {
4646
+ return ZSTD_compressContinue_public(cctx, dst, dstCapacity, src, srcSize);
4647
+ }
4180
4648
 
4181
- size_t ZSTD_getBlockSize(const ZSTD_CCtx* cctx)
4649
+ static size_t ZSTD_getBlockSize_deprecated(const ZSTD_CCtx* cctx)
4182
4650
  {
4183
4651
  ZSTD_compressionParameters const cParams = cctx->appliedParams.cParams;
4184
4652
  assert(!ZSTD_checkCParams(cParams));
4185
- return MIN (ZSTD_BLOCKSIZE_MAX, (U32)1 << cParams.windowLog);
4653
+ return MIN(cctx->appliedParams.maxBlockSize, (size_t)1 << cParams.windowLog);
4186
4654
  }
4187
4655
 
4188
- size_t ZSTD_compressBlock(ZSTD_CCtx* cctx, void* dst, size_t dstCapacity, const void* src, size_t srcSize)
4656
+ /* NOTE: Must just wrap ZSTD_getBlockSize_deprecated() */
4657
+ size_t ZSTD_getBlockSize(const ZSTD_CCtx* cctx)
4658
+ {
4659
+ return ZSTD_getBlockSize_deprecated(cctx);
4660
+ }
4661
+
4662
+ /* NOTE: Must just wrap ZSTD_compressBlock_deprecated() */
4663
+ size_t ZSTD_compressBlock_deprecated(ZSTD_CCtx* cctx, void* dst, size_t dstCapacity, const void* src, size_t srcSize)
4189
4664
  {
4190
4665
  DEBUGLOG(5, "ZSTD_compressBlock: srcSize = %u", (unsigned)srcSize);
4191
- { size_t const blockSizeMax = ZSTD_getBlockSize(cctx);
4666
+ { size_t const blockSizeMax = ZSTD_getBlockSize_deprecated(cctx);
4192
4667
  RETURN_ERROR_IF(srcSize > blockSizeMax, srcSize_wrong, "input is larger than a block"); }
4193
4668
 
4194
4669
  return ZSTD_compressContinue_internal(cctx, dst, dstCapacity, src, srcSize, 0 /* frame mode */, 0 /* last chunk */);
4195
4670
  }
4196
4671
 
4672
+ /* NOTE: Must just wrap ZSTD_compressBlock_deprecated() */
4673
+ size_t ZSTD_compressBlock(ZSTD_CCtx* cctx, void* dst, size_t dstCapacity, const void* src, size_t srcSize)
4674
+ {
4675
+ return ZSTD_compressBlock_deprecated(cctx, dst, dstCapacity, src, srcSize);
4676
+ }
4677
+
4197
4678
  /*! ZSTD_loadDictionaryContent() :
4198
4679
  * @return : 0, or an error code
4199
4680
  */
@@ -4202,25 +4683,36 @@ static size_t ZSTD_loadDictionaryContent(ZSTD_matchState_t* ms,
4202
4683
  ZSTD_cwksp* ws,
4203
4684
  ZSTD_CCtx_params const* params,
4204
4685
  const void* src, size_t srcSize,
4205
- ZSTD_dictTableLoadMethod_e dtlm)
4686
+ ZSTD_dictTableLoadMethod_e dtlm,
4687
+ ZSTD_tableFillPurpose_e tfp)
4206
4688
  {
4207
4689
  const BYTE* ip = (const BYTE*) src;
4208
4690
  const BYTE* const iend = ip + srcSize;
4209
4691
  int const loadLdmDict = params->ldmParams.enableLdm == ZSTD_ps_enable && ls != NULL;
4210
4692
 
4211
- /* Assert that we the ms params match the params we're being given */
4693
+ /* Assert that the ms params match the params we're being given */
4212
4694
  ZSTD_assertEqualCParams(params->cParams, ms->cParams);
4213
4695
 
4214
- if (srcSize > ZSTD_CHUNKSIZE_MAX) {
4696
+ { /* Ensure large dictionaries can't cause index overflow */
4697
+
4215
4698
  /* Allow the dictionary to set indices up to exactly ZSTD_CURRENT_MAX.
4216
4699
  * Dictionaries right at the edge will immediately trigger overflow
4217
4700
  * correction, but I don't want to insert extra constraints here.
4218
4701
  */
4219
- U32 const maxDictSize = ZSTD_CURRENT_MAX - 1;
4220
- /* We must have cleared our windows when our source is this large. */
4221
- assert(ZSTD_window_isEmpty(ms->window));
4222
- if (loadLdmDict)
4223
- assert(ZSTD_window_isEmpty(ls->window));
4702
+ U32 maxDictSize = ZSTD_CURRENT_MAX - ZSTD_WINDOW_START_INDEX;
4703
+
4704
+ int const CDictTaggedIndices = ZSTD_CDictIndicesAreTagged(&params->cParams);
4705
+ if (CDictTaggedIndices && tfp == ZSTD_tfp_forCDict) {
4706
+ /* Some dictionary matchfinders in zstd use "short cache",
4707
+ * which treats the lower ZSTD_SHORT_CACHE_TAG_BITS of each
4708
+ * CDict hashtable entry as a tag rather than as part of an index.
4709
+ * When short cache is used, we need to truncate the dictionary
4710
+ * so that its indices don't overlap with the tag. */
4711
+ U32 const shortCacheMaxDictSize = (1u << (32 - ZSTD_SHORT_CACHE_TAG_BITS)) - ZSTD_WINDOW_START_INDEX;
4712
+ maxDictSize = MIN(maxDictSize, shortCacheMaxDictSize);
4713
+ assert(!loadLdmDict);
4714
+ }
4715
+
4224
4716
  /* If the dictionary is too large, only load the suffix of the dictionary. */
4225
4717
  if (srcSize > maxDictSize) {
4226
4718
  ip = iend - maxDictSize;
@@ -4229,30 +4721,46 @@ static size_t ZSTD_loadDictionaryContent(ZSTD_matchState_t* ms,
4229
4721
  }
4230
4722
  }
4231
4723
 
4232
- DEBUGLOG(4, "ZSTD_loadDictionaryContent(): useRowMatchFinder=%d", (int)params->useRowMatchFinder);
4724
+ if (srcSize > ZSTD_CHUNKSIZE_MAX) {
4725
+ /* We must have cleared our windows when our source is this large. */
4726
+ assert(ZSTD_window_isEmpty(ms->window));
4727
+ if (loadLdmDict) assert(ZSTD_window_isEmpty(ls->window));
4728
+ }
4233
4729
  ZSTD_window_update(&ms->window, src, srcSize, /* forceNonContiguous */ 0);
4234
- ms->loadedDictEnd = params->forceWindow ? 0 : (U32)(iend - ms->window.base);
4235
- ms->forceNonContiguous = params->deterministicRefPrefix;
4236
4730
 
4237
- if (loadLdmDict) {
4731
+ DEBUGLOG(4, "ZSTD_loadDictionaryContent(): useRowMatchFinder=%d", (int)params->useRowMatchFinder);
4732
+
4733
+ if (loadLdmDict) { /* Load the entire dict into LDM matchfinders. */
4238
4734
  ZSTD_window_update(&ls->window, src, srcSize, /* forceNonContiguous */ 0);
4239
4735
  ls->loadedDictEnd = params->forceWindow ? 0 : (U32)(iend - ls->window.base);
4736
+ ZSTD_ldm_fillHashTable(ls, ip, iend, &params->ldmParams);
4737
+ }
4738
+
4739
+ /* If the dict is larger than we can reasonably index in our tables, only load the suffix. */
4740
+ if (params->cParams.strategy < ZSTD_btultra) {
4741
+ U32 maxDictSize = 8U << MIN(MAX(params->cParams.hashLog, params->cParams.chainLog), 28);
4742
+ if (srcSize > maxDictSize) {
4743
+ ip = iend - maxDictSize;
4744
+ src = ip;
4745
+ srcSize = maxDictSize;
4746
+ }
4240
4747
  }
4241
4748
 
4749
+ ms->nextToUpdate = (U32)(ip - ms->window.base);
4750
+ ms->loadedDictEnd = params->forceWindow ? 0 : (U32)(iend - ms->window.base);
4751
+ ms->forceNonContiguous = params->deterministicRefPrefix;
4752
+
4242
4753
  if (srcSize <= HASH_READ_SIZE) return 0;
4243
4754
 
4244
4755
  ZSTD_overflowCorrectIfNeeded(ms, ws, params, ip, iend);
4245
4756
 
4246
- if (loadLdmDict)
4247
- ZSTD_ldm_fillHashTable(ls, ip, iend, &params->ldmParams);
4248
-
4249
4757
  switch(params->cParams.strategy)
4250
4758
  {
4251
4759
  case ZSTD_fast:
4252
- ZSTD_fillHashTable(ms, iend, dtlm);
4760
+ ZSTD_fillHashTable(ms, iend, dtlm, tfp);
4253
4761
  break;
4254
4762
  case ZSTD_dfast:
4255
- ZSTD_fillDoubleHashTable(ms, iend, dtlm);
4763
+ ZSTD_fillDoubleHashTable(ms, iend, dtlm, tfp);
4256
4764
  break;
4257
4765
 
4258
4766
  case ZSTD_greedy:
@@ -4265,7 +4773,7 @@ static size_t ZSTD_loadDictionaryContent(ZSTD_matchState_t* ms,
4265
4773
  } else {
4266
4774
  assert(params->useRowMatchFinder != ZSTD_ps_auto);
4267
4775
  if (params->useRowMatchFinder == ZSTD_ps_enable) {
4268
- size_t const tagTableSize = ((size_t)1 << params->cParams.hashLog) * sizeof(U16);
4776
+ size_t const tagTableSize = ((size_t)1 << params->cParams.hashLog);
4269
4777
  ZSTD_memset(ms->tagTable, 0, tagTableSize);
4270
4778
  ZSTD_row_update(ms, iend-HASH_READ_SIZE);
4271
4779
  DEBUGLOG(4, "Using row-based hash table for lazy dict");
@@ -4418,6 +4926,7 @@ static size_t ZSTD_loadZstdDictionary(ZSTD_compressedBlockState_t* bs,
4418
4926
  ZSTD_CCtx_params const* params,
4419
4927
  const void* dict, size_t dictSize,
4420
4928
  ZSTD_dictTableLoadMethod_e dtlm,
4929
+ ZSTD_tableFillPurpose_e tfp,
4421
4930
  void* workspace)
4422
4931
  {
4423
4932
  const BYTE* dictPtr = (const BYTE*)dict;
@@ -4436,7 +4945,7 @@ static size_t ZSTD_loadZstdDictionary(ZSTD_compressedBlockState_t* bs,
4436
4945
  {
4437
4946
  size_t const dictContentSize = (size_t)(dictEnd - dictPtr);
4438
4947
  FORWARD_IF_ERROR(ZSTD_loadDictionaryContent(
4439
- ms, NULL, ws, params, dictPtr, dictContentSize, dtlm), "");
4948
+ ms, NULL, ws, params, dictPtr, dictContentSize, dtlm, tfp), "");
4440
4949
  }
4441
4950
  return dictID;
4442
4951
  }
@@ -4452,6 +4961,7 @@ ZSTD_compress_insertDictionary(ZSTD_compressedBlockState_t* bs,
4452
4961
  const void* dict, size_t dictSize,
4453
4962
  ZSTD_dictContentType_e dictContentType,
4454
4963
  ZSTD_dictTableLoadMethod_e dtlm,
4964
+ ZSTD_tableFillPurpose_e tfp,
4455
4965
  void* workspace)
4456
4966
  {
4457
4967
  DEBUGLOG(4, "ZSTD_compress_insertDictionary (dictSize=%u)", (U32)dictSize);
@@ -4464,13 +4974,13 @@ ZSTD_compress_insertDictionary(ZSTD_compressedBlockState_t* bs,
4464
4974
 
4465
4975
  /* dict restricted modes */
4466
4976
  if (dictContentType == ZSTD_dct_rawContent)
4467
- return ZSTD_loadDictionaryContent(ms, ls, ws, params, dict, dictSize, dtlm);
4977
+ return ZSTD_loadDictionaryContent(ms, ls, ws, params, dict, dictSize, dtlm, tfp);
4468
4978
 
4469
4979
  if (MEM_readLE32(dict) != ZSTD_MAGIC_DICTIONARY) {
4470
4980
  if (dictContentType == ZSTD_dct_auto) {
4471
4981
  DEBUGLOG(4, "raw content dictionary detected");
4472
4982
  return ZSTD_loadDictionaryContent(
4473
- ms, ls, ws, params, dict, dictSize, dtlm);
4983
+ ms, ls, ws, params, dict, dictSize, dtlm, tfp);
4474
4984
  }
4475
4985
  RETURN_ERROR_IF(dictContentType == ZSTD_dct_fullDict, dictionary_wrong, "");
4476
4986
  assert(0); /* impossible */
@@ -4478,13 +4988,14 @@ ZSTD_compress_insertDictionary(ZSTD_compressedBlockState_t* bs,
4478
4988
 
4479
4989
  /* dict as full zstd dictionary */
4480
4990
  return ZSTD_loadZstdDictionary(
4481
- bs, ms, ws, params, dict, dictSize, dtlm, workspace);
4991
+ bs, ms, ws, params, dict, dictSize, dtlm, tfp, workspace);
4482
4992
  }
4483
4993
 
4484
4994
  #define ZSTD_USE_CDICT_PARAMS_SRCSIZE_CUTOFF (128 KB)
4485
4995
  #define ZSTD_USE_CDICT_PARAMS_DICTSIZE_MULTIPLIER (6ULL)
4486
4996
 
4487
4997
  /*! ZSTD_compressBegin_internal() :
4998
+ * Assumption : either @dict OR @cdict (or none) is non-NULL, never both
4488
4999
  * @return : 0, or an error code */
4489
5000
  static size_t ZSTD_compressBegin_internal(ZSTD_CCtx* cctx,
4490
5001
  const void* dict, size_t dictSize,
@@ -4520,11 +5031,11 @@ static size_t ZSTD_compressBegin_internal(ZSTD_CCtx* cctx,
4520
5031
  cctx->blockState.prevCBlock, &cctx->blockState.matchState,
4521
5032
  &cctx->ldmState, &cctx->workspace, &cctx->appliedParams, cdict->dictContent,
4522
5033
  cdict->dictContentSize, cdict->dictContentType, dtlm,
4523
- cctx->entropyWorkspace)
5034
+ ZSTD_tfp_forCCtx, cctx->entropyWorkspace)
4524
5035
  : ZSTD_compress_insertDictionary(
4525
5036
  cctx->blockState.prevCBlock, &cctx->blockState.matchState,
4526
5037
  &cctx->ldmState, &cctx->workspace, &cctx->appliedParams, dict, dictSize,
4527
- dictContentType, dtlm, cctx->entropyWorkspace);
5038
+ dictContentType, dtlm, ZSTD_tfp_forCCtx, cctx->entropyWorkspace);
4528
5039
  FORWARD_IF_ERROR(dictID, "ZSTD_compress_insertDictionary failed");
4529
5040
  assert(dictID <= UINT_MAX);
4530
5041
  cctx->dictID = (U32)dictID;
@@ -4565,11 +5076,11 @@ size_t ZSTD_compressBegin_advanced(ZSTD_CCtx* cctx,
4565
5076
  &cctxParams, pledgedSrcSize);
4566
5077
  }
4567
5078
 
4568
- size_t ZSTD_compressBegin_usingDict(ZSTD_CCtx* cctx, const void* dict, size_t dictSize, int compressionLevel)
5079
+ static size_t
5080
+ ZSTD_compressBegin_usingDict_deprecated(ZSTD_CCtx* cctx, const void* dict, size_t dictSize, int compressionLevel)
4569
5081
  {
4570
5082
  ZSTD_CCtx_params cctxParams;
4571
- {
4572
- ZSTD_parameters const params = ZSTD_getParams_internal(compressionLevel, ZSTD_CONTENTSIZE_UNKNOWN, dictSize, ZSTD_cpm_noAttachDict);
5083
+ { ZSTD_parameters const params = ZSTD_getParams_internal(compressionLevel, ZSTD_CONTENTSIZE_UNKNOWN, dictSize, ZSTD_cpm_noAttachDict);
4573
5084
  ZSTD_CCtxParams_init_internal(&cctxParams, &params, (compressionLevel == 0) ? ZSTD_CLEVEL_DEFAULT : compressionLevel);
4574
5085
  }
4575
5086
  DEBUGLOG(4, "ZSTD_compressBegin_usingDict (dictSize=%u)", (unsigned)dictSize);
@@ -4577,9 +5088,15 @@ size_t ZSTD_compressBegin_usingDict(ZSTD_CCtx* cctx, const void* dict, size_t di
4577
5088
  &cctxParams, ZSTD_CONTENTSIZE_UNKNOWN, ZSTDb_not_buffered);
4578
5089
  }
4579
5090
 
5091
+ size_t
5092
+ ZSTD_compressBegin_usingDict(ZSTD_CCtx* cctx, const void* dict, size_t dictSize, int compressionLevel)
5093
+ {
5094
+ return ZSTD_compressBegin_usingDict_deprecated(cctx, dict, dictSize, compressionLevel);
5095
+ }
5096
+
4580
5097
  size_t ZSTD_compressBegin(ZSTD_CCtx* cctx, int compressionLevel)
4581
5098
  {
4582
- return ZSTD_compressBegin_usingDict(cctx, NULL, 0, compressionLevel);
5099
+ return ZSTD_compressBegin_usingDict_deprecated(cctx, NULL, 0, compressionLevel);
4583
5100
  }
4584
5101
 
4585
5102
 
@@ -4649,9 +5166,9 @@ void ZSTD_CCtx_trace(ZSTD_CCtx* cctx, size_t extraCSize)
4649
5166
  #endif
4650
5167
  }
4651
5168
 
4652
- size_t ZSTD_compressEnd (ZSTD_CCtx* cctx,
4653
- void* dst, size_t dstCapacity,
4654
- const void* src, size_t srcSize)
5169
+ size_t ZSTD_compressEnd_public(ZSTD_CCtx* cctx,
5170
+ void* dst, size_t dstCapacity,
5171
+ const void* src, size_t srcSize)
4655
5172
  {
4656
5173
  size_t endResult;
4657
5174
  size_t const cSize = ZSTD_compressContinue_internal(cctx,
@@ -4675,6 +5192,14 @@ size_t ZSTD_compressEnd (ZSTD_CCtx* cctx,
4675
5192
  return cSize + endResult;
4676
5193
  }
4677
5194
 
5195
+ /* NOTE: Must just wrap ZSTD_compressEnd_public() */
5196
+ size_t ZSTD_compressEnd(ZSTD_CCtx* cctx,
5197
+ void* dst, size_t dstCapacity,
5198
+ const void* src, size_t srcSize)
5199
+ {
5200
+ return ZSTD_compressEnd_public(cctx, dst, dstCapacity, src, srcSize);
5201
+ }
5202
+
4678
5203
  size_t ZSTD_compress_advanced (ZSTD_CCtx* cctx,
4679
5204
  void* dst, size_t dstCapacity,
4680
5205
  const void* src, size_t srcSize,
@@ -4703,7 +5228,7 @@ size_t ZSTD_compress_advanced_internal(
4703
5228
  FORWARD_IF_ERROR( ZSTD_compressBegin_internal(cctx,
4704
5229
  dict, dictSize, ZSTD_dct_auto, ZSTD_dtlm_fast, NULL,
4705
5230
  params, srcSize, ZSTDb_not_buffered) , "");
4706
- return ZSTD_compressEnd(cctx, dst, dstCapacity, src, srcSize);
5231
+ return ZSTD_compressEnd_public(cctx, dst, dstCapacity, src, srcSize);
4707
5232
  }
4708
5233
 
4709
5234
  size_t ZSTD_compress_usingDict(ZSTD_CCtx* cctx,
@@ -4828,7 +5353,7 @@ static size_t ZSTD_initCDict_internal(
4828
5353
  { size_t const dictID = ZSTD_compress_insertDictionary(
4829
5354
  &cdict->cBlockState, &cdict->matchState, NULL, &cdict->workspace,
4830
5355
  &params, cdict->dictContent, cdict->dictContentSize,
4831
- dictContentType, ZSTD_dtlm_full, cdict->entropyWorkspace);
5356
+ dictContentType, ZSTD_dtlm_full, ZSTD_tfp_forCDict, cdict->entropyWorkspace);
4832
5357
  FORWARD_IF_ERROR(dictID, "ZSTD_compress_insertDictionary failed");
4833
5358
  assert(dictID <= (size_t)(U32)-1);
4834
5359
  cdict->dictID = (U32)dictID;
@@ -5025,6 +5550,7 @@ const ZSTD_CDict* ZSTD_initStaticCDict(
5025
5550
  params.cParams = cParams;
5026
5551
  params.useRowMatchFinder = useRowMatchFinder;
5027
5552
  cdict->useRowMatchFinder = useRowMatchFinder;
5553
+ cdict->compressionLevel = ZSTD_NO_CLEVEL;
5028
5554
 
5029
5555
  if (ZSTD_isError( ZSTD_initCDict_internal(cdict,
5030
5556
  dict, dictSize,
@@ -5104,12 +5630,17 @@ size_t ZSTD_compressBegin_usingCDict_advanced(
5104
5630
 
5105
5631
  /* ZSTD_compressBegin_usingCDict() :
5106
5632
  * cdict must be != NULL */
5107
- size_t ZSTD_compressBegin_usingCDict(ZSTD_CCtx* cctx, const ZSTD_CDict* cdict)
5633
+ size_t ZSTD_compressBegin_usingCDict_deprecated(ZSTD_CCtx* cctx, const ZSTD_CDict* cdict)
5108
5634
  {
5109
5635
  ZSTD_frameParameters const fParams = { 0 /*content*/, 0 /*checksum*/, 0 /*noDictID*/ };
5110
5636
  return ZSTD_compressBegin_usingCDict_internal(cctx, cdict, fParams, ZSTD_CONTENTSIZE_UNKNOWN);
5111
5637
  }
5112
5638
 
5639
+ size_t ZSTD_compressBegin_usingCDict(ZSTD_CCtx* cctx, const ZSTD_CDict* cdict)
5640
+ {
5641
+ return ZSTD_compressBegin_usingCDict_deprecated(cctx, cdict);
5642
+ }
5643
+
5113
5644
  /*! ZSTD_compress_usingCDict_internal():
5114
5645
  * Implementation of various ZSTD_compress_usingCDict* functions.
5115
5646
  */
@@ -5119,7 +5650,7 @@ static size_t ZSTD_compress_usingCDict_internal(ZSTD_CCtx* cctx,
5119
5650
  const ZSTD_CDict* cdict, ZSTD_frameParameters fParams)
5120
5651
  {
5121
5652
  FORWARD_IF_ERROR(ZSTD_compressBegin_usingCDict_internal(cctx, cdict, fParams, srcSize), ""); /* will check if cdict != NULL */
5122
- return ZSTD_compressEnd(cctx, dst, dstCapacity, src, srcSize);
5653
+ return ZSTD_compressEnd_public(cctx, dst, dstCapacity, src, srcSize);
5123
5654
  }
5124
5655
 
5125
5656
  /*! ZSTD_compress_usingCDict_advanced():
@@ -5316,30 +5847,41 @@ size_t ZSTD_initCStream(ZSTD_CStream* zcs, int compressionLevel)
5316
5847
 
5317
5848
  static size_t ZSTD_nextInputSizeHint(const ZSTD_CCtx* cctx)
5318
5849
  {
5319
- size_t hintInSize = cctx->inBuffTarget - cctx->inBuffPos;
5320
- if (hintInSize==0) hintInSize = cctx->blockSize;
5321
- return hintInSize;
5850
+ if (cctx->appliedParams.inBufferMode == ZSTD_bm_stable) {
5851
+ return cctx->blockSize - cctx->stableIn_notConsumed;
5852
+ }
5853
+ assert(cctx->appliedParams.inBufferMode == ZSTD_bm_buffered);
5854
+ { size_t hintInSize = cctx->inBuffTarget - cctx->inBuffPos;
5855
+ if (hintInSize==0) hintInSize = cctx->blockSize;
5856
+ return hintInSize;
5857
+ }
5322
5858
  }
5323
5859
 
5324
5860
  /** ZSTD_compressStream_generic():
5325
5861
  * internal function for all *compressStream*() variants
5326
- * non-static, because can be called from zstdmt_compress.c
5327
- * @return : hint size for next input */
5862
+ * @return : hint size for next input to complete ongoing block */
5328
5863
  static size_t ZSTD_compressStream_generic(ZSTD_CStream* zcs,
5329
5864
  ZSTD_outBuffer* output,
5330
5865
  ZSTD_inBuffer* input,
5331
5866
  ZSTD_EndDirective const flushMode)
5332
5867
  {
5333
- const char* const istart = (const char*)input->src;
5334
- const char* const iend = input->size != 0 ? istart + input->size : istart;
5335
- const char* ip = input->pos != 0 ? istart + input->pos : istart;
5336
- char* const ostart = (char*)output->dst;
5337
- char* const oend = output->size != 0 ? ostart + output->size : ostart;
5338
- char* op = output->pos != 0 ? ostart + output->pos : ostart;
5868
+ const char* const istart = (assert(input != NULL), (const char*)input->src);
5869
+ const char* const iend = (istart != NULL) ? istart + input->size : istart;
5870
+ const char* ip = (istart != NULL) ? istart + input->pos : istart;
5871
+ char* const ostart = (assert(output != NULL), (char*)output->dst);
5872
+ char* const oend = (ostart != NULL) ? ostart + output->size : ostart;
5873
+ char* op = (ostart != NULL) ? ostart + output->pos : ostart;
5339
5874
  U32 someMoreWork = 1;
5340
5875
 
5341
5876
  /* check expectations */
5342
- DEBUGLOG(5, "ZSTD_compressStream_generic, flush=%u", (unsigned)flushMode);
5877
+ DEBUGLOG(5, "ZSTD_compressStream_generic, flush=%i, srcSize = %zu", (int)flushMode, input->size - input->pos);
5878
+ assert(zcs != NULL);
5879
+ if (zcs->appliedParams.inBufferMode == ZSTD_bm_stable) {
5880
+ assert(input->pos >= zcs->stableIn_notConsumed);
5881
+ input->pos -= zcs->stableIn_notConsumed;
5882
+ ip -= zcs->stableIn_notConsumed;
5883
+ zcs->stableIn_notConsumed = 0;
5884
+ }
5343
5885
  if (zcs->appliedParams.inBufferMode == ZSTD_bm_buffered) {
5344
5886
  assert(zcs->inBuff != NULL);
5345
5887
  assert(zcs->inBuffSize > 0);
@@ -5348,8 +5890,10 @@ static size_t ZSTD_compressStream_generic(ZSTD_CStream* zcs,
5348
5890
  assert(zcs->outBuff != NULL);
5349
5891
  assert(zcs->outBuffSize > 0);
5350
5892
  }
5351
- assert(output->pos <= output->size);
5893
+ if (input->src == NULL) assert(input->size == 0);
5352
5894
  assert(input->pos <= input->size);
5895
+ if (output->dst == NULL) assert(output->size == 0);
5896
+ assert(output->pos <= output->size);
5353
5897
  assert((U32)flushMode <= (U32)ZSTD_e_end);
5354
5898
 
5355
5899
  while (someMoreWork) {
@@ -5364,7 +5908,7 @@ static size_t ZSTD_compressStream_generic(ZSTD_CStream* zcs,
5364
5908
  || zcs->appliedParams.outBufferMode == ZSTD_bm_stable) /* OR we are allowed to return dstSizeTooSmall */
5365
5909
  && (zcs->inBuffPos == 0) ) {
5366
5910
  /* shortcut to compression pass directly into output buffer */
5367
- size_t const cSize = ZSTD_compressEnd(zcs,
5911
+ size_t const cSize = ZSTD_compressEnd_public(zcs,
5368
5912
  op, oend-op, ip, iend-ip);
5369
5913
  DEBUGLOG(4, "ZSTD_compressEnd : cSize=%u", (unsigned)cSize);
5370
5914
  FORWARD_IF_ERROR(cSize, "ZSTD_compressEnd failed");
@@ -5381,8 +5925,7 @@ static size_t ZSTD_compressStream_generic(ZSTD_CStream* zcs,
5381
5925
  zcs->inBuff + zcs->inBuffPos, toLoad,
5382
5926
  ip, iend-ip);
5383
5927
  zcs->inBuffPos += loaded;
5384
- if (loaded != 0)
5385
- ip += loaded;
5928
+ if (ip) ip += loaded;
5386
5929
  if ( (flushMode == ZSTD_e_continue)
5387
5930
  && (zcs->inBuffPos < zcs->inBuffTarget) ) {
5388
5931
  /* not enough input to fill full block : stop here */
@@ -5393,6 +5936,20 @@ static size_t ZSTD_compressStream_generic(ZSTD_CStream* zcs,
5393
5936
  /* empty */
5394
5937
  someMoreWork = 0; break;
5395
5938
  }
5939
+ } else {
5940
+ assert(zcs->appliedParams.inBufferMode == ZSTD_bm_stable);
5941
+ if ( (flushMode == ZSTD_e_continue)
5942
+ && ( (size_t)(iend - ip) < zcs->blockSize) ) {
5943
+ /* can't compress a full block : stop here */
5944
+ zcs->stableIn_notConsumed = (size_t)(iend - ip);
5945
+ ip = iend; /* pretend to have consumed input */
5946
+ someMoreWork = 0; break;
5947
+ }
5948
+ if ( (flushMode == ZSTD_e_flush)
5949
+ && (ip == iend) ) {
5950
+ /* empty */
5951
+ someMoreWork = 0; break;
5952
+ }
5396
5953
  }
5397
5954
  /* compress current block (note : this stage cannot be stopped in the middle) */
5398
5955
  DEBUGLOG(5, "stream compression stage (flushMode==%u)", flushMode);
@@ -5400,9 +5957,8 @@ static size_t ZSTD_compressStream_generic(ZSTD_CStream* zcs,
5400
5957
  void* cDst;
5401
5958
  size_t cSize;
5402
5959
  size_t oSize = oend-op;
5403
- size_t const iSize = inputBuffered
5404
- ? zcs->inBuffPos - zcs->inToCompress
5405
- : MIN((size_t)(iend - ip), zcs->blockSize);
5960
+ size_t const iSize = inputBuffered ? zcs->inBuffPos - zcs->inToCompress
5961
+ : MIN((size_t)(iend - ip), zcs->blockSize);
5406
5962
  if (oSize >= ZSTD_compressBound(iSize) || zcs->appliedParams.outBufferMode == ZSTD_bm_stable)
5407
5963
  cDst = op; /* compress into output buffer, to skip flush stage */
5408
5964
  else
@@ -5410,9 +5966,9 @@ static size_t ZSTD_compressStream_generic(ZSTD_CStream* zcs,
5410
5966
  if (inputBuffered) {
5411
5967
  unsigned const lastBlock = (flushMode == ZSTD_e_end) && (ip==iend);
5412
5968
  cSize = lastBlock ?
5413
- ZSTD_compressEnd(zcs, cDst, oSize,
5969
+ ZSTD_compressEnd_public(zcs, cDst, oSize,
5414
5970
  zcs->inBuff + zcs->inToCompress, iSize) :
5415
- ZSTD_compressContinue(zcs, cDst, oSize,
5971
+ ZSTD_compressContinue_public(zcs, cDst, oSize,
5416
5972
  zcs->inBuff + zcs->inToCompress, iSize);
5417
5973
  FORWARD_IF_ERROR(cSize, "%s", lastBlock ? "ZSTD_compressEnd failed" : "ZSTD_compressContinue failed");
5418
5974
  zcs->frameEnded = lastBlock;
@@ -5425,19 +5981,16 @@ static size_t ZSTD_compressStream_generic(ZSTD_CStream* zcs,
5425
5981
  if (!lastBlock)
5426
5982
  assert(zcs->inBuffTarget <= zcs->inBuffSize);
5427
5983
  zcs->inToCompress = zcs->inBuffPos;
5428
- } else {
5429
- unsigned const lastBlock = (ip + iSize == iend);
5430
- assert(flushMode == ZSTD_e_end /* Already validated */);
5984
+ } else { /* !inputBuffered, hence ZSTD_bm_stable */
5985
+ unsigned const lastBlock = (flushMode == ZSTD_e_end) && (ip + iSize == iend);
5431
5986
  cSize = lastBlock ?
5432
- ZSTD_compressEnd(zcs, cDst, oSize, ip, iSize) :
5433
- ZSTD_compressContinue(zcs, cDst, oSize, ip, iSize);
5987
+ ZSTD_compressEnd_public(zcs, cDst, oSize, ip, iSize) :
5988
+ ZSTD_compressContinue_public(zcs, cDst, oSize, ip, iSize);
5434
5989
  /* Consume the input prior to error checking to mirror buffered mode. */
5435
- if (iSize > 0)
5436
- ip += iSize;
5990
+ if (ip) ip += iSize;
5437
5991
  FORWARD_IF_ERROR(cSize, "%s", lastBlock ? "ZSTD_compressEnd failed" : "ZSTD_compressContinue failed");
5438
5992
  zcs->frameEnded = lastBlock;
5439
- if (lastBlock)
5440
- assert(ip == iend);
5993
+ if (lastBlock) assert(ip == iend);
5441
5994
  }
5442
5995
  if (cDst == op) { /* no need to flush */
5443
5996
  op += cSize;
@@ -5513,8 +6066,10 @@ size_t ZSTD_compressStream(ZSTD_CStream* zcs, ZSTD_outBuffer* output, ZSTD_inBuf
5513
6066
  /* After a compression call set the expected input/output buffer.
5514
6067
  * This is validated at the start of the next compression call.
5515
6068
  */
5516
- static void ZSTD_setBufferExpectations(ZSTD_CCtx* cctx, ZSTD_outBuffer const* output, ZSTD_inBuffer const* input)
6069
+ static void
6070
+ ZSTD_setBufferExpectations(ZSTD_CCtx* cctx, const ZSTD_outBuffer* output, const ZSTD_inBuffer* input)
5517
6071
  {
6072
+ DEBUGLOG(5, "ZSTD_setBufferExpectations (for advanced stable in/out modes)");
5518
6073
  if (cctx->appliedParams.inBufferMode == ZSTD_bm_stable) {
5519
6074
  cctx->expectedInBuffer = *input;
5520
6075
  }
@@ -5533,22 +6088,22 @@ static size_t ZSTD_checkBufferStability(ZSTD_CCtx const* cctx,
5533
6088
  {
5534
6089
  if (cctx->appliedParams.inBufferMode == ZSTD_bm_stable) {
5535
6090
  ZSTD_inBuffer const expect = cctx->expectedInBuffer;
5536
- if (expect.src != input->src || expect.pos != input->pos || expect.size != input->size)
5537
- RETURN_ERROR(srcBuffer_wrong, "ZSTD_c_stableInBuffer enabled but input differs!");
5538
- if (endOp != ZSTD_e_end)
5539
- RETURN_ERROR(srcBuffer_wrong, "ZSTD_c_stableInBuffer can only be used with ZSTD_e_end!");
6091
+ if (expect.src != input->src || expect.pos != input->pos)
6092
+ RETURN_ERROR(stabilityCondition_notRespected, "ZSTD_c_stableInBuffer enabled but input differs!");
5540
6093
  }
6094
+ (void)endOp;
5541
6095
  if (cctx->appliedParams.outBufferMode == ZSTD_bm_stable) {
5542
6096
  size_t const outBufferSize = output->size - output->pos;
5543
6097
  if (cctx->expectedOutBufferSize != outBufferSize)
5544
- RETURN_ERROR(dstBuffer_wrong, "ZSTD_c_stableOutBuffer enabled but output size differs!");
6098
+ RETURN_ERROR(stabilityCondition_notRespected, "ZSTD_c_stableOutBuffer enabled but output size differs!");
5545
6099
  }
5546
6100
  return 0;
5547
6101
  }
5548
6102
 
5549
6103
  static size_t ZSTD_CCtx_init_compressStream2(ZSTD_CCtx* cctx,
5550
6104
  ZSTD_EndDirective endOp,
5551
- size_t inSize) {
6105
+ size_t inSize)
6106
+ {
5552
6107
  ZSTD_CCtx_params params = cctx->requestedParams;
5553
6108
  ZSTD_prefixDict const prefixDict = cctx->prefixDict;
5554
6109
  FORWARD_IF_ERROR( ZSTD_initLocalDict(cctx) , ""); /* Init the local dict if present. */
@@ -5562,9 +6117,9 @@ static size_t ZSTD_CCtx_init_compressStream2(ZSTD_CCtx* cctx,
5562
6117
  params.compressionLevel = cctx->cdict->compressionLevel;
5563
6118
  }
5564
6119
  DEBUGLOG(4, "ZSTD_compressStream2 : transparent init stage");
5565
- if (endOp == ZSTD_e_end) cctx->pledgedSrcSizePlusOne = inSize + 1; /* auto-fix pledgedSrcSize */
5566
- {
5567
- size_t const dictSize = prefixDict.dict
6120
+ if (endOp == ZSTD_e_end) cctx->pledgedSrcSizePlusOne = inSize + 1; /* auto-determine pledgedSrcSize */
6121
+
6122
+ { size_t const dictSize = prefixDict.dict
5568
6123
  ? prefixDict.dictSize
5569
6124
  : (cctx->cdict ? cctx->cdict->dictContentSize : 0);
5570
6125
  ZSTD_cParamMode_e const mode = ZSTD_getCParamMode(cctx->cdict, &params, cctx->pledgedSrcSizePlusOne - 1);
@@ -5576,8 +6131,18 @@ static size_t ZSTD_CCtx_init_compressStream2(ZSTD_CCtx* cctx,
5576
6131
  params.useBlockSplitter = ZSTD_resolveBlockSplitterMode(params.useBlockSplitter, &params.cParams);
5577
6132
  params.ldmParams.enableLdm = ZSTD_resolveEnableLdm(params.ldmParams.enableLdm, &params.cParams);
5578
6133
  params.useRowMatchFinder = ZSTD_resolveRowMatchFinderMode(params.useRowMatchFinder, &params.cParams);
6134
+ params.validateSequences = ZSTD_resolveExternalSequenceValidation(params.validateSequences);
6135
+ params.maxBlockSize = ZSTD_resolveMaxBlockSize(params.maxBlockSize);
6136
+ params.searchForExternalRepcodes = ZSTD_resolveExternalRepcodeSearch(params.searchForExternalRepcodes, params.compressionLevel);
5579
6137
 
5580
6138
  #ifdef ZSTD_MULTITHREAD
6139
+ /* If external matchfinder is enabled, make sure to fail before checking job size (for consistency) */
6140
+ RETURN_ERROR_IF(
6141
+ params.useSequenceProducer == 1 && params.nbWorkers >= 1,
6142
+ parameter_combination_unsupported,
6143
+ "External sequence producer isn't supported with nbWorkers >= 1"
6144
+ );
6145
+
5581
6146
  if ((cctx->pledgedSrcSizePlusOne-1) <= ZSTDMT_JOBSIZE_MIN) {
5582
6147
  params.nbWorkers = 0; /* do not invoke multi-threading when src size is too small */
5583
6148
  }
@@ -5605,7 +6170,7 @@ static size_t ZSTD_CCtx_init_compressStream2(ZSTD_CCtx* cctx,
5605
6170
  cctx->streamStage = zcss_load;
5606
6171
  cctx->appliedParams = params;
5607
6172
  } else
5608
- #endif
6173
+ #endif /* ZSTD_MULTITHREAD */
5609
6174
  { U64 const pledgedSrcSize = cctx->pledgedSrcSizePlusOne - 1;
5610
6175
  assert(!ZSTD_isError(ZSTD_checkCParams(params.cParams)));
5611
6176
  FORWARD_IF_ERROR( ZSTD_compressBegin_internal(cctx,
@@ -5631,6 +6196,8 @@ static size_t ZSTD_CCtx_init_compressStream2(ZSTD_CCtx* cctx,
5631
6196
  return 0;
5632
6197
  }
5633
6198
 
6199
+ /* @return provides a minimum amount of data remaining to be flushed from internal buffers
6200
+ */
5634
6201
  size_t ZSTD_compressStream2( ZSTD_CCtx* cctx,
5635
6202
  ZSTD_outBuffer* output,
5636
6203
  ZSTD_inBuffer* input,
@@ -5645,8 +6212,27 @@ size_t ZSTD_compressStream2( ZSTD_CCtx* cctx,
5645
6212
 
5646
6213
  /* transparent initialization stage */
5647
6214
  if (cctx->streamStage == zcss_init) {
5648
- FORWARD_IF_ERROR(ZSTD_CCtx_init_compressStream2(cctx, endOp, input->size), "CompressStream2 initialization failed");
5649
- ZSTD_setBufferExpectations(cctx, output, input); /* Set initial buffer expectations now that we've initialized */
6215
+ size_t const inputSize = input->size - input->pos; /* no obligation to start from pos==0 */
6216
+ size_t const totalInputSize = inputSize + cctx->stableIn_notConsumed;
6217
+ if ( (cctx->requestedParams.inBufferMode == ZSTD_bm_stable) /* input is presumed stable, across invocations */
6218
+ && (endOp == ZSTD_e_continue) /* no flush requested, more input to come */
6219
+ && (totalInputSize < ZSTD_BLOCKSIZE_MAX) ) { /* not even reached one block yet */
6220
+ if (cctx->stableIn_notConsumed) { /* not the first time */
6221
+ /* check stable source guarantees */
6222
+ RETURN_ERROR_IF(input->src != cctx->expectedInBuffer.src, stabilityCondition_notRespected, "stableInBuffer condition not respected: wrong src pointer");
6223
+ RETURN_ERROR_IF(input->pos != cctx->expectedInBuffer.size, stabilityCondition_notRespected, "stableInBuffer condition not respected: externally modified pos");
6224
+ }
6225
+ /* pretend input was consumed, to give a sense forward progress */
6226
+ input->pos = input->size;
6227
+ /* save stable inBuffer, for later control, and flush/end */
6228
+ cctx->expectedInBuffer = *input;
6229
+ /* but actually input wasn't consumed, so keep track of position from where compression shall resume */
6230
+ cctx->stableIn_notConsumed += inputSize;
6231
+ /* don't initialize yet, wait for the first block of flush() order, for better parameters adaptation */
6232
+ return ZSTD_FRAMEHEADERSIZE_MIN(cctx->requestedParams.format); /* at least some header to produce */
6233
+ }
6234
+ FORWARD_IF_ERROR(ZSTD_CCtx_init_compressStream2(cctx, endOp, totalInputSize), "compressStream2 initialization failed");
6235
+ ZSTD_setBufferExpectations(cctx, output, input); /* Set initial buffer expectations now that we've initialized */
5650
6236
  }
5651
6237
  /* end of transparent initialization stage */
5652
6238
 
@@ -5659,6 +6245,13 @@ size_t ZSTD_compressStream2( ZSTD_CCtx* cctx,
5659
6245
  ZSTDMT_updateCParams_whileCompressing(cctx->mtctx, &cctx->requestedParams);
5660
6246
  cctx->cParamsChanged = 0;
5661
6247
  }
6248
+ if (cctx->stableIn_notConsumed) {
6249
+ assert(cctx->appliedParams.inBufferMode == ZSTD_bm_stable);
6250
+ /* some early data was skipped - make it available for consumption */
6251
+ assert(input->pos >= cctx->stableIn_notConsumed);
6252
+ input->pos -= cctx->stableIn_notConsumed;
6253
+ cctx->stableIn_notConsumed = 0;
6254
+ }
5662
6255
  for (;;) {
5663
6256
  size_t const ipos = input->pos;
5664
6257
  size_t const opos = output->pos;
@@ -5697,7 +6290,7 @@ size_t ZSTD_compressStream2( ZSTD_CCtx* cctx,
5697
6290
  ZSTD_setBufferExpectations(cctx, output, input);
5698
6291
  return flushMin;
5699
6292
  }
5700
- #endif
6293
+ #endif /* ZSTD_MULTITHREAD */
5701
6294
  FORWARD_IF_ERROR( ZSTD_compressStream_generic(cctx, output, input, endOp) , "");
5702
6295
  DEBUGLOG(5, "completed ZSTD_compressStream2");
5703
6296
  ZSTD_setBufferExpectations(cctx, output, input);
@@ -5710,13 +6303,20 @@ size_t ZSTD_compressStream2_simpleArgs (
5710
6303
  const void* src, size_t srcSize, size_t* srcPos,
5711
6304
  ZSTD_EndDirective endOp)
5712
6305
  {
5713
- ZSTD_outBuffer output = { dst, dstCapacity, *dstPos };
5714
- ZSTD_inBuffer input = { src, srcSize, *srcPos };
6306
+ ZSTD_outBuffer output;
6307
+ ZSTD_inBuffer input;
6308
+ output.dst = dst;
6309
+ output.size = dstCapacity;
6310
+ output.pos = *dstPos;
6311
+ input.src = src;
6312
+ input.size = srcSize;
6313
+ input.pos = *srcPos;
5715
6314
  /* ZSTD_compressStream2() will check validity of dstPos and srcPos */
5716
- size_t const cErr = ZSTD_compressStream2(cctx, &output, &input, endOp);
5717
- *dstPos = output.pos;
5718
- *srcPos = input.pos;
5719
- return cErr;
6315
+ { size_t const cErr = ZSTD_compressStream2(cctx, &output, &input, endOp);
6316
+ *dstPos = output.pos;
6317
+ *srcPos = input.pos;
6318
+ return cErr;
6319
+ }
5720
6320
  }
5721
6321
 
5722
6322
  size_t ZSTD_compress2(ZSTD_CCtx* cctx,
@@ -5739,6 +6339,7 @@ size_t ZSTD_compress2(ZSTD_CCtx* cctx,
5739
6339
  /* Reset to the original values. */
5740
6340
  cctx->requestedParams.inBufferMode = originalInBufferMode;
5741
6341
  cctx->requestedParams.outBufferMode = originalOutBufferMode;
6342
+
5742
6343
  FORWARD_IF_ERROR(result, "ZSTD_compressStream2_simpleArgs failed");
5743
6344
  if (result != 0) { /* compression not completed, due to lack of output space */
5744
6345
  assert(oPos == dstCapacity);
@@ -5749,64 +6350,61 @@ size_t ZSTD_compress2(ZSTD_CCtx* cctx,
5749
6350
  }
5750
6351
  }
5751
6352
 
5752
- typedef struct {
5753
- U32 idx; /* Index in array of ZSTD_Sequence */
5754
- U32 posInSequence; /* Position within sequence at idx */
5755
- size_t posInSrc; /* Number of bytes given by sequences provided so far */
5756
- } ZSTD_sequencePosition;
5757
-
5758
6353
  /* ZSTD_validateSequence() :
5759
6354
  * @offCode : is presumed to follow format required by ZSTD_storeSeq()
5760
6355
  * @returns a ZSTD error code if sequence is not valid
5761
6356
  */
5762
6357
  static size_t
5763
- ZSTD_validateSequence(U32 offCode, U32 matchLength,
5764
- size_t posInSrc, U32 windowLog, size_t dictSize)
6358
+ ZSTD_validateSequence(U32 offCode, U32 matchLength, U32 minMatch,
6359
+ size_t posInSrc, U32 windowLog, size_t dictSize, int useSequenceProducer)
5765
6360
  {
5766
- U32 const windowSize = 1 << windowLog;
5767
- /* posInSrc represents the amount of data the the decoder would decode up to this point.
6361
+ U32 const windowSize = 1u << windowLog;
6362
+ /* posInSrc represents the amount of data the decoder would decode up to this point.
5768
6363
  * As long as the amount of data decoded is less than or equal to window size, offsets may be
5769
6364
  * larger than the total length of output decoded in order to reference the dict, even larger than
5770
6365
  * window size. After output surpasses windowSize, we're limited to windowSize offsets again.
5771
6366
  */
5772
6367
  size_t const offsetBound = posInSrc > windowSize ? (size_t)windowSize : posInSrc + (size_t)dictSize;
5773
- RETURN_ERROR_IF(offCode > STORE_OFFSET(offsetBound), corruption_detected, "Offset too large!");
5774
- RETURN_ERROR_IF(matchLength < MINMATCH, corruption_detected, "Matchlength too small");
6368
+ size_t const matchLenLowerBound = (minMatch == 3 || useSequenceProducer) ? 3 : 4;
6369
+ RETURN_ERROR_IF(offCode > OFFSET_TO_OFFBASE(offsetBound), externalSequences_invalid, "Offset too large!");
6370
+ /* Validate maxNbSeq is large enough for the given matchLength and minMatch */
6371
+ RETURN_ERROR_IF(matchLength < matchLenLowerBound, externalSequences_invalid, "Matchlength too small for the minMatch");
5775
6372
  return 0;
5776
6373
  }
5777
6374
 
5778
6375
  /* Returns an offset code, given a sequence's raw offset, the ongoing repcode array, and whether litLength == 0 */
5779
- static U32 ZSTD_finalizeOffCode(U32 rawOffset, const U32 rep[ZSTD_REP_NUM], U32 ll0)
6376
+ static U32 ZSTD_finalizeOffBase(U32 rawOffset, const U32 rep[ZSTD_REP_NUM], U32 ll0)
5780
6377
  {
5781
- U32 offCode = STORE_OFFSET(rawOffset);
6378
+ U32 offBase = OFFSET_TO_OFFBASE(rawOffset);
5782
6379
 
5783
6380
  if (!ll0 && rawOffset == rep[0]) {
5784
- offCode = STORE_REPCODE_1;
6381
+ offBase = REPCODE1_TO_OFFBASE;
5785
6382
  } else if (rawOffset == rep[1]) {
5786
- offCode = STORE_REPCODE(2 - ll0);
6383
+ offBase = REPCODE_TO_OFFBASE(2 - ll0);
5787
6384
  } else if (rawOffset == rep[2]) {
5788
- offCode = STORE_REPCODE(3 - ll0);
6385
+ offBase = REPCODE_TO_OFFBASE(3 - ll0);
5789
6386
  } else if (ll0 && rawOffset == rep[0] - 1) {
5790
- offCode = STORE_REPCODE_3;
6387
+ offBase = REPCODE3_TO_OFFBASE;
5791
6388
  }
5792
- return offCode;
6389
+ return offBase;
5793
6390
  }
5794
6391
 
5795
- /* Returns 0 on success, and a ZSTD_error otherwise. This function scans through an array of
5796
- * ZSTD_Sequence, storing the sequences it finds, until it reaches a block delimiter.
5797
- */
5798
- static size_t
6392
+ size_t
5799
6393
  ZSTD_copySequencesToSeqStoreExplicitBlockDelim(ZSTD_CCtx* cctx,
5800
6394
  ZSTD_sequencePosition* seqPos,
5801
6395
  const ZSTD_Sequence* const inSeqs, size_t inSeqsSize,
5802
- const void* src, size_t blockSize)
6396
+ const void* src, size_t blockSize,
6397
+ ZSTD_paramSwitch_e externalRepSearch)
5803
6398
  {
5804
6399
  U32 idx = seqPos->idx;
6400
+ U32 const startIdx = idx;
5805
6401
  BYTE const* ip = (BYTE const*)(src);
5806
6402
  const BYTE* const iend = ip + blockSize;
5807
6403
  repcodes_t updatedRepcodes;
5808
6404
  U32 dictSize;
5809
6405
 
6406
+ DEBUGLOG(5, "ZSTD_copySequencesToSeqStoreExplicitBlockDelim (blockSize = %zu)", blockSize);
6407
+
5810
6408
  if (cctx->cdict) {
5811
6409
  dictSize = (U32)cctx->cdict->dictContentSize;
5812
6410
  } else if (cctx->prefixDict.dict) {
@@ -5815,25 +6413,55 @@ ZSTD_copySequencesToSeqStoreExplicitBlockDelim(ZSTD_CCtx* cctx,
5815
6413
  dictSize = 0;
5816
6414
  }
5817
6415
  ZSTD_memcpy(updatedRepcodes.rep, cctx->blockState.prevCBlock->rep, sizeof(repcodes_t));
5818
- for (; (inSeqs[idx].matchLength != 0 || inSeqs[idx].offset != 0) && idx < inSeqsSize; ++idx) {
6416
+ for (; idx < inSeqsSize && (inSeqs[idx].matchLength != 0 || inSeqs[idx].offset != 0); ++idx) {
5819
6417
  U32 const litLength = inSeqs[idx].litLength;
5820
- U32 const ll0 = (litLength == 0);
5821
6418
  U32 const matchLength = inSeqs[idx].matchLength;
5822
- U32 const offCode = ZSTD_finalizeOffCode(inSeqs[idx].offset, updatedRepcodes.rep, ll0);
5823
- ZSTD_updateRep(updatedRepcodes.rep, offCode, ll0);
6419
+ U32 offBase;
5824
6420
 
5825
- DEBUGLOG(6, "Storing sequence: (of: %u, ml: %u, ll: %u)", offCode, matchLength, litLength);
6421
+ if (externalRepSearch == ZSTD_ps_disable) {
6422
+ offBase = OFFSET_TO_OFFBASE(inSeqs[idx].offset);
6423
+ } else {
6424
+ U32 const ll0 = (litLength == 0);
6425
+ offBase = ZSTD_finalizeOffBase(inSeqs[idx].offset, updatedRepcodes.rep, ll0);
6426
+ ZSTD_updateRep(updatedRepcodes.rep, offBase, ll0);
6427
+ }
6428
+
6429
+ DEBUGLOG(6, "Storing sequence: (of: %u, ml: %u, ll: %u)", offBase, matchLength, litLength);
5826
6430
  if (cctx->appliedParams.validateSequences) {
5827
6431
  seqPos->posInSrc += litLength + matchLength;
5828
- FORWARD_IF_ERROR(ZSTD_validateSequence(offCode, matchLength, seqPos->posInSrc,
5829
- cctx->appliedParams.cParams.windowLog, dictSize),
6432
+ FORWARD_IF_ERROR(ZSTD_validateSequence(offBase, matchLength, cctx->appliedParams.cParams.minMatch, seqPos->posInSrc,
6433
+ cctx->appliedParams.cParams.windowLog, dictSize, cctx->appliedParams.useSequenceProducer),
5830
6434
  "Sequence validation failed");
5831
6435
  }
5832
- RETURN_ERROR_IF(idx - seqPos->idx > cctx->seqStore.maxNbSeq, memory_allocation,
6436
+ RETURN_ERROR_IF(idx - seqPos->idx >= cctx->seqStore.maxNbSeq, externalSequences_invalid,
5833
6437
  "Not enough memory allocated. Try adjusting ZSTD_c_minMatch.");
5834
- ZSTD_storeSeq(&cctx->seqStore, litLength, ip, iend, offCode, matchLength);
6438
+ ZSTD_storeSeq(&cctx->seqStore, litLength, ip, iend, offBase, matchLength);
5835
6439
  ip += matchLength + litLength;
5836
6440
  }
6441
+
6442
+ /* If we skipped repcode search while parsing, we need to update repcodes now */
6443
+ assert(externalRepSearch != ZSTD_ps_auto);
6444
+ assert(idx >= startIdx);
6445
+ if (externalRepSearch == ZSTD_ps_disable && idx != startIdx) {
6446
+ U32* const rep = updatedRepcodes.rep;
6447
+ U32 lastSeqIdx = idx - 1; /* index of last non-block-delimiter sequence */
6448
+
6449
+ if (lastSeqIdx >= startIdx + 2) {
6450
+ rep[2] = inSeqs[lastSeqIdx - 2].offset;
6451
+ rep[1] = inSeqs[lastSeqIdx - 1].offset;
6452
+ rep[0] = inSeqs[lastSeqIdx].offset;
6453
+ } else if (lastSeqIdx == startIdx + 1) {
6454
+ rep[2] = rep[0];
6455
+ rep[1] = inSeqs[lastSeqIdx - 1].offset;
6456
+ rep[0] = inSeqs[lastSeqIdx].offset;
6457
+ } else {
6458
+ assert(lastSeqIdx == startIdx);
6459
+ rep[2] = rep[1];
6460
+ rep[1] = rep[0];
6461
+ rep[0] = inSeqs[lastSeqIdx].offset;
6462
+ }
6463
+ }
6464
+
5837
6465
  ZSTD_memcpy(cctx->blockState.nextCBlock->rep, updatedRepcodes.rep, sizeof(repcodes_t));
5838
6466
 
5839
6467
  if (inSeqs[idx].litLength) {
@@ -5842,26 +6470,15 @@ ZSTD_copySequencesToSeqStoreExplicitBlockDelim(ZSTD_CCtx* cctx,
5842
6470
  ip += inSeqs[idx].litLength;
5843
6471
  seqPos->posInSrc += inSeqs[idx].litLength;
5844
6472
  }
5845
- RETURN_ERROR_IF(ip != iend, corruption_detected, "Blocksize doesn't agree with block delimiter!");
6473
+ RETURN_ERROR_IF(ip != iend, externalSequences_invalid, "Blocksize doesn't agree with block delimiter!");
5846
6474
  seqPos->idx = idx+1;
5847
6475
  return 0;
5848
6476
  }
5849
6477
 
5850
- /* Returns the number of bytes to move the current read position back by. Only non-zero
5851
- * if we ended up splitting a sequence. Otherwise, it may return a ZSTD error if something
5852
- * went wrong.
5853
- *
5854
- * This function will attempt to scan through blockSize bytes represented by the sequences
5855
- * in inSeqs, storing any (partial) sequences.
5856
- *
5857
- * Occasionally, we may want to change the actual number of bytes we consumed from inSeqs to
5858
- * avoid splitting a match, or to avoid splitting a match such that it would produce a match
5859
- * smaller than MINMATCH. In this case, we return the number of bytes that we didn't read from this block.
5860
- */
5861
- static size_t
6478
+ size_t
5862
6479
  ZSTD_copySequencesToSeqStoreNoBlockDelim(ZSTD_CCtx* cctx, ZSTD_sequencePosition* seqPos,
5863
6480
  const ZSTD_Sequence* const inSeqs, size_t inSeqsSize,
5864
- const void* src, size_t blockSize)
6481
+ const void* src, size_t blockSize, ZSTD_paramSwitch_e externalRepSearch)
5865
6482
  {
5866
6483
  U32 idx = seqPos->idx;
5867
6484
  U32 startPosInSequence = seqPos->posInSequence;
@@ -5873,6 +6490,9 @@ ZSTD_copySequencesToSeqStoreNoBlockDelim(ZSTD_CCtx* cctx, ZSTD_sequencePosition*
5873
6490
  U32 bytesAdjustment = 0;
5874
6491
  U32 finalMatchSplit = 0;
5875
6492
 
6493
+ /* TODO(embg) support fast parsing mode in noBlockDelim mode */
6494
+ (void)externalRepSearch;
6495
+
5876
6496
  if (cctx->cdict) {
5877
6497
  dictSize = cctx->cdict->dictContentSize;
5878
6498
  } else if (cctx->prefixDict.dict) {
@@ -5880,7 +6500,7 @@ ZSTD_copySequencesToSeqStoreNoBlockDelim(ZSTD_CCtx* cctx, ZSTD_sequencePosition*
5880
6500
  } else {
5881
6501
  dictSize = 0;
5882
6502
  }
5883
- DEBUGLOG(5, "ZSTD_copySequencesToSeqStore: idx: %u PIS: %u blockSize: %zu", idx, startPosInSequence, blockSize);
6503
+ DEBUGLOG(5, "ZSTD_copySequencesToSeqStoreNoBlockDelim: idx: %u PIS: %u blockSize: %zu", idx, startPosInSequence, blockSize);
5884
6504
  DEBUGLOG(5, "Start seq: idx: %u (of: %u ml: %u ll: %u)", idx, inSeqs[idx].offset, inSeqs[idx].matchLength, inSeqs[idx].litLength);
5885
6505
  ZSTD_memcpy(updatedRepcodes.rep, cctx->blockState.prevCBlock->rep, sizeof(repcodes_t));
5886
6506
  while (endPosInSequence && idx < inSeqsSize && !finalMatchSplit) {
@@ -5888,7 +6508,7 @@ ZSTD_copySequencesToSeqStoreNoBlockDelim(ZSTD_CCtx* cctx, ZSTD_sequencePosition*
5888
6508
  U32 litLength = currSeq.litLength;
5889
6509
  U32 matchLength = currSeq.matchLength;
5890
6510
  U32 const rawOffset = currSeq.offset;
5891
- U32 offCode;
6511
+ U32 offBase;
5892
6512
 
5893
6513
  /* Modify the sequence depending on where endPosInSequence lies */
5894
6514
  if (endPosInSequence >= currSeq.litLength + currSeq.matchLength) {
@@ -5902,7 +6522,6 @@ ZSTD_copySequencesToSeqStoreNoBlockDelim(ZSTD_CCtx* cctx, ZSTD_sequencePosition*
5902
6522
  /* Move to the next sequence */
5903
6523
  endPosInSequence -= currSeq.litLength + currSeq.matchLength;
5904
6524
  startPosInSequence = 0;
5905
- idx++;
5906
6525
  } else {
5907
6526
  /* This is the final (partial) sequence we're adding from inSeqs, and endPosInSequence
5908
6527
  does not reach the end of the match. So, we have to split the sequence */
@@ -5942,21 +6561,23 @@ ZSTD_copySequencesToSeqStoreNoBlockDelim(ZSTD_CCtx* cctx, ZSTD_sequencePosition*
5942
6561
  }
5943
6562
  /* Check if this offset can be represented with a repcode */
5944
6563
  { U32 const ll0 = (litLength == 0);
5945
- offCode = ZSTD_finalizeOffCode(rawOffset, updatedRepcodes.rep, ll0);
5946
- ZSTD_updateRep(updatedRepcodes.rep, offCode, ll0);
6564
+ offBase = ZSTD_finalizeOffBase(rawOffset, updatedRepcodes.rep, ll0);
6565
+ ZSTD_updateRep(updatedRepcodes.rep, offBase, ll0);
5947
6566
  }
5948
6567
 
5949
6568
  if (cctx->appliedParams.validateSequences) {
5950
6569
  seqPos->posInSrc += litLength + matchLength;
5951
- FORWARD_IF_ERROR(ZSTD_validateSequence(offCode, matchLength, seqPos->posInSrc,
5952
- cctx->appliedParams.cParams.windowLog, dictSize),
6570
+ FORWARD_IF_ERROR(ZSTD_validateSequence(offBase, matchLength, cctx->appliedParams.cParams.minMatch, seqPos->posInSrc,
6571
+ cctx->appliedParams.cParams.windowLog, dictSize, cctx->appliedParams.useSequenceProducer),
5953
6572
  "Sequence validation failed");
5954
6573
  }
5955
- DEBUGLOG(6, "Storing sequence: (of: %u, ml: %u, ll: %u)", offCode, matchLength, litLength);
5956
- RETURN_ERROR_IF(idx - seqPos->idx > cctx->seqStore.maxNbSeq, memory_allocation,
6574
+ DEBUGLOG(6, "Storing sequence: (of: %u, ml: %u, ll: %u)", offBase, matchLength, litLength);
6575
+ RETURN_ERROR_IF(idx - seqPos->idx >= cctx->seqStore.maxNbSeq, externalSequences_invalid,
5957
6576
  "Not enough memory allocated. Try adjusting ZSTD_c_minMatch.");
5958
- ZSTD_storeSeq(&cctx->seqStore, litLength, ip, iend, offCode, matchLength);
6577
+ ZSTD_storeSeq(&cctx->seqStore, litLength, ip, iend, offBase, matchLength);
5959
6578
  ip += matchLength + litLength;
6579
+ if (!finalMatchSplit)
6580
+ idx++; /* Next Sequence */
5960
6581
  }
5961
6582
  DEBUGLOG(5, "Ending seq: idx: %u (of: %u ml: %u ll: %u)", idx, inSeqs[idx].offset, inSeqs[idx].matchLength, inSeqs[idx].litLength);
5962
6583
  assert(idx == inSeqsSize || endPosInSequence <= inSeqs[idx].litLength + inSeqs[idx].matchLength);
@@ -5979,7 +6600,7 @@ ZSTD_copySequencesToSeqStoreNoBlockDelim(ZSTD_CCtx* cctx, ZSTD_sequencePosition*
5979
6600
 
5980
6601
  typedef size_t (*ZSTD_sequenceCopier) (ZSTD_CCtx* cctx, ZSTD_sequencePosition* seqPos,
5981
6602
  const ZSTD_Sequence* const inSeqs, size_t inSeqsSize,
5982
- const void* src, size_t blockSize);
6603
+ const void* src, size_t blockSize, ZSTD_paramSwitch_e externalRepSearch);
5983
6604
  static ZSTD_sequenceCopier ZSTD_selectSequenceCopier(ZSTD_sequenceFormat_e mode)
5984
6605
  {
5985
6606
  ZSTD_sequenceCopier sequenceCopier = NULL;
@@ -5993,6 +6614,57 @@ static ZSTD_sequenceCopier ZSTD_selectSequenceCopier(ZSTD_sequenceFormat_e mode)
5993
6614
  return sequenceCopier;
5994
6615
  }
5995
6616
 
6617
+ /* Discover the size of next block by searching for the delimiter.
6618
+ * Note that a block delimiter **must** exist in this mode,
6619
+ * otherwise it's an input error.
6620
+ * The block size retrieved will be later compared to ensure it remains within bounds */
6621
+ static size_t
6622
+ blockSize_explicitDelimiter(const ZSTD_Sequence* inSeqs, size_t inSeqsSize, ZSTD_sequencePosition seqPos)
6623
+ {
6624
+ int end = 0;
6625
+ size_t blockSize = 0;
6626
+ size_t spos = seqPos.idx;
6627
+ DEBUGLOG(6, "blockSize_explicitDelimiter : seq %zu / %zu", spos, inSeqsSize);
6628
+ assert(spos <= inSeqsSize);
6629
+ while (spos < inSeqsSize) {
6630
+ end = (inSeqs[spos].offset == 0);
6631
+ blockSize += inSeqs[spos].litLength + inSeqs[spos].matchLength;
6632
+ if (end) {
6633
+ if (inSeqs[spos].matchLength != 0)
6634
+ RETURN_ERROR(externalSequences_invalid, "delimiter format error : both matchlength and offset must be == 0");
6635
+ break;
6636
+ }
6637
+ spos++;
6638
+ }
6639
+ if (!end)
6640
+ RETURN_ERROR(externalSequences_invalid, "Reached end of sequences without finding a block delimiter");
6641
+ return blockSize;
6642
+ }
6643
+
6644
+ /* More a "target" block size */
6645
+ static size_t blockSize_noDelimiter(size_t blockSize, size_t remaining)
6646
+ {
6647
+ int const lastBlock = (remaining <= blockSize);
6648
+ return lastBlock ? remaining : blockSize;
6649
+ }
6650
+
6651
+ static size_t determine_blockSize(ZSTD_sequenceFormat_e mode,
6652
+ size_t blockSize, size_t remaining,
6653
+ const ZSTD_Sequence* inSeqs, size_t inSeqsSize, ZSTD_sequencePosition seqPos)
6654
+ {
6655
+ DEBUGLOG(6, "determine_blockSize : remainingSize = %zu", remaining);
6656
+ if (mode == ZSTD_sf_noBlockDelimiters)
6657
+ return blockSize_noDelimiter(blockSize, remaining);
6658
+ { size_t const explicitBlockSize = blockSize_explicitDelimiter(inSeqs, inSeqsSize, seqPos);
6659
+ FORWARD_IF_ERROR(explicitBlockSize, "Error while determining block size with explicit delimiters");
6660
+ if (explicitBlockSize > blockSize)
6661
+ RETURN_ERROR(externalSequences_invalid, "sequences incorrectly define a too large block");
6662
+ if (explicitBlockSize > remaining)
6663
+ RETURN_ERROR(externalSequences_invalid, "sequences define a frame longer than source");
6664
+ return explicitBlockSize;
6665
+ }
6666
+ }
6667
+
5996
6668
  /* Compress, block-by-block, all of the sequences given.
5997
6669
  *
5998
6670
  * Returns the cumulative size of all compressed blocks (including their headers),
@@ -6005,9 +6677,6 @@ ZSTD_compressSequences_internal(ZSTD_CCtx* cctx,
6005
6677
  const void* src, size_t srcSize)
6006
6678
  {
6007
6679
  size_t cSize = 0;
6008
- U32 lastBlock;
6009
- size_t blockSize;
6010
- size_t compressedSeqsSize;
6011
6680
  size_t remaining = srcSize;
6012
6681
  ZSTD_sequencePosition seqPos = {0, 0, 0};
6013
6682
 
@@ -6027,22 +6696,29 @@ ZSTD_compressSequences_internal(ZSTD_CCtx* cctx,
6027
6696
  }
6028
6697
 
6029
6698
  while (remaining) {
6699
+ size_t compressedSeqsSize;
6030
6700
  size_t cBlockSize;
6031
6701
  size_t additionalByteAdjustment;
6032
- lastBlock = remaining <= cctx->blockSize;
6033
- blockSize = lastBlock ? (U32)remaining : (U32)cctx->blockSize;
6702
+ size_t blockSize = determine_blockSize(cctx->appliedParams.blockDelimiters,
6703
+ cctx->blockSize, remaining,
6704
+ inSeqs, inSeqsSize, seqPos);
6705
+ U32 const lastBlock = (blockSize == remaining);
6706
+ FORWARD_IF_ERROR(blockSize, "Error while trying to determine block size");
6707
+ assert(blockSize <= remaining);
6034
6708
  ZSTD_resetSeqStore(&cctx->seqStore);
6035
- DEBUGLOG(4, "Working on new block. Blocksize: %zu", blockSize);
6709
+ DEBUGLOG(5, "Working on new block. Blocksize: %zu (total:%zu)", blockSize, (ip - (const BYTE*)src) + blockSize);
6036
6710
 
6037
- additionalByteAdjustment = sequenceCopier(cctx, &seqPos, inSeqs, inSeqsSize, ip, blockSize);
6711
+ additionalByteAdjustment = sequenceCopier(cctx, &seqPos, inSeqs, inSeqsSize, ip, blockSize, cctx->appliedParams.searchForExternalRepcodes);
6038
6712
  FORWARD_IF_ERROR(additionalByteAdjustment, "Bad sequence copy");
6039
6713
  blockSize -= additionalByteAdjustment;
6040
6714
 
6041
6715
  /* If blocks are too small, emit as a nocompress block */
6042
- if (blockSize < MIN_CBLOCK_SIZE+ZSTD_blockHeaderSize+1) {
6716
+ /* TODO: See 3090. We reduced MIN_CBLOCK_SIZE from 3 to 2 so to compensate we are adding
6717
+ * additional 1. We need to revisit and change this logic to be more consistent */
6718
+ if (blockSize < MIN_CBLOCK_SIZE+ZSTD_blockHeaderSize+1+1) {
6043
6719
  cBlockSize = ZSTD_noCompressBlock(op, dstCapacity, ip, blockSize, lastBlock);
6044
6720
  FORWARD_IF_ERROR(cBlockSize, "Nocompress block failed");
6045
- DEBUGLOG(4, "Block too small, writing out nocompress block: cSize: %zu", cBlockSize);
6721
+ DEBUGLOG(5, "Block too small, writing out nocompress block: cSize: %zu", cBlockSize);
6046
6722
  cSize += cBlockSize;
6047
6723
  ip += blockSize;
6048
6724
  op += cBlockSize;
@@ -6051,6 +6727,7 @@ ZSTD_compressSequences_internal(ZSTD_CCtx* cctx,
6051
6727
  continue;
6052
6728
  }
6053
6729
 
6730
+ RETURN_ERROR_IF(dstCapacity < ZSTD_blockHeaderSize, dstSize_tooSmall, "not enough dstCapacity to write a new compressed block");
6054
6731
  compressedSeqsSize = ZSTD_entropyCompressSeqStore(&cctx->seqStore,
6055
6732
  &cctx->blockState.prevCBlock->entropy, &cctx->blockState.nextCBlock->entropy,
6056
6733
  &cctx->appliedParams,
@@ -6059,11 +6736,11 @@ ZSTD_compressSequences_internal(ZSTD_CCtx* cctx,
6059
6736
  cctx->entropyWorkspace, ENTROPY_WORKSPACE_SIZE /* statically allocated in resetCCtx */,
6060
6737
  cctx->bmi2);
6061
6738
  FORWARD_IF_ERROR(compressedSeqsSize, "Compressing sequences of block failed");
6062
- DEBUGLOG(4, "Compressed sequences size: %zu", compressedSeqsSize);
6739
+ DEBUGLOG(5, "Compressed sequences size: %zu", compressedSeqsSize);
6063
6740
 
6064
6741
  if (!cctx->isFirstBlock &&
6065
6742
  ZSTD_maybeRLE(&cctx->seqStore) &&
6066
- ZSTD_isRLE((BYTE const*)src, srcSize)) {
6743
+ ZSTD_isRLE(ip, blockSize)) {
6067
6744
  /* We don't want to emit our first block as a RLE even if it qualifies because
6068
6745
  * doing so will cause the decoder (cli only) to throw a "should consume all input error."
6069
6746
  * This is only an issue for zstd <= v1.4.3
@@ -6074,12 +6751,12 @@ ZSTD_compressSequences_internal(ZSTD_CCtx* cctx,
6074
6751
  if (compressedSeqsSize == 0) {
6075
6752
  /* ZSTD_noCompressBlock writes the block header as well */
6076
6753
  cBlockSize = ZSTD_noCompressBlock(op, dstCapacity, ip, blockSize, lastBlock);
6077
- FORWARD_IF_ERROR(cBlockSize, "Nocompress block failed");
6078
- DEBUGLOG(4, "Writing out nocompress block, size: %zu", cBlockSize);
6754
+ FORWARD_IF_ERROR(cBlockSize, "ZSTD_noCompressBlock failed");
6755
+ DEBUGLOG(5, "Writing out nocompress block, size: %zu", cBlockSize);
6079
6756
  } else if (compressedSeqsSize == 1) {
6080
6757
  cBlockSize = ZSTD_rleCompressBlock(op, dstCapacity, *ip, blockSize, lastBlock);
6081
- FORWARD_IF_ERROR(cBlockSize, "RLE compress block failed");
6082
- DEBUGLOG(4, "Writing out RLE block, size: %zu", cBlockSize);
6758
+ FORWARD_IF_ERROR(cBlockSize, "ZSTD_rleCompressBlock failed");
6759
+ DEBUGLOG(5, "Writing out RLE block, size: %zu", cBlockSize);
6083
6760
  } else {
6084
6761
  U32 cBlockHeader;
6085
6762
  /* Error checking and repcodes update */
@@ -6091,11 +6768,10 @@ ZSTD_compressSequences_internal(ZSTD_CCtx* cctx,
6091
6768
  cBlockHeader = lastBlock + (((U32)bt_compressed)<<1) + (U32)(compressedSeqsSize << 3);
6092
6769
  MEM_writeLE24(op, cBlockHeader);
6093
6770
  cBlockSize = ZSTD_blockHeaderSize + compressedSeqsSize;
6094
- DEBUGLOG(4, "Writing out compressed block, size: %zu", cBlockSize);
6771
+ DEBUGLOG(5, "Writing out compressed block, size: %zu", cBlockSize);
6095
6772
  }
6096
6773
 
6097
6774
  cSize += cBlockSize;
6098
- DEBUGLOG(4, "cSize running total: %zu", cSize);
6099
6775
 
6100
6776
  if (lastBlock) {
6101
6777
  break;
@@ -6106,12 +6782,15 @@ ZSTD_compressSequences_internal(ZSTD_CCtx* cctx,
6106
6782
  dstCapacity -= cBlockSize;
6107
6783
  cctx->isFirstBlock = 0;
6108
6784
  }
6785
+ DEBUGLOG(5, "cSize running total: %zu (remaining dstCapacity=%zu)", cSize, dstCapacity);
6109
6786
  }
6110
6787
 
6788
+ DEBUGLOG(4, "cSize final total: %zu", cSize);
6111
6789
  return cSize;
6112
6790
  }
6113
6791
 
6114
- size_t ZSTD_compressSequences(ZSTD_CCtx* const cctx, void* dst, size_t dstCapacity,
6792
+ size_t ZSTD_compressSequences(ZSTD_CCtx* cctx,
6793
+ void* dst, size_t dstCapacity,
6115
6794
  const ZSTD_Sequence* inSeqs, size_t inSeqsSize,
6116
6795
  const void* src, size_t srcSize)
6117
6796
  {
@@ -6121,7 +6800,7 @@ size_t ZSTD_compressSequences(ZSTD_CCtx* const cctx, void* dst, size_t dstCapaci
6121
6800
  size_t frameHeaderSize = 0;
6122
6801
 
6123
6802
  /* Transparent initialization stage, same as compressStream2() */
6124
- DEBUGLOG(3, "ZSTD_compressSequences()");
6803
+ DEBUGLOG(4, "ZSTD_compressSequences (dstCapacity=%zu)", dstCapacity);
6125
6804
  assert(cctx != NULL);
6126
6805
  FORWARD_IF_ERROR(ZSTD_CCtx_init_compressStream2(cctx, ZSTD_e_end, srcSize), "CCtx initialization failed");
6127
6806
  /* Begin writing output, starting with frame header */
@@ -6149,26 +6828,34 @@ size_t ZSTD_compressSequences(ZSTD_CCtx* const cctx, void* dst, size_t dstCapaci
6149
6828
  cSize += 4;
6150
6829
  }
6151
6830
 
6152
- DEBUGLOG(3, "Final compressed size: %zu", cSize);
6831
+ DEBUGLOG(4, "Final compressed size: %zu", cSize);
6153
6832
  return cSize;
6154
6833
  }
6155
6834
 
6156
6835
  /*====== Finalize ======*/
6157
6836
 
6837
+ static ZSTD_inBuffer inBuffer_forEndFlush(const ZSTD_CStream* zcs)
6838
+ {
6839
+ const ZSTD_inBuffer nullInput = { NULL, 0, 0 };
6840
+ const int stableInput = (zcs->appliedParams.inBufferMode == ZSTD_bm_stable);
6841
+ return stableInput ? zcs->expectedInBuffer : nullInput;
6842
+ }
6843
+
6158
6844
  /*! ZSTD_flushStream() :
6159
6845
  * @return : amount of data remaining to flush */
6160
6846
  size_t ZSTD_flushStream(ZSTD_CStream* zcs, ZSTD_outBuffer* output)
6161
6847
  {
6162
- ZSTD_inBuffer input = { NULL, 0, 0 };
6848
+ ZSTD_inBuffer input = inBuffer_forEndFlush(zcs);
6849
+ input.size = input.pos; /* do not ingest more input during flush */
6163
6850
  return ZSTD_compressStream2(zcs, output, &input, ZSTD_e_flush);
6164
6851
  }
6165
6852
 
6166
6853
 
6167
6854
  size_t ZSTD_endStream(ZSTD_CStream* zcs, ZSTD_outBuffer* output)
6168
6855
  {
6169
- ZSTD_inBuffer input = { NULL, 0, 0 };
6856
+ ZSTD_inBuffer input = inBuffer_forEndFlush(zcs);
6170
6857
  size_t const remainingToFlush = ZSTD_compressStream2(zcs, output, &input, ZSTD_e_end);
6171
- FORWARD_IF_ERROR( remainingToFlush , "ZSTD_compressStream2 failed");
6858
+ FORWARD_IF_ERROR(remainingToFlush , "ZSTD_compressStream2(,,ZSTD_e_end) failed");
6172
6859
  if (zcs->appliedParams.nbWorkers > 0) return remainingToFlush; /* minimal estimation */
6173
6860
  /* single thread mode : attempt to calculate remaining to flush more precisely */
6174
6861
  { size_t const lastBlockSize = zcs->frameEnded ? 0 : ZSTD_BLOCKHEADERSIZE;
@@ -6290,7 +6977,7 @@ static ZSTD_compressionParameters ZSTD_getCParams_internal(int compressionLevel,
6290
6977
  cp.targetLength = (unsigned)(-clampedCompressionLevel);
6291
6978
  }
6292
6979
  /* refine parameters based on srcSize & dictSize */
6293
- return ZSTD_adjustCParams_internal(cp, srcSizeHint, dictSize, mode);
6980
+ return ZSTD_adjustCParams_internal(cp, srcSizeHint, dictSize, mode, ZSTD_ps_auto);
6294
6981
  }
6295
6982
  }
6296
6983
 
@@ -6325,3 +7012,21 @@ ZSTD_parameters ZSTD_getParams(int compressionLevel, unsigned long long srcSizeH
6325
7012
  if (srcSizeHint == 0) srcSizeHint = ZSTD_CONTENTSIZE_UNKNOWN;
6326
7013
  return ZSTD_getParams_internal(compressionLevel, srcSizeHint, dictSize, ZSTD_cpm_unknown);
6327
7014
  }
7015
+
7016
+ void ZSTD_registerSequenceProducer(
7017
+ ZSTD_CCtx* zc, void* mState,
7018
+ ZSTD_sequenceProducer_F* mFinder
7019
+ ) {
7020
+ if (mFinder != NULL) {
7021
+ ZSTD_externalMatchCtx emctx;
7022
+ emctx.mState = mState;
7023
+ emctx.mFinder = mFinder;
7024
+ emctx.seqBuffer = NULL;
7025
+ emctx.seqBufferCapacity = 0;
7026
+ zc->externalMatchCtx = emctx;
7027
+ zc->requestedParams.useSequenceProducer = 1;
7028
+ } else {
7029
+ ZSTD_memset(&zc->externalMatchCtx, 0, sizeof(zc->externalMatchCtx));
7030
+ zc->requestedParams.useSequenceProducer = 0;
7031
+ }
7032
+ }