zstdlib 0.7.0-x86-mingw32 → 0.10.0-x86-mingw32

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (138) hide show
  1. checksums.yaml +4 -4
  2. data/CHANGES.md +20 -0
  3. data/README.md +7 -1
  4. data/Rakefile +38 -8
  5. data/ext/{zstdlib → zstdlib_c}/extconf.rb +11 -6
  6. data/ext/{zstdlib → zstdlib_c}/ruby/zlib-2.2/zstdlib.c +2 -2
  7. data/ext/{zstdlib → zstdlib_c}/ruby/zlib-2.3/zstdlib.c +2 -2
  8. data/ext/{zstdlib → zstdlib_c}/ruby/zlib-2.4/zstdlib.c +2 -2
  9. data/ext/{zstdlib → zstdlib_c}/ruby/zlib-2.5/zstdlib.c +2 -2
  10. data/ext/{zstdlib → zstdlib_c}/ruby/zlib-2.6/zstdlib.c +2 -2
  11. data/ext/{zstdlib → zstdlib_c}/ruby/zlib-2.7/zstdlib.c +2 -2
  12. data/ext/zstdlib_c/ruby/zlib-3.0/zstdlib.c +4994 -0
  13. data/ext/zstdlib_c/ruby/zlib-3.1/zstdlib.c +5076 -0
  14. data/ext/{zstdlib/zlib-1.2.11 → zstdlib_c/zlib-1.2.12}/adler32.c +0 -0
  15. data/ext/{zstdlib/zlib-1.2.11 → zstdlib_c/zlib-1.2.12}/compress.c +0 -0
  16. data/ext/zstdlib_c/zlib-1.2.12/crc32.c +1116 -0
  17. data/ext/zstdlib_c/zlib-1.2.12/crc32.h +9446 -0
  18. data/ext/{zstdlib/zlib-1.2.11 → zstdlib_c/zlib-1.2.12}/deflate.c +78 -30
  19. data/ext/{zstdlib/zlib-1.2.11 → zstdlib_c/zlib-1.2.12}/deflate.h +12 -15
  20. data/ext/{zstdlib/zlib-1.2.11 → zstdlib_c/zlib-1.2.12}/gzclose.c +0 -0
  21. data/ext/{zstdlib/zlib-1.2.11 → zstdlib_c/zlib-1.2.12}/gzguts.h +3 -2
  22. data/ext/{zstdlib/zlib-1.2.11 → zstdlib_c/zlib-1.2.12}/gzlib.c +5 -3
  23. data/ext/{zstdlib/zlib-1.2.11 → zstdlib_c/zlib-1.2.12}/gzread.c +5 -7
  24. data/ext/{zstdlib/zlib-1.2.11 → zstdlib_c/zlib-1.2.12}/gzwrite.c +25 -13
  25. data/ext/{zstdlib/zlib-1.2.11 → zstdlib_c/zlib-1.2.12}/infback.c +2 -1
  26. data/ext/{zstdlib/zlib-1.2.11 → zstdlib_c/zlib-1.2.12}/inffast.c +14 -14
  27. data/ext/{zstdlib/zlib-1.2.11 → zstdlib_c/zlib-1.2.12}/inffast.h +0 -0
  28. data/ext/{zstdlib/zlib-1.2.11 → zstdlib_c/zlib-1.2.12}/inffixed.h +0 -0
  29. data/ext/{zstdlib/zlib-1.2.11 → zstdlib_c/zlib-1.2.12}/inflate.c +39 -8
  30. data/ext/{zstdlib/zlib-1.2.11 → zstdlib_c/zlib-1.2.12}/inflate.h +3 -2
  31. data/ext/{zstdlib/zlib-1.2.11 → zstdlib_c/zlib-1.2.12}/inftrees.c +3 -3
  32. data/ext/{zstdlib/zlib-1.2.11 → zstdlib_c/zlib-1.2.12}/inftrees.h +0 -0
  33. data/ext/{zstdlib/zlib-1.2.11 → zstdlib_c/zlib-1.2.12}/trees.c +27 -48
  34. data/ext/{zstdlib/zlib-1.2.11 → zstdlib_c/zlib-1.2.12}/trees.h +0 -0
  35. data/ext/{zstdlib/zlib-1.2.11 → zstdlib_c/zlib-1.2.12}/uncompr.c +0 -0
  36. data/ext/{zstdlib/zlib-1.2.11 → zstdlib_c/zlib-1.2.12}/zconf.h +0 -0
  37. data/ext/{zstdlib/zlib-1.2.11 → zstdlib_c/zlib-1.2.12}/zlib.h +123 -100
  38. data/ext/{zstdlib/zlib-1.2.11 → zstdlib_c/zlib-1.2.12}/zutil.c +2 -2
  39. data/ext/{zstdlib/zlib-1.2.11 → zstdlib_c/zlib-1.2.12}/zutil.h +12 -9
  40. data/ext/{zstdlib → zstdlib_c}/zlib.mk +0 -0
  41. data/ext/{zstdlib → zstdlib_c}/zlibwrapper/zlibwrapper.c +1 -5
  42. data/ext/{zstdlib → zstdlib_c}/zlibwrapper.mk +0 -0
  43. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/common/bitstream.h +46 -22
  44. data/ext/zstdlib_c/zstd-1.5.2/lib/common/compiler.h +335 -0
  45. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/common/cpu.h +1 -3
  46. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/common/debug.c +1 -1
  47. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/common/debug.h +12 -19
  48. data/ext/zstdlib_c/zstd-1.5.2/lib/common/entropy_common.c +368 -0
  49. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/common/error_private.c +2 -1
  50. data/ext/zstdlib_c/zstd-1.5.2/lib/common/error_private.h +159 -0
  51. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/common/fse.h +41 -12
  52. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/common/fse_decompress.c +139 -22
  53. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/common/huf.h +47 -23
  54. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/common/mem.h +87 -98
  55. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/common/pool.c +34 -23
  56. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/common/pool.h +4 -4
  57. data/ext/zstdlib_c/zstd-1.5.2/lib/common/portability_macros.h +137 -0
  58. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/common/threading.c +6 -5
  59. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/common/threading.h +0 -0
  60. data/ext/zstdlib_c/zstd-1.5.2/lib/common/xxhash.c +24 -0
  61. data/ext/zstdlib_c/zstd-1.5.2/lib/common/xxhash.h +5686 -0
  62. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/common/zstd_common.c +10 -10
  63. data/ext/zstdlib_c/zstd-1.5.2/lib/common/zstd_deps.h +111 -0
  64. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/common/zstd_internal.h +191 -145
  65. data/ext/zstdlib_c/zstd-1.5.2/lib/common/zstd_trace.h +163 -0
  66. data/ext/zstdlib_c/zstd-1.5.2/lib/compress/clevels.h +134 -0
  67. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/compress/fse_compress.c +89 -46
  68. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/compress/hist.c +27 -29
  69. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/compress/hist.h +2 -2
  70. data/ext/zstdlib_c/zstd-1.5.2/lib/compress/huf_compress.c +1370 -0
  71. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/compress/zstd_compress.c +2917 -868
  72. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/compress/zstd_compress_internal.h +458 -125
  73. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/compress/zstd_compress_literals.c +12 -11
  74. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/compress/zstd_compress_literals.h +4 -2
  75. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/compress/zstd_compress_sequences.c +41 -18
  76. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/compress/zstd_compress_sequences.h +1 -1
  77. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/compress/zstd_compress_superblock.c +26 -298
  78. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/compress/zstd_compress_superblock.h +1 -1
  79. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/compress/zstd_cwksp.h +234 -83
  80. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/compress/zstd_double_fast.c +313 -138
  81. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/compress/zstd_double_fast.h +1 -1
  82. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/compress/zstd_fast.c +329 -150
  83. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/compress/zstd_fast.h +1 -1
  84. data/ext/zstdlib_c/zstd-1.5.2/lib/compress/zstd_lazy.c +2104 -0
  85. data/ext/zstdlib_c/zstd-1.5.2/lib/compress/zstd_lazy.h +125 -0
  86. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/compress/zstd_ldm.c +321 -216
  87. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/compress/zstd_ldm.h +9 -2
  88. data/ext/zstdlib_c/zstd-1.5.2/lib/compress/zstd_ldm_geartab.h +106 -0
  89. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/compress/zstd_opt.c +412 -166
  90. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/compress/zstd_opt.h +1 -1
  91. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/compress/zstdmt_compress.c +169 -453
  92. data/ext/zstdlib_c/zstd-1.5.2/lib/compress/zstdmt_compress.h +113 -0
  93. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/decompress/huf_decompress.c +1044 -403
  94. data/ext/zstdlib_c/zstd-1.5.2/lib/decompress/huf_decompress_amd64.S +585 -0
  95. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/decompress/zstd_ddict.c +9 -9
  96. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/decompress/zstd_ddict.h +2 -2
  97. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/decompress/zstd_decompress.c +450 -105
  98. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/decompress/zstd_decompress_block.c +913 -273
  99. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/decompress/zstd_decompress_block.h +14 -5
  100. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/decompress/zstd_decompress_internal.h +59 -12
  101. data/ext/zstdlib_c/zstd-1.5.2/lib/zdict.h +452 -0
  102. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/lib/zstd.h +699 -214
  103. data/ext/{zstdlib/zstd-1.4.5/lib/common → zstdlib_c/zstd-1.5.2/lib}/zstd_errors.h +2 -1
  104. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/zlibWrapper/gzclose.c +0 -0
  105. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/zlibWrapper/gzcompatibility.h +1 -1
  106. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/zlibWrapper/gzguts.h +0 -0
  107. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/zlibWrapper/gzlib.c +0 -0
  108. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/zlibWrapper/gzread.c +0 -0
  109. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/zlibWrapper/gzwrite.c +0 -0
  110. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/zlibWrapper/zstd_zlibwrapper.c +133 -44
  111. data/ext/{zstdlib/zstd-1.4.5 → zstdlib_c/zstd-1.5.2}/zlibWrapper/zstd_zlibwrapper.h +1 -1
  112. data/ext/zstdlib_c/zstd.mk +15 -0
  113. data/lib/2.4/zstdlib_c.so +0 -0
  114. data/lib/2.5/zstdlib_c.so +0 -0
  115. data/lib/2.6/zstdlib_c.so +0 -0
  116. data/lib/2.7/zstdlib_c.so +0 -0
  117. data/lib/3.0/zstdlib_c.so +0 -0
  118. data/lib/3.1/zstdlib_c.so +0 -0
  119. data/lib/zstdlib.rb +2 -2
  120. metadata +125 -116
  121. data/ext/zstdlib/zlib-1.2.11/crc32.c +0 -442
  122. data/ext/zstdlib/zlib-1.2.11/crc32.h +0 -441
  123. data/ext/zstdlib/zstd-1.4.5/lib/common/compiler.h +0 -175
  124. data/ext/zstdlib/zstd-1.4.5/lib/common/entropy_common.c +0 -216
  125. data/ext/zstdlib/zstd-1.4.5/lib/common/error_private.h +0 -80
  126. data/ext/zstdlib/zstd-1.4.5/lib/common/xxhash.c +0 -864
  127. data/ext/zstdlib/zstd-1.4.5/lib/common/xxhash.h +0 -285
  128. data/ext/zstdlib/zstd-1.4.5/lib/compress/huf_compress.c +0 -798
  129. data/ext/zstdlib/zstd-1.4.5/lib/compress/zstd_lazy.c +0 -1138
  130. data/ext/zstdlib/zstd-1.4.5/lib/compress/zstd_lazy.h +0 -67
  131. data/ext/zstdlib/zstd-1.4.5/lib/compress/zstdmt_compress.h +0 -192
  132. data/ext/zstdlib/zstd.mk +0 -14
  133. data/lib/2.2/zstdlib.so +0 -0
  134. data/lib/2.3/zstdlib.so +0 -0
  135. data/lib/2.4/zstdlib.so +0 -0
  136. data/lib/2.5/zstdlib.so +0 -0
  137. data/lib/2.6/zstdlib.so +0 -0
  138. data/lib/2.7/zstdlib.so +0 -0
@@ -1,5 +1,5 @@
1
1
  /*
2
- * Copyright (c) 2016-2020, Yann Collet, Facebook, Inc.
2
+ * Copyright (c) Yann Collet, Facebook, Inc.
3
3
  * All rights reserved.
4
4
  *
5
5
  * This source code is licensed under both the BSD-style license (found in the
@@ -20,8 +20,7 @@
20
20
 
21
21
 
22
22
  /* ====== Dependencies ====== */
23
- #include <string.h> /* memcpy, memset */
24
- #include <limits.h> /* INT_MAX, UINT_MAX */
23
+ #include "../common/zstd_deps.h" /* ZSTD_memcpy, ZSTD_memset, INT_MAX, UINT_MAX */
25
24
  #include "../common/mem.h" /* MEM_STATIC */
26
25
  #include "../common/pool.h" /* threadpool */
27
26
  #include "../common/threading.h" /* mutex */
@@ -103,14 +102,13 @@ typedef struct ZSTDMT_bufferPool_s {
103
102
  buffer_t bTable[1]; /* variable size */
104
103
  } ZSTDMT_bufferPool;
105
104
 
106
- static ZSTDMT_bufferPool* ZSTDMT_createBufferPool(unsigned nbWorkers, ZSTD_customMem cMem)
105
+ static ZSTDMT_bufferPool* ZSTDMT_createBufferPool(unsigned maxNbBuffers, ZSTD_customMem cMem)
107
106
  {
108
- unsigned const maxNbBuffers = 2*nbWorkers + 3;
109
- ZSTDMT_bufferPool* const bufPool = (ZSTDMT_bufferPool*)ZSTD_calloc(
107
+ ZSTDMT_bufferPool* const bufPool = (ZSTDMT_bufferPool*)ZSTD_customCalloc(
110
108
  sizeof(ZSTDMT_bufferPool) + (maxNbBuffers-1) * sizeof(buffer_t), cMem);
111
109
  if (bufPool==NULL) return NULL;
112
110
  if (ZSTD_pthread_mutex_init(&bufPool->poolMutex, NULL)) {
113
- ZSTD_free(bufPool, cMem);
111
+ ZSTD_customFree(bufPool, cMem);
114
112
  return NULL;
115
113
  }
116
114
  bufPool->bufferSize = 64 KB;
@@ -127,10 +125,10 @@ static void ZSTDMT_freeBufferPool(ZSTDMT_bufferPool* bufPool)
127
125
  if (!bufPool) return; /* compatibility with free on NULL */
128
126
  for (u=0; u<bufPool->totalBuffers; u++) {
129
127
  DEBUGLOG(4, "free buffer %2u (address:%08X)", u, (U32)(size_t)bufPool->bTable[u].start);
130
- ZSTD_free(bufPool->bTable[u].start, bufPool->cMem);
128
+ ZSTD_customFree(bufPool->bTable[u].start, bufPool->cMem);
131
129
  }
132
130
  ZSTD_pthread_mutex_destroy(&bufPool->poolMutex);
133
- ZSTD_free(bufPool, bufPool->cMem);
131
+ ZSTD_customFree(bufPool, bufPool->cMem);
134
132
  }
135
133
 
136
134
  /* only works at initialization, not during compression */
@@ -161,9 +159,8 @@ static void ZSTDMT_setBufferSize(ZSTDMT_bufferPool* const bufPool, size_t const
161
159
  }
162
160
 
163
161
 
164
- static ZSTDMT_bufferPool* ZSTDMT_expandBufferPool(ZSTDMT_bufferPool* srcBufPool, U32 nbWorkers)
162
+ static ZSTDMT_bufferPool* ZSTDMT_expandBufferPool(ZSTDMT_bufferPool* srcBufPool, unsigned maxNbBuffers)
165
163
  {
166
- unsigned const maxNbBuffers = 2*nbWorkers + 3;
167
164
  if (srcBufPool==NULL) return NULL;
168
165
  if (srcBufPool->totalBuffers >= maxNbBuffers) /* good enough */
169
166
  return srcBufPool;
@@ -172,7 +169,7 @@ static ZSTDMT_bufferPool* ZSTDMT_expandBufferPool(ZSTDMT_bufferPool* srcBufPool,
172
169
  size_t const bSize = srcBufPool->bufferSize; /* forward parameters */
173
170
  ZSTDMT_bufferPool* newBufPool;
174
171
  ZSTDMT_freeBufferPool(srcBufPool);
175
- newBufPool = ZSTDMT_createBufferPool(nbWorkers, cMem);
172
+ newBufPool = ZSTDMT_createBufferPool(maxNbBuffers, cMem);
176
173
  if (newBufPool==NULL) return newBufPool;
177
174
  ZSTDMT_setBufferSize(newBufPool, bSize);
178
175
  return newBufPool;
@@ -201,13 +198,13 @@ static buffer_t ZSTDMT_getBuffer(ZSTDMT_bufferPool* bufPool)
201
198
  }
202
199
  /* size conditions not respected : scratch this buffer, create new one */
203
200
  DEBUGLOG(5, "ZSTDMT_getBuffer: existing buffer does not meet size conditions => freeing");
204
- ZSTD_free(buf.start, bufPool->cMem);
201
+ ZSTD_customFree(buf.start, bufPool->cMem);
205
202
  }
206
203
  ZSTD_pthread_mutex_unlock(&bufPool->poolMutex);
207
204
  /* create new buffer */
208
205
  DEBUGLOG(5, "ZSTDMT_getBuffer: create a new buffer");
209
206
  { buffer_t buffer;
210
- void* const start = ZSTD_malloc(bSize, bufPool->cMem);
207
+ void* const start = ZSTD_customMalloc(bSize, bufPool->cMem);
211
208
  buffer.start = start; /* note : start can be NULL if malloc fails ! */
212
209
  buffer.capacity = (start==NULL) ? 0 : bSize;
213
210
  if (start==NULL) {
@@ -229,13 +226,13 @@ static buffer_t ZSTDMT_resizeBuffer(ZSTDMT_bufferPool* bufPool, buffer_t buffer)
229
226
  {
230
227
  size_t const bSize = bufPool->bufferSize;
231
228
  if (buffer.capacity < bSize) {
232
- void* const start = ZSTD_malloc(bSize, bufPool->cMem);
229
+ void* const start = ZSTD_customMalloc(bSize, bufPool->cMem);
233
230
  buffer_t newBuffer;
234
231
  newBuffer.start = start;
235
232
  newBuffer.capacity = start == NULL ? 0 : bSize;
236
233
  if (start != NULL) {
237
234
  assert(newBuffer.capacity >= buffer.capacity);
238
- memcpy(newBuffer.start, buffer.start, buffer.capacity);
235
+ ZSTD_memcpy(newBuffer.start, buffer.start, buffer.capacity);
239
236
  DEBUGLOG(5, "ZSTDMT_resizeBuffer: created buffer of size %u", (U32)bSize);
240
237
  return newBuffer;
241
238
  }
@@ -261,13 +258,21 @@ static void ZSTDMT_releaseBuffer(ZSTDMT_bufferPool* bufPool, buffer_t buf)
261
258
  ZSTD_pthread_mutex_unlock(&bufPool->poolMutex);
262
259
  /* Reached bufferPool capacity (should not happen) */
263
260
  DEBUGLOG(5, "ZSTDMT_releaseBuffer: pool capacity reached => freeing ");
264
- ZSTD_free(buf.start, bufPool->cMem);
261
+ ZSTD_customFree(buf.start, bufPool->cMem);
265
262
  }
266
263
 
264
+ /* We need 2 output buffers per worker since each dstBuff must be flushed after it is released.
265
+ * The 3 additional buffers are as follows:
266
+ * 1 buffer for input loading
267
+ * 1 buffer for "next input" when submitting current one
268
+ * 1 buffer stuck in queue */
269
+ #define BUF_POOL_MAX_NB_BUFFERS(nbWorkers) 2*nbWorkers + 3
267
270
 
268
- /* ===== Seq Pool Wrapper ====== */
271
+ /* After a worker releases its rawSeqStore, it is immediately ready for reuse.
272
+ * So we only need one seq buffer per worker. */
273
+ #define SEQ_POOL_MAX_NB_BUFFERS(nbWorkers) nbWorkers
269
274
 
270
- static rawSeqStore_t kNullRawSeqStore = {NULL, 0, 0, 0};
275
+ /* ===== Seq Pool Wrapper ====== */
271
276
 
272
277
  typedef ZSTDMT_bufferPool ZSTDMT_seqPool;
273
278
 
@@ -278,7 +283,7 @@ static size_t ZSTDMT_sizeof_seqPool(ZSTDMT_seqPool* seqPool)
278
283
 
279
284
  static rawSeqStore_t bufferToSeq(buffer_t buffer)
280
285
  {
281
- rawSeqStore_t seq = {NULL, 0, 0, 0};
286
+ rawSeqStore_t seq = kNullRawSeqStore;
282
287
  seq.seq = (rawSeq*)buffer.start;
283
288
  seq.capacity = buffer.capacity / sizeof(rawSeq);
284
289
  return seq;
@@ -319,7 +324,7 @@ static void ZSTDMT_setNbSeq(ZSTDMT_seqPool* const seqPool, size_t const nbSeq)
319
324
 
320
325
  static ZSTDMT_seqPool* ZSTDMT_createSeqPool(unsigned nbWorkers, ZSTD_customMem cMem)
321
326
  {
322
- ZSTDMT_seqPool* const seqPool = ZSTDMT_createBufferPool(nbWorkers, cMem);
327
+ ZSTDMT_seqPool* const seqPool = ZSTDMT_createBufferPool(SEQ_POOL_MAX_NB_BUFFERS(nbWorkers), cMem);
323
328
  if (seqPool == NULL) return NULL;
324
329
  ZSTDMT_setNbSeq(seqPool, 0);
325
330
  return seqPool;
@@ -332,7 +337,7 @@ static void ZSTDMT_freeSeqPool(ZSTDMT_seqPool* seqPool)
332
337
 
333
338
  static ZSTDMT_seqPool* ZSTDMT_expandSeqPool(ZSTDMT_seqPool* pool, U32 nbWorkers)
334
339
  {
335
- return ZSTDMT_expandBufferPool(pool, nbWorkers);
340
+ return ZSTDMT_expandBufferPool(pool, SEQ_POOL_MAX_NB_BUFFERS(nbWorkers));
336
341
  }
337
342
 
338
343
 
@@ -354,7 +359,7 @@ static void ZSTDMT_freeCCtxPool(ZSTDMT_CCtxPool* pool)
354
359
  for (cid=0; cid<pool->totalCCtx; cid++)
355
360
  ZSTD_freeCCtx(pool->cctx[cid]); /* note : compatible with free on NULL */
356
361
  ZSTD_pthread_mutex_destroy(&pool->poolMutex);
357
- ZSTD_free(pool, pool->cMem);
362
+ ZSTD_customFree(pool, pool->cMem);
358
363
  }
359
364
 
360
365
  /* ZSTDMT_createCCtxPool() :
@@ -362,12 +367,12 @@ static void ZSTDMT_freeCCtxPool(ZSTDMT_CCtxPool* pool)
362
367
  static ZSTDMT_CCtxPool* ZSTDMT_createCCtxPool(int nbWorkers,
363
368
  ZSTD_customMem cMem)
364
369
  {
365
- ZSTDMT_CCtxPool* const cctxPool = (ZSTDMT_CCtxPool*) ZSTD_calloc(
370
+ ZSTDMT_CCtxPool* const cctxPool = (ZSTDMT_CCtxPool*) ZSTD_customCalloc(
366
371
  sizeof(ZSTDMT_CCtxPool) + (nbWorkers-1)*sizeof(ZSTD_CCtx*), cMem);
367
372
  assert(nbWorkers > 0);
368
373
  if (!cctxPool) return NULL;
369
374
  if (ZSTD_pthread_mutex_init(&cctxPool->poolMutex, NULL)) {
370
- ZSTD_free(cctxPool, cMem);
375
+ ZSTD_customFree(cctxPool, cMem);
371
376
  return NULL;
372
377
  }
373
378
  cctxPool->cMem = cMem;
@@ -470,54 +475,52 @@ ZSTDMT_serialState_reset(serialState_t* serialState,
470
475
  ZSTD_dictContentType_e dictContentType)
471
476
  {
472
477
  /* Adjust parameters */
473
- if (params.ldmParams.enableLdm) {
478
+ if (params.ldmParams.enableLdm == ZSTD_ps_enable) {
474
479
  DEBUGLOG(4, "LDM window size = %u KB", (1U << params.cParams.windowLog) >> 10);
475
480
  ZSTD_ldm_adjustParameters(&params.ldmParams, &params.cParams);
476
481
  assert(params.ldmParams.hashLog >= params.ldmParams.bucketSizeLog);
477
482
  assert(params.ldmParams.hashRateLog < 32);
478
- serialState->ldmState.hashPower =
479
- ZSTD_rollingHash_primePower(params.ldmParams.minMatchLength);
480
483
  } else {
481
- memset(&params.ldmParams, 0, sizeof(params.ldmParams));
484
+ ZSTD_memset(&params.ldmParams, 0, sizeof(params.ldmParams));
482
485
  }
483
486
  serialState->nextJobID = 0;
484
487
  if (params.fParams.checksumFlag)
485
488
  XXH64_reset(&serialState->xxhState, 0);
486
- if (params.ldmParams.enableLdm) {
489
+ if (params.ldmParams.enableLdm == ZSTD_ps_enable) {
487
490
  ZSTD_customMem cMem = params.customMem;
488
491
  unsigned const hashLog = params.ldmParams.hashLog;
489
492
  size_t const hashSize = ((size_t)1 << hashLog) * sizeof(ldmEntry_t);
490
493
  unsigned const bucketLog =
491
494
  params.ldmParams.hashLog - params.ldmParams.bucketSizeLog;
492
- size_t const bucketSize = (size_t)1 << bucketLog;
493
495
  unsigned const prevBucketLog =
494
496
  serialState->params.ldmParams.hashLog -
495
497
  serialState->params.ldmParams.bucketSizeLog;
498
+ size_t const numBuckets = (size_t)1 << bucketLog;
496
499
  /* Size the seq pool tables */
497
500
  ZSTDMT_setNbSeq(seqPool, ZSTD_ldm_getMaxNbSeq(params.ldmParams, jobSize));
498
501
  /* Reset the window */
499
502
  ZSTD_window_init(&serialState->ldmState.window);
500
503
  /* Resize tables and output space if necessary. */
501
504
  if (serialState->ldmState.hashTable == NULL || serialState->params.ldmParams.hashLog < hashLog) {
502
- ZSTD_free(serialState->ldmState.hashTable, cMem);
503
- serialState->ldmState.hashTable = (ldmEntry_t*)ZSTD_malloc(hashSize, cMem);
505
+ ZSTD_customFree(serialState->ldmState.hashTable, cMem);
506
+ serialState->ldmState.hashTable = (ldmEntry_t*)ZSTD_customMalloc(hashSize, cMem);
504
507
  }
505
508
  if (serialState->ldmState.bucketOffsets == NULL || prevBucketLog < bucketLog) {
506
- ZSTD_free(serialState->ldmState.bucketOffsets, cMem);
507
- serialState->ldmState.bucketOffsets = (BYTE*)ZSTD_malloc(bucketSize, cMem);
509
+ ZSTD_customFree(serialState->ldmState.bucketOffsets, cMem);
510
+ serialState->ldmState.bucketOffsets = (BYTE*)ZSTD_customMalloc(numBuckets, cMem);
508
511
  }
509
512
  if (!serialState->ldmState.hashTable || !serialState->ldmState.bucketOffsets)
510
513
  return 1;
511
514
  /* Zero the tables */
512
- memset(serialState->ldmState.hashTable, 0, hashSize);
513
- memset(serialState->ldmState.bucketOffsets, 0, bucketSize);
515
+ ZSTD_memset(serialState->ldmState.hashTable, 0, hashSize);
516
+ ZSTD_memset(serialState->ldmState.bucketOffsets, 0, numBuckets);
514
517
 
515
518
  /* Update window state and fill hash table with dict */
516
519
  serialState->ldmState.loadedDictEnd = 0;
517
520
  if (dictSize > 0) {
518
521
  if (dictContentType == ZSTD_dct_rawContent) {
519
522
  BYTE const* const dictEnd = (const BYTE*)dict + dictSize;
520
- ZSTD_window_update(&serialState->ldmState.window, dict, dictSize);
523
+ ZSTD_window_update(&serialState->ldmState.window, dict, dictSize, /* forceNonContiguous */ 0);
521
524
  ZSTD_ldm_fillHashTable(&serialState->ldmState, (const BYTE*)dict, dictEnd, &params.ldmParams);
522
525
  serialState->ldmState.loadedDictEnd = params.forceWindow ? 0 : (U32)(dictEnd - serialState->ldmState.window.base);
523
526
  } else {
@@ -537,7 +540,7 @@ ZSTDMT_serialState_reset(serialState_t* serialState,
537
540
  static int ZSTDMT_serialState_init(serialState_t* serialState)
538
541
  {
539
542
  int initError = 0;
540
- memset(serialState, 0, sizeof(*serialState));
543
+ ZSTD_memset(serialState, 0, sizeof(*serialState));
541
544
  initError |= ZSTD_pthread_mutex_init(&serialState->mutex, NULL);
542
545
  initError |= ZSTD_pthread_cond_init(&serialState->cond, NULL);
543
546
  initError |= ZSTD_pthread_mutex_init(&serialState->ldmWindowMutex, NULL);
@@ -552,8 +555,8 @@ static void ZSTDMT_serialState_free(serialState_t* serialState)
552
555
  ZSTD_pthread_cond_destroy(&serialState->cond);
553
556
  ZSTD_pthread_mutex_destroy(&serialState->ldmWindowMutex);
554
557
  ZSTD_pthread_cond_destroy(&serialState->ldmWindowCond);
555
- ZSTD_free(serialState->ldmState.hashTable, cMem);
556
- ZSTD_free(serialState->ldmState.bucketOffsets, cMem);
558
+ ZSTD_customFree(serialState->ldmState.hashTable, cMem);
559
+ ZSTD_customFree(serialState->ldmState.bucketOffsets, cMem);
557
560
  }
558
561
 
559
562
  static void ZSTDMT_serialState_update(serialState_t* serialState,
@@ -569,12 +572,12 @@ static void ZSTDMT_serialState_update(serialState_t* serialState,
569
572
  /* A future job may error and skip our job */
570
573
  if (serialState->nextJobID == jobID) {
571
574
  /* It is now our turn, do any processing necessary */
572
- if (serialState->params.ldmParams.enableLdm) {
575
+ if (serialState->params.ldmParams.enableLdm == ZSTD_ps_enable) {
573
576
  size_t error;
574
577
  assert(seqStore.seq != NULL && seqStore.pos == 0 &&
575
578
  seqStore.size == 0 && seqStore.capacity > 0);
576
579
  assert(src.size <= serialState->params.jobSize);
577
- ZSTD_window_update(&serialState->ldmState.window, src.start, src.size);
580
+ ZSTD_window_update(&serialState->ldmState.window, src.start, src.size, /* forceNonContiguous */ 0);
578
581
  error = ZSTD_ldm_generateSequences(
579
582
  &serialState->ldmState, &seqStore,
580
583
  &serialState->params.ldmParams, src.start, src.size);
@@ -599,7 +602,7 @@ static void ZSTDMT_serialState_update(serialState_t* serialState,
599
602
  if (seqStore.size > 0) {
600
603
  size_t const err = ZSTD_referenceExternalSequences(
601
604
  jobCCtx, seqStore.seq, seqStore.size);
602
- assert(serialState->params.ldmParams.enableLdm);
605
+ assert(serialState->params.ldmParams.enableLdm == ZSTD_ps_enable);
603
606
  assert(!ZSTD_isError(err));
604
607
  (void)err;
605
608
  }
@@ -677,7 +680,7 @@ static void ZSTDMT_compressionJob(void* jobDescription)
677
680
  if (dstBuff.start==NULL) JOB_ERROR(ERROR(memory_allocation));
678
681
  job->dstBuff = dstBuff; /* this value can be read in ZSTDMT_flush, when it copies the whole job */
679
682
  }
680
- if (jobParams.ldmParams.enableLdm && rawSeqStore.seq == NULL)
683
+ if (jobParams.ldmParams.enableLdm == ZSTD_ps_enable && rawSeqStore.seq == NULL)
681
684
  JOB_ERROR(ERROR(memory_allocation));
682
685
 
683
686
  /* Don't compute the checksum for chunks, since we compute it externally,
@@ -685,7 +688,9 @@ static void ZSTDMT_compressionJob(void* jobDescription)
685
688
  */
686
689
  if (job->jobID != 0) jobParams.fParams.checksumFlag = 0;
687
690
  /* Don't run LDM for the chunks, since we handle it externally */
688
- jobParams.ldmParams.enableLdm = 0;
691
+ jobParams.ldmParams.enableLdm = ZSTD_ps_disable;
692
+ /* Correct nbWorkers to 0. */
693
+ jobParams.nbWorkers = 0;
689
694
 
690
695
 
691
696
  /* init */
@@ -698,6 +703,10 @@ static void ZSTDMT_compressionJob(void* jobDescription)
698
703
  { size_t const forceWindowError = ZSTD_CCtxParams_setParameter(&jobParams, ZSTD_c_forceMaxWindow, !job->firstJob);
699
704
  if (ZSTD_isError(forceWindowError)) JOB_ERROR(forceWindowError);
700
705
  }
706
+ if (!job->firstJob) {
707
+ size_t const err = ZSTD_CCtxParams_setParameter(&jobParams, ZSTD_c_deterministicRefPrefix, 0);
708
+ if (ZSTD_isError(err)) JOB_ERROR(err);
709
+ }
701
710
  { size_t const initError = ZSTD_compressBegin_advanced_internal(cctx,
702
711
  job->prefix.start, job->prefix.size, ZSTD_dct_rawContent, /* load dictionary in "content-only" mode (no header analysis) */
703
712
  ZSTD_dtlm_fast,
@@ -753,6 +762,13 @@ static void ZSTDMT_compressionJob(void* jobDescription)
753
762
  if (ZSTD_isError(cSize)) JOB_ERROR(cSize);
754
763
  lastCBlockSize = cSize;
755
764
  } }
765
+ if (!job->firstJob) {
766
+ /* Double check that we don't have an ext-dict, because then our
767
+ * repcode invalidation doesn't work.
768
+ */
769
+ assert(!ZSTD_window_hasExtDict(cctx->blockState.matchState.window));
770
+ }
771
+ ZSTD_CCtx_trace(cctx, 0);
756
772
 
757
773
  _endJob:
758
774
  ZSTDMT_serialState_ensureFinished(job->serial, job->jobID, job->cSize);
@@ -799,6 +815,15 @@ typedef struct {
799
815
  static const roundBuff_t kNullRoundBuff = {NULL, 0, 0};
800
816
 
801
817
  #define RSYNC_LENGTH 32
818
+ /* Don't create chunks smaller than the zstd block size.
819
+ * This stops us from regressing compression ratio too much,
820
+ * and ensures our output fits in ZSTD_compressBound().
821
+ *
822
+ * If this is shrunk < ZSTD_BLOCKSIZELOG_MIN then
823
+ * ZSTD_COMPRESSBOUND() will need to be updated.
824
+ */
825
+ #define RSYNC_MIN_BLOCK_LOG ZSTD_BLOCKSIZELOG_MAX
826
+ #define RSYNC_MIN_BLOCK_SIZE (1<<RSYNC_MIN_BLOCK_LOG)
802
827
 
803
828
  typedef struct {
804
829
  U64 hash;
@@ -820,7 +845,6 @@ struct ZSTDMT_CCtx_s {
820
845
  roundBuff_t roundBuff;
821
846
  serialState_t serial;
822
847
  rsyncState_t rsync;
823
- unsigned singleBlockingThread;
824
848
  unsigned jobIDMask;
825
849
  unsigned doneJobID;
826
850
  unsigned nextJobID;
@@ -832,6 +856,7 @@ struct ZSTDMT_CCtx_s {
832
856
  ZSTD_customMem cMem;
833
857
  ZSTD_CDict* cdictLocal;
834
858
  const ZSTD_CDict* cdict;
859
+ unsigned providedFactory: 1;
835
860
  };
836
861
 
837
862
  static void ZSTDMT_freeJobsTable(ZSTDMT_jobDescription* jobTable, U32 nbJobs, ZSTD_customMem cMem)
@@ -842,7 +867,7 @@ static void ZSTDMT_freeJobsTable(ZSTDMT_jobDescription* jobTable, U32 nbJobs, ZS
842
867
  ZSTD_pthread_mutex_destroy(&jobTable[jobNb].job_mutex);
843
868
  ZSTD_pthread_cond_destroy(&jobTable[jobNb].job_cond);
844
869
  }
845
- ZSTD_free(jobTable, cMem);
870
+ ZSTD_customFree(jobTable, cMem);
846
871
  }
847
872
 
848
873
  /* ZSTDMT_allocJobsTable()
@@ -854,7 +879,7 @@ static ZSTDMT_jobDescription* ZSTDMT_createJobsTable(U32* nbJobsPtr, ZSTD_custom
854
879
  U32 const nbJobs = 1 << nbJobsLog2;
855
880
  U32 jobNb;
856
881
  ZSTDMT_jobDescription* const jobTable = (ZSTDMT_jobDescription*)
857
- ZSTD_calloc(nbJobs * sizeof(ZSTDMT_jobDescription), cMem);
882
+ ZSTD_customCalloc(nbJobs * sizeof(ZSTDMT_jobDescription), cMem);
858
883
  int initError = 0;
859
884
  if (jobTable==NULL) return NULL;
860
885
  *nbJobsPtr = nbJobs;
@@ -885,12 +910,12 @@ static size_t ZSTDMT_expandJobsTable (ZSTDMT_CCtx* mtctx, U32 nbWorkers) {
885
910
 
886
911
  /* ZSTDMT_CCtxParam_setNbWorkers():
887
912
  * Internal use only */
888
- size_t ZSTDMT_CCtxParam_setNbWorkers(ZSTD_CCtx_params* params, unsigned nbWorkers)
913
+ static size_t ZSTDMT_CCtxParam_setNbWorkers(ZSTD_CCtx_params* params, unsigned nbWorkers)
889
914
  {
890
915
  return ZSTD_CCtxParams_setParameter(params, ZSTD_c_nbWorkers, (int)nbWorkers);
891
916
  }
892
917
 
893
- MEM_STATIC ZSTDMT_CCtx* ZSTDMT_createCCtx_advanced_internal(unsigned nbWorkers, ZSTD_customMem cMem)
918
+ MEM_STATIC ZSTDMT_CCtx* ZSTDMT_createCCtx_advanced_internal(unsigned nbWorkers, ZSTD_customMem cMem, ZSTD_threadPool* pool)
894
919
  {
895
920
  ZSTDMT_CCtx* mtctx;
896
921
  U32 nbJobs = nbWorkers + 2;
@@ -903,16 +928,23 @@ MEM_STATIC ZSTDMT_CCtx* ZSTDMT_createCCtx_advanced_internal(unsigned nbWorkers,
903
928
  /* invalid custom allocator */
904
929
  return NULL;
905
930
 
906
- mtctx = (ZSTDMT_CCtx*) ZSTD_calloc(sizeof(ZSTDMT_CCtx), cMem);
931
+ mtctx = (ZSTDMT_CCtx*) ZSTD_customCalloc(sizeof(ZSTDMT_CCtx), cMem);
907
932
  if (!mtctx) return NULL;
908
933
  ZSTDMT_CCtxParam_setNbWorkers(&mtctx->params, nbWorkers);
909
934
  mtctx->cMem = cMem;
910
935
  mtctx->allJobsCompleted = 1;
911
- mtctx->factory = POOL_create_advanced(nbWorkers, 0, cMem);
936
+ if (pool != NULL) {
937
+ mtctx->factory = pool;
938
+ mtctx->providedFactory = 1;
939
+ }
940
+ else {
941
+ mtctx->factory = POOL_create_advanced(nbWorkers, 0, cMem);
942
+ mtctx->providedFactory = 0;
943
+ }
912
944
  mtctx->jobs = ZSTDMT_createJobsTable(&nbJobs, cMem);
913
945
  assert(nbJobs > 0); assert((nbJobs & (nbJobs - 1)) == 0); /* ensure nbJobs is a power of 2 */
914
946
  mtctx->jobIDMask = nbJobs - 1;
915
- mtctx->bufPool = ZSTDMT_createBufferPool(nbWorkers, cMem);
947
+ mtctx->bufPool = ZSTDMT_createBufferPool(BUF_POOL_MAX_NB_BUFFERS(nbWorkers), cMem);
916
948
  mtctx->cctxPool = ZSTDMT_createCCtxPool(nbWorkers, cMem);
917
949
  mtctx->seqPool = ZSTDMT_createSeqPool(nbWorkers, cMem);
918
950
  initError = ZSTDMT_serialState_init(&mtctx->serial);
@@ -925,22 +957,18 @@ MEM_STATIC ZSTDMT_CCtx* ZSTDMT_createCCtx_advanced_internal(unsigned nbWorkers,
925
957
  return mtctx;
926
958
  }
927
959
 
928
- ZSTDMT_CCtx* ZSTDMT_createCCtx_advanced(unsigned nbWorkers, ZSTD_customMem cMem)
960
+ ZSTDMT_CCtx* ZSTDMT_createCCtx_advanced(unsigned nbWorkers, ZSTD_customMem cMem, ZSTD_threadPool* pool)
929
961
  {
930
962
  #ifdef ZSTD_MULTITHREAD
931
- return ZSTDMT_createCCtx_advanced_internal(nbWorkers, cMem);
963
+ return ZSTDMT_createCCtx_advanced_internal(nbWorkers, cMem, pool);
932
964
  #else
933
965
  (void)nbWorkers;
934
966
  (void)cMem;
967
+ (void)pool;
935
968
  return NULL;
936
969
  #endif
937
970
  }
938
971
 
939
- ZSTDMT_CCtx* ZSTDMT_createCCtx(unsigned nbWorkers)
940
- {
941
- return ZSTDMT_createCCtx_advanced(nbWorkers, ZSTD_defaultCMem);
942
- }
943
-
944
972
 
945
973
  /* ZSTDMT_releaseAllJobResources() :
946
974
  * note : ensure all workers are killed first ! */
@@ -957,7 +985,7 @@ static void ZSTDMT_releaseAllJobResources(ZSTDMT_CCtx* mtctx)
957
985
  ZSTDMT_releaseBuffer(mtctx->bufPool, mtctx->jobs[jobID].dstBuff);
958
986
 
959
987
  /* Clear the job description, but keep the mutex/cond */
960
- memset(&mtctx->jobs[jobID], 0, sizeof(mtctx->jobs[jobID]));
988
+ ZSTD_memset(&mtctx->jobs[jobID], 0, sizeof(mtctx->jobs[jobID]));
961
989
  mtctx->jobs[jobID].job_mutex = mutex;
962
990
  mtctx->jobs[jobID].job_cond = cond;
963
991
  }
@@ -984,7 +1012,8 @@ static void ZSTDMT_waitForAllJobsCompleted(ZSTDMT_CCtx* mtctx)
984
1012
  size_t ZSTDMT_freeCCtx(ZSTDMT_CCtx* mtctx)
985
1013
  {
986
1014
  if (mtctx==NULL) return 0; /* compatible with free on NULL */
987
- POOL_free(mtctx->factory); /* stop and free worker threads */
1015
+ if (!mtctx->providedFactory)
1016
+ POOL_free(mtctx->factory); /* stop and free worker threads */
988
1017
  ZSTDMT_releaseAllJobResources(mtctx); /* release job resources into pools first */
989
1018
  ZSTDMT_freeJobsTable(mtctx->jobs, mtctx->jobIDMask+1, mtctx->cMem);
990
1019
  ZSTDMT_freeBufferPool(mtctx->bufPool);
@@ -993,8 +1022,8 @@ size_t ZSTDMT_freeCCtx(ZSTDMT_CCtx* mtctx)
993
1022
  ZSTDMT_serialState_free(&mtctx->serial);
994
1023
  ZSTD_freeCDict(mtctx->cdictLocal);
995
1024
  if (mtctx->roundBuff.buffer)
996
- ZSTD_free(mtctx->roundBuff.buffer, mtctx->cMem);
997
- ZSTD_free(mtctx, mtctx->cMem);
1025
+ ZSTD_customFree(mtctx->roundBuff.buffer, mtctx->cMem);
1026
+ ZSTD_customFree(mtctx, mtctx->cMem);
998
1027
  return 0;
999
1028
  }
1000
1029
 
@@ -1011,65 +1040,6 @@ size_t ZSTDMT_sizeof_CCtx(ZSTDMT_CCtx* mtctx)
1011
1040
  + mtctx->roundBuff.capacity;
1012
1041
  }
1013
1042
 
1014
- /* Internal only */
1015
- size_t
1016
- ZSTDMT_CCtxParam_setMTCtxParameter(ZSTD_CCtx_params* params,
1017
- ZSTDMT_parameter parameter,
1018
- int value)
1019
- {
1020
- DEBUGLOG(4, "ZSTDMT_CCtxParam_setMTCtxParameter");
1021
- switch(parameter)
1022
- {
1023
- case ZSTDMT_p_jobSize :
1024
- DEBUGLOG(4, "ZSTDMT_CCtxParam_setMTCtxParameter : set jobSize to %i", value);
1025
- return ZSTD_CCtxParams_setParameter(params, ZSTD_c_jobSize, value);
1026
- case ZSTDMT_p_overlapLog :
1027
- DEBUGLOG(4, "ZSTDMT_p_overlapLog : %i", value);
1028
- return ZSTD_CCtxParams_setParameter(params, ZSTD_c_overlapLog, value);
1029
- case ZSTDMT_p_rsyncable :
1030
- DEBUGLOG(4, "ZSTD_p_rsyncable : %i", value);
1031
- return ZSTD_CCtxParams_setParameter(params, ZSTD_c_rsyncable, value);
1032
- default :
1033
- return ERROR(parameter_unsupported);
1034
- }
1035
- }
1036
-
1037
- size_t ZSTDMT_setMTCtxParameter(ZSTDMT_CCtx* mtctx, ZSTDMT_parameter parameter, int value)
1038
- {
1039
- DEBUGLOG(4, "ZSTDMT_setMTCtxParameter");
1040
- return ZSTDMT_CCtxParam_setMTCtxParameter(&mtctx->params, parameter, value);
1041
- }
1042
-
1043
- size_t ZSTDMT_getMTCtxParameter(ZSTDMT_CCtx* mtctx, ZSTDMT_parameter parameter, int* value)
1044
- {
1045
- switch (parameter) {
1046
- case ZSTDMT_p_jobSize:
1047
- return ZSTD_CCtxParams_getParameter(&mtctx->params, ZSTD_c_jobSize, value);
1048
- case ZSTDMT_p_overlapLog:
1049
- return ZSTD_CCtxParams_getParameter(&mtctx->params, ZSTD_c_overlapLog, value);
1050
- case ZSTDMT_p_rsyncable:
1051
- return ZSTD_CCtxParams_getParameter(&mtctx->params, ZSTD_c_rsyncable, value);
1052
- default:
1053
- return ERROR(parameter_unsupported);
1054
- }
1055
- }
1056
-
1057
- /* Sets parameters relevant to the compression job,
1058
- * initializing others to default values. */
1059
- static ZSTD_CCtx_params ZSTDMT_initJobCCtxParams(const ZSTD_CCtx_params* params)
1060
- {
1061
- ZSTD_CCtx_params jobParams = *params;
1062
- /* Clear parameters related to multithreading */
1063
- jobParams.forceWindow = 0;
1064
- jobParams.nbWorkers = 0;
1065
- jobParams.jobSize = 0;
1066
- jobParams.overlapLog = 0;
1067
- jobParams.rsyncable = 0;
1068
- memset(&jobParams.ldmParams, 0, sizeof(ldmParams_t));
1069
- memset(&jobParams.customMem, 0, sizeof(ZSTD_customMem));
1070
- return jobParams;
1071
- }
1072
-
1073
1043
 
1074
1044
  /* ZSTDMT_resize() :
1075
1045
  * @return : error code if fails, 0 on success */
@@ -1077,7 +1047,7 @@ static size_t ZSTDMT_resize(ZSTDMT_CCtx* mtctx, unsigned nbWorkers)
1077
1047
  {
1078
1048
  if (POOL_resize(mtctx->factory, nbWorkers)) return ERROR(memory_allocation);
1079
1049
  FORWARD_IF_ERROR( ZSTDMT_expandJobsTable(mtctx, nbWorkers) , "");
1080
- mtctx->bufPool = ZSTDMT_expandBufferPool(mtctx->bufPool, nbWorkers);
1050
+ mtctx->bufPool = ZSTDMT_expandBufferPool(mtctx->bufPool, BUF_POOL_MAX_NB_BUFFERS(nbWorkers));
1081
1051
  if (mtctx->bufPool == NULL) return ERROR(memory_allocation);
1082
1052
  mtctx->cctxPool = ZSTDMT_expandCCtxPool(mtctx->cctxPool, nbWorkers);
1083
1053
  if (mtctx->cctxPool == NULL) return ERROR(memory_allocation);
@@ -1098,7 +1068,7 @@ void ZSTDMT_updateCParams_whileCompressing(ZSTDMT_CCtx* mtctx, const ZSTD_CCtx_p
1098
1068
  DEBUGLOG(5, "ZSTDMT_updateCParams_whileCompressing (level:%i)",
1099
1069
  compressionLevel);
1100
1070
  mtctx->params.compressionLevel = compressionLevel;
1101
- { ZSTD_compressionParameters cParams = ZSTD_getCParamsFromCCtxParams(cctxParams, ZSTD_CONTENTSIZE_UNKNOWN, 0);
1071
+ { ZSTD_compressionParameters cParams = ZSTD_getCParamsFromCCtxParams(cctxParams, ZSTD_CONTENTSIZE_UNKNOWN, 0, ZSTD_cpm_noAttachDict);
1102
1072
  cParams.windowLog = saved_wlog;
1103
1073
  mtctx->params.cParams = cParams;
1104
1074
  }
@@ -1182,11 +1152,11 @@ size_t ZSTDMT_toFlushNow(ZSTDMT_CCtx* mtctx)
1182
1152
  static unsigned ZSTDMT_computeTargetJobLog(const ZSTD_CCtx_params* params)
1183
1153
  {
1184
1154
  unsigned jobLog;
1185
- if (params->ldmParams.enableLdm) {
1155
+ if (params->ldmParams.enableLdm == ZSTD_ps_enable) {
1186
1156
  /* In Long Range Mode, the windowLog is typically oversized.
1187
1157
  * In which case, it's preferable to determine the jobSize
1188
- * based on chainLog instead. */
1189
- jobLog = MAX(21, params->cParams.chainLog + 4);
1158
+ * based on cycleLog instead. */
1159
+ jobLog = MAX(21, ZSTD_cycleLog(params->cParams.chainLog, params->cParams.strategy) + 3);
1190
1160
  } else {
1191
1161
  jobLog = MAX(20, params->cParams.windowLog + 2);
1192
1162
  }
@@ -1226,7 +1196,7 @@ static size_t ZSTDMT_computeOverlapSize(const ZSTD_CCtx_params* params)
1226
1196
  int const overlapRLog = 9 - ZSTDMT_overlapLog(params->overlapLog, params->cParams.strategy);
1227
1197
  int ovLog = (overlapRLog >= 8) ? 0 : (params->cParams.windowLog - overlapRLog);
1228
1198
  assert(0 <= overlapRLog && overlapRLog <= 8);
1229
- if (params->ldmParams.enableLdm) {
1199
+ if (params->ldmParams.enableLdm == ZSTD_ps_enable) {
1230
1200
  /* In Long Range Mode, the windowLog is typically oversized.
1231
1201
  * In which case, it's preferable to determine the jobSize
1232
1202
  * based on chainLog instead.
@@ -1240,174 +1210,6 @@ static size_t ZSTDMT_computeOverlapSize(const ZSTD_CCtx_params* params)
1240
1210
  return (ovLog==0) ? 0 : (size_t)1 << ovLog;
1241
1211
  }
1242
1212
 
1243
- static unsigned
1244
- ZSTDMT_computeNbJobs(const ZSTD_CCtx_params* params, size_t srcSize, unsigned nbWorkers)
1245
- {
1246
- assert(nbWorkers>0);
1247
- { size_t const jobSizeTarget = (size_t)1 << ZSTDMT_computeTargetJobLog(params);
1248
- size_t const jobMaxSize = jobSizeTarget << 2;
1249
- size_t const passSizeMax = jobMaxSize * nbWorkers;
1250
- unsigned const multiplier = (unsigned)(srcSize / passSizeMax) + 1;
1251
- unsigned const nbJobsLarge = multiplier * nbWorkers;
1252
- unsigned const nbJobsMax = (unsigned)(srcSize / jobSizeTarget) + 1;
1253
- unsigned const nbJobsSmall = MIN(nbJobsMax, nbWorkers);
1254
- return (multiplier>1) ? nbJobsLarge : nbJobsSmall;
1255
- } }
1256
-
1257
- /* ZSTDMT_compress_advanced_internal() :
1258
- * This is a blocking function : it will only give back control to caller after finishing its compression job.
1259
- */
1260
- static size_t
1261
- ZSTDMT_compress_advanced_internal(
1262
- ZSTDMT_CCtx* mtctx,
1263
- void* dst, size_t dstCapacity,
1264
- const void* src, size_t srcSize,
1265
- const ZSTD_CDict* cdict,
1266
- ZSTD_CCtx_params params)
1267
- {
1268
- ZSTD_CCtx_params const jobParams = ZSTDMT_initJobCCtxParams(&params);
1269
- size_t const overlapSize = ZSTDMT_computeOverlapSize(&params);
1270
- unsigned const nbJobs = ZSTDMT_computeNbJobs(&params, srcSize, params.nbWorkers);
1271
- size_t const proposedJobSize = (srcSize + (nbJobs-1)) / nbJobs;
1272
- size_t const avgJobSize = (((proposedJobSize-1) & 0x1FFFF) < 0x7FFF) ? proposedJobSize + 0xFFFF : proposedJobSize; /* avoid too small last block */
1273
- const char* const srcStart = (const char*)src;
1274
- size_t remainingSrcSize = srcSize;
1275
- unsigned const compressWithinDst = (dstCapacity >= ZSTD_compressBound(srcSize)) ? nbJobs : (unsigned)(dstCapacity / ZSTD_compressBound(avgJobSize)); /* presumes avgJobSize >= 256 KB, which should be the case */
1276
- size_t frameStartPos = 0, dstBufferPos = 0;
1277
- assert(jobParams.nbWorkers == 0);
1278
- assert(mtctx->cctxPool->totalCCtx == params.nbWorkers);
1279
-
1280
- params.jobSize = (U32)avgJobSize;
1281
- DEBUGLOG(4, "ZSTDMT_compress_advanced_internal: nbJobs=%2u (rawSize=%u bytes; fixedSize=%u) ",
1282
- nbJobs, (U32)proposedJobSize, (U32)avgJobSize);
1283
-
1284
- if ((nbJobs==1) | (params.nbWorkers<=1)) { /* fallback to single-thread mode : this is a blocking invocation anyway */
1285
- ZSTD_CCtx* const cctx = mtctx->cctxPool->cctx[0];
1286
- DEBUGLOG(4, "ZSTDMT_compress_advanced_internal: fallback to single-thread mode");
1287
- if (cdict) return ZSTD_compress_usingCDict_advanced(cctx, dst, dstCapacity, src, srcSize, cdict, jobParams.fParams);
1288
- return ZSTD_compress_advanced_internal(cctx, dst, dstCapacity, src, srcSize, NULL, 0, &jobParams);
1289
- }
1290
-
1291
- assert(avgJobSize >= 256 KB); /* condition for ZSTD_compressBound(A) + ZSTD_compressBound(B) <= ZSTD_compressBound(A+B), required to compress directly into Dst (no additional buffer) */
1292
- ZSTDMT_setBufferSize(mtctx->bufPool, ZSTD_compressBound(avgJobSize) );
1293
- /* LDM doesn't even try to load the dictionary in single-ingestion mode */
1294
- if (ZSTDMT_serialState_reset(&mtctx->serial, mtctx->seqPool, params, avgJobSize, NULL, 0, ZSTD_dct_auto))
1295
- return ERROR(memory_allocation);
1296
-
1297
- FORWARD_IF_ERROR( ZSTDMT_expandJobsTable(mtctx, nbJobs) , ""); /* only expands if necessary */
1298
-
1299
- { unsigned u;
1300
- for (u=0; u<nbJobs; u++) {
1301
- size_t const jobSize = MIN(remainingSrcSize, avgJobSize);
1302
- size_t const dstBufferCapacity = ZSTD_compressBound(jobSize);
1303
- buffer_t const dstAsBuffer = { (char*)dst + dstBufferPos, dstBufferCapacity };
1304
- buffer_t const dstBuffer = u < compressWithinDst ? dstAsBuffer : g_nullBuffer;
1305
- size_t dictSize = u ? overlapSize : 0;
1306
-
1307
- mtctx->jobs[u].prefix.start = srcStart + frameStartPos - dictSize;
1308
- mtctx->jobs[u].prefix.size = dictSize;
1309
- mtctx->jobs[u].src.start = srcStart + frameStartPos;
1310
- mtctx->jobs[u].src.size = jobSize; assert(jobSize > 0); /* avoid job.src.size == 0 */
1311
- mtctx->jobs[u].consumed = 0;
1312
- mtctx->jobs[u].cSize = 0;
1313
- mtctx->jobs[u].cdict = (u==0) ? cdict : NULL;
1314
- mtctx->jobs[u].fullFrameSize = srcSize;
1315
- mtctx->jobs[u].params = jobParams;
1316
- /* do not calculate checksum within sections, but write it in header for first section */
1317
- mtctx->jobs[u].dstBuff = dstBuffer;
1318
- mtctx->jobs[u].cctxPool = mtctx->cctxPool;
1319
- mtctx->jobs[u].bufPool = mtctx->bufPool;
1320
- mtctx->jobs[u].seqPool = mtctx->seqPool;
1321
- mtctx->jobs[u].serial = &mtctx->serial;
1322
- mtctx->jobs[u].jobID = u;
1323
- mtctx->jobs[u].firstJob = (u==0);
1324
- mtctx->jobs[u].lastJob = (u==nbJobs-1);
1325
-
1326
- DEBUGLOG(5, "ZSTDMT_compress_advanced_internal: posting job %u (%u bytes)", u, (U32)jobSize);
1327
- DEBUG_PRINTHEX(6, mtctx->jobs[u].prefix.start, 12);
1328
- POOL_add(mtctx->factory, ZSTDMT_compressionJob, &mtctx->jobs[u]);
1329
-
1330
- frameStartPos += jobSize;
1331
- dstBufferPos += dstBufferCapacity;
1332
- remainingSrcSize -= jobSize;
1333
- } }
1334
-
1335
- /* collect result */
1336
- { size_t error = 0, dstPos = 0;
1337
- unsigned jobID;
1338
- for (jobID=0; jobID<nbJobs; jobID++) {
1339
- DEBUGLOG(5, "waiting for job %u ", jobID);
1340
- ZSTD_PTHREAD_MUTEX_LOCK(&mtctx->jobs[jobID].job_mutex);
1341
- while (mtctx->jobs[jobID].consumed < mtctx->jobs[jobID].src.size) {
1342
- DEBUGLOG(5, "waiting for jobCompleted signal from job %u", jobID);
1343
- ZSTD_pthread_cond_wait(&mtctx->jobs[jobID].job_cond, &mtctx->jobs[jobID].job_mutex);
1344
- }
1345
- ZSTD_pthread_mutex_unlock(&mtctx->jobs[jobID].job_mutex);
1346
- DEBUGLOG(5, "ready to write job %u ", jobID);
1347
-
1348
- { size_t const cSize = mtctx->jobs[jobID].cSize;
1349
- if (ZSTD_isError(cSize)) error = cSize;
1350
- if ((!error) && (dstPos + cSize > dstCapacity)) error = ERROR(dstSize_tooSmall);
1351
- if (jobID) { /* note : job 0 is written directly at dst, which is correct position */
1352
- if (!error)
1353
- memmove((char*)dst + dstPos, mtctx->jobs[jobID].dstBuff.start, cSize); /* may overlap when job compressed within dst */
1354
- if (jobID >= compressWithinDst) { /* job compressed into its own buffer, which must be released */
1355
- DEBUGLOG(5, "releasing buffer %u>=%u", jobID, compressWithinDst);
1356
- ZSTDMT_releaseBuffer(mtctx->bufPool, mtctx->jobs[jobID].dstBuff);
1357
- } }
1358
- mtctx->jobs[jobID].dstBuff = g_nullBuffer;
1359
- mtctx->jobs[jobID].cSize = 0;
1360
- dstPos += cSize ;
1361
- }
1362
- } /* for (jobID=0; jobID<nbJobs; jobID++) */
1363
-
1364
- DEBUGLOG(4, "checksumFlag : %u ", params.fParams.checksumFlag);
1365
- if (params.fParams.checksumFlag) {
1366
- U32 const checksum = (U32)XXH64_digest(&mtctx->serial.xxhState);
1367
- if (dstPos + 4 > dstCapacity) {
1368
- error = ERROR(dstSize_tooSmall);
1369
- } else {
1370
- DEBUGLOG(4, "writing checksum : %08X \n", checksum);
1371
- MEM_writeLE32((char*)dst + dstPos, checksum);
1372
- dstPos += 4;
1373
- } }
1374
-
1375
- if (!error) DEBUGLOG(4, "compressed size : %u ", (U32)dstPos);
1376
- return error ? error : dstPos;
1377
- }
1378
- }
1379
-
1380
- size_t ZSTDMT_compress_advanced(ZSTDMT_CCtx* mtctx,
1381
- void* dst, size_t dstCapacity,
1382
- const void* src, size_t srcSize,
1383
- const ZSTD_CDict* cdict,
1384
- ZSTD_parameters params,
1385
- int overlapLog)
1386
- {
1387
- ZSTD_CCtx_params cctxParams = mtctx->params;
1388
- cctxParams.cParams = params.cParams;
1389
- cctxParams.fParams = params.fParams;
1390
- assert(ZSTD_OVERLAPLOG_MIN <= overlapLog && overlapLog <= ZSTD_OVERLAPLOG_MAX);
1391
- cctxParams.overlapLog = overlapLog;
1392
- return ZSTDMT_compress_advanced_internal(mtctx,
1393
- dst, dstCapacity,
1394
- src, srcSize,
1395
- cdict, cctxParams);
1396
- }
1397
-
1398
-
1399
- size_t ZSTDMT_compressCCtx(ZSTDMT_CCtx* mtctx,
1400
- void* dst, size_t dstCapacity,
1401
- const void* src, size_t srcSize,
1402
- int compressionLevel)
1403
- {
1404
- ZSTD_parameters params = ZSTD_getParams(compressionLevel, srcSize, 0);
1405
- int const overlapLog = ZSTDMT_overlapLog_default(params.cParams.strategy);
1406
- params.fParams.contentSizeFlag = 1;
1407
- return ZSTDMT_compress_advanced(mtctx, dst, dstCapacity, src, srcSize, NULL, params, overlapLog);
1408
- }
1409
-
1410
-
1411
1213
  /* ====================================== */
1412
1214
  /* ======= Streaming API ======= */
1413
1215
  /* ====================================== */
@@ -1432,16 +1234,6 @@ size_t ZSTDMT_initCStream_internal(
1432
1234
  if (params.jobSize != 0 && params.jobSize < ZSTDMT_JOBSIZE_MIN) params.jobSize = ZSTDMT_JOBSIZE_MIN;
1433
1235
  if (params.jobSize > (size_t)ZSTDMT_JOBSIZE_MAX) params.jobSize = (size_t)ZSTDMT_JOBSIZE_MAX;
1434
1236
 
1435
- mtctx->singleBlockingThread = (pledgedSrcSize <= ZSTDMT_JOBSIZE_MIN); /* do not trigger multi-threading when srcSize is too small */
1436
- if (mtctx->singleBlockingThread) {
1437
- ZSTD_CCtx_params const singleThreadParams = ZSTDMT_initJobCCtxParams(&params);
1438
- DEBUGLOG(5, "ZSTDMT_initCStream_internal: switch to single blocking thread mode");
1439
- assert(singleThreadParams.nbWorkers == 0);
1440
- return ZSTD_initCStream_internal(mtctx->cctxPool->cctx[0],
1441
- dict, dictSize, cdict,
1442
- &singleThreadParams, pledgedSrcSize);
1443
- }
1444
-
1445
1237
  DEBUGLOG(4, "ZSTDMT_initCStream_internal: %u workers", params.nbWorkers);
1446
1238
 
1447
1239
  if (mtctx->allJobsCompleted == 0) { /* previous compression not correctly finished */
@@ -1475,9 +1267,11 @@ size_t ZSTDMT_initCStream_internal(
1475
1267
 
1476
1268
  if (params.rsyncable) {
1477
1269
  /* Aim for the targetsectionSize as the average job size. */
1478
- U32 const jobSizeMB = (U32)(mtctx->targetSectionSize >> 20);
1479
- U32 const rsyncBits = ZSTD_highbit32(jobSizeMB) + 20;
1480
- assert(jobSizeMB >= 1);
1270
+ U32 const jobSizeKB = (U32)(mtctx->targetSectionSize >> 10);
1271
+ U32 const rsyncBits = (assert(jobSizeKB >= 1), ZSTD_highbit32(jobSizeKB) + 10);
1272
+ /* We refuse to create jobs < RSYNC_MIN_BLOCK_SIZE bytes, so make sure our
1273
+ * expected job size is at least 4x larger. */
1274
+ assert(rsyncBits >= RSYNC_MIN_BLOCK_LOG + 2);
1481
1275
  DEBUGLOG(4, "rsyncLog = %u", rsyncBits);
1482
1276
  mtctx->rsync.hash = 0;
1483
1277
  mtctx->rsync.hitMask = (1ULL << rsyncBits) - 1;
@@ -1489,7 +1283,7 @@ size_t ZSTDMT_initCStream_internal(
1489
1283
  ZSTDMT_setBufferSize(mtctx->bufPool, ZSTD_compressBound(mtctx->targetSectionSize));
1490
1284
  {
1491
1285
  /* If ldm is enabled we need windowSize space. */
1492
- size_t const windowSize = mtctx->params.ldmParams.enableLdm ? (1U << mtctx->params.cParams.windowLog) : 0;
1286
+ size_t const windowSize = mtctx->params.ldmParams.enableLdm == ZSTD_ps_enable ? (1U << mtctx->params.cParams.windowLog) : 0;
1493
1287
  /* Two buffers of slack, plus extra space for the overlap
1494
1288
  * This is the minimum slack that LDM works with. One extra because
1495
1289
  * flush might waste up to targetSectionSize-1 bytes. Another extra
@@ -1504,8 +1298,8 @@ size_t ZSTDMT_initCStream_internal(
1504
1298
  size_t const capacity = MAX(windowSize, sectionsSize) + slackSize;
1505
1299
  if (mtctx->roundBuff.capacity < capacity) {
1506
1300
  if (mtctx->roundBuff.buffer)
1507
- ZSTD_free(mtctx->roundBuff.buffer, mtctx->cMem);
1508
- mtctx->roundBuff.buffer = (BYTE*)ZSTD_malloc(capacity, mtctx->cMem);
1301
+ ZSTD_customFree(mtctx->roundBuff.buffer, mtctx->cMem);
1302
+ mtctx->roundBuff.buffer = (BYTE*)ZSTD_customMalloc(capacity, mtctx->cMem);
1509
1303
  if (mtctx->roundBuff.buffer == NULL) {
1510
1304
  mtctx->roundBuff.capacity = 0;
1511
1305
  return ERROR(memory_allocation);
@@ -1530,53 +1324,6 @@ size_t ZSTDMT_initCStream_internal(
1530
1324
  return 0;
1531
1325
  }
1532
1326
 
1533
- size_t ZSTDMT_initCStream_advanced(ZSTDMT_CCtx* mtctx,
1534
- const void* dict, size_t dictSize,
1535
- ZSTD_parameters params,
1536
- unsigned long long pledgedSrcSize)
1537
- {
1538
- ZSTD_CCtx_params cctxParams = mtctx->params; /* retrieve sticky params */
1539
- DEBUGLOG(4, "ZSTDMT_initCStream_advanced (pledgedSrcSize=%u)", (U32)pledgedSrcSize);
1540
- cctxParams.cParams = params.cParams;
1541
- cctxParams.fParams = params.fParams;
1542
- return ZSTDMT_initCStream_internal(mtctx, dict, dictSize, ZSTD_dct_auto, NULL,
1543
- cctxParams, pledgedSrcSize);
1544
- }
1545
-
1546
- size_t ZSTDMT_initCStream_usingCDict(ZSTDMT_CCtx* mtctx,
1547
- const ZSTD_CDict* cdict,
1548
- ZSTD_frameParameters fParams,
1549
- unsigned long long pledgedSrcSize)
1550
- {
1551
- ZSTD_CCtx_params cctxParams = mtctx->params;
1552
- if (cdict==NULL) return ERROR(dictionary_wrong); /* method incompatible with NULL cdict */
1553
- cctxParams.cParams = ZSTD_getCParamsFromCDict(cdict);
1554
- cctxParams.fParams = fParams;
1555
- return ZSTDMT_initCStream_internal(mtctx, NULL, 0 /*dictSize*/, ZSTD_dct_auto, cdict,
1556
- cctxParams, pledgedSrcSize);
1557
- }
1558
-
1559
-
1560
- /* ZSTDMT_resetCStream() :
1561
- * pledgedSrcSize can be zero == unknown (for the time being)
1562
- * prefer using ZSTD_CONTENTSIZE_UNKNOWN,
1563
- * as `0` might mean "empty" in the future */
1564
- size_t ZSTDMT_resetCStream(ZSTDMT_CCtx* mtctx, unsigned long long pledgedSrcSize)
1565
- {
1566
- if (!pledgedSrcSize) pledgedSrcSize = ZSTD_CONTENTSIZE_UNKNOWN;
1567
- return ZSTDMT_initCStream_internal(mtctx, NULL, 0, ZSTD_dct_auto, 0, mtctx->params,
1568
- pledgedSrcSize);
1569
- }
1570
-
1571
- size_t ZSTDMT_initCStream(ZSTDMT_CCtx* mtctx, int compressionLevel) {
1572
- ZSTD_parameters const params = ZSTD_getParams(compressionLevel, ZSTD_CONTENTSIZE_UNKNOWN, 0);
1573
- ZSTD_CCtx_params cctxParams = mtctx->params; /* retrieve sticky params */
1574
- DEBUGLOG(4, "ZSTDMT_initCStream (cLevel=%i)", compressionLevel);
1575
- cctxParams.cParams = params.cParams;
1576
- cctxParams.fParams = params.fParams;
1577
- return ZSTDMT_initCStream_internal(mtctx, NULL, 0, ZSTD_dct_auto, NULL, cctxParams, ZSTD_CONTENTSIZE_UNKNOWN);
1578
- }
1579
-
1580
1327
 
1581
1328
  /* ZSTDMT_writeLastEmptyBlock()
1582
1329
  * Write a single empty block with an end-of-frame to finish a frame.
@@ -1740,7 +1487,7 @@ static size_t ZSTDMT_flushProduced(ZSTDMT_CCtx* mtctx, ZSTD_outBuffer* output, u
1740
1487
  assert(cSize >= mtctx->jobs[wJobID].dstFlushed);
1741
1488
  assert(mtctx->jobs[wJobID].dstBuff.start != NULL);
1742
1489
  if (toFlush > 0) {
1743
- memcpy((char*)output->dst + output->pos,
1490
+ ZSTD_memcpy((char*)output->dst + output->pos,
1744
1491
  (const char*)mtctx->jobs[wJobID].dstBuff.start + mtctx->jobs[wJobID].dstFlushed,
1745
1492
  toFlush);
1746
1493
  }
@@ -1811,17 +1558,21 @@ static range_t ZSTDMT_getInputDataInUse(ZSTDMT_CCtx* mtctx)
1811
1558
  static int ZSTDMT_isOverlapped(buffer_t buffer, range_t range)
1812
1559
  {
1813
1560
  BYTE const* const bufferStart = (BYTE const*)buffer.start;
1814
- BYTE const* const bufferEnd = bufferStart + buffer.capacity;
1815
1561
  BYTE const* const rangeStart = (BYTE const*)range.start;
1816
- BYTE const* const rangeEnd = range.size != 0 ? rangeStart + range.size : rangeStart;
1817
1562
 
1818
1563
  if (rangeStart == NULL || bufferStart == NULL)
1819
1564
  return 0;
1820
- /* Empty ranges cannot overlap */
1821
- if (bufferStart == bufferEnd || rangeStart == rangeEnd)
1822
- return 0;
1823
1565
 
1824
- return bufferStart < rangeEnd && rangeStart < bufferEnd;
1566
+ {
1567
+ BYTE const* const bufferEnd = bufferStart + buffer.capacity;
1568
+ BYTE const* const rangeEnd = rangeStart + range.size;
1569
+
1570
+ /* Empty ranges cannot overlap */
1571
+ if (bufferStart == bufferEnd || rangeStart == rangeEnd)
1572
+ return 0;
1573
+
1574
+ return bufferStart < rangeEnd && rangeStart < bufferEnd;
1575
+ }
1825
1576
  }
1826
1577
 
1827
1578
  static int ZSTDMT_doesOverlapWindow(buffer_t buffer, ZSTD_window_t window)
@@ -1848,7 +1599,7 @@ static int ZSTDMT_doesOverlapWindow(buffer_t buffer, ZSTD_window_t window)
1848
1599
 
1849
1600
  static void ZSTDMT_waitForLdmComplete(ZSTDMT_CCtx* mtctx, buffer_t buffer)
1850
1601
  {
1851
- if (mtctx->params.ldmParams.enableLdm) {
1602
+ if (mtctx->params.ldmParams.enableLdm == ZSTD_ps_enable) {
1852
1603
  ZSTD_pthread_mutex_t* mutex = &mtctx->serial.ldmWindowMutex;
1853
1604
  DEBUGLOG(5, "ZSTDMT_waitForLdmComplete");
1854
1605
  DEBUGLOG(5, "source [0x%zx, 0x%zx)",
@@ -1894,7 +1645,7 @@ static int ZSTDMT_tryGetInputRange(ZSTDMT_CCtx* mtctx)
1894
1645
  return 0;
1895
1646
  }
1896
1647
  ZSTDMT_waitForLdmComplete(mtctx, buffer);
1897
- memmove(start, mtctx->inBuff.prefix.start, prefixSize);
1648
+ ZSTD_memmove(start, mtctx->inBuff.prefix.start, prefixSize);
1898
1649
  mtctx->inBuff.prefix.start = start;
1899
1650
  mtctx->roundBuff.pos = prefixSize;
1900
1651
  }
@@ -1951,6 +1702,11 @@ findSynchronizationPoint(ZSTDMT_CCtx const* mtctx, ZSTD_inBuffer const input)
1951
1702
  if (!mtctx->params.rsyncable)
1952
1703
  /* Rsync is disabled. */
1953
1704
  return syncPoint;
1705
+ if (mtctx->inBuff.filled + input.size - input.pos < RSYNC_MIN_BLOCK_SIZE)
1706
+ /* We don't emit synchronization points if it would produce too small blocks.
1707
+ * We don't have enough input to find a synchronization point, so don't look.
1708
+ */
1709
+ return syncPoint;
1954
1710
  if (mtctx->inBuff.filled + syncPoint.toLoad < RSYNC_LENGTH)
1955
1711
  /* Not enough to compute the hash.
1956
1712
  * We will miss any synchronization points in this RSYNC_LENGTH byte
@@ -1961,23 +1717,41 @@ findSynchronizationPoint(ZSTDMT_CCtx const* mtctx, ZSTD_inBuffer const input)
1961
1717
  */
1962
1718
  return syncPoint;
1963
1719
  /* Initialize the loop variables. */
1964
- if (mtctx->inBuff.filled >= RSYNC_LENGTH) {
1965
- /* We have enough bytes buffered to initialize the hash.
1720
+ if (mtctx->inBuff.filled < RSYNC_MIN_BLOCK_SIZE) {
1721
+ /* We don't need to scan the first RSYNC_MIN_BLOCK_SIZE positions
1722
+ * because they can't possibly be a sync point. So we can start
1723
+ * part way through the input buffer.
1724
+ */
1725
+ pos = RSYNC_MIN_BLOCK_SIZE - mtctx->inBuff.filled;
1726
+ if (pos >= RSYNC_LENGTH) {
1727
+ prev = istart + pos - RSYNC_LENGTH;
1728
+ hash = ZSTD_rollingHash_compute(prev, RSYNC_LENGTH);
1729
+ } else {
1730
+ assert(mtctx->inBuff.filled >= RSYNC_LENGTH);
1731
+ prev = (BYTE const*)mtctx->inBuff.buffer.start + mtctx->inBuff.filled - RSYNC_LENGTH;
1732
+ hash = ZSTD_rollingHash_compute(prev + pos, (RSYNC_LENGTH - pos));
1733
+ hash = ZSTD_rollingHash_append(hash, istart, pos);
1734
+ }
1735
+ } else {
1736
+ /* We have enough bytes buffered to initialize the hash,
1737
+ * and are have processed enough bytes to find a sync point.
1966
1738
  * Start scanning at the beginning of the input.
1967
1739
  */
1740
+ assert(mtctx->inBuff.filled >= RSYNC_MIN_BLOCK_SIZE);
1741
+ assert(RSYNC_MIN_BLOCK_SIZE >= RSYNC_LENGTH);
1968
1742
  pos = 0;
1969
1743
  prev = (BYTE const*)mtctx->inBuff.buffer.start + mtctx->inBuff.filled - RSYNC_LENGTH;
1970
1744
  hash = ZSTD_rollingHash_compute(prev, RSYNC_LENGTH);
1971
- } else {
1972
- /* We don't have enough bytes buffered to initialize the hash, but
1973
- * we know we have at least RSYNC_LENGTH bytes total.
1974
- * Start scanning after the first RSYNC_LENGTH bytes less the bytes
1975
- * already buffered.
1976
- */
1977
- pos = RSYNC_LENGTH - mtctx->inBuff.filled;
1978
- prev = (BYTE const*)mtctx->inBuff.buffer.start - pos;
1979
- hash = ZSTD_rollingHash_compute(mtctx->inBuff.buffer.start, mtctx->inBuff.filled);
1980
- hash = ZSTD_rollingHash_append(hash, istart, pos);
1745
+ if ((hash & hitMask) == hitMask) {
1746
+ /* We're already at a sync point so don't load any more until
1747
+ * we're able to flush this sync point.
1748
+ * This likely happened because the job table was full so we
1749
+ * couldn't add our job.
1750
+ */
1751
+ syncPoint.toLoad = 0;
1752
+ syncPoint.flush = 1;
1753
+ return syncPoint;
1754
+ }
1981
1755
  }
1982
1756
  /* Starting with the hash of the previous RSYNC_LENGTH bytes, roll
1983
1757
  * through the input. If we hit a synchronization point, then cut the
@@ -1989,8 +1763,9 @@ findSynchronizationPoint(ZSTDMT_CCtx const* mtctx, ZSTD_inBuffer const input)
1989
1763
  */
1990
1764
  for (; pos < syncPoint.toLoad; ++pos) {
1991
1765
  BYTE const toRemove = pos < RSYNC_LENGTH ? prev[pos] : istart[pos - RSYNC_LENGTH];
1992
- /* if (pos >= RSYNC_LENGTH) assert(ZSTD_rollingHash_compute(istart + pos - RSYNC_LENGTH, RSYNC_LENGTH) == hash); */
1766
+ assert(pos < RSYNC_LENGTH || ZSTD_rollingHash_compute(istart + pos - RSYNC_LENGTH, RSYNC_LENGTH) == hash);
1993
1767
  hash = ZSTD_rollingHash_rotate(hash, toRemove, istart[pos], primePower);
1768
+ assert(mtctx->inBuff.filled + pos >= RSYNC_MIN_BLOCK_SIZE);
1994
1769
  if ((hash & hitMask) == hitMask) {
1995
1770
  syncPoint.toLoad = pos + 1;
1996
1771
  syncPoint.flush = 1;
@@ -2022,34 +1797,11 @@ size_t ZSTDMT_compressStream_generic(ZSTDMT_CCtx* mtctx,
2022
1797
  assert(output->pos <= output->size);
2023
1798
  assert(input->pos <= input->size);
2024
1799
 
2025
- if (mtctx->singleBlockingThread) { /* delegate to single-thread (synchronous) */
2026
- return ZSTD_compressStream2(mtctx->cctxPool->cctx[0], output, input, endOp);
2027
- }
2028
-
2029
1800
  if ((mtctx->frameEnded) && (endOp==ZSTD_e_continue)) {
2030
1801
  /* current frame being ended. Only flush/end are allowed */
2031
1802
  return ERROR(stage_wrong);
2032
1803
  }
2033
1804
 
2034
- /* single-pass shortcut (note : synchronous-mode) */
2035
- if ( (!mtctx->params.rsyncable) /* rsyncable mode is disabled */
2036
- && (mtctx->nextJobID == 0) /* just started */
2037
- && (mtctx->inBuff.filled == 0) /* nothing buffered */
2038
- && (!mtctx->jobReady) /* no job already created */
2039
- && (endOp == ZSTD_e_end) /* end order */
2040
- && (output->size - output->pos >= ZSTD_compressBound(input->size - input->pos)) ) { /* enough space in dst */
2041
- size_t const cSize = ZSTDMT_compress_advanced_internal(mtctx,
2042
- (char*)output->dst + output->pos, output->size - output->pos,
2043
- (const char*)input->src + input->pos, input->size - input->pos,
2044
- mtctx->cdict, mtctx->params);
2045
- if (ZSTD_isError(cSize)) return cSize;
2046
- input->pos = input->size;
2047
- output->pos += cSize;
2048
- mtctx->allJobsCompleted = 1;
2049
- mtctx->frameEnded = 1;
2050
- return 0;
2051
- }
2052
-
2053
1805
  /* fill input buffer */
2054
1806
  if ( (!mtctx->jobReady)
2055
1807
  && (input->size > input->pos) ) { /* support NULL input */
@@ -2072,13 +1824,21 @@ size_t ZSTDMT_compressStream_generic(ZSTDMT_CCtx* mtctx,
2072
1824
  assert(mtctx->inBuff.buffer.capacity >= mtctx->targetSectionSize);
2073
1825
  DEBUGLOG(5, "ZSTDMT_compressStream_generic: adding %u bytes on top of %u to buffer of size %u",
2074
1826
  (U32)syncPoint.toLoad, (U32)mtctx->inBuff.filled, (U32)mtctx->targetSectionSize);
2075
- memcpy((char*)mtctx->inBuff.buffer.start + mtctx->inBuff.filled, (const char*)input->src + input->pos, syncPoint.toLoad);
1827
+ ZSTD_memcpy((char*)mtctx->inBuff.buffer.start + mtctx->inBuff.filled, (const char*)input->src + input->pos, syncPoint.toLoad);
2076
1828
  input->pos += syncPoint.toLoad;
2077
1829
  mtctx->inBuff.filled += syncPoint.toLoad;
2078
1830
  forwardInputProgress = syncPoint.toLoad>0;
2079
1831
  }
2080
- if ((input->pos < input->size) && (endOp == ZSTD_e_end))
2081
- endOp = ZSTD_e_flush; /* can't end now : not all input consumed */
1832
+ }
1833
+ if ((input->pos < input->size) && (endOp == ZSTD_e_end)) {
1834
+ /* Can't end yet because the input is not fully consumed.
1835
+ * We are in one of these cases:
1836
+ * - mtctx->inBuff is NULL & empty: we couldn't get an input buffer so don't create a new job.
1837
+ * - We filled the input buffer: flush this job but don't end the frame.
1838
+ * - We hit a synchronization point: flush this job but don't end the frame.
1839
+ */
1840
+ assert(mtctx->inBuff.filled == 0 || mtctx->inBuff.filled == mtctx->targetSectionSize || mtctx->params.rsyncable);
1841
+ endOp = ZSTD_e_flush;
2082
1842
  }
2083
1843
 
2084
1844
  if ( (mtctx->jobReady)
@@ -2097,47 +1857,3 @@ size_t ZSTDMT_compressStream_generic(ZSTDMT_CCtx* mtctx,
2097
1857
  return remainingToFlush;
2098
1858
  }
2099
1859
  }
2100
-
2101
-
2102
- size_t ZSTDMT_compressStream(ZSTDMT_CCtx* mtctx, ZSTD_outBuffer* output, ZSTD_inBuffer* input)
2103
- {
2104
- FORWARD_IF_ERROR( ZSTDMT_compressStream_generic(mtctx, output, input, ZSTD_e_continue) , "");
2105
-
2106
- /* recommended next input size : fill current input buffer */
2107
- return mtctx->targetSectionSize - mtctx->inBuff.filled; /* note : could be zero when input buffer is fully filled and no more availability to create new job */
2108
- }
2109
-
2110
-
2111
- static size_t ZSTDMT_flushStream_internal(ZSTDMT_CCtx* mtctx, ZSTD_outBuffer* output, ZSTD_EndDirective endFrame)
2112
- {
2113
- size_t const srcSize = mtctx->inBuff.filled;
2114
- DEBUGLOG(5, "ZSTDMT_flushStream_internal");
2115
-
2116
- if ( mtctx->jobReady /* one job ready for a worker to pick up */
2117
- || (srcSize > 0) /* still some data within input buffer */
2118
- || ((endFrame==ZSTD_e_end) && !mtctx->frameEnded)) { /* need a last 0-size block to end frame */
2119
- DEBUGLOG(5, "ZSTDMT_flushStream_internal : create a new job (%u bytes, end:%u)",
2120
- (U32)srcSize, (U32)endFrame);
2121
- FORWARD_IF_ERROR( ZSTDMT_createCompressionJob(mtctx, srcSize, endFrame) , "");
2122
- }
2123
-
2124
- /* check if there is any data available to flush */
2125
- return ZSTDMT_flushProduced(mtctx, output, 1 /* blockToFlush */, endFrame);
2126
- }
2127
-
2128
-
2129
- size_t ZSTDMT_flushStream(ZSTDMT_CCtx* mtctx, ZSTD_outBuffer* output)
2130
- {
2131
- DEBUGLOG(5, "ZSTDMT_flushStream");
2132
- if (mtctx->singleBlockingThread)
2133
- return ZSTD_flushStream(mtctx->cctxPool->cctx[0], output);
2134
- return ZSTDMT_flushStream_internal(mtctx, output, ZSTD_e_flush);
2135
- }
2136
-
2137
- size_t ZSTDMT_endStream(ZSTDMT_CCtx* mtctx, ZSTD_outBuffer* output)
2138
- {
2139
- DEBUGLOG(4, "ZSTDMT_endStream");
2140
- if (mtctx->singleBlockingThread)
2141
- return ZSTD_endStream(mtctx->cctxPool->cctx[0], output);
2142
- return ZSTDMT_flushStream_internal(mtctx, output, ZSTD_e_end);
2143
- }