zstd-ruby 1.4.0.0 → 1.4.9.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (95) hide show
  1. checksums.yaml +4 -4
  2. data/.github/workflows/ruby.yml +35 -0
  3. data/README.md +2 -2
  4. data/ext/zstdruby/libzstd/Makefile +274 -107
  5. data/ext/zstdruby/libzstd/README.md +75 -16
  6. data/ext/zstdruby/libzstd/common/bitstream.h +59 -51
  7. data/ext/zstdruby/libzstd/common/compiler.h +154 -5
  8. data/ext/zstdruby/libzstd/common/cpu.h +1 -3
  9. data/ext/zstdruby/libzstd/common/debug.c +11 -31
  10. data/ext/zstdruby/libzstd/common/debug.h +22 -49
  11. data/ext/zstdruby/libzstd/common/entropy_common.c +201 -75
  12. data/ext/zstdruby/libzstd/common/error_private.c +3 -1
  13. data/ext/zstdruby/libzstd/common/error_private.h +7 -3
  14. data/ext/zstdruby/libzstd/common/fse.h +50 -42
  15. data/ext/zstdruby/libzstd/common/fse_decompress.c +134 -50
  16. data/ext/zstdruby/libzstd/common/huf.h +41 -38
  17. data/ext/zstdruby/libzstd/common/mem.h +68 -22
  18. data/ext/zstdruby/libzstd/common/pool.c +30 -20
  19. data/ext/zstdruby/libzstd/common/pool.h +3 -3
  20. data/ext/zstdruby/libzstd/common/threading.c +51 -4
  21. data/ext/zstdruby/libzstd/common/threading.h +36 -4
  22. data/ext/zstdruby/libzstd/common/xxhash.c +39 -89
  23. data/ext/zstdruby/libzstd/common/xxhash.h +12 -32
  24. data/ext/zstdruby/libzstd/common/zstd_common.c +10 -10
  25. data/ext/zstdruby/libzstd/common/zstd_deps.h +111 -0
  26. data/ext/zstdruby/libzstd/common/zstd_errors.h +3 -1
  27. data/ext/zstdruby/libzstd/common/zstd_internal.h +231 -72
  28. data/ext/zstdruby/libzstd/common/zstd_trace.c +42 -0
  29. data/ext/zstdruby/libzstd/common/zstd_trace.h +152 -0
  30. data/ext/zstdruby/libzstd/compress/fse_compress.c +47 -63
  31. data/ext/zstdruby/libzstd/compress/hist.c +41 -63
  32. data/ext/zstdruby/libzstd/compress/hist.h +13 -33
  33. data/ext/zstdruby/libzstd/compress/huf_compress.c +288 -172
  34. data/ext/zstdruby/libzstd/compress/zstd_compress.c +2504 -1626
  35. data/ext/zstdruby/libzstd/compress/zstd_compress_internal.h +446 -85
  36. data/ext/zstdruby/libzstd/compress/zstd_compress_literals.c +158 -0
  37. data/ext/zstdruby/libzstd/compress/zstd_compress_literals.h +29 -0
  38. data/ext/zstdruby/libzstd/compress/zstd_compress_sequences.c +433 -0
  39. data/ext/zstdruby/libzstd/compress/zstd_compress_sequences.h +54 -0
  40. data/ext/zstdruby/libzstd/compress/zstd_compress_superblock.c +849 -0
  41. data/ext/zstdruby/libzstd/compress/zstd_compress_superblock.h +32 -0
  42. data/ext/zstdruby/libzstd/compress/zstd_cwksp.h +561 -0
  43. data/ext/zstdruby/libzstd/compress/zstd_double_fast.c +82 -60
  44. data/ext/zstdruby/libzstd/compress/zstd_double_fast.h +2 -2
  45. data/ext/zstdruby/libzstd/compress/zstd_fast.c +106 -80
  46. data/ext/zstdruby/libzstd/compress/zstd_fast.h +2 -2
  47. data/ext/zstdruby/libzstd/compress/zstd_lazy.c +411 -105
  48. data/ext/zstdruby/libzstd/compress/zstd_lazy.h +21 -1
  49. data/ext/zstdruby/libzstd/compress/zstd_ldm.c +296 -207
  50. data/ext/zstdruby/libzstd/compress/zstd_ldm.h +14 -3
  51. data/ext/zstdruby/libzstd/compress/zstd_ldm_geartab.h +103 -0
  52. data/ext/zstdruby/libzstd/compress/zstd_opt.c +260 -148
  53. data/ext/zstdruby/libzstd/compress/zstd_opt.h +1 -1
  54. data/ext/zstdruby/libzstd/compress/zstdmt_compress.c +153 -440
  55. data/ext/zstdruby/libzstd/compress/zstdmt_compress.h +29 -110
  56. data/ext/zstdruby/libzstd/decompress/huf_decompress.c +356 -238
  57. data/ext/zstdruby/libzstd/decompress/zstd_ddict.c +20 -16
  58. data/ext/zstdruby/libzstd/decompress/zstd_ddict.h +3 -3
  59. data/ext/zstdruby/libzstd/decompress/zstd_decompress.c +641 -238
  60. data/ext/zstdruby/libzstd/decompress/zstd_decompress_block.c +600 -371
  61. data/ext/zstdruby/libzstd/decompress/zstd_decompress_block.h +8 -5
  62. data/ext/zstdruby/libzstd/decompress/zstd_decompress_internal.h +40 -9
  63. data/ext/zstdruby/libzstd/deprecated/zbuff.h +9 -8
  64. data/ext/zstdruby/libzstd/deprecated/zbuff_common.c +2 -2
  65. data/ext/zstdruby/libzstd/deprecated/zbuff_compress.c +1 -1
  66. data/ext/zstdruby/libzstd/deprecated/zbuff_decompress.c +1 -1
  67. data/ext/zstdruby/libzstd/dictBuilder/cover.c +197 -78
  68. data/ext/zstdruby/libzstd/dictBuilder/cover.h +52 -7
  69. data/ext/zstdruby/libzstd/dictBuilder/divsufsort.c +1 -1
  70. data/ext/zstdruby/libzstd/dictBuilder/fastcover.c +84 -66
  71. data/ext/zstdruby/libzstd/dictBuilder/zdict.c +58 -36
  72. data/ext/zstdruby/libzstd/dictBuilder/zdict.h +60 -31
  73. data/ext/zstdruby/libzstd/dll/example/Makefile +2 -1
  74. data/ext/zstdruby/libzstd/dll/example/README.md +16 -22
  75. data/ext/zstdruby/libzstd/legacy/zstd_legacy.h +8 -4
  76. data/ext/zstdruby/libzstd/legacy/zstd_v01.c +115 -111
  77. data/ext/zstdruby/libzstd/legacy/zstd_v01.h +1 -1
  78. data/ext/zstdruby/libzstd/legacy/zstd_v02.c +28 -14
  79. data/ext/zstdruby/libzstd/legacy/zstd_v02.h +1 -1
  80. data/ext/zstdruby/libzstd/legacy/zstd_v03.c +28 -14
  81. data/ext/zstdruby/libzstd/legacy/zstd_v03.h +1 -1
  82. data/ext/zstdruby/libzstd/legacy/zstd_v04.c +36 -19
  83. data/ext/zstdruby/libzstd/legacy/zstd_v04.h +1 -1
  84. data/ext/zstdruby/libzstd/legacy/zstd_v05.c +122 -107
  85. data/ext/zstdruby/libzstd/legacy/zstd_v05.h +2 -2
  86. data/ext/zstdruby/libzstd/legacy/zstd_v06.c +29 -23
  87. data/ext/zstdruby/libzstd/legacy/zstd_v06.h +1 -1
  88. data/ext/zstdruby/libzstd/legacy/zstd_v07.c +34 -24
  89. data/ext/zstdruby/libzstd/legacy/zstd_v07.h +1 -1
  90. data/ext/zstdruby/libzstd/libzstd.pc.in +2 -1
  91. data/ext/zstdruby/libzstd/zstd.h +655 -118
  92. data/lib/zstd-ruby/version.rb +1 -1
  93. data/zstd-ruby.gemspec +1 -1
  94. metadata +20 -10
  95. data/.travis.yml +0 -14
@@ -1,40 +1,20 @@
1
1
  /* ******************************************************************
2
- hist : Histogram functions
3
- part of Finite State Entropy project
4
- Copyright (C) 2013-present, Yann Collet.
5
-
6
- BSD 2-Clause License (http://www.opensource.org/licenses/bsd-license.php)
7
-
8
- Redistribution and use in source and binary forms, with or without
9
- modification, are permitted provided that the following conditions are
10
- met:
11
-
12
- * Redistributions of source code must retain the above copyright
13
- notice, this list of conditions and the following disclaimer.
14
- * Redistributions in binary form must reproduce the above
15
- copyright notice, this list of conditions and the following disclaimer
16
- in the documentation and/or other materials provided with the
17
- distribution.
18
-
19
- THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
20
- "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
21
- LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
22
- A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
23
- OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
24
- SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
25
- LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
26
- DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
27
- THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
28
- (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
29
- OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
30
-
31
- You can contact the author at :
32
- - FSE source repository : https://github.com/Cyan4973/FiniteStateEntropy
33
- - Public forum : https://groups.google.com/forum/#!forum/lz4c
2
+ * hist : Histogram functions
3
+ * part of Finite State Entropy project
4
+ * Copyright (c) 2013-2021, Yann Collet, Facebook, Inc.
5
+ *
6
+ * You can contact the author at :
7
+ * - FSE source repository : https://github.com/Cyan4973/FiniteStateEntropy
8
+ * - Public forum : https://groups.google.com/forum/#!forum/lz4c
9
+ *
10
+ * This source code is licensed under both the BSD-style license (found in the
11
+ * LICENSE file in the root directory of this source tree) and the GPLv2 (found
12
+ * in the COPYING file in the root directory of this source tree).
13
+ * You may select, at your option, one of the above-listed licenses.
34
14
  ****************************************************************** */
35
15
 
36
16
  /* --- dependencies --- */
37
- #include <stddef.h> /* size_t */
17
+ #include "../common/zstd_deps.h" /* size_t */
38
18
 
39
19
 
40
20
  /* --- simple histogram functions --- */
@@ -1,35 +1,15 @@
1
1
  /* ******************************************************************
2
- Huffman encoder, part of New Generation Entropy library
3
- Copyright (C) 2013-2016, Yann Collet.
4
-
5
- BSD 2-Clause License (http://www.opensource.org/licenses/bsd-license.php)
6
-
7
- Redistribution and use in source and binary forms, with or without
8
- modification, are permitted provided that the following conditions are
9
- met:
10
-
11
- * Redistributions of source code must retain the above copyright
12
- notice, this list of conditions and the following disclaimer.
13
- * Redistributions in binary form must reproduce the above
14
- copyright notice, this list of conditions and the following disclaimer
15
- in the documentation and/or other materials provided with the
16
- distribution.
17
-
18
- THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
19
- "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
20
- LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
21
- A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
22
- OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
23
- SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
24
- LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
25
- DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
26
- THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
27
- (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
28
- OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
29
-
30
- You can contact the author at :
31
- - FSE+HUF source repository : https://github.com/Cyan4973/FiniteStateEntropy
32
- - Public forum : https://groups.google.com/forum/#!forum/lz4c
2
+ * Huffman encoder, part of New Generation Entropy library
3
+ * Copyright (c) 2013-2021, Yann Collet, Facebook, Inc.
4
+ *
5
+ * You can contact the author at :
6
+ * - FSE+HUF source repository : https://github.com/Cyan4973/FiniteStateEntropy
7
+ * - Public forum : https://groups.google.com/forum/#!forum/lz4c
8
+ *
9
+ * This source code is licensed under both the BSD-style license (found in the
10
+ * LICENSE file in the root directory of this source tree) and the GPLv2 (found
11
+ * in the COPYING file in the root directory of this source tree).
12
+ * You may select, at your option, one of the above-listed licenses.
33
13
  ****************************************************************** */
34
14
 
35
15
  /* **************************************************************
@@ -43,16 +23,15 @@
43
23
  /* **************************************************************
44
24
  * Includes
45
25
  ****************************************************************/
46
- #include <string.h> /* memcpy, memset */
47
- #include <stdio.h> /* printf (debug) */
48
- #include "compiler.h"
49
- #include "bitstream.h"
26
+ #include "../common/zstd_deps.h" /* ZSTD_memcpy, ZSTD_memset */
27
+ #include "../common/compiler.h"
28
+ #include "../common/bitstream.h"
50
29
  #include "hist.h"
51
30
  #define FSE_STATIC_LINKING_ONLY /* FSE_optimalTableLog_internal */
52
- #include "fse.h" /* header compression */
31
+ #include "../common/fse.h" /* header compression */
53
32
  #define HUF_STATIC_LINKING_ONLY
54
- #include "huf.h"
55
- #include "error_private.h"
33
+ #include "../common/huf.h"
34
+ #include "../common/error_private.h"
56
35
 
57
36
 
58
37
  /* **************************************************************
@@ -60,8 +39,6 @@
60
39
  ****************************************************************/
61
40
  #define HUF_isError ERR_isError
62
41
  #define HUF_STATIC_ASSERT(c) DEBUG_STATIC_ASSERT(c) /* use only *after* variable declarations */
63
- #define CHECK_V_F(e, f) size_t const e = f; if (ERR_isError(e)) return e
64
- #define CHECK_F(f) { CHECK_V_F(_var_err__, f); }
65
42
 
66
43
 
67
44
  /* **************************************************************
@@ -92,7 +69,7 @@ static size_t HUF_compressWeights (void* dst, size_t dstSize, const void* weight
92
69
  U32 tableLog = MAX_FSE_TABLELOG_FOR_HUFF_HEADER;
93
70
 
94
71
  FSE_CTable CTable[FSE_CTABLE_SIZE_U32(MAX_FSE_TABLELOG_FOR_HUFF_HEADER, HUF_TABLELOG_MAX)];
95
- BYTE scratchBuffer[1<<MAX_FSE_TABLELOG_FOR_HUFF_HEADER];
72
+ U32 scratchBuffer[FSE_BUILD_CTABLE_WORKSPACE_SIZE_U32(HUF_TABLELOG_MAX, MAX_FSE_TABLELOG_FOR_HUFF_HEADER)];
96
73
 
97
74
  unsigned count[HUF_TABLELOG_MAX+1];
98
75
  S16 norm[HUF_TABLELOG_MAX+1];
@@ -107,29 +84,24 @@ static size_t HUF_compressWeights (void* dst, size_t dstSize, const void* weight
107
84
  }
108
85
 
109
86
  tableLog = FSE_optimalTableLog(tableLog, wtSize, maxSymbolValue);
110
- CHECK_F( FSE_normalizeCount(norm, tableLog, count, wtSize, maxSymbolValue) );
87
+ CHECK_F( FSE_normalizeCount(norm, tableLog, count, wtSize, maxSymbolValue, /* useLowProbCount */ 0) );
111
88
 
112
89
  /* Write table description header */
113
- { CHECK_V_F(hSize, FSE_writeNCount(op, oend-op, norm, maxSymbolValue, tableLog) );
90
+ { CHECK_V_F(hSize, FSE_writeNCount(op, (size_t)(oend-op), norm, maxSymbolValue, tableLog) );
114
91
  op += hSize;
115
92
  }
116
93
 
117
94
  /* Compress */
118
95
  CHECK_F( FSE_buildCTable_wksp(CTable, norm, maxSymbolValue, tableLog, scratchBuffer, sizeof(scratchBuffer)) );
119
- { CHECK_V_F(cSize, FSE_compress_usingCTable(op, oend - op, weightTable, wtSize, CTable) );
96
+ { CHECK_V_F(cSize, FSE_compress_usingCTable(op, (size_t)(oend - op), weightTable, wtSize, CTable) );
120
97
  if (cSize == 0) return 0; /* not enough space for compressed data */
121
98
  op += cSize;
122
99
  }
123
100
 
124
- return op-ostart;
101
+ return (size_t)(op-ostart);
125
102
  }
126
103
 
127
104
 
128
- struct HUF_CElt_s {
129
- U16 val;
130
- BYTE nbBits;
131
- }; /* typedef'd to HUF_CElt within "huf.h" */
132
-
133
105
  /*! HUF_writeCTable() :
134
106
  `CTable` : Huffman tree to save, using huf representation.
135
107
  @return : size of saved CTable */
@@ -169,7 +141,7 @@ size_t HUF_writeCTable (void* dst, size_t maxDstSize,
169
141
  }
170
142
 
171
143
 
172
- size_t HUF_readCTable (HUF_CElt* CTable, unsigned* maxSymbolValuePtr, const void* src, size_t srcSize)
144
+ size_t HUF_readCTable (HUF_CElt* CTable, unsigned* maxSymbolValuePtr, const void* src, size_t srcSize, unsigned* hasZeroWeights)
173
145
  {
174
146
  BYTE huffWeight[HUF_SYMBOLVALUE_MAX + 1]; /* init not required, even though some static analyzer may complain */
175
147
  U32 rankVal[HUF_TABLELOG_ABSOLUTEMAX + 1]; /* large enough for values from 0 to 16 */
@@ -178,6 +150,7 @@ size_t HUF_readCTable (HUF_CElt* CTable, unsigned* maxSymbolValuePtr, const void
178
150
 
179
151
  /* get symbol weights */
180
152
  CHECK_V_F(readSize, HUF_readStats(huffWeight, HUF_SYMBOLVALUE_MAX+1, rankVal, &nbSymbols, &tableLog, src, srcSize));
153
+ *hasZeroWeights = (rankVal[0] > 0);
181
154
 
182
155
  /* check result */
183
156
  if (tableLog > HUF_TABLELOG_MAX) return ERROR(tableLog_tooLarge);
@@ -186,15 +159,15 @@ size_t HUF_readCTable (HUF_CElt* CTable, unsigned* maxSymbolValuePtr, const void
186
159
  /* Prepare base value per rank */
187
160
  { U32 n, nextRankStart = 0;
188
161
  for (n=1; n<=tableLog; n++) {
189
- U32 current = nextRankStart;
162
+ U32 curr = nextRankStart;
190
163
  nextRankStart += (rankVal[n] << (n-1));
191
- rankVal[n] = current;
164
+ rankVal[n] = curr;
192
165
  } }
193
166
 
194
167
  /* fill nbBits */
195
168
  { U32 n; for (n=0; n<nbSymbols; n++) {
196
169
  const U32 w = huffWeight[n];
197
- CTable[n].nbBits = (BYTE)(tableLog + 1 - w);
170
+ CTable[n].nbBits = (BYTE)(tableLog + 1 - w) & -(w != 0);
198
171
  } }
199
172
 
200
173
  /* fill val */
@@ -232,106 +205,202 @@ typedef struct nodeElt_s {
232
205
  BYTE nbBits;
233
206
  } nodeElt;
234
207
 
208
+ /**
209
+ * HUF_setMaxHeight():
210
+ * Enforces maxNbBits on the Huffman tree described in huffNode.
211
+ *
212
+ * It sets all nodes with nbBits > maxNbBits to be maxNbBits. Then it adjusts
213
+ * the tree to so that it is a valid canonical Huffman tree.
214
+ *
215
+ * @pre The sum of the ranks of each symbol == 2^largestBits,
216
+ * where largestBits == huffNode[lastNonNull].nbBits.
217
+ * @post The sum of the ranks of each symbol == 2^largestBits,
218
+ * where largestBits is the return value <= maxNbBits.
219
+ *
220
+ * @param huffNode The Huffman tree modified in place to enforce maxNbBits.
221
+ * @param lastNonNull The symbol with the lowest count in the Huffman tree.
222
+ * @param maxNbBits The maximum allowed number of bits, which the Huffman tree
223
+ * may not respect. After this function the Huffman tree will
224
+ * respect maxNbBits.
225
+ * @return The maximum number of bits of the Huffman tree after adjustment,
226
+ * necessarily no more than maxNbBits.
227
+ */
235
228
  static U32 HUF_setMaxHeight(nodeElt* huffNode, U32 lastNonNull, U32 maxNbBits)
236
229
  {
237
230
  const U32 largestBits = huffNode[lastNonNull].nbBits;
238
- if (largestBits <= maxNbBits) return largestBits; /* early exit : no elt > maxNbBits */
231
+ /* early exit : no elt > maxNbBits, so the tree is already valid. */
232
+ if (largestBits <= maxNbBits) return largestBits;
239
233
 
240
234
  /* there are several too large elements (at least >= 2) */
241
235
  { int totalCost = 0;
242
236
  const U32 baseCost = 1 << (largestBits - maxNbBits);
243
- U32 n = lastNonNull;
237
+ int n = (int)lastNonNull;
244
238
 
239
+ /* Adjust any ranks > maxNbBits to maxNbBits.
240
+ * Compute totalCost, which is how far the sum of the ranks is
241
+ * we are over 2^largestBits after adjust the offending ranks.
242
+ */
245
243
  while (huffNode[n].nbBits > maxNbBits) {
246
244
  totalCost += baseCost - (1 << (largestBits - huffNode[n].nbBits));
247
245
  huffNode[n].nbBits = (BYTE)maxNbBits;
248
- n --;
249
- } /* n stops at huffNode[n].nbBits <= maxNbBits */
250
- while (huffNode[n].nbBits == maxNbBits) n--; /* n end at index of smallest symbol using < maxNbBits */
246
+ n--;
247
+ }
248
+ /* n stops at huffNode[n].nbBits <= maxNbBits */
249
+ assert(huffNode[n].nbBits <= maxNbBits);
250
+ /* n end at index of smallest symbol using < maxNbBits */
251
+ while (huffNode[n].nbBits == maxNbBits) --n;
251
252
 
252
- /* renorm totalCost */
253
- totalCost >>= (largestBits - maxNbBits); /* note : totalCost is necessarily a multiple of baseCost */
253
+ /* renorm totalCost from 2^largestBits to 2^maxNbBits
254
+ * note : totalCost is necessarily a multiple of baseCost */
255
+ assert((totalCost & (baseCost - 1)) == 0);
256
+ totalCost >>= (largestBits - maxNbBits);
257
+ assert(totalCost > 0);
254
258
 
255
259
  /* repay normalized cost */
256
260
  { U32 const noSymbol = 0xF0F0F0F0;
257
261
  U32 rankLast[HUF_TABLELOG_MAX+2];
258
- int pos;
259
262
 
260
- /* Get pos of last (smallest) symbol per rank */
261
- memset(rankLast, 0xF0, sizeof(rankLast));
263
+ /* Get pos of last (smallest = lowest cum. count) symbol per rank */
264
+ ZSTD_memset(rankLast, 0xF0, sizeof(rankLast));
262
265
  { U32 currentNbBits = maxNbBits;
266
+ int pos;
263
267
  for (pos=n ; pos >= 0; pos--) {
264
268
  if (huffNode[pos].nbBits >= currentNbBits) continue;
265
269
  currentNbBits = huffNode[pos].nbBits; /* < maxNbBits */
266
- rankLast[maxNbBits-currentNbBits] = pos;
270
+ rankLast[maxNbBits-currentNbBits] = (U32)pos;
267
271
  } }
268
272
 
269
273
  while (totalCost > 0) {
270
- U32 nBitsToDecrease = BIT_highbit32(totalCost) + 1;
274
+ /* Try to reduce the next power of 2 above totalCost because we
275
+ * gain back half the rank.
276
+ */
277
+ U32 nBitsToDecrease = BIT_highbit32((U32)totalCost) + 1;
271
278
  for ( ; nBitsToDecrease > 1; nBitsToDecrease--) {
272
- U32 highPos = rankLast[nBitsToDecrease];
273
- U32 lowPos = rankLast[nBitsToDecrease-1];
279
+ U32 const highPos = rankLast[nBitsToDecrease];
280
+ U32 const lowPos = rankLast[nBitsToDecrease-1];
274
281
  if (highPos == noSymbol) continue;
282
+ /* Decrease highPos if no symbols of lowPos or if it is
283
+ * not cheaper to remove 2 lowPos than highPos.
284
+ */
275
285
  if (lowPos == noSymbol) break;
276
286
  { U32 const highTotal = huffNode[highPos].count;
277
287
  U32 const lowTotal = 2 * huffNode[lowPos].count;
278
288
  if (highTotal <= lowTotal) break;
279
289
  } }
280
290
  /* only triggered when no more rank 1 symbol left => find closest one (note : there is necessarily at least one !) */
291
+ assert(rankLast[nBitsToDecrease] != noSymbol || nBitsToDecrease == 1);
281
292
  /* HUF_MAX_TABLELOG test just to please gcc 5+; but it should not be necessary */
282
293
  while ((nBitsToDecrease<=HUF_TABLELOG_MAX) && (rankLast[nBitsToDecrease] == noSymbol))
283
- nBitsToDecrease ++;
294
+ nBitsToDecrease++;
295
+ assert(rankLast[nBitsToDecrease] != noSymbol);
296
+ /* Increase the number of bits to gain back half the rank cost. */
284
297
  totalCost -= 1 << (nBitsToDecrease-1);
298
+ huffNode[rankLast[nBitsToDecrease]].nbBits++;
299
+
300
+ /* Fix up the new rank.
301
+ * If the new rank was empty, this symbol is now its smallest.
302
+ * Otherwise, this symbol will be the largest in the new rank so no adjustment.
303
+ */
285
304
  if (rankLast[nBitsToDecrease-1] == noSymbol)
286
- rankLast[nBitsToDecrease-1] = rankLast[nBitsToDecrease]; /* this rank is no longer empty */
287
- huffNode[rankLast[nBitsToDecrease]].nbBits ++;
305
+ rankLast[nBitsToDecrease-1] = rankLast[nBitsToDecrease];
306
+ /* Fix up the old rank.
307
+ * If the symbol was at position 0, meaning it was the highest weight symbol in the tree,
308
+ * it must be the only symbol in its rank, so the old rank now has no symbols.
309
+ * Otherwise, since the Huffman nodes are sorted by count, the previous position is now
310
+ * the smallest node in the rank. If the previous position belongs to a different rank,
311
+ * then the rank is now empty.
312
+ */
288
313
  if (rankLast[nBitsToDecrease] == 0) /* special case, reached largest symbol */
289
314
  rankLast[nBitsToDecrease] = noSymbol;
290
315
  else {
291
316
  rankLast[nBitsToDecrease]--;
292
317
  if (huffNode[rankLast[nBitsToDecrease]].nbBits != maxNbBits-nBitsToDecrease)
293
318
  rankLast[nBitsToDecrease] = noSymbol; /* this rank is now empty */
294
- } } /* while (totalCost > 0) */
295
-
319
+ }
320
+ } /* while (totalCost > 0) */
321
+
322
+ /* If we've removed too much weight, then we have to add it back.
323
+ * To avoid overshooting again, we only adjust the smallest rank.
324
+ * We take the largest nodes from the lowest rank 0 and move them
325
+ * to rank 1. There's guaranteed to be enough rank 0 symbols because
326
+ * TODO.
327
+ */
296
328
  while (totalCost < 0) { /* Sometimes, cost correction overshoot */
297
- if (rankLast[1] == noSymbol) { /* special case : no rank 1 symbol (using maxNbBits-1); let's create one from largest rank 0 (using maxNbBits) */
329
+ /* special case : no rank 1 symbol (using maxNbBits-1);
330
+ * let's create one from largest rank 0 (using maxNbBits).
331
+ */
332
+ if (rankLast[1] == noSymbol) {
298
333
  while (huffNode[n].nbBits == maxNbBits) n--;
299
334
  huffNode[n+1].nbBits--;
300
- rankLast[1] = n+1;
335
+ assert(n >= 0);
336
+ rankLast[1] = (U32)(n+1);
301
337
  totalCost++;
302
338
  continue;
303
339
  }
304
340
  huffNode[ rankLast[1] + 1 ].nbBits--;
305
341
  rankLast[1]++;
306
342
  totalCost ++;
307
- } } } /* there are several too large elements (at least >= 2) */
343
+ }
344
+ } /* repay normalized cost */
345
+ } /* there are several too large elements (at least >= 2) */
308
346
 
309
347
  return maxNbBits;
310
348
  }
311
349
 
312
-
313
350
  typedef struct {
314
351
  U32 base;
315
- U32 current;
352
+ U32 curr;
316
353
  } rankPos;
317
354
 
318
- static void HUF_sort(nodeElt* huffNode, const unsigned* count, U32 maxSymbolValue)
319
- {
320
- rankPos rank[32];
321
- U32 n;
355
+ typedef nodeElt huffNodeTable[HUF_CTABLE_WORKSPACE_SIZE_U32];
356
+
357
+ #define RANK_POSITION_TABLE_SIZE 32
322
358
 
323
- memset(rank, 0, sizeof(rank));
324
- for (n=0; n<=maxSymbolValue; n++) {
325
- U32 r = BIT_highbit32(count[n] + 1);
326
- rank[r].base ++;
359
+ typedef struct {
360
+ huffNodeTable huffNodeTbl;
361
+ rankPos rankPosition[RANK_POSITION_TABLE_SIZE];
362
+ } HUF_buildCTable_wksp_tables;
363
+
364
+ /**
365
+ * HUF_sort():
366
+ * Sorts the symbols [0, maxSymbolValue] by count[symbol] in decreasing order.
367
+ *
368
+ * @param[out] huffNode Sorted symbols by decreasing count. Only members `.count` and `.byte` are filled.
369
+ * Must have (maxSymbolValue + 1) entries.
370
+ * @param[in] count Histogram of the symbols.
371
+ * @param[in] maxSymbolValue Maximum symbol value.
372
+ * @param rankPosition This is a scratch workspace. Must have RANK_POSITION_TABLE_SIZE entries.
373
+ */
374
+ static void HUF_sort(nodeElt* huffNode, const unsigned* count, U32 maxSymbolValue, rankPos* rankPosition)
375
+ {
376
+ int n;
377
+ int const maxSymbolValue1 = (int)maxSymbolValue + 1;
378
+
379
+ /* Compute base and set curr to base.
380
+ * For symbol s let lowerRank = BIT_highbit32(count[n]+1) and rank = lowerRank + 1.
381
+ * Then 2^lowerRank <= count[n]+1 <= 2^rank.
382
+ * We attribute each symbol to lowerRank's base value, because we want to know where
383
+ * each rank begins in the output, so for rank R we want to count ranks R+1 and above.
384
+ */
385
+ ZSTD_memset(rankPosition, 0, sizeof(*rankPosition) * RANK_POSITION_TABLE_SIZE);
386
+ for (n = 0; n < maxSymbolValue1; ++n) {
387
+ U32 lowerRank = BIT_highbit32(count[n] + 1);
388
+ rankPosition[lowerRank].base++;
327
389
  }
328
- for (n=30; n>0; n--) rank[n-1].base += rank[n].base;
329
- for (n=0; n<32; n++) rank[n].current = rank[n].base;
330
- for (n=0; n<=maxSymbolValue; n++) {
390
+ assert(rankPosition[RANK_POSITION_TABLE_SIZE - 1].base == 0);
391
+ for (n = RANK_POSITION_TABLE_SIZE - 1; n > 0; --n) {
392
+ rankPosition[n-1].base += rankPosition[n].base;
393
+ rankPosition[n-1].curr = rankPosition[n-1].base;
394
+ }
395
+ /* Sort */
396
+ for (n = 0; n < maxSymbolValue1; ++n) {
331
397
  U32 const c = count[n];
332
398
  U32 const r = BIT_highbit32(c+1) + 1;
333
- U32 pos = rank[r].current++;
334
- while ((pos > rank[r].base) && (c > huffNode[pos-1].count)) {
399
+ U32 pos = rankPosition[r].curr++;
400
+ /* Insert into the correct position in the rank.
401
+ * We have at most 256 symbols, so this insertion should be fine.
402
+ */
403
+ while ((pos > rankPosition[r].base) && (c > huffNode[pos-1].count)) {
335
404
  huffNode[pos] = huffNode[pos-1];
336
405
  pos--;
337
406
  }
@@ -343,45 +412,40 @@ static void HUF_sort(nodeElt* huffNode, const unsigned* count, U32 maxSymbolValu
343
412
 
344
413
  /** HUF_buildCTable_wksp() :
345
414
  * Same as HUF_buildCTable(), but using externally allocated scratch buffer.
346
- * `workSpace` must be aligned on 4-bytes boundaries, and be at least as large as a table of HUF_CTABLE_WORKSPACE_SIZE_U32 unsigned.
415
+ * `workSpace` must be aligned on 4-bytes boundaries, and be at least as large as sizeof(HUF_buildCTable_wksp_tables).
347
416
  */
348
417
  #define STARTNODE (HUF_SYMBOLVALUE_MAX+1)
349
- typedef nodeElt huffNodeTable[HUF_CTABLE_WORKSPACE_SIZE_U32];
350
- size_t HUF_buildCTable_wksp (HUF_CElt* tree, const unsigned* count, U32 maxSymbolValue, U32 maxNbBits, void* workSpace, size_t wkspSize)
418
+
419
+ /* HUF_buildTree():
420
+ * Takes the huffNode array sorted by HUF_sort() and builds an unlimited-depth Huffman tree.
421
+ *
422
+ * @param huffNode The array sorted by HUF_sort(). Builds the Huffman tree in this array.
423
+ * @param maxSymbolValue The maximum symbol value.
424
+ * @return The smallest node in the Huffman tree (by count).
425
+ */
426
+ static int HUF_buildTree(nodeElt* huffNode, U32 maxSymbolValue)
351
427
  {
352
- nodeElt* const huffNode0 = (nodeElt*)workSpace;
353
- nodeElt* const huffNode = huffNode0+1;
354
- U32 n, nonNullRank;
428
+ nodeElt* const huffNode0 = huffNode - 1;
429
+ int nonNullRank;
355
430
  int lowS, lowN;
356
- U16 nodeNb = STARTNODE;
357
- U32 nodeRoot;
358
-
359
- /* safety checks */
360
- if (((size_t)workSpace & 3) != 0) return ERROR(GENERIC); /* must be aligned on 4-bytes boundaries */
361
- if (wkspSize < sizeof(huffNodeTable)) return ERROR(workSpace_tooSmall);
362
- if (maxNbBits == 0) maxNbBits = HUF_TABLELOG_DEFAULT;
363
- if (maxSymbolValue > HUF_SYMBOLVALUE_MAX) return ERROR(maxSymbolValue_tooLarge);
364
- memset(huffNode0, 0, sizeof(huffNodeTable));
365
-
366
- /* sort, decreasing order */
367
- HUF_sort(huffNode, count, maxSymbolValue);
368
-
431
+ int nodeNb = STARTNODE;
432
+ int n, nodeRoot;
369
433
  /* init for parents */
370
- nonNullRank = maxSymbolValue;
434
+ nonNullRank = (int)maxSymbolValue;
371
435
  while(huffNode[nonNullRank].count == 0) nonNullRank--;
372
436
  lowS = nonNullRank; nodeRoot = nodeNb + lowS - 1; lowN = nodeNb;
373
437
  huffNode[nodeNb].count = huffNode[lowS].count + huffNode[lowS-1].count;
374
- huffNode[lowS].parent = huffNode[lowS-1].parent = nodeNb;
438
+ huffNode[lowS].parent = huffNode[lowS-1].parent = (U16)nodeNb;
375
439
  nodeNb++; lowS-=2;
376
440
  for (n=nodeNb; n<=nodeRoot; n++) huffNode[n].count = (U32)(1U<<30);
377
441
  huffNode0[0].count = (U32)(1U<<31); /* fake entry, strong barrier */
378
442
 
379
443
  /* create parents */
380
444
  while (nodeNb <= nodeRoot) {
381
- U32 n1 = (huffNode[lowS].count < huffNode[lowN].count) ? lowS-- : lowN++;
382
- U32 n2 = (huffNode[lowS].count < huffNode[lowN].count) ? lowS-- : lowN++;
445
+ int const n1 = (huffNode[lowS].count < huffNode[lowN].count) ? lowS-- : lowN++;
446
+ int const n2 = (huffNode[lowS].count < huffNode[lowN].count) ? lowS-- : lowN++;
383
447
  huffNode[nodeNb].count = huffNode[n1].count + huffNode[n2].count;
384
- huffNode[n1].parent = huffNode[n2].parent = nodeNb;
448
+ huffNode[n1].parent = huffNode[n2].parent = (U16)nodeNb;
385
449
  nodeNb++;
386
450
  }
387
451
 
@@ -392,42 +456,73 @@ size_t HUF_buildCTable_wksp (HUF_CElt* tree, const unsigned* count, U32 maxSymbo
392
456
  for (n=0; n<=nonNullRank; n++)
393
457
  huffNode[n].nbBits = huffNode[ huffNode[n].parent ].nbBits + 1;
394
458
 
395
- /* enforce maxTableLog */
396
- maxNbBits = HUF_setMaxHeight(huffNode, nonNullRank, maxNbBits);
397
-
398
- /* fill result into tree (val, nbBits) */
399
- { U16 nbPerRank[HUF_TABLELOG_MAX+1] = {0};
400
- U16 valPerRank[HUF_TABLELOG_MAX+1] = {0};
401
- if (maxNbBits > HUF_TABLELOG_MAX) return ERROR(GENERIC); /* check fit into table */
402
- for (n=0; n<=nonNullRank; n++)
403
- nbPerRank[huffNode[n].nbBits]++;
404
- /* determine stating value per rank */
405
- { U16 min = 0;
406
- for (n=maxNbBits; n>0; n--) {
407
- valPerRank[n] = min; /* get starting value within each rank */
408
- min += nbPerRank[n];
409
- min >>= 1;
410
- } }
411
- for (n=0; n<=maxSymbolValue; n++)
412
- tree[huffNode[n].byte].nbBits = huffNode[n].nbBits; /* push nbBits per symbol, symbol order */
413
- for (n=0; n<=maxSymbolValue; n++)
414
- tree[n].val = valPerRank[tree[n].nbBits]++; /* assign value within rank, symbol order */
415
- }
416
-
417
- return maxNbBits;
459
+ return nonNullRank;
418
460
  }
419
461
 
420
- /** HUF_buildCTable() :
421
- * @return : maxNbBits
422
- * Note : count is used before tree is written, so they can safely overlap
462
+ /**
463
+ * HUF_buildCTableFromTree():
464
+ * Build the CTable given the Huffman tree in huffNode.
465
+ *
466
+ * @param[out] CTable The output Huffman CTable.
467
+ * @param huffNode The Huffman tree.
468
+ * @param nonNullRank The last and smallest node in the Huffman tree.
469
+ * @param maxSymbolValue The maximum symbol value.
470
+ * @param maxNbBits The exact maximum number of bits used in the Huffman tree.
423
471
  */
424
- size_t HUF_buildCTable (HUF_CElt* tree, const unsigned* count, unsigned maxSymbolValue, unsigned maxNbBits)
472
+ static void HUF_buildCTableFromTree(HUF_CElt* CTable, nodeElt const* huffNode, int nonNullRank, U32 maxSymbolValue, U32 maxNbBits)
425
473
  {
426
- huffNodeTable nodeTable;
427
- return HUF_buildCTable_wksp(tree, count, maxSymbolValue, maxNbBits, nodeTable, sizeof(nodeTable));
474
+ /* fill result into ctable (val, nbBits) */
475
+ int n;
476
+ U16 nbPerRank[HUF_TABLELOG_MAX+1] = {0};
477
+ U16 valPerRank[HUF_TABLELOG_MAX+1] = {0};
478
+ int const alphabetSize = (int)(maxSymbolValue + 1);
479
+ for (n=0; n<=nonNullRank; n++)
480
+ nbPerRank[huffNode[n].nbBits]++;
481
+ /* determine starting value per rank */
482
+ { U16 min = 0;
483
+ for (n=(int)maxNbBits; n>0; n--) {
484
+ valPerRank[n] = min; /* get starting value within each rank */
485
+ min += nbPerRank[n];
486
+ min >>= 1;
487
+ } }
488
+ for (n=0; n<alphabetSize; n++)
489
+ CTable[huffNode[n].byte].nbBits = huffNode[n].nbBits; /* push nbBits per symbol, symbol order */
490
+ for (n=0; n<alphabetSize; n++)
491
+ CTable[n].val = valPerRank[CTable[n].nbBits]++; /* assign value within rank, symbol order */
428
492
  }
429
493
 
430
- static size_t HUF_estimateCompressedSize(HUF_CElt* CTable, const unsigned* count, unsigned maxSymbolValue)
494
+ size_t HUF_buildCTable_wksp (HUF_CElt* tree, const unsigned* count, U32 maxSymbolValue, U32 maxNbBits, void* workSpace, size_t wkspSize)
495
+ {
496
+ HUF_buildCTable_wksp_tables* const wksp_tables = (HUF_buildCTable_wksp_tables*)workSpace;
497
+ nodeElt* const huffNode0 = wksp_tables->huffNodeTbl;
498
+ nodeElt* const huffNode = huffNode0+1;
499
+ int nonNullRank;
500
+
501
+ /* safety checks */
502
+ if (((size_t)workSpace & 3) != 0) return ERROR(GENERIC); /* must be aligned on 4-bytes boundaries */
503
+ if (wkspSize < sizeof(HUF_buildCTable_wksp_tables))
504
+ return ERROR(workSpace_tooSmall);
505
+ if (maxNbBits == 0) maxNbBits = HUF_TABLELOG_DEFAULT;
506
+ if (maxSymbolValue > HUF_SYMBOLVALUE_MAX)
507
+ return ERROR(maxSymbolValue_tooLarge);
508
+ ZSTD_memset(huffNode0, 0, sizeof(huffNodeTable));
509
+
510
+ /* sort, decreasing order */
511
+ HUF_sort(huffNode, count, maxSymbolValue, wksp_tables->rankPosition);
512
+
513
+ /* build tree */
514
+ nonNullRank = HUF_buildTree(huffNode, maxSymbolValue);
515
+
516
+ /* enforce maxTableLog */
517
+ maxNbBits = HUF_setMaxHeight(huffNode, (U32)nonNullRank, maxNbBits);
518
+ if (maxNbBits > HUF_TABLELOG_MAX) return ERROR(GENERIC); /* check fit into table */
519
+
520
+ HUF_buildCTableFromTree(tree, huffNode, nonNullRank, maxSymbolValue, maxNbBits);
521
+
522
+ return maxNbBits;
523
+ }
524
+
525
+ size_t HUF_estimateCompressedSize(const HUF_CElt* CTable, const unsigned* count, unsigned maxSymbolValue)
431
526
  {
432
527
  size_t nbBits = 0;
433
528
  int s;
@@ -437,7 +532,7 @@ static size_t HUF_estimateCompressedSize(HUF_CElt* CTable, const unsigned* count
437
532
  return nbBits >> 3;
438
533
  }
439
534
 
440
- static int HUF_validateCTable(const HUF_CElt* CTable, const unsigned* count, unsigned maxSymbolValue) {
535
+ int HUF_validateCTable(const HUF_CElt* CTable, const unsigned* count, unsigned maxSymbolValue) {
441
536
  int bad = 0;
442
537
  int s;
443
538
  for (s = 0; s <= (int)maxSymbolValue; ++s) {
@@ -476,7 +571,7 @@ HUF_compress1X_usingCTable_internal_body(void* dst, size_t dstSize,
476
571
 
477
572
  /* init */
478
573
  if (dstSize < 8) return 0; /* not enough space to compress */
479
- { size_t const initErr = BIT_initCStream(&bitC, op, oend-op);
574
+ { size_t const initErr = BIT_initCStream(&bitC, op, (size_t)(oend-op));
480
575
  if (HUF_isError(initErr)) return 0; }
481
576
 
482
577
  n = srcSize & ~3; /* join to mod 4 */
@@ -573,7 +668,8 @@ HUF_compress4X_usingCTable_internal(void* dst, size_t dstSize,
573
668
  if (srcSize < 12) return 0; /* no saving possible : too small input */
574
669
  op += 6; /* jumpTable */
575
670
 
576
- { CHECK_V_F(cSize, HUF_compress1X_usingCTable_internal(op, oend-op, ip, segmentSize, CTable, bmi2) );
671
+ assert(op <= oend);
672
+ { CHECK_V_F(cSize, HUF_compress1X_usingCTable_internal(op, (size_t)(oend-op), ip, segmentSize, CTable, bmi2) );
577
673
  if (cSize==0) return 0;
578
674
  assert(cSize <= 65535);
579
675
  MEM_writeLE16(ostart, (U16)cSize);
@@ -581,7 +677,8 @@ HUF_compress4X_usingCTable_internal(void* dst, size_t dstSize,
581
677
  }
582
678
 
583
679
  ip += segmentSize;
584
- { CHECK_V_F(cSize, HUF_compress1X_usingCTable_internal(op, oend-op, ip, segmentSize, CTable, bmi2) );
680
+ assert(op <= oend);
681
+ { CHECK_V_F(cSize, HUF_compress1X_usingCTable_internal(op, (size_t)(oend-op), ip, segmentSize, CTable, bmi2) );
585
682
  if (cSize==0) return 0;
586
683
  assert(cSize <= 65535);
587
684
  MEM_writeLE16(ostart+2, (U16)cSize);
@@ -589,7 +686,8 @@ HUF_compress4X_usingCTable_internal(void* dst, size_t dstSize,
589
686
  }
590
687
 
591
688
  ip += segmentSize;
592
- { CHECK_V_F(cSize, HUF_compress1X_usingCTable_internal(op, oend-op, ip, segmentSize, CTable, bmi2) );
689
+ assert(op <= oend);
690
+ { CHECK_V_F(cSize, HUF_compress1X_usingCTable_internal(op, (size_t)(oend-op), ip, segmentSize, CTable, bmi2) );
593
691
  if (cSize==0) return 0;
594
692
  assert(cSize <= 65535);
595
693
  MEM_writeLE16(ostart+4, (U16)cSize);
@@ -597,12 +695,14 @@ HUF_compress4X_usingCTable_internal(void* dst, size_t dstSize,
597
695
  }
598
696
 
599
697
  ip += segmentSize;
600
- { CHECK_V_F(cSize, HUF_compress1X_usingCTable_internal(op, oend-op, ip, iend-ip, CTable, bmi2) );
698
+ assert(op <= oend);
699
+ assert(ip <= iend);
700
+ { CHECK_V_F(cSize, HUF_compress1X_usingCTable_internal(op, (size_t)(oend-op), ip, (size_t)(iend-ip), CTable, bmi2) );
601
701
  if (cSize==0) return 0;
602
702
  op += cSize;
603
703
  }
604
704
 
605
- return op-ostart;
705
+ return (size_t)(op-ostart);
606
706
  }
607
707
 
608
708
  size_t HUF_compress4X_usingCTable(void* dst, size_t dstSize, const void* src, size_t srcSize, const HUF_CElt* CTable)
@@ -618,40 +718,44 @@ static size_t HUF_compressCTable_internal(
618
718
  HUF_nbStreams_e nbStreams, const HUF_CElt* CTable, const int bmi2)
619
719
  {
620
720
  size_t const cSize = (nbStreams==HUF_singleStream) ?
621
- HUF_compress1X_usingCTable_internal(op, oend - op, src, srcSize, CTable, bmi2) :
622
- HUF_compress4X_usingCTable_internal(op, oend - op, src, srcSize, CTable, bmi2);
721
+ HUF_compress1X_usingCTable_internal(op, (size_t)(oend - op), src, srcSize, CTable, bmi2) :
722
+ HUF_compress4X_usingCTable_internal(op, (size_t)(oend - op), src, srcSize, CTable, bmi2);
623
723
  if (HUF_isError(cSize)) { return cSize; }
624
724
  if (cSize==0) { return 0; } /* uncompressible */
625
725
  op += cSize;
626
726
  /* check compressibility */
727
+ assert(op >= ostart);
627
728
  if ((size_t)(op-ostart) >= srcSize-1) { return 0; }
628
- return op-ostart;
729
+ return (size_t)(op-ostart);
629
730
  }
630
731
 
631
732
  typedef struct {
632
733
  unsigned count[HUF_SYMBOLVALUE_MAX + 1];
633
734
  HUF_CElt CTable[HUF_SYMBOLVALUE_MAX + 1];
634
- huffNodeTable nodeTable;
735
+ HUF_buildCTable_wksp_tables buildCTable_wksp;
635
736
  } HUF_compress_tables_t;
636
737
 
637
738
  /* HUF_compress_internal() :
638
- * `workSpace` must a table of at least HUF_WORKSPACE_SIZE_U32 unsigned */
739
+ * `workSpace_align4` must be aligned on 4-bytes boundaries,
740
+ * and occupies the same space as a table of HUF_WORKSPACE_SIZE_U32 unsigned */
639
741
  static size_t
640
742
  HUF_compress_internal (void* dst, size_t dstSize,
641
743
  const void* src, size_t srcSize,
642
744
  unsigned maxSymbolValue, unsigned huffLog,
643
745
  HUF_nbStreams_e nbStreams,
644
- void* workSpace, size_t wkspSize,
746
+ void* workSpace_align4, size_t wkspSize,
645
747
  HUF_CElt* oldHufTable, HUF_repeat* repeat, int preferRepeat,
646
748
  const int bmi2)
647
749
  {
648
- HUF_compress_tables_t* const table = (HUF_compress_tables_t*)workSpace;
750
+ HUF_compress_tables_t* const table = (HUF_compress_tables_t*)workSpace_align4;
649
751
  BYTE* const ostart = (BYTE*)dst;
650
752
  BYTE* const oend = ostart + dstSize;
651
753
  BYTE* op = ostart;
652
754
 
755
+ HUF_STATIC_ASSERT(sizeof(*table) <= HUF_WORKSPACE_SIZE);
756
+ assert(((size_t)workSpace_align4 & 3) == 0); /* must be aligned on 4-bytes boundaries */
757
+
653
758
  /* checks & inits */
654
- if (((size_t)workSpace & 3) != 0) return ERROR(GENERIC); /* must be aligned on 4-bytes boundaries */
655
759
  if (wkspSize < HUF_WORKSPACE_SIZE) return ERROR(workSpace_tooSmall);
656
760
  if (!srcSize) return 0; /* Uncompressed */
657
761
  if (!dstSize) return 0; /* cannot fit anything within dst budget */
@@ -669,7 +773,7 @@ HUF_compress_internal (void* dst, size_t dstSize,
669
773
  }
670
774
 
671
775
  /* Scan input and build symbol stats */
672
- { CHECK_V_F(largest, HIST_count_wksp (table->count, &maxSymbolValue, (const BYTE*)src, srcSize, workSpace, wkspSize) );
776
+ { CHECK_V_F(largest, HIST_count_wksp (table->count, &maxSymbolValue, (const BYTE*)src, srcSize, workSpace_align4, wkspSize) );
673
777
  if (largest == srcSize) { *ostart = ((const BYTE*)src)[0]; return 1; } /* single symbol, rle */
674
778
  if (largest <= (srcSize >> 7)+4) return 0; /* heuristic : probably not compressible enough */
675
779
  }
@@ -691,11 +795,11 @@ HUF_compress_internal (void* dst, size_t dstSize,
691
795
  huffLog = HUF_optimalTableLog(huffLog, srcSize, maxSymbolValue);
692
796
  { size_t const maxBits = HUF_buildCTable_wksp(table->CTable, table->count,
693
797
  maxSymbolValue, huffLog,
694
- table->nodeTable, sizeof(table->nodeTable));
798
+ &table->buildCTable_wksp, sizeof(table->buildCTable_wksp));
695
799
  CHECK_F(maxBits);
696
800
  huffLog = (U32)maxBits;
697
801
  /* Zero unused symbols in CTable, so we can check it for validity */
698
- memset(table->CTable + (maxSymbolValue + 1), 0,
802
+ ZSTD_memset(table->CTable + (maxSymbolValue + 1), 0,
699
803
  sizeof(table->CTable) - ((maxSymbolValue + 1) * sizeof(HUF_CElt)));
700
804
  }
701
805
 
@@ -716,7 +820,7 @@ HUF_compress_internal (void* dst, size_t dstSize,
716
820
  op += hSize;
717
821
  if (repeat) { *repeat = HUF_repeat_none; }
718
822
  if (oldHufTable)
719
- memcpy(oldHufTable, table->CTable, sizeof(table->CTable)); /* Save new table */
823
+ ZSTD_memcpy(oldHufTable, table->CTable, sizeof(table->CTable)); /* Save new table */
720
824
  }
721
825
  return HUF_compressCTable_internal(ostart, op, oend,
722
826
  src, srcSize,
@@ -747,14 +851,6 @@ size_t HUF_compress1X_repeat (void* dst, size_t dstSize,
747
851
  repeat, preferRepeat, bmi2);
748
852
  }
749
853
 
750
- size_t HUF_compress1X (void* dst, size_t dstSize,
751
- const void* src, size_t srcSize,
752
- unsigned maxSymbolValue, unsigned huffLog)
753
- {
754
- unsigned workSpace[HUF_WORKSPACE_SIZE_U32];
755
- return HUF_compress1X_wksp(dst, dstSize, src, srcSize, maxSymbolValue, huffLog, workSpace, sizeof(workSpace));
756
- }
757
-
758
854
  /* HUF_compress4X_repeat():
759
855
  * compress input using 4 streams.
760
856
  * provide workspace to generate compression tables */
@@ -784,6 +880,25 @@ size_t HUF_compress4X_repeat (void* dst, size_t dstSize,
784
880
  hufTable, repeat, preferRepeat, bmi2);
785
881
  }
786
882
 
883
+ #ifndef ZSTD_NO_UNUSED_FUNCTIONS
884
+ /** HUF_buildCTable() :
885
+ * @return : maxNbBits
886
+ * Note : count is used before tree is written, so they can safely overlap
887
+ */
888
+ size_t HUF_buildCTable (HUF_CElt* tree, const unsigned* count, unsigned maxSymbolValue, unsigned maxNbBits)
889
+ {
890
+ HUF_buildCTable_wksp_tables workspace;
891
+ return HUF_buildCTable_wksp(tree, count, maxSymbolValue, maxNbBits, &workspace, sizeof(workspace));
892
+ }
893
+
894
+ size_t HUF_compress1X (void* dst, size_t dstSize,
895
+ const void* src, size_t srcSize,
896
+ unsigned maxSymbolValue, unsigned huffLog)
897
+ {
898
+ unsigned workSpace[HUF_WORKSPACE_SIZE_U32];
899
+ return HUF_compress1X_wksp(dst, dstSize, src, srcSize, maxSymbolValue, huffLog, workSpace, sizeof(workSpace));
900
+ }
901
+
787
902
  size_t HUF_compress2 (void* dst, size_t dstSize,
788
903
  const void* src, size_t srcSize,
789
904
  unsigned maxSymbolValue, unsigned huffLog)
@@ -796,3 +911,4 @@ size_t HUF_compress (void* dst, size_t maxDstSize, const void* src, size_t srcSi
796
911
  {
797
912
  return HUF_compress2(dst, maxDstSize, src, srcSize, 255, HUF_TABLELOG_DEFAULT);
798
913
  }
914
+ #endif