zstd-ruby 1.5.2.2 → 1.5.4.0
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/README.md +15 -3
- data/ext/zstdruby/common.h +7 -0
- data/ext/zstdruby/libzstd/common/bits.h +175 -0
- data/ext/zstdruby/libzstd/common/bitstream.h +18 -59
- data/ext/zstdruby/libzstd/common/compiler.h +22 -3
- data/ext/zstdruby/libzstd/common/cpu.h +1 -1
- data/ext/zstdruby/libzstd/common/debug.c +1 -1
- data/ext/zstdruby/libzstd/common/debug.h +1 -1
- data/ext/zstdruby/libzstd/common/entropy_common.c +12 -40
- data/ext/zstdruby/libzstd/common/error_private.c +9 -2
- data/ext/zstdruby/libzstd/common/error_private.h +1 -1
- data/ext/zstdruby/libzstd/common/fse.h +5 -83
- data/ext/zstdruby/libzstd/common/fse_decompress.c +7 -99
- data/ext/zstdruby/libzstd/common/huf.h +65 -156
- data/ext/zstdruby/libzstd/common/mem.h +39 -46
- data/ext/zstdruby/libzstd/common/pool.c +26 -10
- data/ext/zstdruby/libzstd/common/pool.h +7 -1
- data/ext/zstdruby/libzstd/common/portability_macros.h +22 -3
- data/ext/zstdruby/libzstd/common/threading.c +68 -14
- data/ext/zstdruby/libzstd/common/threading.h +5 -10
- data/ext/zstdruby/libzstd/common/xxhash.c +2 -2
- data/ext/zstdruby/libzstd/common/xxhash.h +8 -8
- data/ext/zstdruby/libzstd/common/zstd_common.c +1 -1
- data/ext/zstdruby/libzstd/common/zstd_deps.h +1 -1
- data/ext/zstdruby/libzstd/common/zstd_internal.h +17 -113
- data/ext/zstdruby/libzstd/common/zstd_trace.h +3 -3
- data/ext/zstdruby/libzstd/compress/clevels.h +1 -1
- data/ext/zstdruby/libzstd/compress/fse_compress.c +7 -124
- data/ext/zstdruby/libzstd/compress/hist.c +1 -1
- data/ext/zstdruby/libzstd/compress/hist.h +1 -1
- data/ext/zstdruby/libzstd/compress/huf_compress.c +234 -169
- data/ext/zstdruby/libzstd/compress/zstd_compress.c +1055 -455
- data/ext/zstdruby/libzstd/compress/zstd_compress_internal.h +165 -145
- data/ext/zstdruby/libzstd/compress/zstd_compress_literals.c +115 -39
- data/ext/zstdruby/libzstd/compress/zstd_compress_literals.h +16 -8
- data/ext/zstdruby/libzstd/compress/zstd_compress_sequences.c +3 -3
- data/ext/zstdruby/libzstd/compress/zstd_compress_sequences.h +1 -1
- data/ext/zstdruby/libzstd/compress/zstd_compress_superblock.c +25 -21
- data/ext/zstdruby/libzstd/compress/zstd_compress_superblock.h +1 -1
- data/ext/zstdruby/libzstd/compress/zstd_cwksp.h +5 -3
- data/ext/zstdruby/libzstd/compress/zstd_double_fast.c +95 -33
- data/ext/zstdruby/libzstd/compress/zstd_double_fast.h +3 -2
- data/ext/zstdruby/libzstd/compress/zstd_fast.c +433 -148
- data/ext/zstdruby/libzstd/compress/zstd_fast.h +3 -2
- data/ext/zstdruby/libzstd/compress/zstd_lazy.c +306 -283
- data/ext/zstdruby/libzstd/compress/zstd_lazy.h +4 -2
- data/ext/zstdruby/libzstd/compress/zstd_ldm.c +5 -5
- data/ext/zstdruby/libzstd/compress/zstd_ldm.h +1 -1
- data/ext/zstdruby/libzstd/compress/zstd_ldm_geartab.h +1 -1
- data/ext/zstdruby/libzstd/compress/zstd_opt.c +104 -80
- data/ext/zstdruby/libzstd/compress/zstd_opt.h +1 -1
- data/ext/zstdruby/libzstd/compress/zstdmt_compress.c +12 -5
- data/ext/zstdruby/libzstd/compress/zstdmt_compress.h +1 -1
- data/ext/zstdruby/libzstd/decompress/huf_decompress.c +434 -441
- data/ext/zstdruby/libzstd/decompress/huf_decompress_amd64.S +30 -39
- data/ext/zstdruby/libzstd/decompress/zstd_ddict.c +3 -4
- data/ext/zstdruby/libzstd/decompress/zstd_ddict.h +1 -1
- data/ext/zstdruby/libzstd/decompress/zstd_decompress.c +164 -42
- data/ext/zstdruby/libzstd/decompress/zstd_decompress_block.c +186 -65
- data/ext/zstdruby/libzstd/decompress/zstd_decompress_block.h +1 -1
- data/ext/zstdruby/libzstd/decompress/zstd_decompress_internal.h +4 -2
- data/ext/zstdruby/libzstd/dictBuilder/cover.c +19 -15
- data/ext/zstdruby/libzstd/dictBuilder/cover.h +1 -1
- data/ext/zstdruby/libzstd/dictBuilder/fastcover.c +2 -2
- data/ext/zstdruby/libzstd/dictBuilder/zdict.c +9 -87
- data/ext/zstdruby/libzstd/zdict.h +53 -31
- data/ext/zstdruby/libzstd/zstd.h +489 -90
- data/ext/zstdruby/libzstd/zstd_errors.h +27 -8
- data/ext/zstdruby/main.c +4 -0
- data/ext/zstdruby/streaming_compress.c +1 -7
- data/ext/zstdruby/zstdruby.c +110 -26
- data/lib/zstd-ruby/version.rb +1 -1
- data/lib/zstd-ruby.rb +0 -1
- metadata +7 -6
@@ -1,5 +1,5 @@
|
|
1
1
|
/*
|
2
|
-
* Copyright (c)
|
2
|
+
* Copyright (c) Meta Platforms, Inc. and affiliates.
|
3
3
|
* All rights reserved.
|
4
4
|
*
|
5
5
|
* This source code is licensed under both the BSD-style license (found in the
|
@@ -13,11 +13,36 @@
|
|
13
13
|
***************************************/
|
14
14
|
#include "zstd_compress_literals.h"
|
15
15
|
|
16
|
+
|
17
|
+
/* **************************************************************
|
18
|
+
* Debug Traces
|
19
|
+
****************************************************************/
|
20
|
+
#if DEBUGLEVEL >= 2
|
21
|
+
|
22
|
+
static size_t showHexa(const void* src, size_t srcSize)
|
23
|
+
{
|
24
|
+
const BYTE* const ip = (const BYTE*)src;
|
25
|
+
size_t u;
|
26
|
+
for (u=0; u<srcSize; u++) {
|
27
|
+
RAWLOG(5, " %02X", ip[u]); (void)ip;
|
28
|
+
}
|
29
|
+
RAWLOG(5, " \n");
|
30
|
+
return srcSize;
|
31
|
+
}
|
32
|
+
|
33
|
+
#endif
|
34
|
+
|
35
|
+
|
36
|
+
/* **************************************************************
|
37
|
+
* Literals compression - special cases
|
38
|
+
****************************************************************/
|
16
39
|
size_t ZSTD_noCompressLiterals (void* dst, size_t dstCapacity, const void* src, size_t srcSize)
|
17
40
|
{
|
18
41
|
BYTE* const ostart = (BYTE*)dst;
|
19
42
|
U32 const flSize = 1 + (srcSize>31) + (srcSize>4095);
|
20
43
|
|
44
|
+
DEBUGLOG(5, "ZSTD_noCompressLiterals: srcSize=%zu, dstCapacity=%zu", srcSize, dstCapacity);
|
45
|
+
|
21
46
|
RETURN_ERROR_IF(srcSize + flSize > dstCapacity, dstSize_tooSmall, "");
|
22
47
|
|
23
48
|
switch(flSize)
|
@@ -36,16 +61,30 @@ size_t ZSTD_noCompressLiterals (void* dst, size_t dstCapacity, const void* src,
|
|
36
61
|
}
|
37
62
|
|
38
63
|
ZSTD_memcpy(ostart + flSize, src, srcSize);
|
39
|
-
DEBUGLOG(5, "Raw literals: %u -> %u", (U32)srcSize, (U32)(srcSize + flSize));
|
64
|
+
DEBUGLOG(5, "Raw (uncompressed) literals: %u -> %u", (U32)srcSize, (U32)(srcSize + flSize));
|
40
65
|
return srcSize + flSize;
|
41
66
|
}
|
42
67
|
|
68
|
+
static int allBytesIdentical(const void* src, size_t srcSize)
|
69
|
+
{
|
70
|
+
assert(srcSize >= 1);
|
71
|
+
assert(src != NULL);
|
72
|
+
{ const BYTE b = ((const BYTE*)src)[0];
|
73
|
+
size_t p;
|
74
|
+
for (p=1; p<srcSize; p++) {
|
75
|
+
if (((const BYTE*)src)[p] != b) return 0;
|
76
|
+
}
|
77
|
+
return 1;
|
78
|
+
}
|
79
|
+
}
|
80
|
+
|
43
81
|
size_t ZSTD_compressRleLiteralsBlock (void* dst, size_t dstCapacity, const void* src, size_t srcSize)
|
44
82
|
{
|
45
83
|
BYTE* const ostart = (BYTE*)dst;
|
46
84
|
U32 const flSize = 1 + (srcSize>31) + (srcSize>4095);
|
47
85
|
|
48
|
-
(
|
86
|
+
assert(dstCapacity >= 4); (void)dstCapacity;
|
87
|
+
assert(allBytesIdentical(src, srcSize));
|
49
88
|
|
50
89
|
switch(flSize)
|
51
90
|
{
|
@@ -63,28 +102,51 @@ size_t ZSTD_compressRleLiteralsBlock (void* dst, size_t dstCapacity, const void*
|
|
63
102
|
}
|
64
103
|
|
65
104
|
ostart[flSize] = *(const BYTE*)src;
|
66
|
-
DEBUGLOG(5, "RLE
|
105
|
+
DEBUGLOG(5, "RLE : Repeated Literal (%02X: %u times) -> %u bytes encoded", ((const BYTE*)src)[0], (U32)srcSize, (U32)flSize + 1);
|
67
106
|
return flSize+1;
|
68
107
|
}
|
69
108
|
|
70
|
-
|
71
|
-
|
72
|
-
|
73
|
-
|
74
|
-
|
75
|
-
|
76
|
-
|
77
|
-
|
109
|
+
/* ZSTD_minLiteralsToCompress() :
|
110
|
+
* returns minimal amount of literals
|
111
|
+
* for literal compression to even be attempted.
|
112
|
+
* Minimum is made tighter as compression strategy increases.
|
113
|
+
*/
|
114
|
+
static size_t
|
115
|
+
ZSTD_minLiteralsToCompress(ZSTD_strategy strategy, HUF_repeat huf_repeat)
|
116
|
+
{
|
117
|
+
assert((int)strategy >= 0);
|
118
|
+
assert((int)strategy <= 9);
|
119
|
+
/* btultra2 : min 8 bytes;
|
120
|
+
* then 2x larger for each successive compression strategy
|
121
|
+
* max threshold 64 bytes */
|
122
|
+
{ int const shift = MIN(9-(int)strategy, 3);
|
123
|
+
size_t const mintc = (huf_repeat == HUF_repeat_valid) ? 6 : (size_t)8 << shift;
|
124
|
+
DEBUGLOG(7, "minLiteralsToCompress = %zu", mintc);
|
125
|
+
return mintc;
|
126
|
+
}
|
127
|
+
}
|
128
|
+
|
129
|
+
size_t ZSTD_compressLiterals (
|
130
|
+
void* dst, size_t dstCapacity,
|
131
|
+
const void* src, size_t srcSize,
|
132
|
+
void* entropyWorkspace, size_t entropyWorkspaceSize,
|
133
|
+
const ZSTD_hufCTables_t* prevHuf,
|
134
|
+
ZSTD_hufCTables_t* nextHuf,
|
135
|
+
ZSTD_strategy strategy,
|
136
|
+
int disableLiteralCompression,
|
137
|
+
int suspectUncompressible,
|
138
|
+
int bmi2)
|
78
139
|
{
|
79
|
-
size_t const minGain = ZSTD_minGain(srcSize, strategy);
|
80
140
|
size_t const lhSize = 3 + (srcSize >= 1 KB) + (srcSize >= 16 KB);
|
81
141
|
BYTE* const ostart = (BYTE*)dst;
|
82
142
|
U32 singleStream = srcSize < 256;
|
83
143
|
symbolEncodingType_e hType = set_compressed;
|
84
144
|
size_t cLitSize;
|
85
145
|
|
86
|
-
DEBUGLOG(5,"ZSTD_compressLiterals (disableLiteralCompression=%i srcSize=%u)",
|
87
|
-
disableLiteralCompression, (U32)srcSize);
|
146
|
+
DEBUGLOG(5,"ZSTD_compressLiterals (disableLiteralCompression=%i, srcSize=%u, dstCapacity=%zu)",
|
147
|
+
disableLiteralCompression, (U32)srcSize, dstCapacity);
|
148
|
+
|
149
|
+
DEBUGLOG(6, "Completed literals listing (%zu bytes)", showHexa(src, srcSize));
|
88
150
|
|
89
151
|
/* Prepare nextEntropy assuming reusing the existing table */
|
90
152
|
ZSTD_memcpy(nextHuf, prevHuf, sizeof(*prevHuf));
|
@@ -92,40 +154,51 @@ size_t ZSTD_compressLiterals (ZSTD_hufCTables_t const* prevHuf,
|
|
92
154
|
if (disableLiteralCompression)
|
93
155
|
return ZSTD_noCompressLiterals(dst, dstCapacity, src, srcSize);
|
94
156
|
|
95
|
-
/* small
|
96
|
-
|
97
|
-
|
98
|
-
if (srcSize <= minLitSize) return ZSTD_noCompressLiterals(dst, dstCapacity, src, srcSize);
|
99
|
-
}
|
157
|
+
/* if too small, don't even attempt compression (speed opt) */
|
158
|
+
if (srcSize < ZSTD_minLiteralsToCompress(strategy, prevHuf->repeatMode))
|
159
|
+
return ZSTD_noCompressLiterals(dst, dstCapacity, src, srcSize);
|
100
160
|
|
101
161
|
RETURN_ERROR_IF(dstCapacity < lhSize+1, dstSize_tooSmall, "not enough space for compression");
|
102
162
|
{ HUF_repeat repeat = prevHuf->repeatMode;
|
103
|
-
int const
|
163
|
+
int const flags = 0
|
164
|
+
| (bmi2 ? HUF_flags_bmi2 : 0)
|
165
|
+
| (strategy < ZSTD_lazy && srcSize <= 1024 ? HUF_flags_preferRepeat : 0)
|
166
|
+
| (strategy >= HUF_OPTIMAL_DEPTH_THRESHOLD ? HUF_flags_optimalDepth : 0)
|
167
|
+
| (suspectUncompressible ? HUF_flags_suspectUncompressible : 0);
|
168
|
+
|
169
|
+
typedef size_t (*huf_compress_f)(void*, size_t, const void*, size_t, unsigned, unsigned, void*, size_t, HUF_CElt*, HUF_repeat*, int);
|
170
|
+
huf_compress_f huf_compress;
|
104
171
|
if (repeat == HUF_repeat_valid && lhSize == 3) singleStream = 1;
|
105
|
-
|
106
|
-
|
107
|
-
|
108
|
-
|
109
|
-
|
110
|
-
|
111
|
-
|
112
|
-
|
113
|
-
(HUF_CElt*)nextHuf->CTable, &repeat, preferRepeat, bmi2, suspectUncompressible);
|
172
|
+
huf_compress = singleStream ? HUF_compress1X_repeat : HUF_compress4X_repeat;
|
173
|
+
cLitSize = huf_compress(ostart+lhSize, dstCapacity-lhSize,
|
174
|
+
src, srcSize,
|
175
|
+
HUF_SYMBOLVALUE_MAX, LitHufLog,
|
176
|
+
entropyWorkspace, entropyWorkspaceSize,
|
177
|
+
(HUF_CElt*)nextHuf->CTable,
|
178
|
+
&repeat, flags);
|
179
|
+
DEBUGLOG(5, "%zu literals compressed into %zu bytes (before header)", srcSize, cLitSize);
|
114
180
|
if (repeat != HUF_repeat_none) {
|
115
181
|
/* reused the existing table */
|
116
|
-
DEBUGLOG(5, "
|
182
|
+
DEBUGLOG(5, "reusing statistics from previous huffman block");
|
117
183
|
hType = set_repeat;
|
118
184
|
}
|
119
185
|
}
|
120
186
|
|
121
|
-
|
122
|
-
|
123
|
-
|
124
|
-
|
187
|
+
{ size_t const minGain = ZSTD_minGain(srcSize, strategy);
|
188
|
+
if ((cLitSize==0) || (cLitSize >= srcSize - minGain) || ERR_isError(cLitSize)) {
|
189
|
+
ZSTD_memcpy(nextHuf, prevHuf, sizeof(*prevHuf));
|
190
|
+
return ZSTD_noCompressLiterals(dst, dstCapacity, src, srcSize);
|
191
|
+
} }
|
125
192
|
if (cLitSize==1) {
|
126
|
-
|
127
|
-
|
128
|
-
|
193
|
+
/* A return value of 1 signals that the alphabet consists of a single symbol.
|
194
|
+
* However, in some rare circumstances, it could be the compressed size (a single byte).
|
195
|
+
* For that outcome to have a chance to happen, it's necessary that `srcSize < 8`.
|
196
|
+
* (it's also necessary to not generate statistics).
|
197
|
+
* Therefore, in such a case, actively check that all bytes are identical. */
|
198
|
+
if ((srcSize >= 8) || allBytesIdentical(src, srcSize)) {
|
199
|
+
ZSTD_memcpy(nextHuf, prevHuf, sizeof(*prevHuf));
|
200
|
+
return ZSTD_compressRleLiteralsBlock(dst, dstCapacity, src, srcSize);
|
201
|
+
} }
|
129
202
|
|
130
203
|
if (hType == set_compressed) {
|
131
204
|
/* using a newly constructed table */
|
@@ -136,16 +209,19 @@ size_t ZSTD_compressLiterals (ZSTD_hufCTables_t const* prevHuf,
|
|
136
209
|
switch(lhSize)
|
137
210
|
{
|
138
211
|
case 3: /* 2 - 2 - 10 - 10 */
|
139
|
-
|
212
|
+
if (!singleStream) assert(srcSize >= MIN_LITERALS_FOR_4_STREAMS);
|
213
|
+
{ U32 const lhc = hType + ((U32)(!singleStream) << 2) + ((U32)srcSize<<4) + ((U32)cLitSize<<14);
|
140
214
|
MEM_writeLE24(ostart, lhc);
|
141
215
|
break;
|
142
216
|
}
|
143
217
|
case 4: /* 2 - 2 - 14 - 14 */
|
218
|
+
assert(srcSize >= MIN_LITERALS_FOR_4_STREAMS);
|
144
219
|
{ U32 const lhc = hType + (2 << 2) + ((U32)srcSize<<4) + ((U32)cLitSize<<18);
|
145
220
|
MEM_writeLE32(ostart, lhc);
|
146
221
|
break;
|
147
222
|
}
|
148
223
|
case 5: /* 2 - 2 - 18 - 18 */
|
224
|
+
assert(srcSize >= MIN_LITERALS_FOR_4_STREAMS);
|
149
225
|
{ U32 const lhc = hType + (3 << 2) + ((U32)srcSize<<4) + ((U32)cLitSize<<22);
|
150
226
|
MEM_writeLE32(ostart, lhc);
|
151
227
|
ostart[4] = (BYTE)(cLitSize >> 10);
|
@@ -1,5 +1,5 @@
|
|
1
1
|
/*
|
2
|
-
* Copyright (c)
|
2
|
+
* Copyright (c) Meta Platforms, Inc. and affiliates.
|
3
3
|
* All rights reserved.
|
4
4
|
*
|
5
5
|
* This source code is licensed under both the BSD-style license (found in the
|
@@ -16,16 +16,24 @@
|
|
16
16
|
|
17
17
|
size_t ZSTD_noCompressLiterals (void* dst, size_t dstCapacity, const void* src, size_t srcSize);
|
18
18
|
|
19
|
+
/* ZSTD_compressRleLiteralsBlock() :
|
20
|
+
* Conditions :
|
21
|
+
* - All bytes in @src are identical
|
22
|
+
* - dstCapacity >= 4 */
|
19
23
|
size_t ZSTD_compressRleLiteralsBlock (void* dst, size_t dstCapacity, const void* src, size_t srcSize);
|
20
24
|
|
21
|
-
/*
|
22
|
-
|
23
|
-
|
24
|
-
|
25
|
-
|
25
|
+
/* ZSTD_compressLiterals():
|
26
|
+
* @entropyWorkspace: must be aligned on 4-bytes boundaries
|
27
|
+
* @entropyWorkspaceSize : must be >= HUF_WORKSPACE_SIZE
|
28
|
+
* @suspectUncompressible: sampling checks, to potentially skip huffman coding
|
29
|
+
*/
|
30
|
+
size_t ZSTD_compressLiterals (void* dst, size_t dstCapacity,
|
26
31
|
const void* src, size_t srcSize,
|
27
32
|
void* entropyWorkspace, size_t entropyWorkspaceSize,
|
28
|
-
const
|
29
|
-
|
33
|
+
const ZSTD_hufCTables_t* prevHuf,
|
34
|
+
ZSTD_hufCTables_t* nextHuf,
|
35
|
+
ZSTD_strategy strategy, int disableLiteralCompression,
|
36
|
+
int suspectUncompressible,
|
37
|
+
int bmi2);
|
30
38
|
|
31
39
|
#endif /* ZSTD_COMPRESS_LITERALS_H */
|
@@ -1,5 +1,5 @@
|
|
1
1
|
/*
|
2
|
-
* Copyright (c)
|
2
|
+
* Copyright (c) Meta Platforms, Inc. and affiliates.
|
3
3
|
* All rights reserved.
|
4
4
|
*
|
5
5
|
* This source code is licensed under both the BSD-style license (found in the
|
@@ -58,7 +58,7 @@ static unsigned ZSTD_useLowProbCount(size_t const nbSeq)
|
|
58
58
|
{
|
59
59
|
/* Heuristic: This should cover most blocks <= 16K and
|
60
60
|
* start to fade out after 16K to about 32K depending on
|
61
|
-
*
|
61
|
+
* compressibility.
|
62
62
|
*/
|
63
63
|
return nbSeq >= 2048;
|
64
64
|
}
|
@@ -166,7 +166,7 @@ ZSTD_selectEncodingType(
|
|
166
166
|
if (mostFrequent == nbSeq) {
|
167
167
|
*repeatMode = FSE_repeat_none;
|
168
168
|
if (isDefaultAllowed && nbSeq <= 2) {
|
169
|
-
/* Prefer set_basic over set_rle when there are 2 or
|
169
|
+
/* Prefer set_basic over set_rle when there are 2 or fewer symbols,
|
170
170
|
* since RLE uses 1 byte, but set_basic uses 5-6 bits per symbol.
|
171
171
|
* If basic encoding isn't possible, always choose RLE.
|
172
172
|
*/
|
@@ -1,5 +1,5 @@
|
|
1
1
|
/*
|
2
|
-
* Copyright (c)
|
2
|
+
* Copyright (c) Meta Platforms, Inc. and affiliates.
|
3
3
|
* All rights reserved.
|
4
4
|
*
|
5
5
|
* This source code is licensed under both the BSD-style license (found in the
|
@@ -36,13 +36,14 @@
|
|
36
36
|
* If it is set_compressed, first sub-block's literals section will be Treeless_Literals_Block
|
37
37
|
* and the following sub-blocks' literals sections will be Treeless_Literals_Block.
|
38
38
|
* @return : compressed size of literals section of a sub-block
|
39
|
-
* Or 0 if
|
39
|
+
* Or 0 if unable to compress.
|
40
40
|
* Or error code */
|
41
|
-
static size_t
|
42
|
-
|
43
|
-
|
44
|
-
|
45
|
-
|
41
|
+
static size_t
|
42
|
+
ZSTD_compressSubBlock_literal(const HUF_CElt* hufTable,
|
43
|
+
const ZSTD_hufCTablesMetadata_t* hufMetadata,
|
44
|
+
const BYTE* literals, size_t litSize,
|
45
|
+
void* dst, size_t dstSize,
|
46
|
+
const int bmi2, int writeEntropy, int* entropyWritten)
|
46
47
|
{
|
47
48
|
size_t const header = writeEntropy ? 200 : 0;
|
48
49
|
size_t const lhSize = 3 + (litSize >= (1 KB - header)) + (litSize >= (16 KB - header));
|
@@ -53,8 +54,6 @@ static size_t ZSTD_compressSubBlock_literal(const HUF_CElt* hufTable,
|
|
53
54
|
symbolEncodingType_e hType = writeEntropy ? hufMetadata->hType : set_repeat;
|
54
55
|
size_t cLitSize = 0;
|
55
56
|
|
56
|
-
(void)bmi2; /* TODO bmi2... */
|
57
|
-
|
58
57
|
DEBUGLOG(5, "ZSTD_compressSubBlock_literal (litSize=%zu, lhSize=%zu, writeEntropy=%d)", litSize, lhSize, writeEntropy);
|
59
58
|
|
60
59
|
*entropyWritten = 0;
|
@@ -76,9 +75,9 @@ static size_t ZSTD_compressSubBlock_literal(const HUF_CElt* hufTable,
|
|
76
75
|
DEBUGLOG(5, "ZSTD_compressSubBlock_literal (hSize=%zu)", hufMetadata->hufDesSize);
|
77
76
|
}
|
78
77
|
|
79
|
-
|
80
|
-
|
81
|
-
: HUF_compress4X_usingCTable(op, oend-op, literals, litSize, hufTable);
|
78
|
+
{ int const flags = bmi2 ? HUF_flags_bmi2 : 0;
|
79
|
+
const size_t cSize = singleStream ? HUF_compress1X_usingCTable(op, oend-op, literals, litSize, hufTable, flags)
|
80
|
+
: HUF_compress4X_usingCTable(op, oend-op, literals, litSize, hufTable, flags);
|
82
81
|
op += cSize;
|
83
82
|
cLitSize += cSize;
|
84
83
|
if (cSize == 0 || ERR_isError(cSize)) {
|
@@ -126,7 +125,11 @@ static size_t ZSTD_compressSubBlock_literal(const HUF_CElt* hufTable,
|
|
126
125
|
return op-ostart;
|
127
126
|
}
|
128
127
|
|
129
|
-
static size_t
|
128
|
+
static size_t
|
129
|
+
ZSTD_seqDecompressedSize(seqStore_t const* seqStore,
|
130
|
+
const seqDef* sequences, size_t nbSeq,
|
131
|
+
size_t litSize, int lastSequence)
|
132
|
+
{
|
130
133
|
const seqDef* const sstart = sequences;
|
131
134
|
const seqDef* const send = sequences + nbSeq;
|
132
135
|
const seqDef* sp = sstart;
|
@@ -156,13 +159,14 @@ static size_t ZSTD_seqDecompressedSize(seqStore_t const* seqStore, const seqDef*
|
|
156
159
|
* @return : compressed size of sequences section of a sub-block
|
157
160
|
* Or 0 if it is unable to compress
|
158
161
|
* Or error code. */
|
159
|
-
static size_t
|
160
|
-
|
161
|
-
|
162
|
-
|
163
|
-
|
164
|
-
|
165
|
-
|
162
|
+
static size_t
|
163
|
+
ZSTD_compressSubBlock_sequences(const ZSTD_fseCTables_t* fseTables,
|
164
|
+
const ZSTD_fseCTablesMetadata_t* fseMetadata,
|
165
|
+
const seqDef* sequences, size_t nbSeq,
|
166
|
+
const BYTE* llCode, const BYTE* mlCode, const BYTE* ofCode,
|
167
|
+
const ZSTD_CCtx_params* cctxParams,
|
168
|
+
void* dst, size_t dstCapacity,
|
169
|
+
const int bmi2, int writeEntropy, int* entropyWritten)
|
166
170
|
{
|
167
171
|
const int longOffsets = cctxParams->cParams.windowLog > STREAM_ACCUMULATOR_MIN;
|
168
172
|
BYTE* const ostart = (BYTE*)dst;
|
@@ -539,7 +543,7 @@ static size_t ZSTD_compressSubBlock_multi(const seqStore_t* seqStorePtr,
|
|
539
543
|
repcodes_t rep;
|
540
544
|
ZSTD_memcpy(&rep, prevCBlock->rep, sizeof(rep));
|
541
545
|
for (seq = sstart; seq < sp; ++seq) {
|
542
|
-
ZSTD_updateRep(rep.rep, seq->offBase
|
546
|
+
ZSTD_updateRep(rep.rep, seq->offBase, ZSTD_getSequenceLength(seqStorePtr, seq).litLength == 0);
|
543
547
|
}
|
544
548
|
ZSTD_memcpy(nextCBlock->rep, &rep, sizeof(rep));
|
545
549
|
}
|
@@ -1,5 +1,5 @@
|
|
1
1
|
/*
|
2
|
-
* Copyright (c)
|
2
|
+
* Copyright (c) Meta Platforms, Inc. and affiliates.
|
3
3
|
* All rights reserved.
|
4
4
|
*
|
5
5
|
* This source code is licensed under both the BSD-style license (found in the
|
@@ -348,7 +348,9 @@ ZSTD_cwksp_reserve_internal(ZSTD_cwksp* ws, size_t bytes, ZSTD_cwksp_alloc_phase
|
|
348
348
|
if (alloc) {
|
349
349
|
alloc = (BYTE *)alloc + ZSTD_CWKSP_ASAN_REDZONE_SIZE;
|
350
350
|
if (ws->isStatic == ZSTD_cwksp_dynamic_alloc) {
|
351
|
-
|
351
|
+
/* We need to keep the redzone poisoned while unpoisoning the bytes that
|
352
|
+
* are actually allocated. */
|
353
|
+
__asan_unpoison_memory_region(alloc, bytes - 2 * ZSTD_CWKSP_ASAN_REDZONE_SIZE);
|
352
354
|
}
|
353
355
|
}
|
354
356
|
#endif
|
@@ -499,7 +501,7 @@ MEM_STATIC void ZSTD_cwksp_clean_tables(ZSTD_cwksp* ws) {
|
|
499
501
|
assert(ws->tableValidEnd >= ws->objectEnd);
|
500
502
|
assert(ws->tableValidEnd <= ws->allocStart);
|
501
503
|
if (ws->tableValidEnd < ws->tableEnd) {
|
502
|
-
ZSTD_memset(ws->tableValidEnd, 0, (BYTE*)ws->tableEnd - (BYTE*)ws->tableValidEnd);
|
504
|
+
ZSTD_memset(ws->tableValidEnd, 0, (size_t)((BYTE*)ws->tableEnd - (BYTE*)ws->tableValidEnd));
|
503
505
|
}
|
504
506
|
ZSTD_cwksp_mark_tables_clean(ws);
|
505
507
|
}
|