zstdlib 0.4.0-x64-mingw32 → 0.5.0-x64-mingw32
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/CHANGES.md +6 -1
- data/README.md +1 -1
- data/ext/zstdlib/extconf.rb +2 -2
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/bitstream.h +3 -2
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/compiler.h +14 -2
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/cpu.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/debug.c +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/debug.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/entropy_common.c +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/error_private.c +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/error_private.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/fse.h +1 -1
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/fse_decompress.c +2 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/huf.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/mem.h +73 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/pool.c +7 -3
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/pool.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/threading.c +46 -1
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/threading.h +32 -1
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/xxhash.c +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/xxhash.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/zstd_common.c +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/zstd_errors.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/common/zstd_internal.h +32 -55
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/fse_compress.c +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/hist.c +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/hist.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/huf_compress.c +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_compress.c +633 -436
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_compress_internal.h +54 -12
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_compress_literals.c +10 -5
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_compress_literals.h +1 -1
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_compress_sequences.c +3 -3
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_compress_sequences.h +1 -1
- data/ext/zstdlib/zstd-1.4.4/lib/compress/zstd_cwksp.h +535 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_double_fast.c +9 -9
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_double_fast.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_fast.c +30 -39
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_fast.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_lazy.c +5 -5
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_lazy.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_ldm.c +4 -4
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_ldm.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_opt.c +1 -1
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstd_opt.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstdmt_compress.c +32 -26
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/compress/zstdmt_compress.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/decompress/huf_decompress.c +2 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/decompress/zstd_ddict.c +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/decompress/zstd_ddict.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/decompress/zstd_decompress.c +14 -16
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/decompress/zstd_decompress_block.c +144 -146
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/decompress/zstd_decompress_block.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/decompress/zstd_decompress_internal.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/lib/zstd.h +161 -59
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/zlibWrapper/gzclose.c +1 -1
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/zlibWrapper/gzcompatibility.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/zlibWrapper/gzguts.h +0 -0
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/zlibWrapper/gzlib.c +9 -9
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/zlibWrapper/gzread.c +16 -8
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/zlibWrapper/gzwrite.c +8 -8
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/zlibWrapper/zstd_zlibwrapper.c +15 -11
- data/ext/zstdlib/{zstd-1.4.3 → zstd-1.4.4}/zlibWrapper/zstd_zlibwrapper.h +0 -0
- data/lib/2.2/zstdlib.so +0 -0
- data/lib/2.3/zstdlib.so +0 -0
- data/lib/2.4/zstdlib.so +0 -0
- data/lib/2.5/zstdlib.so +0 -0
- data/lib/2.6/zstdlib.so +0 -0
- metadata +62 -61
File without changes
|
@@ -61,7 +61,9 @@
|
|
61
61
|
* Error Management
|
62
62
|
****************************************************************/
|
63
63
|
#define HUF_isError ERR_isError
|
64
|
+
#ifndef CHECK_F
|
64
65
|
#define CHECK_F(f) { size_t const err_ = (f); if (HUF_isError(err_)) return err_; }
|
66
|
+
#endif
|
65
67
|
|
66
68
|
|
67
69
|
/* **************************************************************
|
File without changes
|
File without changes
|
@@ -88,10 +88,7 @@ size_t ZSTD_estimateDCtxSize(void) { return sizeof(ZSTD_DCtx); }
|
|
88
88
|
|
89
89
|
static size_t ZSTD_startingInputLength(ZSTD_format_e format)
|
90
90
|
{
|
91
|
-
size_t const startingInputLength = (format
|
92
|
-
ZSTD_FRAMEHEADERSIZE_PREFIX - ZSTD_FRAMEIDSIZE :
|
93
|
-
ZSTD_FRAMEHEADERSIZE_PREFIX;
|
94
|
-
ZSTD_STATIC_ASSERT(ZSTD_FRAMEHEADERSIZE_PREFIX >= ZSTD_FRAMEIDSIZE);
|
91
|
+
size_t const startingInputLength = ZSTD_FRAMEHEADERSIZE_PREFIX(format);
|
95
92
|
/* only supports formats ZSTD_f_zstd1 and ZSTD_f_zstd1_magicless */
|
96
93
|
assert( (format == ZSTD_f_zstd1) || (format == ZSTD_f_zstd1_magicless) );
|
97
94
|
return startingInputLength;
|
@@ -376,7 +373,7 @@ unsigned long long ZSTD_findDecompressedSize(const void* src, size_t srcSize)
|
|
376
373
|
{
|
377
374
|
unsigned long long totalDstSize = 0;
|
378
375
|
|
379
|
-
while (srcSize >=
|
376
|
+
while (srcSize >= ZSTD_startingInputLength(ZSTD_f_zstd1)) {
|
380
377
|
U32 const magicNumber = MEM_readLE32(src);
|
381
378
|
|
382
379
|
if ((magicNumber & ZSTD_MAGIC_SKIPPABLE_MASK) == ZSTD_MAGIC_SKIPPABLE_START) {
|
@@ -629,11 +626,12 @@ static size_t ZSTD_decompressFrame(ZSTD_DCtx* dctx,
|
|
629
626
|
|
630
627
|
/* check */
|
631
628
|
RETURN_ERROR_IF(
|
632
|
-
remainingSrcSize < ZSTD_FRAMEHEADERSIZE_MIN+ZSTD_blockHeaderSize,
|
629
|
+
remainingSrcSize < ZSTD_FRAMEHEADERSIZE_MIN(dctx->format)+ZSTD_blockHeaderSize,
|
633
630
|
srcSize_wrong);
|
634
631
|
|
635
632
|
/* Frame Header */
|
636
|
-
{ size_t const frameHeaderSize =
|
633
|
+
{ size_t const frameHeaderSize = ZSTD_frameHeaderSize_internal(
|
634
|
+
ip, ZSTD_FRAMEHEADERSIZE_PREFIX(dctx->format), dctx->format);
|
637
635
|
if (ZSTD_isError(frameHeaderSize)) return frameHeaderSize;
|
638
636
|
RETURN_ERROR_IF(remainingSrcSize < frameHeaderSize+ZSTD_blockHeaderSize,
|
639
637
|
srcSize_wrong);
|
@@ -714,7 +712,7 @@ static size_t ZSTD_decompressMultiFrame(ZSTD_DCtx* dctx,
|
|
714
712
|
dictSize = ZSTD_DDict_dictSize(ddict);
|
715
713
|
}
|
716
714
|
|
717
|
-
while (srcSize >=
|
715
|
+
while (srcSize >= ZSTD_startingInputLength(dctx->format)) {
|
718
716
|
|
719
717
|
#if defined(ZSTD_LEGACY_SUPPORT) && (ZSTD_LEGACY_SUPPORT >= 1)
|
720
718
|
if (ZSTD_isLegacy(src, srcSize)) {
|
@@ -1098,7 +1096,7 @@ ZSTD_loadDEntropy(ZSTD_entropyDTables_t* entropy,
|
|
1098
1096
|
size_t const dictContentSize = (size_t)(dictEnd - (dictPtr+12));
|
1099
1097
|
for (i=0; i<3; i++) {
|
1100
1098
|
U32 const rep = MEM_readLE32(dictPtr); dictPtr += 4;
|
1101
|
-
RETURN_ERROR_IF(rep==0 || rep
|
1099
|
+
RETURN_ERROR_IF(rep==0 || rep > dictContentSize,
|
1102
1100
|
dictionary_corrupted);
|
1103
1101
|
entropy->rep[i] = rep;
|
1104
1102
|
} }
|
@@ -1267,7 +1265,7 @@ size_t ZSTD_DCtx_loadDictionary_advanced(ZSTD_DCtx* dctx,
|
|
1267
1265
|
{
|
1268
1266
|
RETURN_ERROR_IF(dctx->streamStage != zdss_init, stage_wrong);
|
1269
1267
|
ZSTD_clearDict(dctx);
|
1270
|
-
if (dict && dictSize
|
1268
|
+
if (dict && dictSize != 0) {
|
1271
1269
|
dctx->ddictLocal = ZSTD_createDDict_advanced(dict, dictSize, dictLoadMethod, dictContentType, dctx->customMem);
|
1272
1270
|
RETURN_ERROR_IF(dctx->ddictLocal == NULL, memory_allocation);
|
1273
1271
|
dctx->ddict = dctx->ddictLocal;
|
@@ -1300,14 +1298,14 @@ size_t ZSTD_DCtx_refPrefix(ZSTD_DCtx* dctx, const void* prefix, size_t prefixSiz
|
|
1300
1298
|
|
1301
1299
|
|
1302
1300
|
/* ZSTD_initDStream_usingDict() :
|
1303
|
-
* return : expected size, aka
|
1301
|
+
* return : expected size, aka ZSTD_startingInputLength().
|
1304
1302
|
* this function cannot fail */
|
1305
1303
|
size_t ZSTD_initDStream_usingDict(ZSTD_DStream* zds, const void* dict, size_t dictSize)
|
1306
1304
|
{
|
1307
1305
|
DEBUGLOG(4, "ZSTD_initDStream_usingDict");
|
1308
1306
|
FORWARD_IF_ERROR( ZSTD_DCtx_reset(zds, ZSTD_reset_session_only) );
|
1309
1307
|
FORWARD_IF_ERROR( ZSTD_DCtx_loadDictionary(zds, dict, dictSize) );
|
1310
|
-
return
|
1308
|
+
return ZSTD_startingInputLength(zds->format);
|
1311
1309
|
}
|
1312
1310
|
|
1313
1311
|
/* note : this variant can't fail */
|
@@ -1324,16 +1322,16 @@ size_t ZSTD_initDStream_usingDDict(ZSTD_DStream* dctx, const ZSTD_DDict* ddict)
|
|
1324
1322
|
{
|
1325
1323
|
FORWARD_IF_ERROR( ZSTD_DCtx_reset(dctx, ZSTD_reset_session_only) );
|
1326
1324
|
FORWARD_IF_ERROR( ZSTD_DCtx_refDDict(dctx, ddict) );
|
1327
|
-
return
|
1325
|
+
return ZSTD_startingInputLength(dctx->format);
|
1328
1326
|
}
|
1329
1327
|
|
1330
1328
|
/* ZSTD_resetDStream() :
|
1331
|
-
* return : expected size, aka
|
1329
|
+
* return : expected size, aka ZSTD_startingInputLength().
|
1332
1330
|
* this function cannot fail */
|
1333
1331
|
size_t ZSTD_resetDStream(ZSTD_DStream* dctx)
|
1334
1332
|
{
|
1335
1333
|
FORWARD_IF_ERROR(ZSTD_DCtx_reset(dctx, ZSTD_reset_session_only));
|
1336
|
-
return
|
1334
|
+
return ZSTD_startingInputLength(dctx->format);
|
1337
1335
|
}
|
1338
1336
|
|
1339
1337
|
|
@@ -1564,7 +1562,7 @@ size_t ZSTD_decompressStream(ZSTD_DStream* zds, ZSTD_outBuffer* output, ZSTD_inB
|
|
1564
1562
|
zds->lhSize += remainingInput;
|
1565
1563
|
}
|
1566
1564
|
input->pos = input->size;
|
1567
|
-
return (MAX(ZSTD_FRAMEHEADERSIZE_MIN, hSize) - zds->lhSize) + ZSTD_blockHeaderSize; /* remaining header bytes + next block header */
|
1565
|
+
return (MAX((size_t)ZSTD_FRAMEHEADERSIZE_MIN(zds->format), hSize) - zds->lhSize) + ZSTD_blockHeaderSize; /* remaining header bytes + next block header */
|
1568
1566
|
}
|
1569
1567
|
assert(ip != NULL);
|
1570
1568
|
memcpy(zds->headerBuffer + zds->lhSize, ip, toLoad); zds->lhSize = hSize; ip += toLoad;
|
@@ -573,38 +573,118 @@ typedef struct {
|
|
573
573
|
size_t pos;
|
574
574
|
} seqState_t;
|
575
575
|
|
576
|
+
/*! ZSTD_overlapCopy8() :
|
577
|
+
* Copies 8 bytes from ip to op and updates op and ip where ip <= op.
|
578
|
+
* If the offset is < 8 then the offset is spread to at least 8 bytes.
|
579
|
+
*
|
580
|
+
* Precondition: *ip <= *op
|
581
|
+
* Postcondition: *op - *op >= 8
|
582
|
+
*/
|
583
|
+
static void ZSTD_overlapCopy8(BYTE** op, BYTE const** ip, size_t offset) {
|
584
|
+
assert(*ip <= *op);
|
585
|
+
if (offset < 8) {
|
586
|
+
/* close range match, overlap */
|
587
|
+
static const U32 dec32table[] = { 0, 1, 2, 1, 4, 4, 4, 4 }; /* added */
|
588
|
+
static const int dec64table[] = { 8, 8, 8, 7, 8, 9,10,11 }; /* subtracted */
|
589
|
+
int const sub2 = dec64table[offset];
|
590
|
+
(*op)[0] = (*ip)[0];
|
591
|
+
(*op)[1] = (*ip)[1];
|
592
|
+
(*op)[2] = (*ip)[2];
|
593
|
+
(*op)[3] = (*ip)[3];
|
594
|
+
*ip += dec32table[offset];
|
595
|
+
ZSTD_copy4(*op+4, *ip);
|
596
|
+
*ip -= sub2;
|
597
|
+
} else {
|
598
|
+
ZSTD_copy8(*op, *ip);
|
599
|
+
}
|
600
|
+
*ip += 8;
|
601
|
+
*op += 8;
|
602
|
+
assert(*op - *ip >= 8);
|
603
|
+
}
|
604
|
+
|
605
|
+
/*! ZSTD_safecopy() :
|
606
|
+
* Specialized version of memcpy() that is allowed to READ up to WILDCOPY_OVERLENGTH past the input buffer
|
607
|
+
* and write up to 16 bytes past oend_w (op >= oend_w is allowed).
|
608
|
+
* This function is only called in the uncommon case where the sequence is near the end of the block. It
|
609
|
+
* should be fast for a single long sequence, but can be slow for several short sequences.
|
610
|
+
*
|
611
|
+
* @param ovtype controls the overlap detection
|
612
|
+
* - ZSTD_no_overlap: The source and destination are guaranteed to be at least WILDCOPY_VECLEN bytes apart.
|
613
|
+
* - ZSTD_overlap_src_before_dst: The src and dst may overlap and may be any distance apart.
|
614
|
+
* The src buffer must be before the dst buffer.
|
615
|
+
*/
|
616
|
+
static void ZSTD_safecopy(BYTE* op, BYTE* const oend_w, BYTE const* ip, ptrdiff_t length, ZSTD_overlap_e ovtype) {
|
617
|
+
ptrdiff_t const diff = op - ip;
|
618
|
+
BYTE* const oend = op + length;
|
576
619
|
|
577
|
-
|
578
|
-
|
579
|
-
|
580
|
-
|
581
|
-
|
582
|
-
|
583
|
-
|
620
|
+
assert((ovtype == ZSTD_no_overlap && (diff <= -8 || diff >= 8 || op >= oend_w)) ||
|
621
|
+
(ovtype == ZSTD_overlap_src_before_dst && diff >= 0));
|
622
|
+
|
623
|
+
if (length < 8) {
|
624
|
+
/* Handle short lengths. */
|
625
|
+
while (op < oend) *op++ = *ip++;
|
626
|
+
return;
|
627
|
+
}
|
628
|
+
if (ovtype == ZSTD_overlap_src_before_dst) {
|
629
|
+
/* Copy 8 bytes and ensure the offset >= 8 when there can be overlap. */
|
630
|
+
assert(length >= 8);
|
631
|
+
ZSTD_overlapCopy8(&op, &ip, diff);
|
632
|
+
assert(op - ip >= 8);
|
633
|
+
assert(op <= oend);
|
634
|
+
}
|
635
|
+
|
636
|
+
if (oend <= oend_w) {
|
637
|
+
/* No risk of overwrite. */
|
638
|
+
ZSTD_wildcopy(op, ip, length, ovtype);
|
639
|
+
return;
|
640
|
+
}
|
641
|
+
if (op <= oend_w) {
|
642
|
+
/* Wildcopy until we get close to the end. */
|
643
|
+
assert(oend > oend_w);
|
644
|
+
ZSTD_wildcopy(op, ip, oend_w - op, ovtype);
|
645
|
+
ip += oend_w - op;
|
646
|
+
op = oend_w;
|
647
|
+
}
|
648
|
+
/* Handle the leftovers. */
|
649
|
+
while (op < oend) *op++ = *ip++;
|
650
|
+
}
|
651
|
+
|
652
|
+
/* ZSTD_execSequenceEnd():
|
653
|
+
* This version handles cases that are near the end of the output buffer. It requires
|
654
|
+
* more careful checks to make sure there is no overflow. By separating out these hard
|
655
|
+
* and unlikely cases, we can speed up the common cases.
|
656
|
+
*
|
657
|
+
* NOTE: This function needs to be fast for a single long sequence, but doesn't need
|
658
|
+
* to be optimized for many small sequences, since those fall into ZSTD_execSequence().
|
659
|
+
*/
|
584
660
|
FORCE_NOINLINE
|
585
|
-
size_t
|
586
|
-
|
587
|
-
|
588
|
-
|
661
|
+
size_t ZSTD_execSequenceEnd(BYTE* op,
|
662
|
+
BYTE* const oend, seq_t sequence,
|
663
|
+
const BYTE** litPtr, const BYTE* const litLimit,
|
664
|
+
const BYTE* const prefixStart, const BYTE* const virtualStart, const BYTE* const dictEnd)
|
589
665
|
{
|
590
666
|
BYTE* const oLitEnd = op + sequence.litLength;
|
591
667
|
size_t const sequenceLength = sequence.litLength + sequence.matchLength;
|
592
668
|
BYTE* const oMatchEnd = op + sequenceLength; /* risk : address space overflow (32-bits) */
|
593
669
|
const BYTE* const iLitEnd = *litPtr + sequence.litLength;
|
594
670
|
const BYTE* match = oLitEnd - sequence.offset;
|
671
|
+
BYTE* const oend_w = oend - WILDCOPY_OVERLENGTH;
|
595
672
|
|
596
|
-
/*
|
597
|
-
|
673
|
+
/* bounds checks */
|
674
|
+
assert(oLitEnd < oMatchEnd);
|
675
|
+
RETURN_ERROR_IF(oMatchEnd > oend, dstSize_tooSmall, "last match must fit within dstBuffer");
|
598
676
|
RETURN_ERROR_IF(iLitEnd > litLimit, corruption_detected, "try to read beyond literal buffer");
|
599
677
|
|
600
678
|
/* copy literals */
|
601
|
-
|
679
|
+
ZSTD_safecopy(op, oend_w, *litPtr, sequence.litLength, ZSTD_no_overlap);
|
680
|
+
op = oLitEnd;
|
681
|
+
*litPtr = iLitEnd;
|
602
682
|
|
603
683
|
/* copy Match */
|
604
|
-
if (sequence.offset > (size_t)(oLitEnd -
|
684
|
+
if (sequence.offset > (size_t)(oLitEnd - prefixStart)) {
|
605
685
|
/* offset beyond prefix */
|
606
|
-
RETURN_ERROR_IF(sequence.offset > (size_t)(oLitEnd -
|
607
|
-
match = dictEnd - (
|
686
|
+
RETURN_ERROR_IF(sequence.offset > (size_t)(oLitEnd - virtualStart), corruption_detected);
|
687
|
+
match = dictEnd - (prefixStart-match);
|
608
688
|
if (match + sequence.matchLength <= dictEnd) {
|
609
689
|
memmove(oLitEnd, match, sequence.matchLength);
|
610
690
|
return sequenceLength;
|
@@ -614,13 +694,12 @@ size_t ZSTD_execSequenceLast7(BYTE* op,
|
|
614
694
|
memmove(oLitEnd, match, length1);
|
615
695
|
op = oLitEnd + length1;
|
616
696
|
sequence.matchLength -= length1;
|
617
|
-
match =
|
697
|
+
match = prefixStart;
|
618
698
|
} }
|
619
|
-
|
699
|
+
ZSTD_safecopy(op, oend_w, match, sequence.matchLength, ZSTD_overlap_src_before_dst);
|
620
700
|
return sequenceLength;
|
621
701
|
}
|
622
702
|
|
623
|
-
|
624
703
|
HINT_INLINE
|
625
704
|
size_t ZSTD_execSequence(BYTE* op,
|
626
705
|
BYTE* const oend, seq_t sequence,
|
@@ -634,20 +713,29 @@ size_t ZSTD_execSequence(BYTE* op,
|
|
634
713
|
const BYTE* const iLitEnd = *litPtr + sequence.litLength;
|
635
714
|
const BYTE* match = oLitEnd - sequence.offset;
|
636
715
|
|
637
|
-
/*
|
638
|
-
|
639
|
-
|
640
|
-
|
641
|
-
|
642
|
-
/*
|
643
|
-
|
644
|
-
|
645
|
-
|
646
|
-
|
716
|
+
/* Errors and uncommon cases handled here. */
|
717
|
+
assert(oLitEnd < oMatchEnd);
|
718
|
+
if (iLitEnd > litLimit || oMatchEnd > oend_w)
|
719
|
+
return ZSTD_execSequenceEnd(op, oend, sequence, litPtr, litLimit, prefixStart, virtualStart, dictEnd);
|
720
|
+
|
721
|
+
/* Assumptions (everything else goes into ZSTD_execSequenceEnd()) */
|
722
|
+
assert(iLitEnd <= litLimit /* Literal length is in bounds */);
|
723
|
+
assert(oLitEnd <= oend_w /* Can wildcopy literals */);
|
724
|
+
assert(oMatchEnd <= oend_w /* Can wildcopy matches */);
|
725
|
+
|
726
|
+
/* Copy Literals:
|
727
|
+
* Split out litLength <= 16 since it is nearly always true. +1.6% on gcc-9.
|
728
|
+
* We likely don't need the full 32-byte wildcopy.
|
729
|
+
*/
|
730
|
+
assert(WILDCOPY_OVERLENGTH >= 16);
|
731
|
+
ZSTD_copy16(op, (*litPtr));
|
732
|
+
if (sequence.litLength > 16) {
|
733
|
+
ZSTD_wildcopy(op+16, (*litPtr)+16, sequence.litLength-16, ZSTD_no_overlap);
|
734
|
+
}
|
647
735
|
op = oLitEnd;
|
648
736
|
*litPtr = iLitEnd; /* update for next sequence */
|
649
737
|
|
650
|
-
/*
|
738
|
+
/* Copy Match */
|
651
739
|
if (sequence.offset > (size_t)(oLitEnd - prefixStart)) {
|
652
740
|
/* offset beyond prefix -> go into extDict */
|
653
741
|
RETURN_ERROR_IF(sequence.offset > (size_t)(oLitEnd - virtualStart), corruption_detected);
|
@@ -662,123 +750,33 @@ size_t ZSTD_execSequence(BYTE* op,
|
|
662
750
|
op = oLitEnd + length1;
|
663
751
|
sequence.matchLength -= length1;
|
664
752
|
match = prefixStart;
|
665
|
-
if (op > oend_w || sequence.matchLength < MINMATCH) {
|
666
|
-
U32 i;
|
667
|
-
for (i = 0; i < sequence.matchLength; ++i) op[i] = match[i];
|
668
|
-
return sequenceLength;
|
669
|
-
}
|
670
753
|
} }
|
671
|
-
/*
|
672
|
-
|
673
|
-
|
674
|
-
|
675
|
-
|
676
|
-
|
677
|
-
|
678
|
-
|
679
|
-
|
680
|
-
|
681
|
-
|
682
|
-
|
683
|
-
|
684
|
-
|
685
|
-
match
|
686
|
-
|
687
|
-
ZSTD_copy8(op, match);
|
688
|
-
}
|
689
|
-
op += 8; match += 8;
|
690
|
-
|
691
|
-
if (oMatchEnd > oend-(16-MINMATCH)) {
|
692
|
-
if (op < oend_w) {
|
693
|
-
ZSTD_wildcopy(op, match, oend_w - op, ZSTD_overlap_src_before_dst);
|
694
|
-
match += oend_w - op;
|
695
|
-
op = oend_w;
|
696
|
-
}
|
697
|
-
while (op < oMatchEnd) *op++ = *match++;
|
698
|
-
} else {
|
699
|
-
ZSTD_wildcopy(op, match, (ptrdiff_t)sequence.matchLength-8, ZSTD_overlap_src_before_dst); /* works even if matchLength < 8 */
|
754
|
+
/* Match within prefix of 1 or more bytes */
|
755
|
+
assert(op <= oMatchEnd);
|
756
|
+
assert(oMatchEnd <= oend_w);
|
757
|
+
assert(match >= prefixStart);
|
758
|
+
assert(sequence.matchLength >= 1);
|
759
|
+
|
760
|
+
/* Nearly all offsets are >= WILDCOPY_VECLEN bytes, which means we can use wildcopy
|
761
|
+
* without overlap checking.
|
762
|
+
*/
|
763
|
+
if (sequence.offset >= WILDCOPY_VECLEN) {
|
764
|
+
/* We bet on a full wildcopy for matches, since we expect matches to be
|
765
|
+
* longer than literals (in general). In silesia, ~10% of matches are longer
|
766
|
+
* than 16 bytes.
|
767
|
+
*/
|
768
|
+
ZSTD_wildcopy(op, match, (ptrdiff_t)sequence.matchLength, ZSTD_no_overlap);
|
769
|
+
return sequenceLength;
|
700
770
|
}
|
701
|
-
|
702
|
-
}
|
703
|
-
|
704
|
-
|
705
|
-
HINT_INLINE
|
706
|
-
size_t ZSTD_execSequenceLong(BYTE* op,
|
707
|
-
BYTE* const oend, seq_t sequence,
|
708
|
-
const BYTE** litPtr, const BYTE* const litLimit,
|
709
|
-
const BYTE* const prefixStart, const BYTE* const dictStart, const BYTE* const dictEnd)
|
710
|
-
{
|
711
|
-
BYTE* const oLitEnd = op + sequence.litLength;
|
712
|
-
size_t const sequenceLength = sequence.litLength + sequence.matchLength;
|
713
|
-
BYTE* const oMatchEnd = op + sequenceLength; /* risk : address space overflow (32-bits) */
|
714
|
-
BYTE* const oend_w = oend - WILDCOPY_OVERLENGTH;
|
715
|
-
const BYTE* const iLitEnd = *litPtr + sequence.litLength;
|
716
|
-
const BYTE* match = sequence.match;
|
771
|
+
assert(sequence.offset < WILDCOPY_VECLEN);
|
717
772
|
|
718
|
-
/*
|
719
|
-
|
720
|
-
RETURN_ERROR_IF(iLitEnd > litLimit, corruption_detected, "over-read beyond lit buffer");
|
721
|
-
if (oLitEnd > oend_w) return ZSTD_execSequenceLast7(op, oend, sequence, litPtr, litLimit, prefixStart, dictStart, dictEnd);
|
722
|
-
|
723
|
-
/* copy Literals */
|
724
|
-
if (sequence.litLength > 8)
|
725
|
-
ZSTD_wildcopy_16min(op, *litPtr, sequence.litLength, ZSTD_no_overlap); /* note : since oLitEnd <= oend-WILDCOPY_OVERLENGTH, no risk of overwrite beyond oend */
|
726
|
-
else
|
727
|
-
ZSTD_copy8(op, *litPtr); /* note : op <= oLitEnd <= oend_w == oend - 8 */
|
773
|
+
/* Copy 8 bytes and spread the offset to be >= 8. */
|
774
|
+
ZSTD_overlapCopy8(&op, &match, sequence.offset);
|
728
775
|
|
729
|
-
|
730
|
-
|
731
|
-
|
732
|
-
|
733
|
-
if (sequence.offset > (size_t)(oLitEnd - prefixStart)) {
|
734
|
-
/* offset beyond prefix */
|
735
|
-
RETURN_ERROR_IF(sequence.offset > (size_t)(oLitEnd - dictStart), corruption_detected);
|
736
|
-
if (match + sequence.matchLength <= dictEnd) {
|
737
|
-
memmove(oLitEnd, match, sequence.matchLength);
|
738
|
-
return sequenceLength;
|
739
|
-
}
|
740
|
-
/* span extDict & currentPrefixSegment */
|
741
|
-
{ size_t const length1 = dictEnd - match;
|
742
|
-
memmove(oLitEnd, match, length1);
|
743
|
-
op = oLitEnd + length1;
|
744
|
-
sequence.matchLength -= length1;
|
745
|
-
match = prefixStart;
|
746
|
-
if (op > oend_w || sequence.matchLength < MINMATCH) {
|
747
|
-
U32 i;
|
748
|
-
for (i = 0; i < sequence.matchLength; ++i) op[i] = match[i];
|
749
|
-
return sequenceLength;
|
750
|
-
}
|
751
|
-
} }
|
752
|
-
assert(op <= oend_w);
|
753
|
-
assert(sequence.matchLength >= MINMATCH);
|
754
|
-
|
755
|
-
/* match within prefix */
|
756
|
-
if (sequence.offset < 8) {
|
757
|
-
/* close range match, overlap */
|
758
|
-
static const U32 dec32table[] = { 0, 1, 2, 1, 4, 4, 4, 4 }; /* added */
|
759
|
-
static const int dec64table[] = { 8, 8, 8, 7, 8, 9,10,11 }; /* subtracted */
|
760
|
-
int const sub2 = dec64table[sequence.offset];
|
761
|
-
op[0] = match[0];
|
762
|
-
op[1] = match[1];
|
763
|
-
op[2] = match[2];
|
764
|
-
op[3] = match[3];
|
765
|
-
match += dec32table[sequence.offset];
|
766
|
-
ZSTD_copy4(op+4, match);
|
767
|
-
match -= sub2;
|
768
|
-
} else {
|
769
|
-
ZSTD_copy8(op, match);
|
770
|
-
}
|
771
|
-
op += 8; match += 8;
|
772
|
-
|
773
|
-
if (oMatchEnd > oend-(16-MINMATCH)) {
|
774
|
-
if (op < oend_w) {
|
775
|
-
ZSTD_wildcopy(op, match, oend_w - op, ZSTD_overlap_src_before_dst);
|
776
|
-
match += oend_w - op;
|
777
|
-
op = oend_w;
|
778
|
-
}
|
779
|
-
while (op < oMatchEnd) *op++ = *match++;
|
780
|
-
} else {
|
781
|
-
ZSTD_wildcopy(op, match, (ptrdiff_t)sequence.matchLength-8, ZSTD_overlap_src_before_dst); /* works even if matchLength < 8 */
|
776
|
+
/* If the match length is > 8 bytes, then continue with the wildcopy. */
|
777
|
+
if (sequence.matchLength > 8) {
|
778
|
+
assert(op < oMatchEnd);
|
779
|
+
ZSTD_wildcopy(op, match, (ptrdiff_t)sequence.matchLength-8, ZSTD_overlap_src_before_dst);
|
782
780
|
}
|
783
781
|
return sequenceLength;
|
784
782
|
}
|
@@ -1098,7 +1096,7 @@ ZSTD_decompressSequencesLong_body(
|
|
1098
1096
|
/* decode and decompress */
|
1099
1097
|
for ( ; (BIT_reloadDStream(&(seqState.DStream)) <= BIT_DStream_completed) && (seqNb<nbSeq) ; seqNb++) {
|
1100
1098
|
seq_t const sequence = ZSTD_decodeSequenceLong(&seqState, isLongOffset);
|
1101
|
-
size_t const oneSeqSize =
|
1099
|
+
size_t const oneSeqSize = ZSTD_execSequence(op, oend, sequences[(seqNb-ADVANCED_SEQS) & STORED_SEQS_MASK], &litPtr, litEnd, prefixStart, dictStart, dictEnd);
|
1102
1100
|
if (ZSTD_isError(oneSeqSize)) return oneSeqSize;
|
1103
1101
|
PREFETCH_L1(sequence.match); PREFETCH_L1(sequence.match + sequence.matchLength - 1); /* note : it's safe to invoke PREFETCH() on any memory address, including invalid ones */
|
1104
1102
|
sequences[seqNb & STORED_SEQS_MASK] = sequence;
|
@@ -1109,7 +1107,7 @@ ZSTD_decompressSequencesLong_body(
|
|
1109
1107
|
/* finish queue */
|
1110
1108
|
seqNb -= seqAdvance;
|
1111
1109
|
for ( ; seqNb<nbSeq ; seqNb++) {
|
1112
|
-
size_t const oneSeqSize =
|
1110
|
+
size_t const oneSeqSize = ZSTD_execSequence(op, oend, sequences[seqNb&STORED_SEQS_MASK], &litPtr, litEnd, prefixStart, dictStart, dictEnd);
|
1113
1111
|
if (ZSTD_isError(oneSeqSize)) return oneSeqSize;
|
1114
1112
|
op += oneSeqSize;
|
1115
1113
|
}
|