zstdlib 0.2.0-x64-mingw32 → 0.3.0-x64-mingw32
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/CHANGES.md +5 -0
- data/ext/zstdlib/extconf.rb +1 -1
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/bitstream.h +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/compiler.h +7 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/cpu.h +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/debug.c +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/debug.h +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/entropy_common.c +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/error_private.c +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/error_private.h +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/fse.h +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/fse_decompress.c +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/huf.h +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/mem.h +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/pool.c +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/pool.h +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/threading.c +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/threading.h +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/xxhash.c +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/xxhash.h +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/zstd_common.c +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/zstd_errors.h +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/common/zstd_internal.h +58 -6
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/fse_compress.c +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/hist.c +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/hist.h +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/huf_compress.c +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/zstd_compress.c +178 -691
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/zstd_compress_internal.h +98 -30
- data/ext/zstdlib/zstd-1.4.2/lib/compress/zstd_compress_literals.c +149 -0
- data/ext/zstdlib/zstd-1.4.2/lib/compress/zstd_compress_literals.h +29 -0
- data/ext/zstdlib/zstd-1.4.2/lib/compress/zstd_compress_sequences.c +415 -0
- data/ext/zstdlib/zstd-1.4.2/lib/compress/zstd_compress_sequences.h +47 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/zstd_double_fast.c +56 -36
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/zstd_double_fast.h +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/zstd_fast.c +35 -14
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/zstd_fast.h +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/zstd_lazy.c +10 -5
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/zstd_lazy.h +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/zstd_ldm.c +1 -1
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/zstd_ldm.h +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/zstd_opt.c +45 -32
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/zstd_opt.h +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/zstdmt_compress.c +18 -7
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/compress/zstdmt_compress.h +1 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/decompress/huf_decompress.c +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/decompress/zstd_ddict.c +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/decompress/zstd_ddict.h +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/decompress/zstd_decompress.c +14 -9
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/decompress/zstd_decompress_block.c +20 -9
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/decompress/zstd_decompress_block.h +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/decompress/zstd_decompress_internal.h +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/lib/zstd.h +53 -21
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/zlibWrapper/gzclose.c +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/zlibWrapper/gzcompatibility.h +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/zlibWrapper/gzguts.h +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/zlibWrapper/gzlib.c +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/zlibWrapper/gzread.c +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/zlibWrapper/gzwrite.c +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/zlibWrapper/zstd_zlibwrapper.c +0 -0
- data/ext/zstdlib/{zstd-1.4.0 → zstd-1.4.2}/zlibWrapper/zstd_zlibwrapper.h +0 -0
- data/lib/2.2/zstdlib.so +0 -0
- data/lib/2.3/zstdlib.so +0 -0
- data/lib/2.4/zstdlib.so +0 -0
- data/lib/2.5/zstdlib.so +0 -0
- data/lib/2.6/zstdlib.so +0 -0
- metadata +61 -57
@@ -33,13 +33,13 @@ extern "C" {
|
|
33
33
|
***************************************/
|
34
34
|
#define kSearchStrength 8
|
35
35
|
#define HASH_READ_SIZE 8
|
36
|
-
#define ZSTD_DUBT_UNSORTED_MARK 1 /* For btlazy2 strategy, index 1
|
36
|
+
#define ZSTD_DUBT_UNSORTED_MARK 1 /* For btlazy2 strategy, index ZSTD_DUBT_UNSORTED_MARK==1 means "unsorted".
|
37
37
|
It could be confused for a real successor at index "1", if sorted as larger than its predecessor.
|
38
38
|
It's not a big deal though : candidate will just be sorted again.
|
39
39
|
Additionally, candidate position 1 will be lost.
|
40
40
|
But candidate 1 cannot hide a large tree of candidates, so it's a minimal loss.
|
41
|
-
The benefit is that ZSTD_DUBT_UNSORTED_MARK cannot be mishandled after table re-use with a different strategy
|
42
|
-
|
41
|
+
The benefit is that ZSTD_DUBT_UNSORTED_MARK cannot be mishandled after table re-use with a different strategy.
|
42
|
+
This constant is required by ZSTD_compressBlock_btlazy2() and ZSTD_reduceTable_internal() */
|
43
43
|
|
44
44
|
|
45
45
|
/*-*************************************
|
@@ -128,21 +128,20 @@ typedef struct {
|
|
128
128
|
BYTE const* base; /* All regular indexes relative to this position */
|
129
129
|
BYTE const* dictBase; /* extDict indexes relative to this position */
|
130
130
|
U32 dictLimit; /* below that point, need extDict */
|
131
|
-
U32 lowLimit; /* below that point, no more data */
|
131
|
+
U32 lowLimit; /* below that point, no more valid data */
|
132
132
|
} ZSTD_window_t;
|
133
133
|
|
134
134
|
typedef struct ZSTD_matchState_t ZSTD_matchState_t;
|
135
135
|
struct ZSTD_matchState_t {
|
136
136
|
ZSTD_window_t window; /* State for window round buffer management */
|
137
|
-
U32 loadedDictEnd; /* index of end of dictionary */
|
137
|
+
U32 loadedDictEnd; /* index of end of dictionary, within context's referential. When dict referential is copied into active context (i.e. not attached), effectively same value as dictSize, since referential starts from zero */
|
138
138
|
U32 nextToUpdate; /* index from which to continue table update */
|
139
|
-
U32 nextToUpdate3; /* index from which to continue table update */
|
140
139
|
U32 hashLog3; /* dispatch table : larger == faster, more memory */
|
141
140
|
U32* hashTable;
|
142
141
|
U32* hashTable3;
|
143
142
|
U32* chainTable;
|
144
143
|
optState_t opt; /* optimal parser state */
|
145
|
-
const ZSTD_matchState_t
|
144
|
+
const ZSTD_matchState_t* dictMatchState;
|
146
145
|
ZSTD_compressionParameters cParams;
|
147
146
|
};
|
148
147
|
|
@@ -195,6 +194,9 @@ struct ZSTD_CCtx_params_s {
|
|
195
194
|
int compressionLevel;
|
196
195
|
int forceWindow; /* force back-references to respect limit of
|
197
196
|
* 1<<wLog, even for dictionary */
|
197
|
+
size_t targetCBlockSize; /* Tries to fit compressed block size to be around targetCBlockSize.
|
198
|
+
* No target when targetCBlockSize == 0.
|
199
|
+
* There is no guarantee on compressed block size */
|
198
200
|
|
199
201
|
ZSTD_dictAttachPref_e attachDictPref;
|
200
202
|
ZSTD_literalCompressionMode_e literalCompressionMode;
|
@@ -305,6 +307,30 @@ MEM_STATIC U32 ZSTD_MLcode(U32 mlBase)
|
|
305
307
|
return (mlBase > 127) ? ZSTD_highbit32(mlBase) + ML_deltaCode : ML_Code[mlBase];
|
306
308
|
}
|
307
309
|
|
310
|
+
/* ZSTD_cParam_withinBounds:
|
311
|
+
* @return 1 if value is within cParam bounds,
|
312
|
+
* 0 otherwise */
|
313
|
+
MEM_STATIC int ZSTD_cParam_withinBounds(ZSTD_cParameter cParam, int value)
|
314
|
+
{
|
315
|
+
ZSTD_bounds const bounds = ZSTD_cParam_getBounds(cParam);
|
316
|
+
if (ZSTD_isError(bounds.error)) return 0;
|
317
|
+
if (value < bounds.lowerBound) return 0;
|
318
|
+
if (value > bounds.upperBound) return 0;
|
319
|
+
return 1;
|
320
|
+
}
|
321
|
+
|
322
|
+
/* ZSTD_minGain() :
|
323
|
+
* minimum compression required
|
324
|
+
* to generate a compress block or a compressed literals section.
|
325
|
+
* note : use same formula for both situations */
|
326
|
+
MEM_STATIC size_t ZSTD_minGain(size_t srcSize, ZSTD_strategy strat)
|
327
|
+
{
|
328
|
+
U32 const minlog = (strat>=ZSTD_btultra) ? (U32)(strat) - 1 : 6;
|
329
|
+
ZSTD_STATIC_ASSERT(ZSTD_btultra == 8);
|
330
|
+
assert(ZSTD_cParam_withinBounds(ZSTD_c_strategy, strat));
|
331
|
+
return (srcSize >> minlog) + 2;
|
332
|
+
}
|
333
|
+
|
308
334
|
/*! ZSTD_storeSeq() :
|
309
335
|
* Store a sequence (literal length, literals, offset code and match length code) into seqStore_t.
|
310
336
|
* `offsetCode` : distance to match + 3 (values 1-3 are repCodes).
|
@@ -324,7 +350,7 @@ MEM_STATIC void ZSTD_storeSeq(seqStore_t* seqStorePtr, size_t litLength, const v
|
|
324
350
|
/* copy Literals */
|
325
351
|
assert(seqStorePtr->maxNbLit <= 128 KB);
|
326
352
|
assert(seqStorePtr->lit + litLength <= seqStorePtr->litStart + seqStorePtr->maxNbLit);
|
327
|
-
ZSTD_wildcopy(seqStorePtr->lit, literals, litLength);
|
353
|
+
ZSTD_wildcopy(seqStorePtr->lit, literals, litLength, ZSTD_no_overlap);
|
328
354
|
seqStorePtr->lit += litLength;
|
329
355
|
|
330
356
|
/* literal Length */
|
@@ -564,6 +590,9 @@ MEM_STATIC U64 ZSTD_rollingHash_rotate(U64 hash, BYTE toRemove, BYTE toAdd, U64
|
|
564
590
|
/*-*************************************
|
565
591
|
* Round buffer management
|
566
592
|
***************************************/
|
593
|
+
#if (ZSTD_WINDOWLOG_MAX_64 > 31)
|
594
|
+
# error "ZSTD_WINDOWLOG_MAX is too large : would overflow ZSTD_CURRENT_MAX"
|
595
|
+
#endif
|
567
596
|
/* Max current allowed */
|
568
597
|
#define ZSTD_CURRENT_MAX ((3U << 29) + (1U << ZSTD_WINDOWLOG_MAX))
|
569
598
|
/* Maximum chunk size before overflow correction needs to be called again */
|
@@ -675,31 +704,49 @@ MEM_STATIC U32 ZSTD_window_correctOverflow(ZSTD_window_t* window, U32 cycleLog,
|
|
675
704
|
* Updates lowLimit so that:
|
676
705
|
* (srcEnd - base) - lowLimit == maxDist + loadedDictEnd
|
677
706
|
*
|
678
|
-
*
|
679
|
-
* This must be called before a block compression call
|
680
|
-
* source end.
|
707
|
+
* It ensures index is valid as long as index >= lowLimit.
|
708
|
+
* This must be called before a block compression call.
|
681
709
|
*
|
682
|
-
*
|
683
|
-
*
|
684
|
-
*
|
685
|
-
* they cannot be referenced. If loadedDictEndPtr is NULL, we use
|
686
|
-
* loadedDictEnd == 0.
|
710
|
+
* loadedDictEnd is only defined if a dictionary is in use for current compression.
|
711
|
+
* As the name implies, loadedDictEnd represents the index at end of dictionary.
|
712
|
+
* The value lies within context's referential, it can be directly compared to blockEndIdx.
|
687
713
|
*
|
688
|
-
*
|
689
|
-
*
|
690
|
-
* is
|
714
|
+
* If loadedDictEndPtr is NULL, no dictionary is in use, and we use loadedDictEnd == 0.
|
715
|
+
* If loadedDictEndPtr is not NULL, we set it to zero after updating lowLimit.
|
716
|
+
* This is because dictionaries are allowed to be referenced fully
|
717
|
+
* as long as the last byte of the dictionary is in the window.
|
718
|
+
* Once input has progressed beyond window size, dictionary cannot be referenced anymore.
|
719
|
+
*
|
720
|
+
* In normal dict mode, the dictionary lies between lowLimit and dictLimit.
|
721
|
+
* In dictMatchState mode, lowLimit and dictLimit are the same,
|
722
|
+
* and the dictionary is below them.
|
723
|
+
* forceWindow and dictMatchState are therefore incompatible.
|
691
724
|
*/
|
692
725
|
MEM_STATIC void
|
693
726
|
ZSTD_window_enforceMaxDist(ZSTD_window_t* window,
|
694
|
-
|
695
|
-
U32
|
696
|
-
U32*
|
727
|
+
const void* blockEnd,
|
728
|
+
U32 maxDist,
|
729
|
+
U32* loadedDictEndPtr,
|
697
730
|
const ZSTD_matchState_t** dictMatchStatePtr)
|
698
731
|
{
|
699
|
-
U32 const blockEndIdx = (U32)((BYTE const*)
|
700
|
-
U32 loadedDictEnd = (loadedDictEndPtr != NULL) ? *loadedDictEndPtr : 0;
|
701
|
-
DEBUGLOG(5, "ZSTD_window_enforceMaxDist: blockEndIdx=%u, maxDist=%u",
|
702
|
-
(unsigned)blockEndIdx, (unsigned)maxDist);
|
732
|
+
U32 const blockEndIdx = (U32)((BYTE const*)blockEnd - window->base);
|
733
|
+
U32 const loadedDictEnd = (loadedDictEndPtr != NULL) ? *loadedDictEndPtr : 0;
|
734
|
+
DEBUGLOG(5, "ZSTD_window_enforceMaxDist: blockEndIdx=%u, maxDist=%u, loadedDictEnd=%u",
|
735
|
+
(unsigned)blockEndIdx, (unsigned)maxDist, (unsigned)loadedDictEnd);
|
736
|
+
|
737
|
+
/* - When there is no dictionary : loadedDictEnd == 0.
|
738
|
+
In which case, the test (blockEndIdx > maxDist) is merely to avoid
|
739
|
+
overflowing next operation `newLowLimit = blockEndIdx - maxDist`.
|
740
|
+
- When there is a standard dictionary :
|
741
|
+
Index referential is copied from the dictionary,
|
742
|
+
which means it starts from 0.
|
743
|
+
In which case, loadedDictEnd == dictSize,
|
744
|
+
and it makes sense to compare `blockEndIdx > maxDist + dictSize`
|
745
|
+
since `blockEndIdx` also starts from zero.
|
746
|
+
- When there is an attached dictionary :
|
747
|
+
loadedDictEnd is expressed within the referential of the context,
|
748
|
+
so it can be directly compared against blockEndIdx.
|
749
|
+
*/
|
703
750
|
if (blockEndIdx > maxDist + loadedDictEnd) {
|
704
751
|
U32 const newLowLimit = blockEndIdx - maxDist;
|
705
752
|
if (window->lowLimit < newLowLimit) window->lowLimit = newLowLimit;
|
@@ -708,10 +755,31 @@ ZSTD_window_enforceMaxDist(ZSTD_window_t* window,
|
|
708
755
|
(unsigned)window->dictLimit, (unsigned)window->lowLimit);
|
709
756
|
window->dictLimit = window->lowLimit;
|
710
757
|
}
|
711
|
-
|
712
|
-
|
713
|
-
if (dictMatchStatePtr)
|
714
|
-
|
758
|
+
/* On reaching window size, dictionaries are invalidated */
|
759
|
+
if (loadedDictEndPtr) *loadedDictEndPtr = 0;
|
760
|
+
if (dictMatchStatePtr) *dictMatchStatePtr = NULL;
|
761
|
+
}
|
762
|
+
}
|
763
|
+
|
764
|
+
/* Similar to ZSTD_window_enforceMaxDist(),
|
765
|
+
* but only invalidates dictionary
|
766
|
+
* when input progresses beyond window size. */
|
767
|
+
MEM_STATIC void
|
768
|
+
ZSTD_checkDictValidity(ZSTD_window_t* window,
|
769
|
+
const void* blockEnd,
|
770
|
+
U32 maxDist,
|
771
|
+
U32* loadedDictEndPtr,
|
772
|
+
const ZSTD_matchState_t** dictMatchStatePtr)
|
773
|
+
{
|
774
|
+
U32 const blockEndIdx = (U32)((BYTE const*)blockEnd - window->base);
|
775
|
+
U32 const loadedDictEnd = (loadedDictEndPtr != NULL) ? *loadedDictEndPtr : 0;
|
776
|
+
DEBUGLOG(5, "ZSTD_checkDictValidity: blockEndIdx=%u, maxDist=%u, loadedDictEnd=%u",
|
777
|
+
(unsigned)blockEndIdx, (unsigned)maxDist, (unsigned)loadedDictEnd);
|
778
|
+
|
779
|
+
if (loadedDictEnd && (blockEndIdx > maxDist + loadedDictEnd)) {
|
780
|
+
/* On reaching window size, dictionaries are invalidated */
|
781
|
+
if (loadedDictEndPtr) *loadedDictEndPtr = 0;
|
782
|
+
if (dictMatchStatePtr) *dictMatchStatePtr = NULL;
|
715
783
|
}
|
716
784
|
}
|
717
785
|
|
@@ -0,0 +1,149 @@
|
|
1
|
+
/*
|
2
|
+
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
3
|
+
* All rights reserved.
|
4
|
+
*
|
5
|
+
* This source code is licensed under both the BSD-style license (found in the
|
6
|
+
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
7
|
+
* in the COPYING file in the root directory of this source tree).
|
8
|
+
* You may select, at your option, one of the above-listed licenses.
|
9
|
+
*/
|
10
|
+
|
11
|
+
/*-*************************************
|
12
|
+
* Dependencies
|
13
|
+
***************************************/
|
14
|
+
#include "zstd_compress_literals.h"
|
15
|
+
|
16
|
+
size_t ZSTD_noCompressLiterals (void* dst, size_t dstCapacity, const void* src, size_t srcSize)
|
17
|
+
{
|
18
|
+
BYTE* const ostart = (BYTE* const)dst;
|
19
|
+
U32 const flSize = 1 + (srcSize>31) + (srcSize>4095);
|
20
|
+
|
21
|
+
RETURN_ERROR_IF(srcSize + flSize > dstCapacity, dstSize_tooSmall);
|
22
|
+
|
23
|
+
switch(flSize)
|
24
|
+
{
|
25
|
+
case 1: /* 2 - 1 - 5 */
|
26
|
+
ostart[0] = (BYTE)((U32)set_basic + (srcSize<<3));
|
27
|
+
break;
|
28
|
+
case 2: /* 2 - 2 - 12 */
|
29
|
+
MEM_writeLE16(ostart, (U16)((U32)set_basic + (1<<2) + (srcSize<<4)));
|
30
|
+
break;
|
31
|
+
case 3: /* 2 - 2 - 20 */
|
32
|
+
MEM_writeLE32(ostart, (U32)((U32)set_basic + (3<<2) + (srcSize<<4)));
|
33
|
+
break;
|
34
|
+
default: /* not necessary : flSize is {1,2,3} */
|
35
|
+
assert(0);
|
36
|
+
}
|
37
|
+
|
38
|
+
memcpy(ostart + flSize, src, srcSize);
|
39
|
+
return srcSize + flSize;
|
40
|
+
}
|
41
|
+
|
42
|
+
size_t ZSTD_compressRleLiteralsBlock (void* dst, size_t dstCapacity, const void* src, size_t srcSize)
|
43
|
+
{
|
44
|
+
BYTE* const ostart = (BYTE* const)dst;
|
45
|
+
U32 const flSize = 1 + (srcSize>31) + (srcSize>4095);
|
46
|
+
|
47
|
+
(void)dstCapacity; /* dstCapacity already guaranteed to be >=4, hence large enough */
|
48
|
+
|
49
|
+
switch(flSize)
|
50
|
+
{
|
51
|
+
case 1: /* 2 - 1 - 5 */
|
52
|
+
ostart[0] = (BYTE)((U32)set_rle + (srcSize<<3));
|
53
|
+
break;
|
54
|
+
case 2: /* 2 - 2 - 12 */
|
55
|
+
MEM_writeLE16(ostart, (U16)((U32)set_rle + (1<<2) + (srcSize<<4)));
|
56
|
+
break;
|
57
|
+
case 3: /* 2 - 2 - 20 */
|
58
|
+
MEM_writeLE32(ostart, (U32)((U32)set_rle + (3<<2) + (srcSize<<4)));
|
59
|
+
break;
|
60
|
+
default: /* not necessary : flSize is {1,2,3} */
|
61
|
+
assert(0);
|
62
|
+
}
|
63
|
+
|
64
|
+
ostart[flSize] = *(const BYTE*)src;
|
65
|
+
return flSize+1;
|
66
|
+
}
|
67
|
+
|
68
|
+
size_t ZSTD_compressLiterals (ZSTD_hufCTables_t const* prevHuf,
|
69
|
+
ZSTD_hufCTables_t* nextHuf,
|
70
|
+
ZSTD_strategy strategy, int disableLiteralCompression,
|
71
|
+
void* dst, size_t dstCapacity,
|
72
|
+
const void* src, size_t srcSize,
|
73
|
+
void* workspace, size_t wkspSize,
|
74
|
+
const int bmi2)
|
75
|
+
{
|
76
|
+
size_t const minGain = ZSTD_minGain(srcSize, strategy);
|
77
|
+
size_t const lhSize = 3 + (srcSize >= 1 KB) + (srcSize >= 16 KB);
|
78
|
+
BYTE* const ostart = (BYTE*)dst;
|
79
|
+
U32 singleStream = srcSize < 256;
|
80
|
+
symbolEncodingType_e hType = set_compressed;
|
81
|
+
size_t cLitSize;
|
82
|
+
|
83
|
+
DEBUGLOG(5,"ZSTD_compressLiterals (disableLiteralCompression=%i)",
|
84
|
+
disableLiteralCompression);
|
85
|
+
|
86
|
+
/* Prepare nextEntropy assuming reusing the existing table */
|
87
|
+
memcpy(nextHuf, prevHuf, sizeof(*prevHuf));
|
88
|
+
|
89
|
+
if (disableLiteralCompression)
|
90
|
+
return ZSTD_noCompressLiterals(dst, dstCapacity, src, srcSize);
|
91
|
+
|
92
|
+
/* small ? don't even attempt compression (speed opt) */
|
93
|
+
# define COMPRESS_LITERALS_SIZE_MIN 63
|
94
|
+
{ size_t const minLitSize = (prevHuf->repeatMode == HUF_repeat_valid) ? 6 : COMPRESS_LITERALS_SIZE_MIN;
|
95
|
+
if (srcSize <= minLitSize) return ZSTD_noCompressLiterals(dst, dstCapacity, src, srcSize);
|
96
|
+
}
|
97
|
+
|
98
|
+
RETURN_ERROR_IF(dstCapacity < lhSize+1, dstSize_tooSmall, "not enough space for compression");
|
99
|
+
{ HUF_repeat repeat = prevHuf->repeatMode;
|
100
|
+
int const preferRepeat = strategy < ZSTD_lazy ? srcSize <= 1024 : 0;
|
101
|
+
if (repeat == HUF_repeat_valid && lhSize == 3) singleStream = 1;
|
102
|
+
cLitSize = singleStream ? HUF_compress1X_repeat(ostart+lhSize, dstCapacity-lhSize, src, srcSize, 255, 11,
|
103
|
+
workspace, wkspSize, (HUF_CElt*)nextHuf->CTable, &repeat, preferRepeat, bmi2)
|
104
|
+
: HUF_compress4X_repeat(ostart+lhSize, dstCapacity-lhSize, src, srcSize, 255, 11,
|
105
|
+
workspace, wkspSize, (HUF_CElt*)nextHuf->CTable, &repeat, preferRepeat, bmi2);
|
106
|
+
if (repeat != HUF_repeat_none) {
|
107
|
+
/* reused the existing table */
|
108
|
+
hType = set_repeat;
|
109
|
+
}
|
110
|
+
}
|
111
|
+
|
112
|
+
if ((cLitSize==0) | (cLitSize >= srcSize - minGain) | ERR_isError(cLitSize)) {
|
113
|
+
memcpy(nextHuf, prevHuf, sizeof(*prevHuf));
|
114
|
+
return ZSTD_noCompressLiterals(dst, dstCapacity, src, srcSize);
|
115
|
+
}
|
116
|
+
if (cLitSize==1) {
|
117
|
+
memcpy(nextHuf, prevHuf, sizeof(*prevHuf));
|
118
|
+
return ZSTD_compressRleLiteralsBlock(dst, dstCapacity, src, srcSize);
|
119
|
+
}
|
120
|
+
|
121
|
+
if (hType == set_compressed) {
|
122
|
+
/* using a newly constructed table */
|
123
|
+
nextHuf->repeatMode = HUF_repeat_check;
|
124
|
+
}
|
125
|
+
|
126
|
+
/* Build header */
|
127
|
+
switch(lhSize)
|
128
|
+
{
|
129
|
+
case 3: /* 2 - 2 - 10 - 10 */
|
130
|
+
{ U32 const lhc = hType + ((!singleStream) << 2) + ((U32)srcSize<<4) + ((U32)cLitSize<<14);
|
131
|
+
MEM_writeLE24(ostart, lhc);
|
132
|
+
break;
|
133
|
+
}
|
134
|
+
case 4: /* 2 - 2 - 14 - 14 */
|
135
|
+
{ U32 const lhc = hType + (2 << 2) + ((U32)srcSize<<4) + ((U32)cLitSize<<18);
|
136
|
+
MEM_writeLE32(ostart, lhc);
|
137
|
+
break;
|
138
|
+
}
|
139
|
+
case 5: /* 2 - 2 - 18 - 18 */
|
140
|
+
{ U32 const lhc = hType + (3 << 2) + ((U32)srcSize<<4) + ((U32)cLitSize<<22);
|
141
|
+
MEM_writeLE32(ostart, lhc);
|
142
|
+
ostart[4] = (BYTE)(cLitSize >> 10);
|
143
|
+
break;
|
144
|
+
}
|
145
|
+
default: /* not possible : lhSize is {3,4,5} */
|
146
|
+
assert(0);
|
147
|
+
}
|
148
|
+
return lhSize+cLitSize;
|
149
|
+
}
|
@@ -0,0 +1,29 @@
|
|
1
|
+
/*
|
2
|
+
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
3
|
+
* All rights reserved.
|
4
|
+
*
|
5
|
+
* This source code is licensed under both the BSD-style license (found in the
|
6
|
+
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
7
|
+
* in the COPYING file in the root directory of this source tree).
|
8
|
+
* You may select, at your option, one of the above-listed licenses.
|
9
|
+
*/
|
10
|
+
|
11
|
+
#ifndef ZSTD_COMPRESS_LITERALS_H
|
12
|
+
#define ZSTD_COMPRESS_LITERALS_H
|
13
|
+
|
14
|
+
#include "zstd_compress_internal.h" /* ZSTD_hufCTables_t, ZSTD_minGain() */
|
15
|
+
|
16
|
+
|
17
|
+
size_t ZSTD_noCompressLiterals (void* dst, size_t dstCapacity, const void* src, size_t srcSize);
|
18
|
+
|
19
|
+
size_t ZSTD_compressRleLiteralsBlock (void* dst, size_t dstCapacity, const void* src, size_t srcSize);
|
20
|
+
|
21
|
+
size_t ZSTD_compressLiterals (ZSTD_hufCTables_t const* prevHuf,
|
22
|
+
ZSTD_hufCTables_t* nextHuf,
|
23
|
+
ZSTD_strategy strategy, int disableLiteralCompression,
|
24
|
+
void* dst, size_t dstCapacity,
|
25
|
+
const void* src, size_t srcSize,
|
26
|
+
void* workspace, size_t wkspSize,
|
27
|
+
const int bmi2);
|
28
|
+
|
29
|
+
#endif /* ZSTD_COMPRESS_LITERALS_H */
|
@@ -0,0 +1,415 @@
|
|
1
|
+
/*
|
2
|
+
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
3
|
+
* All rights reserved.
|
4
|
+
*
|
5
|
+
* This source code is licensed under both the BSD-style license (found in the
|
6
|
+
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
7
|
+
* in the COPYING file in the root directory of this source tree).
|
8
|
+
* You may select, at your option, one of the above-listed licenses.
|
9
|
+
*/
|
10
|
+
|
11
|
+
/*-*************************************
|
12
|
+
* Dependencies
|
13
|
+
***************************************/
|
14
|
+
#include "zstd_compress_sequences.h"
|
15
|
+
|
16
|
+
/**
|
17
|
+
* -log2(x / 256) lookup table for x in [0, 256).
|
18
|
+
* If x == 0: Return 0
|
19
|
+
* Else: Return floor(-log2(x / 256) * 256)
|
20
|
+
*/
|
21
|
+
static unsigned const kInverseProbabilityLog256[256] = {
|
22
|
+
0, 2048, 1792, 1642, 1536, 1453, 1386, 1329, 1280, 1236, 1197, 1162,
|
23
|
+
1130, 1100, 1073, 1047, 1024, 1001, 980, 960, 941, 923, 906, 889,
|
24
|
+
874, 859, 844, 830, 817, 804, 791, 779, 768, 756, 745, 734,
|
25
|
+
724, 714, 704, 694, 685, 676, 667, 658, 650, 642, 633, 626,
|
26
|
+
618, 610, 603, 595, 588, 581, 574, 567, 561, 554, 548, 542,
|
27
|
+
535, 529, 523, 517, 512, 506, 500, 495, 489, 484, 478, 473,
|
28
|
+
468, 463, 458, 453, 448, 443, 438, 434, 429, 424, 420, 415,
|
29
|
+
411, 407, 402, 398, 394, 390, 386, 382, 377, 373, 370, 366,
|
30
|
+
362, 358, 354, 350, 347, 343, 339, 336, 332, 329, 325, 322,
|
31
|
+
318, 315, 311, 308, 305, 302, 298, 295, 292, 289, 286, 282,
|
32
|
+
279, 276, 273, 270, 267, 264, 261, 258, 256, 253, 250, 247,
|
33
|
+
244, 241, 239, 236, 233, 230, 228, 225, 222, 220, 217, 215,
|
34
|
+
212, 209, 207, 204, 202, 199, 197, 194, 192, 190, 187, 185,
|
35
|
+
182, 180, 178, 175, 173, 171, 168, 166, 164, 162, 159, 157,
|
36
|
+
155, 153, 151, 149, 146, 144, 142, 140, 138, 136, 134, 132,
|
37
|
+
130, 128, 126, 123, 121, 119, 117, 115, 114, 112, 110, 108,
|
38
|
+
106, 104, 102, 100, 98, 96, 94, 93, 91, 89, 87, 85,
|
39
|
+
83, 82, 80, 78, 76, 74, 73, 71, 69, 67, 66, 64,
|
40
|
+
62, 61, 59, 57, 55, 54, 52, 50, 49, 47, 46, 44,
|
41
|
+
42, 41, 39, 37, 36, 34, 33, 31, 30, 28, 26, 25,
|
42
|
+
23, 22, 20, 19, 17, 16, 14, 13, 11, 10, 8, 7,
|
43
|
+
5, 4, 2, 1,
|
44
|
+
};
|
45
|
+
|
46
|
+
static unsigned ZSTD_getFSEMaxSymbolValue(FSE_CTable const* ctable) {
|
47
|
+
void const* ptr = ctable;
|
48
|
+
U16 const* u16ptr = (U16 const*)ptr;
|
49
|
+
U32 const maxSymbolValue = MEM_read16(u16ptr + 1);
|
50
|
+
return maxSymbolValue;
|
51
|
+
}
|
52
|
+
|
53
|
+
/**
|
54
|
+
* Returns the cost in bytes of encoding the normalized count header.
|
55
|
+
* Returns an error if any of the helper functions return an error.
|
56
|
+
*/
|
57
|
+
static size_t ZSTD_NCountCost(unsigned const* count, unsigned const max,
|
58
|
+
size_t const nbSeq, unsigned const FSELog)
|
59
|
+
{
|
60
|
+
BYTE wksp[FSE_NCOUNTBOUND];
|
61
|
+
S16 norm[MaxSeq + 1];
|
62
|
+
const U32 tableLog = FSE_optimalTableLog(FSELog, nbSeq, max);
|
63
|
+
FORWARD_IF_ERROR(FSE_normalizeCount(norm, tableLog, count, nbSeq, max));
|
64
|
+
return FSE_writeNCount(wksp, sizeof(wksp), norm, max, tableLog);
|
65
|
+
}
|
66
|
+
|
67
|
+
/**
|
68
|
+
* Returns the cost in bits of encoding the distribution described by count
|
69
|
+
* using the entropy bound.
|
70
|
+
*/
|
71
|
+
static size_t ZSTD_entropyCost(unsigned const* count, unsigned const max, size_t const total)
|
72
|
+
{
|
73
|
+
unsigned cost = 0;
|
74
|
+
unsigned s;
|
75
|
+
for (s = 0; s <= max; ++s) {
|
76
|
+
unsigned norm = (unsigned)((256 * count[s]) / total);
|
77
|
+
if (count[s] != 0 && norm == 0)
|
78
|
+
norm = 1;
|
79
|
+
assert(count[s] < total);
|
80
|
+
cost += count[s] * kInverseProbabilityLog256[norm];
|
81
|
+
}
|
82
|
+
return cost >> 8;
|
83
|
+
}
|
84
|
+
|
85
|
+
/**
|
86
|
+
* Returns the cost in bits of encoding the distribution in count using ctable.
|
87
|
+
* Returns an error if ctable cannot represent all the symbols in count.
|
88
|
+
*/
|
89
|
+
static size_t ZSTD_fseBitCost(
|
90
|
+
FSE_CTable const* ctable,
|
91
|
+
unsigned const* count,
|
92
|
+
unsigned const max)
|
93
|
+
{
|
94
|
+
unsigned const kAccuracyLog = 8;
|
95
|
+
size_t cost = 0;
|
96
|
+
unsigned s;
|
97
|
+
FSE_CState_t cstate;
|
98
|
+
FSE_initCState(&cstate, ctable);
|
99
|
+
RETURN_ERROR_IF(ZSTD_getFSEMaxSymbolValue(ctable) < max, GENERIC,
|
100
|
+
"Repeat FSE_CTable has maxSymbolValue %u < %u",
|
101
|
+
ZSTD_getFSEMaxSymbolValue(ctable), max);
|
102
|
+
for (s = 0; s <= max; ++s) {
|
103
|
+
unsigned const tableLog = cstate.stateLog;
|
104
|
+
unsigned const badCost = (tableLog + 1) << kAccuracyLog;
|
105
|
+
unsigned const bitCost = FSE_bitCost(cstate.symbolTT, tableLog, s, kAccuracyLog);
|
106
|
+
if (count[s] == 0)
|
107
|
+
continue;
|
108
|
+
RETURN_ERROR_IF(bitCost >= badCost, GENERIC,
|
109
|
+
"Repeat FSE_CTable has Prob[%u] == 0", s);
|
110
|
+
cost += count[s] * bitCost;
|
111
|
+
}
|
112
|
+
return cost >> kAccuracyLog;
|
113
|
+
}
|
114
|
+
|
115
|
+
/**
|
116
|
+
* Returns the cost in bits of encoding the distribution in count using the
|
117
|
+
* table described by norm. The max symbol support by norm is assumed >= max.
|
118
|
+
* norm must be valid for every symbol with non-zero probability in count.
|
119
|
+
*/
|
120
|
+
static size_t ZSTD_crossEntropyCost(short const* norm, unsigned accuracyLog,
|
121
|
+
unsigned const* count, unsigned const max)
|
122
|
+
{
|
123
|
+
unsigned const shift = 8 - accuracyLog;
|
124
|
+
size_t cost = 0;
|
125
|
+
unsigned s;
|
126
|
+
assert(accuracyLog <= 8);
|
127
|
+
for (s = 0; s <= max; ++s) {
|
128
|
+
unsigned const normAcc = norm[s] != -1 ? norm[s] : 1;
|
129
|
+
unsigned const norm256 = normAcc << shift;
|
130
|
+
assert(norm256 > 0);
|
131
|
+
assert(norm256 < 256);
|
132
|
+
cost += count[s] * kInverseProbabilityLog256[norm256];
|
133
|
+
}
|
134
|
+
return cost >> 8;
|
135
|
+
}
|
136
|
+
|
137
|
+
symbolEncodingType_e
|
138
|
+
ZSTD_selectEncodingType(
|
139
|
+
FSE_repeat* repeatMode, unsigned const* count, unsigned const max,
|
140
|
+
size_t const mostFrequent, size_t nbSeq, unsigned const FSELog,
|
141
|
+
FSE_CTable const* prevCTable,
|
142
|
+
short const* defaultNorm, U32 defaultNormLog,
|
143
|
+
ZSTD_defaultPolicy_e const isDefaultAllowed,
|
144
|
+
ZSTD_strategy const strategy)
|
145
|
+
{
|
146
|
+
ZSTD_STATIC_ASSERT(ZSTD_defaultDisallowed == 0 && ZSTD_defaultAllowed != 0);
|
147
|
+
if (mostFrequent == nbSeq) {
|
148
|
+
*repeatMode = FSE_repeat_none;
|
149
|
+
if (isDefaultAllowed && nbSeq <= 2) {
|
150
|
+
/* Prefer set_basic over set_rle when there are 2 or less symbols,
|
151
|
+
* since RLE uses 1 byte, but set_basic uses 5-6 bits per symbol.
|
152
|
+
* If basic encoding isn't possible, always choose RLE.
|
153
|
+
*/
|
154
|
+
DEBUGLOG(5, "Selected set_basic");
|
155
|
+
return set_basic;
|
156
|
+
}
|
157
|
+
DEBUGLOG(5, "Selected set_rle");
|
158
|
+
return set_rle;
|
159
|
+
}
|
160
|
+
if (strategy < ZSTD_lazy) {
|
161
|
+
if (isDefaultAllowed) {
|
162
|
+
size_t const staticFse_nbSeq_max = 1000;
|
163
|
+
size_t const mult = 10 - strategy;
|
164
|
+
size_t const baseLog = 3;
|
165
|
+
size_t const dynamicFse_nbSeq_min = (((size_t)1 << defaultNormLog) * mult) >> baseLog; /* 28-36 for offset, 56-72 for lengths */
|
166
|
+
assert(defaultNormLog >= 5 && defaultNormLog <= 6); /* xx_DEFAULTNORMLOG */
|
167
|
+
assert(mult <= 9 && mult >= 7);
|
168
|
+
if ( (*repeatMode == FSE_repeat_valid)
|
169
|
+
&& (nbSeq < staticFse_nbSeq_max) ) {
|
170
|
+
DEBUGLOG(5, "Selected set_repeat");
|
171
|
+
return set_repeat;
|
172
|
+
}
|
173
|
+
if ( (nbSeq < dynamicFse_nbSeq_min)
|
174
|
+
|| (mostFrequent < (nbSeq >> (defaultNormLog-1))) ) {
|
175
|
+
DEBUGLOG(5, "Selected set_basic");
|
176
|
+
/* The format allows default tables to be repeated, but it isn't useful.
|
177
|
+
* When using simple heuristics to select encoding type, we don't want
|
178
|
+
* to confuse these tables with dictionaries. When running more careful
|
179
|
+
* analysis, we don't need to waste time checking both repeating tables
|
180
|
+
* and default tables.
|
181
|
+
*/
|
182
|
+
*repeatMode = FSE_repeat_none;
|
183
|
+
return set_basic;
|
184
|
+
}
|
185
|
+
}
|
186
|
+
} else {
|
187
|
+
size_t const basicCost = isDefaultAllowed ? ZSTD_crossEntropyCost(defaultNorm, defaultNormLog, count, max) : ERROR(GENERIC);
|
188
|
+
size_t const repeatCost = *repeatMode != FSE_repeat_none ? ZSTD_fseBitCost(prevCTable, count, max) : ERROR(GENERIC);
|
189
|
+
size_t const NCountCost = ZSTD_NCountCost(count, max, nbSeq, FSELog);
|
190
|
+
size_t const compressedCost = (NCountCost << 3) + ZSTD_entropyCost(count, max, nbSeq);
|
191
|
+
|
192
|
+
if (isDefaultAllowed) {
|
193
|
+
assert(!ZSTD_isError(basicCost));
|
194
|
+
assert(!(*repeatMode == FSE_repeat_valid && ZSTD_isError(repeatCost)));
|
195
|
+
}
|
196
|
+
assert(!ZSTD_isError(NCountCost));
|
197
|
+
assert(compressedCost < ERROR(maxCode));
|
198
|
+
DEBUGLOG(5, "Estimated bit costs: basic=%u\trepeat=%u\tcompressed=%u",
|
199
|
+
(unsigned)basicCost, (unsigned)repeatCost, (unsigned)compressedCost);
|
200
|
+
if (basicCost <= repeatCost && basicCost <= compressedCost) {
|
201
|
+
DEBUGLOG(5, "Selected set_basic");
|
202
|
+
assert(isDefaultAllowed);
|
203
|
+
*repeatMode = FSE_repeat_none;
|
204
|
+
return set_basic;
|
205
|
+
}
|
206
|
+
if (repeatCost <= compressedCost) {
|
207
|
+
DEBUGLOG(5, "Selected set_repeat");
|
208
|
+
assert(!ZSTD_isError(repeatCost));
|
209
|
+
return set_repeat;
|
210
|
+
}
|
211
|
+
assert(compressedCost < basicCost && compressedCost < repeatCost);
|
212
|
+
}
|
213
|
+
DEBUGLOG(5, "Selected set_compressed");
|
214
|
+
*repeatMode = FSE_repeat_check;
|
215
|
+
return set_compressed;
|
216
|
+
}
|
217
|
+
|
218
|
+
size_t
|
219
|
+
ZSTD_buildCTable(void* dst, size_t dstCapacity,
|
220
|
+
FSE_CTable* nextCTable, U32 FSELog, symbolEncodingType_e type,
|
221
|
+
unsigned* count, U32 max,
|
222
|
+
const BYTE* codeTable, size_t nbSeq,
|
223
|
+
const S16* defaultNorm, U32 defaultNormLog, U32 defaultMax,
|
224
|
+
const FSE_CTable* prevCTable, size_t prevCTableSize,
|
225
|
+
void* workspace, size_t workspaceSize)
|
226
|
+
{
|
227
|
+
BYTE* op = (BYTE*)dst;
|
228
|
+
const BYTE* const oend = op + dstCapacity;
|
229
|
+
DEBUGLOG(6, "ZSTD_buildCTable (dstCapacity=%u)", (unsigned)dstCapacity);
|
230
|
+
|
231
|
+
switch (type) {
|
232
|
+
case set_rle:
|
233
|
+
FORWARD_IF_ERROR(FSE_buildCTable_rle(nextCTable, (BYTE)max));
|
234
|
+
RETURN_ERROR_IF(dstCapacity==0, dstSize_tooSmall);
|
235
|
+
*op = codeTable[0];
|
236
|
+
return 1;
|
237
|
+
case set_repeat:
|
238
|
+
memcpy(nextCTable, prevCTable, prevCTableSize);
|
239
|
+
return 0;
|
240
|
+
case set_basic:
|
241
|
+
FORWARD_IF_ERROR(FSE_buildCTable_wksp(nextCTable, defaultNorm, defaultMax, defaultNormLog, workspace, workspaceSize)); /* note : could be pre-calculated */
|
242
|
+
return 0;
|
243
|
+
case set_compressed: {
|
244
|
+
S16 norm[MaxSeq + 1];
|
245
|
+
size_t nbSeq_1 = nbSeq;
|
246
|
+
const U32 tableLog = FSE_optimalTableLog(FSELog, nbSeq, max);
|
247
|
+
if (count[codeTable[nbSeq-1]] > 1) {
|
248
|
+
count[codeTable[nbSeq-1]]--;
|
249
|
+
nbSeq_1--;
|
250
|
+
}
|
251
|
+
assert(nbSeq_1 > 1);
|
252
|
+
FORWARD_IF_ERROR(FSE_normalizeCount(norm, tableLog, count, nbSeq_1, max));
|
253
|
+
{ size_t const NCountSize = FSE_writeNCount(op, oend - op, norm, max, tableLog); /* overflow protected */
|
254
|
+
FORWARD_IF_ERROR(NCountSize);
|
255
|
+
FORWARD_IF_ERROR(FSE_buildCTable_wksp(nextCTable, norm, max, tableLog, workspace, workspaceSize));
|
256
|
+
return NCountSize;
|
257
|
+
}
|
258
|
+
}
|
259
|
+
default: assert(0); RETURN_ERROR(GENERIC);
|
260
|
+
}
|
261
|
+
}
|
262
|
+
|
263
|
+
FORCE_INLINE_TEMPLATE size_t
|
264
|
+
ZSTD_encodeSequences_body(
|
265
|
+
void* dst, size_t dstCapacity,
|
266
|
+
FSE_CTable const* CTable_MatchLength, BYTE const* mlCodeTable,
|
267
|
+
FSE_CTable const* CTable_OffsetBits, BYTE const* ofCodeTable,
|
268
|
+
FSE_CTable const* CTable_LitLength, BYTE const* llCodeTable,
|
269
|
+
seqDef const* sequences, size_t nbSeq, int longOffsets)
|
270
|
+
{
|
271
|
+
BIT_CStream_t blockStream;
|
272
|
+
FSE_CState_t stateMatchLength;
|
273
|
+
FSE_CState_t stateOffsetBits;
|
274
|
+
FSE_CState_t stateLitLength;
|
275
|
+
|
276
|
+
RETURN_ERROR_IF(
|
277
|
+
ERR_isError(BIT_initCStream(&blockStream, dst, dstCapacity)),
|
278
|
+
dstSize_tooSmall, "not enough space remaining");
|
279
|
+
DEBUGLOG(6, "available space for bitstream : %i (dstCapacity=%u)",
|
280
|
+
(int)(blockStream.endPtr - blockStream.startPtr),
|
281
|
+
(unsigned)dstCapacity);
|
282
|
+
|
283
|
+
/* first symbols */
|
284
|
+
FSE_initCState2(&stateMatchLength, CTable_MatchLength, mlCodeTable[nbSeq-1]);
|
285
|
+
FSE_initCState2(&stateOffsetBits, CTable_OffsetBits, ofCodeTable[nbSeq-1]);
|
286
|
+
FSE_initCState2(&stateLitLength, CTable_LitLength, llCodeTable[nbSeq-1]);
|
287
|
+
BIT_addBits(&blockStream, sequences[nbSeq-1].litLength, LL_bits[llCodeTable[nbSeq-1]]);
|
288
|
+
if (MEM_32bits()) BIT_flushBits(&blockStream);
|
289
|
+
BIT_addBits(&blockStream, sequences[nbSeq-1].matchLength, ML_bits[mlCodeTable[nbSeq-1]]);
|
290
|
+
if (MEM_32bits()) BIT_flushBits(&blockStream);
|
291
|
+
if (longOffsets) {
|
292
|
+
U32 const ofBits = ofCodeTable[nbSeq-1];
|
293
|
+
int const extraBits = ofBits - MIN(ofBits, STREAM_ACCUMULATOR_MIN-1);
|
294
|
+
if (extraBits) {
|
295
|
+
BIT_addBits(&blockStream, sequences[nbSeq-1].offset, extraBits);
|
296
|
+
BIT_flushBits(&blockStream);
|
297
|
+
}
|
298
|
+
BIT_addBits(&blockStream, sequences[nbSeq-1].offset >> extraBits,
|
299
|
+
ofBits - extraBits);
|
300
|
+
} else {
|
301
|
+
BIT_addBits(&blockStream, sequences[nbSeq-1].offset, ofCodeTable[nbSeq-1]);
|
302
|
+
}
|
303
|
+
BIT_flushBits(&blockStream);
|
304
|
+
|
305
|
+
{ size_t n;
|
306
|
+
for (n=nbSeq-2 ; n<nbSeq ; n--) { /* intentional underflow */
|
307
|
+
BYTE const llCode = llCodeTable[n];
|
308
|
+
BYTE const ofCode = ofCodeTable[n];
|
309
|
+
BYTE const mlCode = mlCodeTable[n];
|
310
|
+
U32 const llBits = LL_bits[llCode];
|
311
|
+
U32 const ofBits = ofCode;
|
312
|
+
U32 const mlBits = ML_bits[mlCode];
|
313
|
+
DEBUGLOG(6, "encoding: litlen:%2u - matchlen:%2u - offCode:%7u",
|
314
|
+
(unsigned)sequences[n].litLength,
|
315
|
+
(unsigned)sequences[n].matchLength + MINMATCH,
|
316
|
+
(unsigned)sequences[n].offset);
|
317
|
+
/* 32b*/ /* 64b*/
|
318
|
+
/* (7)*/ /* (7)*/
|
319
|
+
FSE_encodeSymbol(&blockStream, &stateOffsetBits, ofCode); /* 15 */ /* 15 */
|
320
|
+
FSE_encodeSymbol(&blockStream, &stateMatchLength, mlCode); /* 24 */ /* 24 */
|
321
|
+
if (MEM_32bits()) BIT_flushBits(&blockStream); /* (7)*/
|
322
|
+
FSE_encodeSymbol(&blockStream, &stateLitLength, llCode); /* 16 */ /* 33 */
|
323
|
+
if (MEM_32bits() || (ofBits+mlBits+llBits >= 64-7-(LLFSELog+MLFSELog+OffFSELog)))
|
324
|
+
BIT_flushBits(&blockStream); /* (7)*/
|
325
|
+
BIT_addBits(&blockStream, sequences[n].litLength, llBits);
|
326
|
+
if (MEM_32bits() && ((llBits+mlBits)>24)) BIT_flushBits(&blockStream);
|
327
|
+
BIT_addBits(&blockStream, sequences[n].matchLength, mlBits);
|
328
|
+
if (MEM_32bits() || (ofBits+mlBits+llBits > 56)) BIT_flushBits(&blockStream);
|
329
|
+
if (longOffsets) {
|
330
|
+
int const extraBits = ofBits - MIN(ofBits, STREAM_ACCUMULATOR_MIN-1);
|
331
|
+
if (extraBits) {
|
332
|
+
BIT_addBits(&blockStream, sequences[n].offset, extraBits);
|
333
|
+
BIT_flushBits(&blockStream); /* (7)*/
|
334
|
+
}
|
335
|
+
BIT_addBits(&blockStream, sequences[n].offset >> extraBits,
|
336
|
+
ofBits - extraBits); /* 31 */
|
337
|
+
} else {
|
338
|
+
BIT_addBits(&blockStream, sequences[n].offset, ofBits); /* 31 */
|
339
|
+
}
|
340
|
+
BIT_flushBits(&blockStream); /* (7)*/
|
341
|
+
DEBUGLOG(7, "remaining space : %i", (int)(blockStream.endPtr - blockStream.ptr));
|
342
|
+
} }
|
343
|
+
|
344
|
+
DEBUGLOG(6, "ZSTD_encodeSequences: flushing ML state with %u bits", stateMatchLength.stateLog);
|
345
|
+
FSE_flushCState(&blockStream, &stateMatchLength);
|
346
|
+
DEBUGLOG(6, "ZSTD_encodeSequences: flushing Off state with %u bits", stateOffsetBits.stateLog);
|
347
|
+
FSE_flushCState(&blockStream, &stateOffsetBits);
|
348
|
+
DEBUGLOG(6, "ZSTD_encodeSequences: flushing LL state with %u bits", stateLitLength.stateLog);
|
349
|
+
FSE_flushCState(&blockStream, &stateLitLength);
|
350
|
+
|
351
|
+
{ size_t const streamSize = BIT_closeCStream(&blockStream);
|
352
|
+
RETURN_ERROR_IF(streamSize==0, dstSize_tooSmall, "not enough space");
|
353
|
+
return streamSize;
|
354
|
+
}
|
355
|
+
}
|
356
|
+
|
357
|
+
static size_t
|
358
|
+
ZSTD_encodeSequences_default(
|
359
|
+
void* dst, size_t dstCapacity,
|
360
|
+
FSE_CTable const* CTable_MatchLength, BYTE const* mlCodeTable,
|
361
|
+
FSE_CTable const* CTable_OffsetBits, BYTE const* ofCodeTable,
|
362
|
+
FSE_CTable const* CTable_LitLength, BYTE const* llCodeTable,
|
363
|
+
seqDef const* sequences, size_t nbSeq, int longOffsets)
|
364
|
+
{
|
365
|
+
return ZSTD_encodeSequences_body(dst, dstCapacity,
|
366
|
+
CTable_MatchLength, mlCodeTable,
|
367
|
+
CTable_OffsetBits, ofCodeTable,
|
368
|
+
CTable_LitLength, llCodeTable,
|
369
|
+
sequences, nbSeq, longOffsets);
|
370
|
+
}
|
371
|
+
|
372
|
+
|
373
|
+
#if DYNAMIC_BMI2
|
374
|
+
|
375
|
+
static TARGET_ATTRIBUTE("bmi2") size_t
|
376
|
+
ZSTD_encodeSequences_bmi2(
|
377
|
+
void* dst, size_t dstCapacity,
|
378
|
+
FSE_CTable const* CTable_MatchLength, BYTE const* mlCodeTable,
|
379
|
+
FSE_CTable const* CTable_OffsetBits, BYTE const* ofCodeTable,
|
380
|
+
FSE_CTable const* CTable_LitLength, BYTE const* llCodeTable,
|
381
|
+
seqDef const* sequences, size_t nbSeq, int longOffsets)
|
382
|
+
{
|
383
|
+
return ZSTD_encodeSequences_body(dst, dstCapacity,
|
384
|
+
CTable_MatchLength, mlCodeTable,
|
385
|
+
CTable_OffsetBits, ofCodeTable,
|
386
|
+
CTable_LitLength, llCodeTable,
|
387
|
+
sequences, nbSeq, longOffsets);
|
388
|
+
}
|
389
|
+
|
390
|
+
#endif
|
391
|
+
|
392
|
+
size_t ZSTD_encodeSequences(
|
393
|
+
void* dst, size_t dstCapacity,
|
394
|
+
FSE_CTable const* CTable_MatchLength, BYTE const* mlCodeTable,
|
395
|
+
FSE_CTable const* CTable_OffsetBits, BYTE const* ofCodeTable,
|
396
|
+
FSE_CTable const* CTable_LitLength, BYTE const* llCodeTable,
|
397
|
+
seqDef const* sequences, size_t nbSeq, int longOffsets, int bmi2)
|
398
|
+
{
|
399
|
+
DEBUGLOG(5, "ZSTD_encodeSequences: dstCapacity = %u", (unsigned)dstCapacity);
|
400
|
+
#if DYNAMIC_BMI2
|
401
|
+
if (bmi2) {
|
402
|
+
return ZSTD_encodeSequences_bmi2(dst, dstCapacity,
|
403
|
+
CTable_MatchLength, mlCodeTable,
|
404
|
+
CTable_OffsetBits, ofCodeTable,
|
405
|
+
CTable_LitLength, llCodeTable,
|
406
|
+
sequences, nbSeq, longOffsets);
|
407
|
+
}
|
408
|
+
#endif
|
409
|
+
(void)bmi2;
|
410
|
+
return ZSTD_encodeSequences_default(dst, dstCapacity,
|
411
|
+
CTable_MatchLength, mlCodeTable,
|
412
|
+
CTable_OffsetBits, ofCodeTable,
|
413
|
+
CTable_LitLength, llCodeTable,
|
414
|
+
sequences, nbSeq, longOffsets);
|
415
|
+
}
|