zstd-ruby 0.1.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (71) hide show
  1. checksums.yaml +7 -0
  2. data/.gitignore +19 -0
  3. data/.gitmodules +3 -0
  4. data/.rspec +2 -0
  5. data/.travis.yml +11 -0
  6. data/CODE_OF_CONDUCT.md +74 -0
  7. data/Gemfile +4 -0
  8. data/LICENSE.txt +29 -0
  9. data/README.md +63 -0
  10. data/Rakefile +22 -0
  11. data/bin/console +14 -0
  12. data/bin/setup +8 -0
  13. data/ext/zstdruby/extconf.rb +23 -0
  14. data/ext/zstdruby/libzstd/.gitignore +2 -0
  15. data/ext/zstdruby/libzstd/Makefile +133 -0
  16. data/ext/zstdruby/libzstd/README.md +77 -0
  17. data/ext/zstdruby/libzstd/common/bitstream.h +414 -0
  18. data/ext/zstdruby/libzstd/common/entropy_common.c +227 -0
  19. data/ext/zstdruby/libzstd/common/error_private.c +43 -0
  20. data/ext/zstdruby/libzstd/common/error_private.h +76 -0
  21. data/ext/zstdruby/libzstd/common/fse.h +668 -0
  22. data/ext/zstdruby/libzstd/common/fse_decompress.c +329 -0
  23. data/ext/zstdruby/libzstd/common/huf.h +238 -0
  24. data/ext/zstdruby/libzstd/common/mem.h +372 -0
  25. data/ext/zstdruby/libzstd/common/xxhash.c +867 -0
  26. data/ext/zstdruby/libzstd/common/xxhash.h +309 -0
  27. data/ext/zstdruby/libzstd/common/zstd_common.c +77 -0
  28. data/ext/zstdruby/libzstd/common/zstd_errors.h +60 -0
  29. data/ext/zstdruby/libzstd/common/zstd_internal.h +270 -0
  30. data/ext/zstdruby/libzstd/compress/fse_compress.c +850 -0
  31. data/ext/zstdruby/libzstd/compress/huf_compress.c +609 -0
  32. data/ext/zstdruby/libzstd/compress/zstd_compress.c +3291 -0
  33. data/ext/zstdruby/libzstd/compress/zstd_opt.h +919 -0
  34. data/ext/zstdruby/libzstd/decompress/huf_decompress.c +885 -0
  35. data/ext/zstdruby/libzstd/decompress/zstd_decompress.c +2154 -0
  36. data/ext/zstdruby/libzstd/deprecated/zbuff.h +210 -0
  37. data/ext/zstdruby/libzstd/deprecated/zbuff_compress.c +145 -0
  38. data/ext/zstdruby/libzstd/deprecated/zbuff_decompress.c +74 -0
  39. data/ext/zstdruby/libzstd/dictBuilder/divsufsort.c +1913 -0
  40. data/ext/zstdruby/libzstd/dictBuilder/divsufsort.h +67 -0
  41. data/ext/zstdruby/libzstd/dictBuilder/zdict.c +1012 -0
  42. data/ext/zstdruby/libzstd/dictBuilder/zdict.h +111 -0
  43. data/ext/zstdruby/libzstd/dll/example/Makefile +47 -0
  44. data/ext/zstdruby/libzstd/dll/example/README.md +69 -0
  45. data/ext/zstdruby/libzstd/dll/example/build_package.bat +17 -0
  46. data/ext/zstdruby/libzstd/dll/example/fullbench-dll.sln +25 -0
  47. data/ext/zstdruby/libzstd/dll/example/fullbench-dll.vcxproj +179 -0
  48. data/ext/zstdruby/libzstd/dll/libzstd.def +86 -0
  49. data/ext/zstdruby/libzstd/legacy/zstd_legacy.h +259 -0
  50. data/ext/zstdruby/libzstd/legacy/zstd_v01.c +2095 -0
  51. data/ext/zstdruby/libzstd/legacy/zstd_v01.h +80 -0
  52. data/ext/zstdruby/libzstd/legacy/zstd_v02.c +3518 -0
  53. data/ext/zstdruby/libzstd/legacy/zstd_v02.h +79 -0
  54. data/ext/zstdruby/libzstd/legacy/zstd_v03.c +3159 -0
  55. data/ext/zstdruby/libzstd/legacy/zstd_v03.h +79 -0
  56. data/ext/zstdruby/libzstd/legacy/zstd_v04.c +3795 -0
  57. data/ext/zstdruby/libzstd/legacy/zstd_v04.h +128 -0
  58. data/ext/zstdruby/libzstd/legacy/zstd_v05.c +4056 -0
  59. data/ext/zstdruby/libzstd/legacy/zstd_v05.h +149 -0
  60. data/ext/zstdruby/libzstd/legacy/zstd_v06.c +4167 -0
  61. data/ext/zstdruby/libzstd/legacy/zstd_v06.h +159 -0
  62. data/ext/zstdruby/libzstd/legacy/zstd_v07.c +4540 -0
  63. data/ext/zstdruby/libzstd/legacy/zstd_v07.h +173 -0
  64. data/ext/zstdruby/libzstd/libzstd.pc.in +14 -0
  65. data/ext/zstdruby/libzstd/zstd.h +673 -0
  66. data/ext/zstdruby/zstdruby.c +117 -0
  67. data/ext/zstdruby/zstdruby.h +6 -0
  68. data/lib/zstd-ruby.rb +6 -0
  69. data/lib/zstd-ruby/version.rb +3 -0
  70. data/zstd-ruby.gemspec +37 -0
  71. metadata +170 -0
@@ -0,0 +1,86 @@
1
+ LIBRARY libzstd.dll
2
+ EXPORTS
3
+ ZDICT_getDictID
4
+ ZDICT_getErrorName
5
+ ZDICT_isError
6
+ ZDICT_trainFromBuffer
7
+ ZSTD_CStreamInSize
8
+ ZSTD_CStreamOutSize
9
+ ZSTD_DStreamInSize
10
+ ZSTD_DStreamOutSize
11
+ ZSTD_adjustCParams
12
+ ZSTD_checkCParams
13
+ ZSTD_compress
14
+ ZSTD_compressBegin
15
+ ZSTD_compressBegin_advanced
16
+ ZSTD_compressBegin_usingDict
17
+ ZSTD_compressBlock
18
+ ZSTD_compressBound
19
+ ZSTD_compressCCtx
20
+ ZSTD_compressContinue
21
+ ZSTD_compressEnd
22
+ ZSTD_compressStream
23
+ ZSTD_compress_advanced
24
+ ZSTD_compress_usingCDict
25
+ ZSTD_compress_usingDict
26
+ ZSTD_copyCCtx
27
+ ZSTD_copyDCtx
28
+ ZSTD_createCCtx
29
+ ZSTD_createCCtx_advanced
30
+ ZSTD_createCDict
31
+ ZSTD_createCDict_advanced
32
+ ZSTD_createCStream
33
+ ZSTD_createCStream_advanced
34
+ ZSTD_createDCtx
35
+ ZSTD_createDCtx_advanced
36
+ ZSTD_createDDict
37
+ ZSTD_createDStream
38
+ ZSTD_createDStream_advanced
39
+ ZSTD_decompress
40
+ ZSTD_decompressBegin
41
+ ZSTD_decompressBegin_usingDict
42
+ ZSTD_decompressBlock
43
+ ZSTD_decompressContinue
44
+ ZSTD_decompressDCtx
45
+ ZSTD_decompressStream
46
+ ZSTD_decompress_usingDDict
47
+ ZSTD_decompress_usingDict
48
+ ZSTD_endStream
49
+ ZSTD_estimateCCtxSize
50
+ ZSTD_estimateDCtxSize
51
+ ZSTD_flushStream
52
+ ZSTD_freeCCtx
53
+ ZSTD_freeCDict
54
+ ZSTD_freeCStream
55
+ ZSTD_freeDCtx
56
+ ZSTD_freeDDict
57
+ ZSTD_freeDStream
58
+ ZSTD_getBlockSizeMax
59
+ ZSTD_getCParams
60
+ ZSTD_getDecompressedSize
61
+ ZSTD_getErrorName
62
+ ZSTD_getFrameParams
63
+ ZSTD_getParams
64
+ ZSTD_initCStream
65
+ ZSTD_initCStream_advanced
66
+ ZSTD_initCStream_usingCDict
67
+ ZSTD_initCStream_usingDict
68
+ ZSTD_initDStream
69
+ ZSTD_initDStream_usingDDict
70
+ ZSTD_initDStream_usingDict
71
+ ZSTD_insertBlock
72
+ ZSTD_isError
73
+ ZSTD_isFrame
74
+ ZSTD_maxCLevel
75
+ ZSTD_nextInputType
76
+ ZSTD_nextSrcSizeToDecompress
77
+ ZSTD_resetCStream
78
+ ZSTD_resetDStream
79
+ ZSTD_setDStreamParameter
80
+ ZSTD_sizeof_CCtx
81
+ ZSTD_sizeof_CDict
82
+ ZSTD_sizeof_CStream
83
+ ZSTD_sizeof_DCtx
84
+ ZSTD_sizeof_DDict
85
+ ZSTD_sizeof_DStream
86
+ ZSTD_versionNumber
@@ -0,0 +1,259 @@
1
+ /**
2
+ * Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
3
+ * All rights reserved.
4
+ *
5
+ * This source code is licensed under the BSD-style license found in the
6
+ * LICENSE file in the root directory of this source tree. An additional grant
7
+ * of patent rights can be found in the PATENTS file in the same directory.
8
+ */
9
+
10
+ #ifndef ZSTD_LEGACY_H
11
+ #define ZSTD_LEGACY_H
12
+
13
+ #if defined (__cplusplus)
14
+ extern "C" {
15
+ #endif
16
+
17
+ /* *************************************
18
+ * Includes
19
+ ***************************************/
20
+ #include "mem.h" /* MEM_STATIC */
21
+ #include "error_private.h" /* ERROR */
22
+ #include "zstd.h" /* ZSTD_inBuffer, ZSTD_outBuffer */
23
+ #include "zstd_v01.h"
24
+ #include "zstd_v02.h"
25
+ #include "zstd_v03.h"
26
+ #include "zstd_v04.h"
27
+ #include "zstd_v05.h"
28
+ #include "zstd_v06.h"
29
+ #include "zstd_v07.h"
30
+
31
+
32
+ /** ZSTD_isLegacy() :
33
+ @return : > 0 if supported by legacy decoder. 0 otherwise.
34
+ return value is the version.
35
+ */
36
+ MEM_STATIC unsigned ZSTD_isLegacy(const void* src, size_t srcSize)
37
+ {
38
+ U32 magicNumberLE;
39
+ if (srcSize<4) return 0;
40
+ magicNumberLE = MEM_readLE32(src);
41
+ switch(magicNumberLE)
42
+ {
43
+ case ZSTDv01_magicNumberLE:return 1;
44
+ case ZSTDv02_magicNumber : return 2;
45
+ case ZSTDv03_magicNumber : return 3;
46
+ case ZSTDv04_magicNumber : return 4;
47
+ case ZSTDv05_MAGICNUMBER : return 5;
48
+ case ZSTDv06_MAGICNUMBER : return 6;
49
+ case ZSTDv07_MAGICNUMBER : return 7;
50
+ default : return 0;
51
+ }
52
+ }
53
+
54
+
55
+ MEM_STATIC unsigned long long ZSTD_getDecompressedSize_legacy(const void* src, size_t srcSize)
56
+ {
57
+ U32 const version = ZSTD_isLegacy(src, srcSize);
58
+ if (version < 5) return 0; /* no decompressed size in frame header, or not a legacy format */
59
+ if (version==5) {
60
+ ZSTDv05_parameters fParams;
61
+ size_t const frResult = ZSTDv05_getFrameParams(&fParams, src, srcSize);
62
+ if (frResult != 0) return 0;
63
+ return fParams.srcSize;
64
+ }
65
+ if (version==6) {
66
+ ZSTDv06_frameParams fParams;
67
+ size_t const frResult = ZSTDv06_getFrameParams(&fParams, src, srcSize);
68
+ if (frResult != 0) return 0;
69
+ return fParams.frameContentSize;
70
+ }
71
+ if (version==7) {
72
+ ZSTDv07_frameParams fParams;
73
+ size_t const frResult = ZSTDv07_getFrameParams(&fParams, src, srcSize);
74
+ if (frResult != 0) return 0;
75
+ return fParams.frameContentSize;
76
+ }
77
+ return 0; /* should not be possible */
78
+ }
79
+
80
+
81
+ MEM_STATIC size_t ZSTD_decompressLegacy(
82
+ void* dst, size_t dstCapacity,
83
+ const void* src, size_t compressedSize,
84
+ const void* dict,size_t dictSize)
85
+ {
86
+ U32 const version = ZSTD_isLegacy(src, compressedSize);
87
+ switch(version)
88
+ {
89
+ case 1 :
90
+ return ZSTDv01_decompress(dst, dstCapacity, src, compressedSize);
91
+ case 2 :
92
+ return ZSTDv02_decompress(dst, dstCapacity, src, compressedSize);
93
+ case 3 :
94
+ return ZSTDv03_decompress(dst, dstCapacity, src, compressedSize);
95
+ case 4 :
96
+ return ZSTDv04_decompress(dst, dstCapacity, src, compressedSize);
97
+ case 5 :
98
+ { size_t result;
99
+ ZSTDv05_DCtx* const zd = ZSTDv05_createDCtx();
100
+ if (zd==NULL) return ERROR(memory_allocation);
101
+ result = ZSTDv05_decompress_usingDict(zd, dst, dstCapacity, src, compressedSize, dict, dictSize);
102
+ ZSTDv05_freeDCtx(zd);
103
+ return result;
104
+ }
105
+ case 6 :
106
+ { size_t result;
107
+ ZSTDv06_DCtx* const zd = ZSTDv06_createDCtx();
108
+ if (zd==NULL) return ERROR(memory_allocation);
109
+ result = ZSTDv06_decompress_usingDict(zd, dst, dstCapacity, src, compressedSize, dict, dictSize);
110
+ ZSTDv06_freeDCtx(zd);
111
+ return result;
112
+ }
113
+ case 7 :
114
+ { size_t result;
115
+ ZSTDv07_DCtx* const zd = ZSTDv07_createDCtx();
116
+ if (zd==NULL) return ERROR(memory_allocation);
117
+ result = ZSTDv07_decompress_usingDict(zd, dst, dstCapacity, src, compressedSize, dict, dictSize);
118
+ ZSTDv07_freeDCtx(zd);
119
+ return result;
120
+ }
121
+ default :
122
+ return ERROR(prefix_unknown);
123
+ }
124
+ }
125
+
126
+
127
+ MEM_STATIC size_t ZSTD_freeLegacyStreamContext(void* legacyContext, U32 version)
128
+ {
129
+ switch(version)
130
+ {
131
+ default :
132
+ case 1 :
133
+ case 2 :
134
+ case 3 :
135
+ return ERROR(version_unsupported);
136
+ case 4 : return ZBUFFv04_freeDCtx((ZBUFFv04_DCtx*)legacyContext);
137
+ case 5 : return ZBUFFv05_freeDCtx((ZBUFFv05_DCtx*)legacyContext);
138
+ case 6 : return ZBUFFv06_freeDCtx((ZBUFFv06_DCtx*)legacyContext);
139
+ case 7 : return ZBUFFv07_freeDCtx((ZBUFFv07_DCtx*)legacyContext);
140
+ }
141
+ }
142
+
143
+
144
+ MEM_STATIC size_t ZSTD_initLegacyStream(void** legacyContext, U32 prevVersion, U32 newVersion,
145
+ const void* dict, size_t dictSize)
146
+ {
147
+ if (prevVersion != newVersion) ZSTD_freeLegacyStreamContext(*legacyContext, prevVersion);
148
+ switch(newVersion)
149
+ {
150
+ default :
151
+ case 1 :
152
+ case 2 :
153
+ case 3 :
154
+ return 0;
155
+ case 4 :
156
+ {
157
+ ZBUFFv04_DCtx* dctx = (prevVersion != newVersion) ? ZBUFFv04_createDCtx() : (ZBUFFv04_DCtx*)*legacyContext;
158
+ if (dctx==NULL) return ERROR(memory_allocation);
159
+ ZBUFFv04_decompressInit(dctx);
160
+ ZBUFFv04_decompressWithDictionary(dctx, dict, dictSize);
161
+ *legacyContext = dctx;
162
+ return 0;
163
+ }
164
+ case 5 :
165
+ {
166
+ ZBUFFv05_DCtx* dctx = (prevVersion != newVersion) ? ZBUFFv05_createDCtx() : (ZBUFFv05_DCtx*)*legacyContext;
167
+ if (dctx==NULL) return ERROR(memory_allocation);
168
+ ZBUFFv05_decompressInitDictionary(dctx, dict, dictSize);
169
+ *legacyContext = dctx;
170
+ return 0;
171
+ }
172
+ case 6 :
173
+ {
174
+ ZBUFFv06_DCtx* dctx = (prevVersion != newVersion) ? ZBUFFv06_createDCtx() : (ZBUFFv06_DCtx*)*legacyContext;
175
+ if (dctx==NULL) return ERROR(memory_allocation);
176
+ ZBUFFv06_decompressInitDictionary(dctx, dict, dictSize);
177
+ *legacyContext = dctx;
178
+ return 0;
179
+ }
180
+ case 7 :
181
+ {
182
+ ZBUFFv07_DCtx* dctx = (prevVersion != newVersion) ? ZBUFFv07_createDCtx() : (ZBUFFv07_DCtx*)*legacyContext;
183
+ if (dctx==NULL) return ERROR(memory_allocation);
184
+ ZBUFFv07_decompressInitDictionary(dctx, dict, dictSize);
185
+ *legacyContext = dctx;
186
+ return 0;
187
+ }
188
+ }
189
+ }
190
+
191
+
192
+
193
+ MEM_STATIC size_t ZSTD_decompressLegacyStream(void* legacyContext, U32 version,
194
+ ZSTD_outBuffer* output, ZSTD_inBuffer* input)
195
+ {
196
+ switch(version)
197
+ {
198
+ default :
199
+ case 1 :
200
+ case 2 :
201
+ case 3 :
202
+ return ERROR(version_unsupported);
203
+ case 4 :
204
+ {
205
+ ZBUFFv04_DCtx* dctx = (ZBUFFv04_DCtx*) legacyContext;
206
+ const void* src = (const char*)input->src + input->pos;
207
+ size_t readSize = input->size - input->pos;
208
+ void* dst = (char*)output->dst + output->pos;
209
+ size_t decodedSize = output->size - output->pos;
210
+ size_t const hintSize = ZBUFFv04_decompressContinue(dctx, dst, &decodedSize, src, &readSize);
211
+ output->pos += decodedSize;
212
+ input->pos += readSize;
213
+ return hintSize;
214
+ }
215
+ case 5 :
216
+ {
217
+ ZBUFFv05_DCtx* dctx = (ZBUFFv05_DCtx*) legacyContext;
218
+ const void* src = (const char*)input->src + input->pos;
219
+ size_t readSize = input->size - input->pos;
220
+ void* dst = (char*)output->dst + output->pos;
221
+ size_t decodedSize = output->size - output->pos;
222
+ size_t const hintSize = ZBUFFv05_decompressContinue(dctx, dst, &decodedSize, src, &readSize);
223
+ output->pos += decodedSize;
224
+ input->pos += readSize;
225
+ return hintSize;
226
+ }
227
+ case 6 :
228
+ {
229
+ ZBUFFv06_DCtx* dctx = (ZBUFFv06_DCtx*) legacyContext;
230
+ const void* src = (const char*)input->src + input->pos;
231
+ size_t readSize = input->size - input->pos;
232
+ void* dst = (char*)output->dst + output->pos;
233
+ size_t decodedSize = output->size - output->pos;
234
+ size_t const hintSize = ZBUFFv06_decompressContinue(dctx, dst, &decodedSize, src, &readSize);
235
+ output->pos += decodedSize;
236
+ input->pos += readSize;
237
+ return hintSize;
238
+ }
239
+ case 7 :
240
+ {
241
+ ZBUFFv07_DCtx* dctx = (ZBUFFv07_DCtx*) legacyContext;
242
+ const void* src = (const char*)input->src + input->pos;
243
+ size_t readSize = input->size - input->pos;
244
+ void* dst = (char*)output->dst + output->pos;
245
+ size_t decodedSize = output->size - output->pos;
246
+ size_t const hintSize = ZBUFFv07_decompressContinue(dctx, dst, &decodedSize, src, &readSize);
247
+ output->pos += decodedSize;
248
+ input->pos += readSize;
249
+ return hintSize;
250
+ }
251
+ }
252
+ }
253
+
254
+
255
+ #if defined (__cplusplus)
256
+ }
257
+ #endif
258
+
259
+ #endif /* ZSTD_LEGACY_H */
@@ -0,0 +1,2095 @@
1
+ /**
2
+ * Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
3
+ * All rights reserved.
4
+ *
5
+ * This source code is licensed under the BSD-style license found in the
6
+ * LICENSE file in the root directory of this source tree. An additional grant
7
+ * of patent rights can be found in the PATENTS file in the same directory.
8
+ */
9
+
10
+
11
+ /******************************************
12
+ * Includes
13
+ ******************************************/
14
+ #include <stddef.h> /* size_t, ptrdiff_t */
15
+ #include "zstd_v01.h"
16
+ #include "error_private.h"
17
+
18
+
19
+ /******************************************
20
+ * Static allocation
21
+ ******************************************/
22
+ /* You can statically allocate FSE CTable/DTable as a table of unsigned using below macro */
23
+ #define FSE_DTABLE_SIZE_U32(maxTableLog) (1 + (1<<maxTableLog))
24
+
25
+ /* You can statically allocate Huff0 DTable as a table of unsigned short using below macro */
26
+ #define HUF_DTABLE_SIZE_U16(maxTableLog) (1 + (1<<maxTableLog))
27
+ #define HUF_CREATE_STATIC_DTABLE(DTable, maxTableLog) \
28
+ unsigned short DTable[HUF_DTABLE_SIZE_U16(maxTableLog)] = { maxTableLog }
29
+
30
+
31
+ /******************************************
32
+ * Error Management
33
+ ******************************************/
34
+ #define FSE_LIST_ERRORS(ITEM) \
35
+ ITEM(FSE_OK_NoError) ITEM(FSE_ERROR_GENERIC) \
36
+ ITEM(FSE_ERROR_tableLog_tooLarge) ITEM(FSE_ERROR_maxSymbolValue_tooLarge) ITEM(FSE_ERROR_maxSymbolValue_tooSmall) \
37
+ ITEM(FSE_ERROR_dstSize_tooSmall) ITEM(FSE_ERROR_srcSize_wrong)\
38
+ ITEM(FSE_ERROR_corruptionDetected) \
39
+ ITEM(FSE_ERROR_maxCode)
40
+
41
+ #define FSE_GENERATE_ENUM(ENUM) ENUM,
42
+ typedef enum { FSE_LIST_ERRORS(FSE_GENERATE_ENUM) } FSE_errorCodes; /* enum is exposed, to detect & handle specific errors; compare function result to -enum value */
43
+
44
+
45
+ /******************************************
46
+ * FSE symbol compression API
47
+ ******************************************/
48
+ /*
49
+ This API consists of small unitary functions, which highly benefit from being inlined.
50
+ You will want to enable link-time-optimization to ensure these functions are properly inlined in your binary.
51
+ Visual seems to do it automatically.
52
+ For gcc or clang, you'll need to add -flto flag at compilation and linking stages.
53
+ If none of these solutions is applicable, include "fse.c" directly.
54
+ */
55
+
56
+ typedef unsigned FSE_CTable; /* don't allocate that. It's just a way to be more restrictive than void* */
57
+ typedef unsigned FSE_DTable; /* don't allocate that. It's just a way to be more restrictive than void* */
58
+
59
+ typedef struct
60
+ {
61
+ size_t bitContainer;
62
+ int bitPos;
63
+ char* startPtr;
64
+ char* ptr;
65
+ char* endPtr;
66
+ } FSE_CStream_t;
67
+
68
+ typedef struct
69
+ {
70
+ ptrdiff_t value;
71
+ const void* stateTable;
72
+ const void* symbolTT;
73
+ unsigned stateLog;
74
+ } FSE_CState_t;
75
+
76
+ typedef struct
77
+ {
78
+ size_t bitContainer;
79
+ unsigned bitsConsumed;
80
+ const char* ptr;
81
+ const char* start;
82
+ } FSE_DStream_t;
83
+
84
+ typedef struct
85
+ {
86
+ size_t state;
87
+ const void* table; /* precise table may vary, depending on U16 */
88
+ } FSE_DState_t;
89
+
90
+ typedef enum { FSE_DStream_unfinished = 0,
91
+ FSE_DStream_endOfBuffer = 1,
92
+ FSE_DStream_completed = 2,
93
+ FSE_DStream_tooFar = 3 } FSE_DStream_status; /* result of FSE_reloadDStream() */
94
+ /* 1,2,4,8 would be better for bitmap combinations, but slows down performance a bit ... ?! */
95
+
96
+
97
+ /****************************************************************
98
+ * Tuning parameters
99
+ ****************************************************************/
100
+ /* MEMORY_USAGE :
101
+ * Memory usage formula : N->2^N Bytes (examples : 10 -> 1KB; 12 -> 4KB ; 16 -> 64KB; 20 -> 1MB; etc.)
102
+ * Increasing memory usage improves compression ratio
103
+ * Reduced memory usage can improve speed, due to cache effect
104
+ * Recommended max value is 14, for 16KB, which nicely fits into Intel x86 L1 cache */
105
+ #define FSE_MAX_MEMORY_USAGE 14
106
+ #define FSE_DEFAULT_MEMORY_USAGE 13
107
+
108
+ /* FSE_MAX_SYMBOL_VALUE :
109
+ * Maximum symbol value authorized.
110
+ * Required for proper stack allocation */
111
+ #define FSE_MAX_SYMBOL_VALUE 255
112
+
113
+
114
+ /****************************************************************
115
+ * template functions type & suffix
116
+ ****************************************************************/
117
+ #define FSE_FUNCTION_TYPE BYTE
118
+ #define FSE_FUNCTION_EXTENSION
119
+
120
+
121
+ /****************************************************************
122
+ * Byte symbol type
123
+ ****************************************************************/
124
+ typedef struct
125
+ {
126
+ unsigned short newState;
127
+ unsigned char symbol;
128
+ unsigned char nbBits;
129
+ } FSE_decode_t; /* size == U32 */
130
+
131
+
132
+
133
+ /****************************************************************
134
+ * Compiler specifics
135
+ ****************************************************************/
136
+ #ifdef _MSC_VER /* Visual Studio */
137
+ # define FORCE_INLINE static __forceinline
138
+ # include <intrin.h> /* For Visual 2005 */
139
+ # pragma warning(disable : 4127) /* disable: C4127: conditional expression is constant */
140
+ # pragma warning(disable : 4214) /* disable: C4214: non-int bitfields */
141
+ #else
142
+ # define GCC_VERSION (__GNUC__ * 100 + __GNUC_MINOR__)
143
+ # if defined (__cplusplus) || defined (__STDC_VERSION__) && __STDC_VERSION__ >= 199901L /* C99 */
144
+ # ifdef __GNUC__
145
+ # define FORCE_INLINE static inline __attribute__((always_inline))
146
+ # else
147
+ # define FORCE_INLINE static inline
148
+ # endif
149
+ # else
150
+ # define FORCE_INLINE static
151
+ # endif /* __STDC_VERSION__ */
152
+ #endif
153
+
154
+
155
+ /****************************************************************
156
+ * Includes
157
+ ****************************************************************/
158
+ #include <stdlib.h> /* malloc, free, qsort */
159
+ #include <string.h> /* memcpy, memset */
160
+ #include <stdio.h> /* printf (debug) */
161
+
162
+
163
+ #ifndef MEM_ACCESS_MODULE
164
+ #define MEM_ACCESS_MODULE
165
+ /****************************************************************
166
+ * Basic Types
167
+ *****************************************************************/
168
+ #if defined (__STDC_VERSION__) && __STDC_VERSION__ >= 199901L /* C99 */
169
+ # include <stdint.h>
170
+ typedef uint8_t BYTE;
171
+ typedef uint16_t U16;
172
+ typedef int16_t S16;
173
+ typedef uint32_t U32;
174
+ typedef int32_t S32;
175
+ typedef uint64_t U64;
176
+ typedef int64_t S64;
177
+ #else
178
+ typedef unsigned char BYTE;
179
+ typedef unsigned short U16;
180
+ typedef signed short S16;
181
+ typedef unsigned int U32;
182
+ typedef signed int S32;
183
+ typedef unsigned long long U64;
184
+ typedef signed long long S64;
185
+ #endif
186
+
187
+ #endif /* MEM_ACCESS_MODULE */
188
+
189
+ /****************************************************************
190
+ * Memory I/O
191
+ *****************************************************************/
192
+ /* FSE_FORCE_MEMORY_ACCESS
193
+ * By default, access to unaligned memory is controlled by `memcpy()`, which is safe and portable.
194
+ * Unfortunately, on some target/compiler combinations, the generated assembly is sub-optimal.
195
+ * The below switch allow to select different access method for improved performance.
196
+ * Method 0 (default) : use `memcpy()`. Safe and portable.
197
+ * Method 1 : `__packed` statement. It depends on compiler extension (ie, not portable).
198
+ * This method is safe if your compiler supports it, and *generally* as fast or faster than `memcpy`.
199
+ * Method 2 : direct access. This method is portable but violate C standard.
200
+ * It can generate buggy code on targets generating assembly depending on alignment.
201
+ * But in some circumstances, it's the only known way to get the most performance (ie GCC + ARMv6)
202
+ * See http://fastcompression.blogspot.fr/2015/08/accessing-unaligned-memory.html for details.
203
+ * Prefer these methods in priority order (0 > 1 > 2)
204
+ */
205
+ #ifndef FSE_FORCE_MEMORY_ACCESS /* can be defined externally, on command line for example */
206
+ # if defined(__GNUC__) && ( defined(__ARM_ARCH_6__) || defined(__ARM_ARCH_6J__) || defined(__ARM_ARCH_6K__) || defined(__ARM_ARCH_6Z__) || defined(__ARM_ARCH_6ZK__) || defined(__ARM_ARCH_6T2__) )
207
+ # define FSE_FORCE_MEMORY_ACCESS 2
208
+ # elif (defined(__INTEL_COMPILER) && !defined(WIN32)) || \
209
+ (defined(__GNUC__) && ( defined(__ARM_ARCH_7__) || defined(__ARM_ARCH_7A__) || defined(__ARM_ARCH_7R__) || defined(__ARM_ARCH_7M__) || defined(__ARM_ARCH_7S__) ))
210
+ # define FSE_FORCE_MEMORY_ACCESS 1
211
+ # endif
212
+ #endif
213
+
214
+
215
+ static unsigned FSE_32bits(void)
216
+ {
217
+ return sizeof(void*)==4;
218
+ }
219
+
220
+ static unsigned FSE_isLittleEndian(void)
221
+ {
222
+ const union { U32 i; BYTE c[4]; } one = { 1 }; /* don't use static : performance detrimental */
223
+ return one.c[0];
224
+ }
225
+
226
+ #if defined(FSE_FORCE_MEMORY_ACCESS) && (FSE_FORCE_MEMORY_ACCESS==2)
227
+
228
+ static U16 FSE_read16(const void* memPtr) { return *(const U16*) memPtr; }
229
+ static U32 FSE_read32(const void* memPtr) { return *(const U32*) memPtr; }
230
+ static U64 FSE_read64(const void* memPtr) { return *(const U64*) memPtr; }
231
+
232
+ #elif defined(FSE_FORCE_MEMORY_ACCESS) && (FSE_FORCE_MEMORY_ACCESS==1)
233
+
234
+ /* __pack instructions are safer, but compiler specific, hence potentially problematic for some compilers */
235
+ /* currently only defined for gcc and icc */
236
+ typedef union { U16 u16; U32 u32; U64 u64; } __attribute__((packed)) unalign;
237
+
238
+ static U16 FSE_read16(const void* ptr) { return ((const unalign*)ptr)->u16; }
239
+ static U32 FSE_read32(const void* ptr) { return ((const unalign*)ptr)->u32; }
240
+ static U64 FSE_read64(const void* ptr) { return ((const unalign*)ptr)->u64; }
241
+
242
+ #else
243
+
244
+ static U16 FSE_read16(const void* memPtr)
245
+ {
246
+ U16 val; memcpy(&val, memPtr, sizeof(val)); return val;
247
+ }
248
+
249
+ static U32 FSE_read32(const void* memPtr)
250
+ {
251
+ U32 val; memcpy(&val, memPtr, sizeof(val)); return val;
252
+ }
253
+
254
+ static U64 FSE_read64(const void* memPtr)
255
+ {
256
+ U64 val; memcpy(&val, memPtr, sizeof(val)); return val;
257
+ }
258
+
259
+ #endif // FSE_FORCE_MEMORY_ACCESS
260
+
261
+ static U16 FSE_readLE16(const void* memPtr)
262
+ {
263
+ if (FSE_isLittleEndian())
264
+ return FSE_read16(memPtr);
265
+ else
266
+ {
267
+ const BYTE* p = (const BYTE*)memPtr;
268
+ return (U16)(p[0] + (p[1]<<8));
269
+ }
270
+ }
271
+
272
+ static U32 FSE_readLE32(const void* memPtr)
273
+ {
274
+ if (FSE_isLittleEndian())
275
+ return FSE_read32(memPtr);
276
+ else
277
+ {
278
+ const BYTE* p = (const BYTE*)memPtr;
279
+ return (U32)((U32)p[0] + ((U32)p[1]<<8) + ((U32)p[2]<<16) + ((U32)p[3]<<24));
280
+ }
281
+ }
282
+
283
+
284
+ static U64 FSE_readLE64(const void* memPtr)
285
+ {
286
+ if (FSE_isLittleEndian())
287
+ return FSE_read64(memPtr);
288
+ else
289
+ {
290
+ const BYTE* p = (const BYTE*)memPtr;
291
+ return (U64)((U64)p[0] + ((U64)p[1]<<8) + ((U64)p[2]<<16) + ((U64)p[3]<<24)
292
+ + ((U64)p[4]<<32) + ((U64)p[5]<<40) + ((U64)p[6]<<48) + ((U64)p[7]<<56));
293
+ }
294
+ }
295
+
296
+ static size_t FSE_readLEST(const void* memPtr)
297
+ {
298
+ if (FSE_32bits())
299
+ return (size_t)FSE_readLE32(memPtr);
300
+ else
301
+ return (size_t)FSE_readLE64(memPtr);
302
+ }
303
+
304
+
305
+
306
+ /****************************************************************
307
+ * Constants
308
+ *****************************************************************/
309
+ #define FSE_MAX_TABLELOG (FSE_MAX_MEMORY_USAGE-2)
310
+ #define FSE_MAX_TABLESIZE (1U<<FSE_MAX_TABLELOG)
311
+ #define FSE_MAXTABLESIZE_MASK (FSE_MAX_TABLESIZE-1)
312
+ #define FSE_DEFAULT_TABLELOG (FSE_DEFAULT_MEMORY_USAGE-2)
313
+ #define FSE_MIN_TABLELOG 5
314
+
315
+ #define FSE_TABLELOG_ABSOLUTE_MAX 15
316
+ #if FSE_MAX_TABLELOG > FSE_TABLELOG_ABSOLUTE_MAX
317
+ #error "FSE_MAX_TABLELOG > FSE_TABLELOG_ABSOLUTE_MAX is not supported"
318
+ #endif
319
+
320
+
321
+ /****************************************************************
322
+ * Error Management
323
+ ****************************************************************/
324
+ #define FSE_STATIC_ASSERT(c) { enum { FSE_static_assert = 1/(int)(!!(c)) }; } /* use only *after* variable declarations */
325
+
326
+
327
+ /****************************************************************
328
+ * Complex types
329
+ ****************************************************************/
330
+ typedef struct
331
+ {
332
+ int deltaFindState;
333
+ U32 deltaNbBits;
334
+ } FSE_symbolCompressionTransform; /* total 8 bytes */
335
+
336
+ typedef U32 DTable_max_t[FSE_DTABLE_SIZE_U32(FSE_MAX_TABLELOG)];
337
+
338
+ /****************************************************************
339
+ * Internal functions
340
+ ****************************************************************/
341
+ FORCE_INLINE unsigned FSE_highbit32 (register U32 val)
342
+ {
343
+ # if defined(_MSC_VER) /* Visual */
344
+ unsigned long r;
345
+ _BitScanReverse ( &r, val );
346
+ return (unsigned) r;
347
+ # elif defined(__GNUC__) && (GCC_VERSION >= 304) /* GCC Intrinsic */
348
+ return 31 - __builtin_clz (val);
349
+ # else /* Software version */
350
+ static const unsigned DeBruijnClz[32] = { 0, 9, 1, 10, 13, 21, 2, 29, 11, 14, 16, 18, 22, 25, 3, 30, 8, 12, 20, 28, 15, 17, 24, 7, 19, 27, 23, 6, 26, 5, 4, 31 };
351
+ U32 v = val;
352
+ unsigned r;
353
+ v |= v >> 1;
354
+ v |= v >> 2;
355
+ v |= v >> 4;
356
+ v |= v >> 8;
357
+ v |= v >> 16;
358
+ r = DeBruijnClz[ (U32) (v * 0x07C4ACDDU) >> 27];
359
+ return r;
360
+ # endif
361
+ }
362
+
363
+
364
+ /****************************************************************
365
+ * Templates
366
+ ****************************************************************/
367
+ /*
368
+ designed to be included
369
+ for type-specific functions (template emulation in C)
370
+ Objective is to write these functions only once, for improved maintenance
371
+ */
372
+
373
+ /* safety checks */
374
+ #ifndef FSE_FUNCTION_EXTENSION
375
+ # error "FSE_FUNCTION_EXTENSION must be defined"
376
+ #endif
377
+ #ifndef FSE_FUNCTION_TYPE
378
+ # error "FSE_FUNCTION_TYPE must be defined"
379
+ #endif
380
+
381
+ /* Function names */
382
+ #define FSE_CAT(X,Y) X##Y
383
+ #define FSE_FUNCTION_NAME(X,Y) FSE_CAT(X,Y)
384
+ #define FSE_TYPE_NAME(X,Y) FSE_CAT(X,Y)
385
+
386
+
387
+
388
+ static U32 FSE_tableStep(U32 tableSize) { return (tableSize>>1) + (tableSize>>3) + 3; }
389
+
390
+ #define FSE_DECODE_TYPE FSE_decode_t
391
+
392
+
393
+ typedef struct {
394
+ U16 tableLog;
395
+ U16 fastMode;
396
+ } FSE_DTableHeader; /* sizeof U32 */
397
+
398
+ static size_t FSE_buildDTable
399
+ (FSE_DTable* dt, const short* normalizedCounter, unsigned maxSymbolValue, unsigned tableLog)
400
+ {
401
+ void* ptr = dt;
402
+ FSE_DTableHeader* const DTableH = (FSE_DTableHeader*)ptr;
403
+ FSE_DECODE_TYPE* const tableDecode = (FSE_DECODE_TYPE*)(ptr) + 1; /* because dt is unsigned, 32-bits aligned on 32-bits */
404
+ const U32 tableSize = 1 << tableLog;
405
+ const U32 tableMask = tableSize-1;
406
+ const U32 step = FSE_tableStep(tableSize);
407
+ U16 symbolNext[FSE_MAX_SYMBOL_VALUE+1];
408
+ U32 position = 0;
409
+ U32 highThreshold = tableSize-1;
410
+ const S16 largeLimit= (S16)(1 << (tableLog-1));
411
+ U32 noLarge = 1;
412
+ U32 s;
413
+
414
+ /* Sanity Checks */
415
+ if (maxSymbolValue > FSE_MAX_SYMBOL_VALUE) return (size_t)-FSE_ERROR_maxSymbolValue_tooLarge;
416
+ if (tableLog > FSE_MAX_TABLELOG) return (size_t)-FSE_ERROR_tableLog_tooLarge;
417
+
418
+ /* Init, lay down lowprob symbols */
419
+ DTableH[0].tableLog = (U16)tableLog;
420
+ for (s=0; s<=maxSymbolValue; s++)
421
+ {
422
+ if (normalizedCounter[s]==-1)
423
+ {
424
+ tableDecode[highThreshold--].symbol = (FSE_FUNCTION_TYPE)s;
425
+ symbolNext[s] = 1;
426
+ }
427
+ else
428
+ {
429
+ if (normalizedCounter[s] >= largeLimit) noLarge=0;
430
+ symbolNext[s] = normalizedCounter[s];
431
+ }
432
+ }
433
+
434
+ /* Spread symbols */
435
+ for (s=0; s<=maxSymbolValue; s++)
436
+ {
437
+ int i;
438
+ for (i=0; i<normalizedCounter[s]; i++)
439
+ {
440
+ tableDecode[position].symbol = (FSE_FUNCTION_TYPE)s;
441
+ position = (position + step) & tableMask;
442
+ while (position > highThreshold) position = (position + step) & tableMask; /* lowprob area */
443
+ }
444
+ }
445
+
446
+ if (position!=0) return (size_t)-FSE_ERROR_GENERIC; /* position must reach all cells once, otherwise normalizedCounter is incorrect */
447
+
448
+ /* Build Decoding table */
449
+ {
450
+ U32 i;
451
+ for (i=0; i<tableSize; i++)
452
+ {
453
+ FSE_FUNCTION_TYPE symbol = (FSE_FUNCTION_TYPE)(tableDecode[i].symbol);
454
+ U16 nextState = symbolNext[symbol]++;
455
+ tableDecode[i].nbBits = (BYTE) (tableLog - FSE_highbit32 ((U32)nextState) );
456
+ tableDecode[i].newState = (U16) ( (nextState << tableDecode[i].nbBits) - tableSize);
457
+ }
458
+ }
459
+
460
+ DTableH->fastMode = (U16)noLarge;
461
+ return 0;
462
+ }
463
+
464
+
465
+ /******************************************
466
+ * FSE byte symbol
467
+ ******************************************/
468
+ #ifndef FSE_COMMONDEFS_ONLY
469
+
470
+ static unsigned FSE_isError(size_t code) { return (code > (size_t)(-FSE_ERROR_maxCode)); }
471
+
472
+ static short FSE_abs(short a)
473
+ {
474
+ return a<0? -a : a;
475
+ }
476
+
477
+
478
+ /****************************************************************
479
+ * Header bitstream management
480
+ ****************************************************************/
481
+ static size_t FSE_readNCount (short* normalizedCounter, unsigned* maxSVPtr, unsigned* tableLogPtr,
482
+ const void* headerBuffer, size_t hbSize)
483
+ {
484
+ const BYTE* const istart = (const BYTE*) headerBuffer;
485
+ const BYTE* const iend = istart + hbSize;
486
+ const BYTE* ip = istart;
487
+ int nbBits;
488
+ int remaining;
489
+ int threshold;
490
+ U32 bitStream;
491
+ int bitCount;
492
+ unsigned charnum = 0;
493
+ int previous0 = 0;
494
+
495
+ if (hbSize < 4) return (size_t)-FSE_ERROR_srcSize_wrong;
496
+ bitStream = FSE_readLE32(ip);
497
+ nbBits = (bitStream & 0xF) + FSE_MIN_TABLELOG; /* extract tableLog */
498
+ if (nbBits > FSE_TABLELOG_ABSOLUTE_MAX) return (size_t)-FSE_ERROR_tableLog_tooLarge;
499
+ bitStream >>= 4;
500
+ bitCount = 4;
501
+ *tableLogPtr = nbBits;
502
+ remaining = (1<<nbBits)+1;
503
+ threshold = 1<<nbBits;
504
+ nbBits++;
505
+
506
+ while ((remaining>1) && (charnum<=*maxSVPtr))
507
+ {
508
+ if (previous0)
509
+ {
510
+ unsigned n0 = charnum;
511
+ while ((bitStream & 0xFFFF) == 0xFFFF)
512
+ {
513
+ n0+=24;
514
+ if (ip < iend-5)
515
+ {
516
+ ip+=2;
517
+ bitStream = FSE_readLE32(ip) >> bitCount;
518
+ }
519
+ else
520
+ {
521
+ bitStream >>= 16;
522
+ bitCount+=16;
523
+ }
524
+ }
525
+ while ((bitStream & 3) == 3)
526
+ {
527
+ n0+=3;
528
+ bitStream>>=2;
529
+ bitCount+=2;
530
+ }
531
+ n0 += bitStream & 3;
532
+ bitCount += 2;
533
+ if (n0 > *maxSVPtr) return (size_t)-FSE_ERROR_maxSymbolValue_tooSmall;
534
+ while (charnum < n0) normalizedCounter[charnum++] = 0;
535
+ if ((ip <= iend-7) || (ip + (bitCount>>3) <= iend-4))
536
+ {
537
+ ip += bitCount>>3;
538
+ bitCount &= 7;
539
+ bitStream = FSE_readLE32(ip) >> bitCount;
540
+ }
541
+ else
542
+ bitStream >>= 2;
543
+ }
544
+ {
545
+ const short max = (short)((2*threshold-1)-remaining);
546
+ short count;
547
+
548
+ if ((bitStream & (threshold-1)) < (U32)max)
549
+ {
550
+ count = (short)(bitStream & (threshold-1));
551
+ bitCount += nbBits-1;
552
+ }
553
+ else
554
+ {
555
+ count = (short)(bitStream & (2*threshold-1));
556
+ if (count >= threshold) count -= max;
557
+ bitCount += nbBits;
558
+ }
559
+
560
+ count--; /* extra accuracy */
561
+ remaining -= FSE_abs(count);
562
+ normalizedCounter[charnum++] = count;
563
+ previous0 = !count;
564
+ while (remaining < threshold)
565
+ {
566
+ nbBits--;
567
+ threshold >>= 1;
568
+ }
569
+
570
+ {
571
+ if ((ip <= iend-7) || (ip + (bitCount>>3) <= iend-4))
572
+ {
573
+ ip += bitCount>>3;
574
+ bitCount &= 7;
575
+ }
576
+ else
577
+ {
578
+ bitCount -= (int)(8 * (iend - 4 - ip));
579
+ ip = iend - 4;
580
+ }
581
+ bitStream = FSE_readLE32(ip) >> (bitCount & 31);
582
+ }
583
+ }
584
+ }
585
+ if (remaining != 1) return (size_t)-FSE_ERROR_GENERIC;
586
+ *maxSVPtr = charnum-1;
587
+
588
+ ip += (bitCount+7)>>3;
589
+ if ((size_t)(ip-istart) > hbSize) return (size_t)-FSE_ERROR_srcSize_wrong;
590
+ return ip-istart;
591
+ }
592
+
593
+
594
+ /*********************************************************
595
+ * Decompression (Byte symbols)
596
+ *********************************************************/
597
+ static size_t FSE_buildDTable_rle (FSE_DTable* dt, BYTE symbolValue)
598
+ {
599
+ void* ptr = dt;
600
+ FSE_DTableHeader* const DTableH = (FSE_DTableHeader*)ptr;
601
+ FSE_decode_t* const cell = (FSE_decode_t*)(ptr) + 1; /* because dt is unsigned */
602
+
603
+ DTableH->tableLog = 0;
604
+ DTableH->fastMode = 0;
605
+
606
+ cell->newState = 0;
607
+ cell->symbol = symbolValue;
608
+ cell->nbBits = 0;
609
+
610
+ return 0;
611
+ }
612
+
613
+
614
+ static size_t FSE_buildDTable_raw (FSE_DTable* dt, unsigned nbBits)
615
+ {
616
+ void* ptr = dt;
617
+ FSE_DTableHeader* const DTableH = (FSE_DTableHeader*)ptr;
618
+ FSE_decode_t* const dinfo = (FSE_decode_t*)(ptr) + 1; /* because dt is unsigned */
619
+ const unsigned tableSize = 1 << nbBits;
620
+ const unsigned tableMask = tableSize - 1;
621
+ const unsigned maxSymbolValue = tableMask;
622
+ unsigned s;
623
+
624
+ /* Sanity checks */
625
+ if (nbBits < 1) return (size_t)-FSE_ERROR_GENERIC; /* min size */
626
+
627
+ /* Build Decoding Table */
628
+ DTableH->tableLog = (U16)nbBits;
629
+ DTableH->fastMode = 1;
630
+ for (s=0; s<=maxSymbolValue; s++)
631
+ {
632
+ dinfo[s].newState = 0;
633
+ dinfo[s].symbol = (BYTE)s;
634
+ dinfo[s].nbBits = (BYTE)nbBits;
635
+ }
636
+
637
+ return 0;
638
+ }
639
+
640
+
641
+ /* FSE_initDStream
642
+ * Initialize a FSE_DStream_t.
643
+ * srcBuffer must point at the beginning of an FSE block.
644
+ * The function result is the size of the FSE_block (== srcSize).
645
+ * If srcSize is too small, the function will return an errorCode;
646
+ */
647
+ static size_t FSE_initDStream(FSE_DStream_t* bitD, const void* srcBuffer, size_t srcSize)
648
+ {
649
+ if (srcSize < 1) return (size_t)-FSE_ERROR_srcSize_wrong;
650
+
651
+ if (srcSize >= sizeof(size_t))
652
+ {
653
+ U32 contain32;
654
+ bitD->start = (const char*)srcBuffer;
655
+ bitD->ptr = (const char*)srcBuffer + srcSize - sizeof(size_t);
656
+ bitD->bitContainer = FSE_readLEST(bitD->ptr);
657
+ contain32 = ((const BYTE*)srcBuffer)[srcSize-1];
658
+ if (contain32 == 0) return (size_t)-FSE_ERROR_GENERIC; /* stop bit not present */
659
+ bitD->bitsConsumed = 8 - FSE_highbit32(contain32);
660
+ }
661
+ else
662
+ {
663
+ U32 contain32;
664
+ bitD->start = (const char*)srcBuffer;
665
+ bitD->ptr = bitD->start;
666
+ bitD->bitContainer = *(const BYTE*)(bitD->start);
667
+ switch(srcSize)
668
+ {
669
+ case 7: bitD->bitContainer += (size_t)(((const BYTE*)(bitD->start))[6]) << (sizeof(size_t)*8 - 16);
670
+ case 6: bitD->bitContainer += (size_t)(((const BYTE*)(bitD->start))[5]) << (sizeof(size_t)*8 - 24);
671
+ case 5: bitD->bitContainer += (size_t)(((const BYTE*)(bitD->start))[4]) << (sizeof(size_t)*8 - 32);
672
+ case 4: bitD->bitContainer += (size_t)(((const BYTE*)(bitD->start))[3]) << 24;
673
+ case 3: bitD->bitContainer += (size_t)(((const BYTE*)(bitD->start))[2]) << 16;
674
+ case 2: bitD->bitContainer += (size_t)(((const BYTE*)(bitD->start))[1]) << 8;
675
+ default:;
676
+ }
677
+ contain32 = ((const BYTE*)srcBuffer)[srcSize-1];
678
+ if (contain32 == 0) return (size_t)-FSE_ERROR_GENERIC; /* stop bit not present */
679
+ bitD->bitsConsumed = 8 - FSE_highbit32(contain32);
680
+ bitD->bitsConsumed += (U32)(sizeof(size_t) - srcSize)*8;
681
+ }
682
+
683
+ return srcSize;
684
+ }
685
+
686
+
687
+ /*!FSE_lookBits
688
+ * Provides next n bits from the bitContainer.
689
+ * bitContainer is not modified (bits are still present for next read/look)
690
+ * On 32-bits, maxNbBits==25
691
+ * On 64-bits, maxNbBits==57
692
+ * return : value extracted.
693
+ */
694
+ static size_t FSE_lookBits(FSE_DStream_t* bitD, U32 nbBits)
695
+ {
696
+ const U32 bitMask = sizeof(bitD->bitContainer)*8 - 1;
697
+ return ((bitD->bitContainer << (bitD->bitsConsumed & bitMask)) >> 1) >> ((bitMask-nbBits) & bitMask);
698
+ }
699
+
700
+ static size_t FSE_lookBitsFast(FSE_DStream_t* bitD, U32 nbBits) /* only if nbBits >= 1 !! */
701
+ {
702
+ const U32 bitMask = sizeof(bitD->bitContainer)*8 - 1;
703
+ return (bitD->bitContainer << (bitD->bitsConsumed & bitMask)) >> (((bitMask+1)-nbBits) & bitMask);
704
+ }
705
+
706
+ static void FSE_skipBits(FSE_DStream_t* bitD, U32 nbBits)
707
+ {
708
+ bitD->bitsConsumed += nbBits;
709
+ }
710
+
711
+
712
+ /*!FSE_readBits
713
+ * Read next n bits from the bitContainer.
714
+ * On 32-bits, don't read more than maxNbBits==25
715
+ * On 64-bits, don't read more than maxNbBits==57
716
+ * Use the fast variant *only* if n >= 1.
717
+ * return : value extracted.
718
+ */
719
+ static size_t FSE_readBits(FSE_DStream_t* bitD, U32 nbBits)
720
+ {
721
+ size_t value = FSE_lookBits(bitD, nbBits);
722
+ FSE_skipBits(bitD, nbBits);
723
+ return value;
724
+ }
725
+
726
+ static size_t FSE_readBitsFast(FSE_DStream_t* bitD, U32 nbBits) /* only if nbBits >= 1 !! */
727
+ {
728
+ size_t value = FSE_lookBitsFast(bitD, nbBits);
729
+ FSE_skipBits(bitD, nbBits);
730
+ return value;
731
+ }
732
+
733
+ static unsigned FSE_reloadDStream(FSE_DStream_t* bitD)
734
+ {
735
+ if (bitD->bitsConsumed > (sizeof(bitD->bitContainer)*8)) /* should never happen */
736
+ return FSE_DStream_tooFar;
737
+
738
+ if (bitD->ptr >= bitD->start + sizeof(bitD->bitContainer))
739
+ {
740
+ bitD->ptr -= bitD->bitsConsumed >> 3;
741
+ bitD->bitsConsumed &= 7;
742
+ bitD->bitContainer = FSE_readLEST(bitD->ptr);
743
+ return FSE_DStream_unfinished;
744
+ }
745
+ if (bitD->ptr == bitD->start)
746
+ {
747
+ if (bitD->bitsConsumed < sizeof(bitD->bitContainer)*8) return FSE_DStream_endOfBuffer;
748
+ return FSE_DStream_completed;
749
+ }
750
+ {
751
+ U32 nbBytes = bitD->bitsConsumed >> 3;
752
+ U32 result = FSE_DStream_unfinished;
753
+ if (bitD->ptr - nbBytes < bitD->start)
754
+ {
755
+ nbBytes = (U32)(bitD->ptr - bitD->start); /* ptr > start */
756
+ result = FSE_DStream_endOfBuffer;
757
+ }
758
+ bitD->ptr -= nbBytes;
759
+ bitD->bitsConsumed -= nbBytes*8;
760
+ bitD->bitContainer = FSE_readLEST(bitD->ptr); /* reminder : srcSize > sizeof(bitD) */
761
+ return result;
762
+ }
763
+ }
764
+
765
+
766
+ static void FSE_initDState(FSE_DState_t* DStatePtr, FSE_DStream_t* bitD, const FSE_DTable* dt)
767
+ {
768
+ const void* ptr = dt;
769
+ const FSE_DTableHeader* const DTableH = (const FSE_DTableHeader*)ptr;
770
+ DStatePtr->state = FSE_readBits(bitD, DTableH->tableLog);
771
+ FSE_reloadDStream(bitD);
772
+ DStatePtr->table = dt + 1;
773
+ }
774
+
775
+ static BYTE FSE_decodeSymbol(FSE_DState_t* DStatePtr, FSE_DStream_t* bitD)
776
+ {
777
+ const FSE_decode_t DInfo = ((const FSE_decode_t*)(DStatePtr->table))[DStatePtr->state];
778
+ const U32 nbBits = DInfo.nbBits;
779
+ BYTE symbol = DInfo.symbol;
780
+ size_t lowBits = FSE_readBits(bitD, nbBits);
781
+
782
+ DStatePtr->state = DInfo.newState + lowBits;
783
+ return symbol;
784
+ }
785
+
786
+ static BYTE FSE_decodeSymbolFast(FSE_DState_t* DStatePtr, FSE_DStream_t* bitD)
787
+ {
788
+ const FSE_decode_t DInfo = ((const FSE_decode_t*)(DStatePtr->table))[DStatePtr->state];
789
+ const U32 nbBits = DInfo.nbBits;
790
+ BYTE symbol = DInfo.symbol;
791
+ size_t lowBits = FSE_readBitsFast(bitD, nbBits);
792
+
793
+ DStatePtr->state = DInfo.newState + lowBits;
794
+ return symbol;
795
+ }
796
+
797
+ /* FSE_endOfDStream
798
+ Tells if bitD has reached end of bitStream or not */
799
+
800
+ static unsigned FSE_endOfDStream(const FSE_DStream_t* bitD)
801
+ {
802
+ return ((bitD->ptr == bitD->start) && (bitD->bitsConsumed == sizeof(bitD->bitContainer)*8));
803
+ }
804
+
805
+ static unsigned FSE_endOfDState(const FSE_DState_t* DStatePtr)
806
+ {
807
+ return DStatePtr->state == 0;
808
+ }
809
+
810
+
811
+ FORCE_INLINE size_t FSE_decompress_usingDTable_generic(
812
+ void* dst, size_t maxDstSize,
813
+ const void* cSrc, size_t cSrcSize,
814
+ const FSE_DTable* dt, const unsigned fast)
815
+ {
816
+ BYTE* const ostart = (BYTE*) dst;
817
+ BYTE* op = ostart;
818
+ BYTE* const omax = op + maxDstSize;
819
+ BYTE* const olimit = omax-3;
820
+
821
+ FSE_DStream_t bitD;
822
+ FSE_DState_t state1;
823
+ FSE_DState_t state2;
824
+ size_t errorCode;
825
+
826
+ /* Init */
827
+ errorCode = FSE_initDStream(&bitD, cSrc, cSrcSize); /* replaced last arg by maxCompressed Size */
828
+ if (FSE_isError(errorCode)) return errorCode;
829
+
830
+ FSE_initDState(&state1, &bitD, dt);
831
+ FSE_initDState(&state2, &bitD, dt);
832
+
833
+ #define FSE_GETSYMBOL(statePtr) fast ? FSE_decodeSymbolFast(statePtr, &bitD) : FSE_decodeSymbol(statePtr, &bitD)
834
+
835
+ /* 4 symbols per loop */
836
+ for ( ; (FSE_reloadDStream(&bitD)==FSE_DStream_unfinished) && (op<olimit) ; op+=4)
837
+ {
838
+ op[0] = FSE_GETSYMBOL(&state1);
839
+
840
+ if (FSE_MAX_TABLELOG*2+7 > sizeof(bitD.bitContainer)*8) /* This test must be static */
841
+ FSE_reloadDStream(&bitD);
842
+
843
+ op[1] = FSE_GETSYMBOL(&state2);
844
+
845
+ if (FSE_MAX_TABLELOG*4+7 > sizeof(bitD.bitContainer)*8) /* This test must be static */
846
+ { if (FSE_reloadDStream(&bitD) > FSE_DStream_unfinished) { op+=2; break; } }
847
+
848
+ op[2] = FSE_GETSYMBOL(&state1);
849
+
850
+ if (FSE_MAX_TABLELOG*2+7 > sizeof(bitD.bitContainer)*8) /* This test must be static */
851
+ FSE_reloadDStream(&bitD);
852
+
853
+ op[3] = FSE_GETSYMBOL(&state2);
854
+ }
855
+
856
+ /* tail */
857
+ /* note : FSE_reloadDStream(&bitD) >= FSE_DStream_partiallyFilled; Ends at exactly FSE_DStream_completed */
858
+ while (1)
859
+ {
860
+ if ( (FSE_reloadDStream(&bitD)>FSE_DStream_completed) || (op==omax) || (FSE_endOfDStream(&bitD) && (fast || FSE_endOfDState(&state1))) )
861
+ break;
862
+
863
+ *op++ = FSE_GETSYMBOL(&state1);
864
+
865
+ if ( (FSE_reloadDStream(&bitD)>FSE_DStream_completed) || (op==omax) || (FSE_endOfDStream(&bitD) && (fast || FSE_endOfDState(&state2))) )
866
+ break;
867
+
868
+ *op++ = FSE_GETSYMBOL(&state2);
869
+ }
870
+
871
+ /* end ? */
872
+ if (FSE_endOfDStream(&bitD) && FSE_endOfDState(&state1) && FSE_endOfDState(&state2))
873
+ return op-ostart;
874
+
875
+ if (op==omax) return (size_t)-FSE_ERROR_dstSize_tooSmall; /* dst buffer is full, but cSrc unfinished */
876
+
877
+ return (size_t)-FSE_ERROR_corruptionDetected;
878
+ }
879
+
880
+
881
+ static size_t FSE_decompress_usingDTable(void* dst, size_t originalSize,
882
+ const void* cSrc, size_t cSrcSize,
883
+ const FSE_DTable* dt)
884
+ {
885
+ FSE_DTableHeader DTableH;
886
+ memcpy(&DTableH, dt, sizeof(DTableH)); /* memcpy() into local variable, to avoid strict aliasing warning */
887
+
888
+ /* select fast mode (static) */
889
+ if (DTableH.fastMode) return FSE_decompress_usingDTable_generic(dst, originalSize, cSrc, cSrcSize, dt, 1);
890
+ return FSE_decompress_usingDTable_generic(dst, originalSize, cSrc, cSrcSize, dt, 0);
891
+ }
892
+
893
+
894
+ static size_t FSE_decompress(void* dst, size_t maxDstSize, const void* cSrc, size_t cSrcSize)
895
+ {
896
+ const BYTE* const istart = (const BYTE*)cSrc;
897
+ const BYTE* ip = istart;
898
+ short counting[FSE_MAX_SYMBOL_VALUE+1];
899
+ DTable_max_t dt; /* Static analyzer seems unable to understand this table will be properly initialized later */
900
+ unsigned tableLog;
901
+ unsigned maxSymbolValue = FSE_MAX_SYMBOL_VALUE;
902
+ size_t errorCode;
903
+
904
+ if (cSrcSize<2) return (size_t)-FSE_ERROR_srcSize_wrong; /* too small input size */
905
+
906
+ /* normal FSE decoding mode */
907
+ errorCode = FSE_readNCount (counting, &maxSymbolValue, &tableLog, istart, cSrcSize);
908
+ if (FSE_isError(errorCode)) return errorCode;
909
+ if (errorCode >= cSrcSize) return (size_t)-FSE_ERROR_srcSize_wrong; /* too small input size */
910
+ ip += errorCode;
911
+ cSrcSize -= errorCode;
912
+
913
+ errorCode = FSE_buildDTable (dt, counting, maxSymbolValue, tableLog);
914
+ if (FSE_isError(errorCode)) return errorCode;
915
+
916
+ /* always return, even if it is an error code */
917
+ return FSE_decompress_usingDTable (dst, maxDstSize, ip, cSrcSize, dt);
918
+ }
919
+
920
+
921
+
922
+ /* *******************************************************
923
+ * Huff0 : Huffman block compression
924
+ *********************************************************/
925
+ #define HUF_MAX_SYMBOL_VALUE 255
926
+ #define HUF_DEFAULT_TABLELOG 12 /* used by default, when not specified */
927
+ #define HUF_MAX_TABLELOG 12 /* max possible tableLog; for allocation purpose; can be modified */
928
+ #define HUF_ABSOLUTEMAX_TABLELOG 16 /* absolute limit of HUF_MAX_TABLELOG. Beyond that value, code does not work */
929
+ #if (HUF_MAX_TABLELOG > HUF_ABSOLUTEMAX_TABLELOG)
930
+ # error "HUF_MAX_TABLELOG is too large !"
931
+ #endif
932
+
933
+ typedef struct HUF_CElt_s {
934
+ U16 val;
935
+ BYTE nbBits;
936
+ } HUF_CElt ;
937
+
938
+ typedef struct nodeElt_s {
939
+ U32 count;
940
+ U16 parent;
941
+ BYTE byte;
942
+ BYTE nbBits;
943
+ } nodeElt;
944
+
945
+
946
+ /* *******************************************************
947
+ * Huff0 : Huffman block decompression
948
+ *********************************************************/
949
+ typedef struct {
950
+ BYTE byte;
951
+ BYTE nbBits;
952
+ } HUF_DElt;
953
+
954
+ static size_t HUF_readDTable (U16* DTable, const void* src, size_t srcSize)
955
+ {
956
+ BYTE huffWeight[HUF_MAX_SYMBOL_VALUE + 1];
957
+ U32 rankVal[HUF_ABSOLUTEMAX_TABLELOG + 1]; /* large enough for values from 0 to 16 */
958
+ U32 weightTotal;
959
+ U32 maxBits;
960
+ const BYTE* ip = (const BYTE*) src;
961
+ size_t iSize;
962
+ size_t oSize;
963
+ U32 n;
964
+ U32 nextRankStart;
965
+ void* ptr = DTable+1;
966
+ HUF_DElt* const dt = (HUF_DElt*)ptr;
967
+
968
+ if (!srcSize) return (size_t)-FSE_ERROR_srcSize_wrong;
969
+ iSize = ip[0];
970
+
971
+ FSE_STATIC_ASSERT(sizeof(HUF_DElt) == sizeof(U16)); /* if compilation fails here, assertion is false */
972
+ //memset(huffWeight, 0, sizeof(huffWeight)); /* should not be necessary, but some analyzer complain ... */
973
+ if (iSize >= 128) /* special header */
974
+ {
975
+ if (iSize >= (242)) /* RLE */
976
+ {
977
+ static int l[14] = { 1, 2, 3, 4, 7, 8, 15, 16, 31, 32, 63, 64, 127, 128 };
978
+ oSize = l[iSize-242];
979
+ memset(huffWeight, 1, sizeof(huffWeight));
980
+ iSize = 0;
981
+ }
982
+ else /* Incompressible */
983
+ {
984
+ oSize = iSize - 127;
985
+ iSize = ((oSize+1)/2);
986
+ if (iSize+1 > srcSize) return (size_t)-FSE_ERROR_srcSize_wrong;
987
+ ip += 1;
988
+ for (n=0; n<oSize; n+=2)
989
+ {
990
+ huffWeight[n] = ip[n/2] >> 4;
991
+ huffWeight[n+1] = ip[n/2] & 15;
992
+ }
993
+ }
994
+ }
995
+ else /* header compressed with FSE (normal case) */
996
+ {
997
+ if (iSize+1 > srcSize) return (size_t)-FSE_ERROR_srcSize_wrong;
998
+ oSize = FSE_decompress(huffWeight, HUF_MAX_SYMBOL_VALUE, ip+1, iSize); /* max 255 values decoded, last one is implied */
999
+ if (FSE_isError(oSize)) return oSize;
1000
+ }
1001
+
1002
+ /* collect weight stats */
1003
+ memset(rankVal, 0, sizeof(rankVal));
1004
+ weightTotal = 0;
1005
+ for (n=0; n<oSize; n++)
1006
+ {
1007
+ if (huffWeight[n] >= HUF_ABSOLUTEMAX_TABLELOG) return (size_t)-FSE_ERROR_corruptionDetected;
1008
+ rankVal[huffWeight[n]]++;
1009
+ weightTotal += (1 << huffWeight[n]) >> 1;
1010
+ }
1011
+ if (weightTotal == 0) return (size_t)-FSE_ERROR_corruptionDetected;
1012
+
1013
+ /* get last non-null symbol weight (implied, total must be 2^n) */
1014
+ maxBits = FSE_highbit32(weightTotal) + 1;
1015
+ if (maxBits > DTable[0]) return (size_t)-FSE_ERROR_tableLog_tooLarge; /* DTable is too small */
1016
+ DTable[0] = (U16)maxBits;
1017
+ {
1018
+ U32 total = 1 << maxBits;
1019
+ U32 rest = total - weightTotal;
1020
+ U32 verif = 1 << FSE_highbit32(rest);
1021
+ U32 lastWeight = FSE_highbit32(rest) + 1;
1022
+ if (verif != rest) return (size_t)-FSE_ERROR_corruptionDetected; /* last value must be a clean power of 2 */
1023
+ huffWeight[oSize] = (BYTE)lastWeight;
1024
+ rankVal[lastWeight]++;
1025
+ }
1026
+
1027
+ /* check tree construction validity */
1028
+ if ((rankVal[1] < 2) || (rankVal[1] & 1)) return (size_t)-FSE_ERROR_corruptionDetected; /* by construction : at least 2 elts of rank 1, must be even */
1029
+
1030
+ /* Prepare ranks */
1031
+ nextRankStart = 0;
1032
+ for (n=1; n<=maxBits; n++)
1033
+ {
1034
+ U32 current = nextRankStart;
1035
+ nextRankStart += (rankVal[n] << (n-1));
1036
+ rankVal[n] = current;
1037
+ }
1038
+
1039
+ /* fill DTable */
1040
+ for (n=0; n<=oSize; n++)
1041
+ {
1042
+ const U32 w = huffWeight[n];
1043
+ const U32 length = (1 << w) >> 1;
1044
+ U32 i;
1045
+ HUF_DElt D;
1046
+ D.byte = (BYTE)n; D.nbBits = (BYTE)(maxBits + 1 - w);
1047
+ for (i = rankVal[w]; i < rankVal[w] + length; i++)
1048
+ dt[i] = D;
1049
+ rankVal[w] += length;
1050
+ }
1051
+
1052
+ return iSize+1;
1053
+ }
1054
+
1055
+
1056
+ static BYTE HUF_decodeSymbol(FSE_DStream_t* Dstream, const HUF_DElt* dt, const U32 dtLog)
1057
+ {
1058
+ const size_t val = FSE_lookBitsFast(Dstream, dtLog); /* note : dtLog >= 1 */
1059
+ const BYTE c = dt[val].byte;
1060
+ FSE_skipBits(Dstream, dt[val].nbBits);
1061
+ return c;
1062
+ }
1063
+
1064
+ static size_t HUF_decompress_usingDTable( /* -3% slower when non static */
1065
+ void* dst, size_t maxDstSize,
1066
+ const void* cSrc, size_t cSrcSize,
1067
+ const U16* DTable)
1068
+ {
1069
+ BYTE* const ostart = (BYTE*) dst;
1070
+ BYTE* op = ostart;
1071
+ BYTE* const omax = op + maxDstSize;
1072
+ BYTE* const olimit = omax-15;
1073
+
1074
+ const void* ptr = DTable;
1075
+ const HUF_DElt* const dt = (const HUF_DElt*)(ptr)+1;
1076
+ const U32 dtLog = DTable[0];
1077
+ size_t errorCode;
1078
+ U32 reloadStatus;
1079
+
1080
+ /* Init */
1081
+
1082
+ const U16* jumpTable = (const U16*)cSrc;
1083
+ const size_t length1 = FSE_readLE16(jumpTable);
1084
+ const size_t length2 = FSE_readLE16(jumpTable+1);
1085
+ const size_t length3 = FSE_readLE16(jumpTable+2);
1086
+ const size_t length4 = cSrcSize - 6 - length1 - length2 - length3; // check coherency !!
1087
+ const char* const start1 = (const char*)(cSrc) + 6;
1088
+ const char* const start2 = start1 + length1;
1089
+ const char* const start3 = start2 + length2;
1090
+ const char* const start4 = start3 + length3;
1091
+ FSE_DStream_t bitD1, bitD2, bitD3, bitD4;
1092
+
1093
+ if (length1+length2+length3+6 >= cSrcSize) return (size_t)-FSE_ERROR_srcSize_wrong;
1094
+
1095
+ errorCode = FSE_initDStream(&bitD1, start1, length1);
1096
+ if (FSE_isError(errorCode)) return errorCode;
1097
+ errorCode = FSE_initDStream(&bitD2, start2, length2);
1098
+ if (FSE_isError(errorCode)) return errorCode;
1099
+ errorCode = FSE_initDStream(&bitD3, start3, length3);
1100
+ if (FSE_isError(errorCode)) return errorCode;
1101
+ errorCode = FSE_initDStream(&bitD4, start4, length4);
1102
+ if (FSE_isError(errorCode)) return errorCode;
1103
+
1104
+ reloadStatus=FSE_reloadDStream(&bitD2);
1105
+
1106
+ /* 16 symbols per loop */
1107
+ for ( ; (reloadStatus<FSE_DStream_completed) && (op<olimit); /* D2-3-4 are supposed to be synchronized and finish together */
1108
+ op+=16, reloadStatus = FSE_reloadDStream(&bitD2) | FSE_reloadDStream(&bitD3) | FSE_reloadDStream(&bitD4), FSE_reloadDStream(&bitD1))
1109
+ {
1110
+ #define HUF_DECODE_SYMBOL_0(n, Dstream) \
1111
+ op[n] = HUF_decodeSymbol(&Dstream, dt, dtLog);
1112
+
1113
+ #define HUF_DECODE_SYMBOL_1(n, Dstream) \
1114
+ op[n] = HUF_decodeSymbol(&Dstream, dt, dtLog); \
1115
+ if (FSE_32bits() && (HUF_MAX_TABLELOG>12)) FSE_reloadDStream(&Dstream)
1116
+
1117
+ #define HUF_DECODE_SYMBOL_2(n, Dstream) \
1118
+ op[n] = HUF_decodeSymbol(&Dstream, dt, dtLog); \
1119
+ if (FSE_32bits()) FSE_reloadDStream(&Dstream)
1120
+
1121
+ HUF_DECODE_SYMBOL_1( 0, bitD1);
1122
+ HUF_DECODE_SYMBOL_1( 1, bitD2);
1123
+ HUF_DECODE_SYMBOL_1( 2, bitD3);
1124
+ HUF_DECODE_SYMBOL_1( 3, bitD4);
1125
+ HUF_DECODE_SYMBOL_2( 4, bitD1);
1126
+ HUF_DECODE_SYMBOL_2( 5, bitD2);
1127
+ HUF_DECODE_SYMBOL_2( 6, bitD3);
1128
+ HUF_DECODE_SYMBOL_2( 7, bitD4);
1129
+ HUF_DECODE_SYMBOL_1( 8, bitD1);
1130
+ HUF_DECODE_SYMBOL_1( 9, bitD2);
1131
+ HUF_DECODE_SYMBOL_1(10, bitD3);
1132
+ HUF_DECODE_SYMBOL_1(11, bitD4);
1133
+ HUF_DECODE_SYMBOL_0(12, bitD1);
1134
+ HUF_DECODE_SYMBOL_0(13, bitD2);
1135
+ HUF_DECODE_SYMBOL_0(14, bitD3);
1136
+ HUF_DECODE_SYMBOL_0(15, bitD4);
1137
+ }
1138
+
1139
+ if (reloadStatus!=FSE_DStream_completed) /* not complete : some bitStream might be FSE_DStream_unfinished */
1140
+ return (size_t)-FSE_ERROR_corruptionDetected;
1141
+
1142
+ /* tail */
1143
+ {
1144
+ // bitTail = bitD1; // *much* slower : -20% !??!
1145
+ FSE_DStream_t bitTail;
1146
+ bitTail.ptr = bitD1.ptr;
1147
+ bitTail.bitsConsumed = bitD1.bitsConsumed;
1148
+ bitTail.bitContainer = bitD1.bitContainer; // required in case of FSE_DStream_endOfBuffer
1149
+ bitTail.start = start1;
1150
+ for ( ; (FSE_reloadDStream(&bitTail) < FSE_DStream_completed) && (op<omax) ; op++)
1151
+ {
1152
+ HUF_DECODE_SYMBOL_0(0, bitTail);
1153
+ }
1154
+
1155
+ if (FSE_endOfDStream(&bitTail))
1156
+ return op-ostart;
1157
+ }
1158
+
1159
+ if (op==omax) return (size_t)-FSE_ERROR_dstSize_tooSmall; /* dst buffer is full, but cSrc unfinished */
1160
+
1161
+ return (size_t)-FSE_ERROR_corruptionDetected;
1162
+ }
1163
+
1164
+
1165
+ static size_t HUF_decompress (void* dst, size_t maxDstSize, const void* cSrc, size_t cSrcSize)
1166
+ {
1167
+ HUF_CREATE_STATIC_DTABLE(DTable, HUF_MAX_TABLELOG);
1168
+ const BYTE* ip = (const BYTE*) cSrc;
1169
+ size_t errorCode;
1170
+
1171
+ errorCode = HUF_readDTable (DTable, cSrc, cSrcSize);
1172
+ if (FSE_isError(errorCode)) return errorCode;
1173
+ if (errorCode >= cSrcSize) return (size_t)-FSE_ERROR_srcSize_wrong;
1174
+ ip += errorCode;
1175
+ cSrcSize -= errorCode;
1176
+
1177
+ return HUF_decompress_usingDTable (dst, maxDstSize, ip, cSrcSize, DTable);
1178
+ }
1179
+
1180
+
1181
+ #endif /* FSE_COMMONDEFS_ONLY */
1182
+
1183
+ /*
1184
+ zstd - standard compression library
1185
+ Copyright (C) 2014-2015, Yann Collet.
1186
+
1187
+ BSD 2-Clause License (http://www.opensource.org/licenses/bsd-license.php)
1188
+
1189
+ Redistribution and use in source and binary forms, with or without
1190
+ modification, are permitted provided that the following conditions are
1191
+ met:
1192
+ * Redistributions of source code must retain the above copyright
1193
+ notice, this list of conditions and the following disclaimer.
1194
+ * Redistributions in binary form must reproduce the above
1195
+ copyright notice, this list of conditions and the following disclaimer
1196
+ in the documentation and/or other materials provided with the
1197
+ distribution.
1198
+ THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
1199
+ "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
1200
+ LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
1201
+ A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
1202
+ OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
1203
+ SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
1204
+ LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
1205
+ DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
1206
+ THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
1207
+ (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
1208
+ OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
1209
+
1210
+ You can contact the author at :
1211
+ - zstd source repository : https://github.com/Cyan4973/zstd
1212
+ - ztsd public forum : https://groups.google.com/forum/#!forum/lz4c
1213
+ */
1214
+
1215
+ /****************************************************************
1216
+ * Tuning parameters
1217
+ *****************************************************************/
1218
+ /* MEMORY_USAGE :
1219
+ * Memory usage formula : N->2^N Bytes (examples : 10 -> 1KB; 12 -> 4KB ; 16 -> 64KB; 20 -> 1MB; etc.)
1220
+ * Increasing memory usage improves compression ratio
1221
+ * Reduced memory usage can improve speed, due to cache effect */
1222
+ #define ZSTD_MEMORY_USAGE 17
1223
+
1224
+
1225
+ /**************************************
1226
+ CPU Feature Detection
1227
+ **************************************/
1228
+ /*
1229
+ * Automated efficient unaligned memory access detection
1230
+ * Based on known hardware architectures
1231
+ * This list will be updated thanks to feedbacks
1232
+ */
1233
+ #if defined(CPU_HAS_EFFICIENT_UNALIGNED_MEMORY_ACCESS) \
1234
+ || defined(__ARM_FEATURE_UNALIGNED) \
1235
+ || defined(__i386__) || defined(__x86_64__) \
1236
+ || defined(_M_IX86) || defined(_M_X64) \
1237
+ || defined(__ARM_ARCH_7__) || defined(__ARM_ARCH_8__) \
1238
+ || (defined(_M_ARM) && (_M_ARM >= 7))
1239
+ # define ZSTD_UNALIGNED_ACCESS 1
1240
+ #else
1241
+ # define ZSTD_UNALIGNED_ACCESS 0
1242
+ #endif
1243
+
1244
+
1245
+ /********************************************************
1246
+ * Includes
1247
+ *********************************************************/
1248
+ #include <stdlib.h> /* calloc */
1249
+ #include <string.h> /* memcpy, memmove */
1250
+ #include <stdio.h> /* debug : printf */
1251
+
1252
+
1253
+ /********************************************************
1254
+ * Compiler specifics
1255
+ *********************************************************/
1256
+ #ifdef __AVX2__
1257
+ # include <immintrin.h> /* AVX2 intrinsics */
1258
+ #endif
1259
+
1260
+ #ifdef _MSC_VER /* Visual Studio */
1261
+ # include <intrin.h> /* For Visual 2005 */
1262
+ # pragma warning(disable : 4127) /* disable: C4127: conditional expression is constant */
1263
+ # pragma warning(disable : 4324) /* disable: C4324: padded structure */
1264
+ #endif
1265
+
1266
+
1267
+ #ifndef MEM_ACCESS_MODULE
1268
+ #define MEM_ACCESS_MODULE
1269
+ /********************************************************
1270
+ * Basic Types
1271
+ *********************************************************/
1272
+ #if defined (__STDC_VERSION__) && __STDC_VERSION__ >= 199901L /* C99 */
1273
+ # include <stdint.h>
1274
+ typedef uint8_t BYTE;
1275
+ typedef uint16_t U16;
1276
+ typedef int16_t S16;
1277
+ typedef uint32_t U32;
1278
+ typedef int32_t S32;
1279
+ typedef uint64_t U64;
1280
+ #else
1281
+ typedef unsigned char BYTE;
1282
+ typedef unsigned short U16;
1283
+ typedef signed short S16;
1284
+ typedef unsigned int U32;
1285
+ typedef signed int S32;
1286
+ typedef unsigned long long U64;
1287
+ #endif
1288
+
1289
+ #endif /* MEM_ACCESS_MODULE */
1290
+
1291
+
1292
+ /********************************************************
1293
+ * Constants
1294
+ *********************************************************/
1295
+ static const U32 ZSTD_magicNumber = 0xFD2FB51E; /* 3rd version : seqNb header */
1296
+
1297
+ #define HASH_LOG (ZSTD_MEMORY_USAGE - 2)
1298
+ #define HASH_TABLESIZE (1 << HASH_LOG)
1299
+ #define HASH_MASK (HASH_TABLESIZE - 1)
1300
+
1301
+ #define KNUTH 2654435761
1302
+
1303
+ #define BIT7 128
1304
+ #define BIT6 64
1305
+ #define BIT5 32
1306
+ #define BIT4 16
1307
+
1308
+ #define KB *(1 <<10)
1309
+ #define MB *(1 <<20)
1310
+ #define GB *(1U<<30)
1311
+
1312
+ #define BLOCKSIZE (128 KB) /* define, for static allocation */
1313
+
1314
+ #define WORKPLACESIZE (BLOCKSIZE*3)
1315
+ #define MINMATCH 4
1316
+ #define MLbits 7
1317
+ #define LLbits 6
1318
+ #define Offbits 5
1319
+ #define MaxML ((1<<MLbits )-1)
1320
+ #define MaxLL ((1<<LLbits )-1)
1321
+ #define MaxOff ((1<<Offbits)-1)
1322
+ #define LitFSELog 11
1323
+ #define MLFSELog 10
1324
+ #define LLFSELog 10
1325
+ #define OffFSELog 9
1326
+ #define MAX(a,b) ((a)<(b)?(b):(a))
1327
+ #define MaxSeq MAX(MaxLL, MaxML)
1328
+
1329
+ #define LITERAL_NOENTROPY 63
1330
+ #define COMMAND_NOENTROPY 7 /* to remove */
1331
+
1332
+ static const size_t ZSTD_blockHeaderSize = 3;
1333
+ static const size_t ZSTD_frameHeaderSize = 4;
1334
+
1335
+
1336
+ /********************************************************
1337
+ * Memory operations
1338
+ *********************************************************/
1339
+ static unsigned ZSTD_32bits(void) { return sizeof(void*)==4; }
1340
+
1341
+ static unsigned ZSTD_isLittleEndian(void)
1342
+ {
1343
+ const union { U32 i; BYTE c[4]; } one = { 1 }; /* don't use static : performance detrimental */
1344
+ return one.c[0];
1345
+ }
1346
+
1347
+ static U16 ZSTD_read16(const void* p) { U16 r; memcpy(&r, p, sizeof(r)); return r; }
1348
+
1349
+ static U32 ZSTD_read32(const void* p) { U32 r; memcpy(&r, p, sizeof(r)); return r; }
1350
+
1351
+ static void ZSTD_copy4(void* dst, const void* src) { memcpy(dst, src, 4); }
1352
+
1353
+ static void ZSTD_copy8(void* dst, const void* src) { memcpy(dst, src, 8); }
1354
+
1355
+ #define COPY8(d,s) { ZSTD_copy8(d,s); d+=8; s+=8; }
1356
+
1357
+ static void ZSTD_wildcopy(void* dst, const void* src, ptrdiff_t length)
1358
+ {
1359
+ const BYTE* ip = (const BYTE*)src;
1360
+ BYTE* op = (BYTE*)dst;
1361
+ BYTE* const oend = op + length;
1362
+ while (op < oend) COPY8(op, ip);
1363
+ }
1364
+
1365
+ static U16 ZSTD_readLE16(const void* memPtr)
1366
+ {
1367
+ if (ZSTD_isLittleEndian()) return ZSTD_read16(memPtr);
1368
+ else
1369
+ {
1370
+ const BYTE* p = (const BYTE*)memPtr;
1371
+ return (U16)((U16)p[0] + ((U16)p[1]<<8));
1372
+ }
1373
+ }
1374
+
1375
+
1376
+ static U32 ZSTD_readLE32(const void* memPtr)
1377
+ {
1378
+ if (ZSTD_isLittleEndian())
1379
+ return ZSTD_read32(memPtr);
1380
+ else
1381
+ {
1382
+ const BYTE* p = (const BYTE*)memPtr;
1383
+ return (U32)((U32)p[0] + ((U32)p[1]<<8) + ((U32)p[2]<<16) + ((U32)p[3]<<24));
1384
+ }
1385
+ }
1386
+
1387
+ static U32 ZSTD_readBE32(const void* memPtr)
1388
+ {
1389
+ const BYTE* p = (const BYTE*)memPtr;
1390
+ return (U32)(((U32)p[0]<<24) + ((U32)p[1]<<16) + ((U32)p[2]<<8) + ((U32)p[3]<<0));
1391
+ }
1392
+
1393
+
1394
+ /**************************************
1395
+ * Local structures
1396
+ ***************************************/
1397
+ typedef struct ZSTD_Cctx_s ZSTD_Cctx;
1398
+
1399
+ typedef enum { bt_compressed, bt_raw, bt_rle, bt_end } blockType_t;
1400
+
1401
+ typedef struct
1402
+ {
1403
+ blockType_t blockType;
1404
+ U32 origSize;
1405
+ } blockProperties_t;
1406
+
1407
+ typedef struct {
1408
+ void* buffer;
1409
+ U32* offsetStart;
1410
+ U32* offset;
1411
+ BYTE* offCodeStart;
1412
+ BYTE* offCode;
1413
+ BYTE* litStart;
1414
+ BYTE* lit;
1415
+ BYTE* litLengthStart;
1416
+ BYTE* litLength;
1417
+ BYTE* matchLengthStart;
1418
+ BYTE* matchLength;
1419
+ BYTE* dumpsStart;
1420
+ BYTE* dumps;
1421
+ } seqStore_t;
1422
+
1423
+
1424
+ typedef struct ZSTD_Cctx_s
1425
+ {
1426
+ const BYTE* base;
1427
+ U32 current;
1428
+ U32 nextUpdate;
1429
+ seqStore_t seqStore;
1430
+ #ifdef __AVX2__
1431
+ __m256i hashTable[HASH_TABLESIZE>>3];
1432
+ #else
1433
+ U32 hashTable[HASH_TABLESIZE];
1434
+ #endif
1435
+ BYTE buffer[WORKPLACESIZE];
1436
+ } cctxi_t;
1437
+
1438
+
1439
+
1440
+
1441
+ /**************************************
1442
+ * Error Management
1443
+ **************************************/
1444
+ /* published entry point */
1445
+ unsigned ZSTDv01_isError(size_t code) { return ERR_isError(code); }
1446
+
1447
+
1448
+ /**************************************
1449
+ * Tool functions
1450
+ **************************************/
1451
+ #define ZSTD_VERSION_MAJOR 0 /* for breaking interface changes */
1452
+ #define ZSTD_VERSION_MINOR 1 /* for new (non-breaking) interface capabilities */
1453
+ #define ZSTD_VERSION_RELEASE 3 /* for tweaks, bug-fixes, or development */
1454
+ #define ZSTD_VERSION_NUMBER (ZSTD_VERSION_MAJOR *100*100 + ZSTD_VERSION_MINOR *100 + ZSTD_VERSION_RELEASE)
1455
+
1456
+ /**************************************************************
1457
+ * Decompression code
1458
+ **************************************************************/
1459
+
1460
+ size_t ZSTDv01_getcBlockSize(const void* src, size_t srcSize, blockProperties_t* bpPtr)
1461
+ {
1462
+ const BYTE* const in = (const BYTE* const)src;
1463
+ BYTE headerFlags;
1464
+ U32 cSize;
1465
+
1466
+ if (srcSize < 3) return ERROR(srcSize_wrong);
1467
+
1468
+ headerFlags = *in;
1469
+ cSize = in[2] + (in[1]<<8) + ((in[0] & 7)<<16);
1470
+
1471
+ bpPtr->blockType = (blockType_t)(headerFlags >> 6);
1472
+ bpPtr->origSize = (bpPtr->blockType == bt_rle) ? cSize : 0;
1473
+
1474
+ if (bpPtr->blockType == bt_end) return 0;
1475
+ if (bpPtr->blockType == bt_rle) return 1;
1476
+ return cSize;
1477
+ }
1478
+
1479
+
1480
+ static size_t ZSTD_copyUncompressedBlock(void* dst, size_t maxDstSize, const void* src, size_t srcSize)
1481
+ {
1482
+ if (srcSize > maxDstSize) return ERROR(dstSize_tooSmall);
1483
+ memcpy(dst, src, srcSize);
1484
+ return srcSize;
1485
+ }
1486
+
1487
+
1488
+ static size_t ZSTD_decompressLiterals(void* ctx,
1489
+ void* dst, size_t maxDstSize,
1490
+ const void* src, size_t srcSize)
1491
+ {
1492
+ BYTE* op = (BYTE*)dst;
1493
+ BYTE* const oend = op + maxDstSize;
1494
+ const BYTE* ip = (const BYTE*)src;
1495
+ size_t errorCode;
1496
+ size_t litSize;
1497
+
1498
+ /* check : minimum 2, for litSize, +1, for content */
1499
+ if (srcSize <= 3) return ERROR(corruption_detected);
1500
+
1501
+ litSize = ip[1] + (ip[0]<<8);
1502
+ litSize += ((ip[-3] >> 3) & 7) << 16; // mmmmh....
1503
+ op = oend - litSize;
1504
+
1505
+ (void)ctx;
1506
+ if (litSize > maxDstSize) return ERROR(dstSize_tooSmall);
1507
+ errorCode = HUF_decompress(op, litSize, ip+2, srcSize-2);
1508
+ if (FSE_isError(errorCode)) return ERROR(GENERIC);
1509
+ return litSize;
1510
+ }
1511
+
1512
+
1513
+ size_t ZSTDv01_decodeLiteralsBlock(void* ctx,
1514
+ void* dst, size_t maxDstSize,
1515
+ const BYTE** litStart, size_t* litSize,
1516
+ const void* src, size_t srcSize)
1517
+ {
1518
+ const BYTE* const istart = (const BYTE* const)src;
1519
+ const BYTE* ip = istart;
1520
+ BYTE* const ostart = (BYTE* const)dst;
1521
+ BYTE* const oend = ostart + maxDstSize;
1522
+ blockProperties_t litbp;
1523
+
1524
+ size_t litcSize = ZSTDv01_getcBlockSize(src, srcSize, &litbp);
1525
+ if (ZSTDv01_isError(litcSize)) return litcSize;
1526
+ if (litcSize > srcSize - ZSTD_blockHeaderSize) return ERROR(srcSize_wrong);
1527
+ ip += ZSTD_blockHeaderSize;
1528
+
1529
+ switch(litbp.blockType)
1530
+ {
1531
+ case bt_raw:
1532
+ *litStart = ip;
1533
+ ip += litcSize;
1534
+ *litSize = litcSize;
1535
+ break;
1536
+ case bt_rle:
1537
+ {
1538
+ size_t rleSize = litbp.origSize;
1539
+ if (rleSize>maxDstSize) return ERROR(dstSize_tooSmall);
1540
+ if (!srcSize) return ERROR(srcSize_wrong);
1541
+ memset(oend - rleSize, *ip, rleSize);
1542
+ *litStart = oend - rleSize;
1543
+ *litSize = rleSize;
1544
+ ip++;
1545
+ break;
1546
+ }
1547
+ case bt_compressed:
1548
+ {
1549
+ size_t decodedLitSize = ZSTD_decompressLiterals(ctx, dst, maxDstSize, ip, litcSize);
1550
+ if (ZSTDv01_isError(decodedLitSize)) return decodedLitSize;
1551
+ *litStart = oend - decodedLitSize;
1552
+ *litSize = decodedLitSize;
1553
+ ip += litcSize;
1554
+ break;
1555
+ }
1556
+ case bt_end:
1557
+ default:
1558
+ return ERROR(GENERIC);
1559
+ }
1560
+
1561
+ return ip-istart;
1562
+ }
1563
+
1564
+
1565
+ size_t ZSTDv01_decodeSeqHeaders(int* nbSeq, const BYTE** dumpsPtr, size_t* dumpsLengthPtr,
1566
+ FSE_DTable* DTableLL, FSE_DTable* DTableML, FSE_DTable* DTableOffb,
1567
+ const void* src, size_t srcSize)
1568
+ {
1569
+ const BYTE* const istart = (const BYTE* const)src;
1570
+ const BYTE* ip = istart;
1571
+ const BYTE* const iend = istart + srcSize;
1572
+ U32 LLtype, Offtype, MLtype;
1573
+ U32 LLlog, Offlog, MLlog;
1574
+ size_t dumpsLength;
1575
+
1576
+ /* check */
1577
+ if (srcSize < 5) return ERROR(srcSize_wrong);
1578
+
1579
+ /* SeqHead */
1580
+ *nbSeq = ZSTD_readLE16(ip); ip+=2;
1581
+ LLtype = *ip >> 6;
1582
+ Offtype = (*ip >> 4) & 3;
1583
+ MLtype = (*ip >> 2) & 3;
1584
+ if (*ip & 2)
1585
+ {
1586
+ dumpsLength = ip[2];
1587
+ dumpsLength += ip[1] << 8;
1588
+ ip += 3;
1589
+ }
1590
+ else
1591
+ {
1592
+ dumpsLength = ip[1];
1593
+ dumpsLength += (ip[0] & 1) << 8;
1594
+ ip += 2;
1595
+ }
1596
+ *dumpsPtr = ip;
1597
+ ip += dumpsLength;
1598
+ *dumpsLengthPtr = dumpsLength;
1599
+
1600
+ /* check */
1601
+ if (ip > iend-3) return ERROR(srcSize_wrong); /* min : all 3 are "raw", hence no header, but at least xxLog bits per type */
1602
+
1603
+ /* sequences */
1604
+ {
1605
+ S16 norm[MaxML+1]; /* assumption : MaxML >= MaxLL and MaxOff */
1606
+ size_t headerSize;
1607
+
1608
+ /* Build DTables */
1609
+ switch(LLtype)
1610
+ {
1611
+ case bt_rle :
1612
+ LLlog = 0;
1613
+ FSE_buildDTable_rle(DTableLL, *ip++); break;
1614
+ case bt_raw :
1615
+ LLlog = LLbits;
1616
+ FSE_buildDTable_raw(DTableLL, LLbits); break;
1617
+ default :
1618
+ { U32 max = MaxLL;
1619
+ headerSize = FSE_readNCount(norm, &max, &LLlog, ip, iend-ip);
1620
+ if (FSE_isError(headerSize)) return ERROR(GENERIC);
1621
+ if (LLlog > LLFSELog) return ERROR(corruption_detected);
1622
+ ip += headerSize;
1623
+ FSE_buildDTable(DTableLL, norm, max, LLlog);
1624
+ } }
1625
+
1626
+ switch(Offtype)
1627
+ {
1628
+ case bt_rle :
1629
+ Offlog = 0;
1630
+ if (ip > iend-2) return ERROR(srcSize_wrong); /* min : "raw", hence no header, but at least xxLog bits */
1631
+ FSE_buildDTable_rle(DTableOffb, *ip++); break;
1632
+ case bt_raw :
1633
+ Offlog = Offbits;
1634
+ FSE_buildDTable_raw(DTableOffb, Offbits); break;
1635
+ default :
1636
+ { U32 max = MaxOff;
1637
+ headerSize = FSE_readNCount(norm, &max, &Offlog, ip, iend-ip);
1638
+ if (FSE_isError(headerSize)) return ERROR(GENERIC);
1639
+ if (Offlog > OffFSELog) return ERROR(corruption_detected);
1640
+ ip += headerSize;
1641
+ FSE_buildDTable(DTableOffb, norm, max, Offlog);
1642
+ } }
1643
+
1644
+ switch(MLtype)
1645
+ {
1646
+ case bt_rle :
1647
+ MLlog = 0;
1648
+ if (ip > iend-2) return ERROR(srcSize_wrong); /* min : "raw", hence no header, but at least xxLog bits */
1649
+ FSE_buildDTable_rle(DTableML, *ip++); break;
1650
+ case bt_raw :
1651
+ MLlog = MLbits;
1652
+ FSE_buildDTable_raw(DTableML, MLbits); break;
1653
+ default :
1654
+ { U32 max = MaxML;
1655
+ headerSize = FSE_readNCount(norm, &max, &MLlog, ip, iend-ip);
1656
+ if (FSE_isError(headerSize)) return ERROR(GENERIC);
1657
+ if (MLlog > MLFSELog) return ERROR(corruption_detected);
1658
+ ip += headerSize;
1659
+ FSE_buildDTable(DTableML, norm, max, MLlog);
1660
+ } } }
1661
+
1662
+ return ip-istart;
1663
+ }
1664
+
1665
+
1666
+ typedef struct {
1667
+ size_t litLength;
1668
+ size_t offset;
1669
+ size_t matchLength;
1670
+ } seq_t;
1671
+
1672
+ typedef struct {
1673
+ FSE_DStream_t DStream;
1674
+ FSE_DState_t stateLL;
1675
+ FSE_DState_t stateOffb;
1676
+ FSE_DState_t stateML;
1677
+ size_t prevOffset;
1678
+ const BYTE* dumps;
1679
+ const BYTE* dumpsEnd;
1680
+ } seqState_t;
1681
+
1682
+
1683
+ static void ZSTD_decodeSequence(seq_t* seq, seqState_t* seqState)
1684
+ {
1685
+ size_t litLength;
1686
+ size_t prevOffset;
1687
+ size_t offset;
1688
+ size_t matchLength;
1689
+ const BYTE* dumps = seqState->dumps;
1690
+ const BYTE* const de = seqState->dumpsEnd;
1691
+
1692
+ /* Literal length */
1693
+ litLength = FSE_decodeSymbol(&(seqState->stateLL), &(seqState->DStream));
1694
+ prevOffset = litLength ? seq->offset : seqState->prevOffset;
1695
+ seqState->prevOffset = seq->offset;
1696
+ if (litLength == MaxLL)
1697
+ {
1698
+ U32 add = dumps<de ? *dumps++ : 0;
1699
+ if (add < 255) litLength += add;
1700
+ else
1701
+ {
1702
+ if (dumps<=(de-3))
1703
+ {
1704
+ litLength = ZSTD_readLE32(dumps) & 0xFFFFFF; /* no pb : dumps is always followed by seq tables > 1 byte */
1705
+ dumps += 3;
1706
+ }
1707
+ }
1708
+ }
1709
+
1710
+ /* Offset */
1711
+ {
1712
+ U32 offsetCode, nbBits;
1713
+ offsetCode = FSE_decodeSymbol(&(seqState->stateOffb), &(seqState->DStream));
1714
+ if (ZSTD_32bits()) FSE_reloadDStream(&(seqState->DStream));
1715
+ nbBits = offsetCode - 1;
1716
+ if (offsetCode==0) nbBits = 0; /* cmove */
1717
+ offset = ((size_t)1 << (nbBits & ((sizeof(offset)*8)-1))) + FSE_readBits(&(seqState->DStream), nbBits);
1718
+ if (ZSTD_32bits()) FSE_reloadDStream(&(seqState->DStream));
1719
+ if (offsetCode==0) offset = prevOffset;
1720
+ }
1721
+
1722
+ /* MatchLength */
1723
+ matchLength = FSE_decodeSymbol(&(seqState->stateML), &(seqState->DStream));
1724
+ if (matchLength == MaxML)
1725
+ {
1726
+ U32 add = dumps<de ? *dumps++ : 0;
1727
+ if (add < 255) matchLength += add;
1728
+ else
1729
+ {
1730
+ if (dumps<=(de-3))
1731
+ {
1732
+ matchLength = ZSTD_readLE32(dumps) & 0xFFFFFF; /* no pb : dumps is always followed by seq tables > 1 byte */
1733
+ dumps += 3;
1734
+ }
1735
+ }
1736
+ }
1737
+ matchLength += MINMATCH;
1738
+
1739
+ /* save result */
1740
+ seq->litLength = litLength;
1741
+ seq->offset = offset;
1742
+ seq->matchLength = matchLength;
1743
+ seqState->dumps = dumps;
1744
+ }
1745
+
1746
+
1747
+ static size_t ZSTD_execSequence(BYTE* op,
1748
+ seq_t sequence,
1749
+ const BYTE** litPtr, const BYTE* const litLimit,
1750
+ BYTE* const base, BYTE* const oend)
1751
+ {
1752
+ static const int dec32table[] = {0, 1, 2, 1, 4, 4, 4, 4}; /* added */
1753
+ static const int dec64table[] = {8, 8, 8, 7, 8, 9,10,11}; /* substracted */
1754
+ const BYTE* const ostart = op;
1755
+ const size_t litLength = sequence.litLength;
1756
+ BYTE* const endMatch = op + litLength + sequence.matchLength; /* risk : address space overflow (32-bits) */
1757
+ const BYTE* const litEnd = *litPtr + litLength;
1758
+
1759
+ /* check */
1760
+ if (endMatch > oend) return ERROR(dstSize_tooSmall); /* overwrite beyond dst buffer */
1761
+ if (litEnd > litLimit) return ERROR(corruption_detected);
1762
+ if (sequence.matchLength > (size_t)(*litPtr-op)) return ERROR(dstSize_tooSmall); /* overwrite literal segment */
1763
+
1764
+ /* copy Literals */
1765
+ if (((size_t)(*litPtr - op) < 8) || ((size_t)(oend-litEnd) < 8) || (op+litLength > oend-8))
1766
+ memmove(op, *litPtr, litLength); /* overwrite risk */
1767
+ else
1768
+ ZSTD_wildcopy(op, *litPtr, litLength);
1769
+ op += litLength;
1770
+ *litPtr = litEnd; /* update for next sequence */
1771
+
1772
+ /* check : last match must be at a minimum distance of 8 from end of dest buffer */
1773
+ if (oend-op < 8) return ERROR(dstSize_tooSmall);
1774
+
1775
+ /* copy Match */
1776
+ {
1777
+ const U32 overlapRisk = (((size_t)(litEnd - endMatch)) < 12);
1778
+ const BYTE* match = op - sequence.offset; /* possible underflow at op - offset ? */
1779
+ size_t qutt = 12;
1780
+ U64 saved[2];
1781
+
1782
+ /* check */
1783
+ if (match < base) return ERROR(corruption_detected);
1784
+ if (sequence.offset > (size_t)base) return ERROR(corruption_detected);
1785
+
1786
+ /* save beginning of literal sequence, in case of write overlap */
1787
+ if (overlapRisk)
1788
+ {
1789
+ if ((endMatch + qutt) > oend) qutt = oend-endMatch;
1790
+ memcpy(saved, endMatch, qutt);
1791
+ }
1792
+
1793
+ if (sequence.offset < 8)
1794
+ {
1795
+ const int dec64 = dec64table[sequence.offset];
1796
+ op[0] = match[0];
1797
+ op[1] = match[1];
1798
+ op[2] = match[2];
1799
+ op[3] = match[3];
1800
+ match += dec32table[sequence.offset];
1801
+ ZSTD_copy4(op+4, match);
1802
+ match -= dec64;
1803
+ } else { ZSTD_copy8(op, match); }
1804
+ op += 8; match += 8;
1805
+
1806
+ if (endMatch > oend-(16-MINMATCH))
1807
+ {
1808
+ if (op < oend-8)
1809
+ {
1810
+ ZSTD_wildcopy(op, match, (oend-8) - op);
1811
+ match += (oend-8) - op;
1812
+ op = oend-8;
1813
+ }
1814
+ while (op<endMatch) *op++ = *match++;
1815
+ }
1816
+ else
1817
+ ZSTD_wildcopy(op, match, (ptrdiff_t)sequence.matchLength-8); /* works even if matchLength < 8 */
1818
+
1819
+ /* restore, in case of overlap */
1820
+ if (overlapRisk) memcpy(endMatch, saved, qutt);
1821
+ }
1822
+
1823
+ return endMatch-ostart;
1824
+ }
1825
+
1826
+ typedef struct ZSTDv01_Dctx_s
1827
+ {
1828
+ U32 LLTable[FSE_DTABLE_SIZE_U32(LLFSELog)];
1829
+ U32 OffTable[FSE_DTABLE_SIZE_U32(OffFSELog)];
1830
+ U32 MLTable[FSE_DTABLE_SIZE_U32(MLFSELog)];
1831
+ void* previousDstEnd;
1832
+ void* base;
1833
+ size_t expected;
1834
+ blockType_t bType;
1835
+ U32 phase;
1836
+ } dctx_t;
1837
+
1838
+
1839
+ static size_t ZSTD_decompressSequences(
1840
+ void* ctx,
1841
+ void* dst, size_t maxDstSize,
1842
+ const void* seqStart, size_t seqSize,
1843
+ const BYTE* litStart, size_t litSize)
1844
+ {
1845
+ dctx_t* dctx = (dctx_t*)ctx;
1846
+ const BYTE* ip = (const BYTE*)seqStart;
1847
+ const BYTE* const iend = ip + seqSize;
1848
+ BYTE* const ostart = (BYTE* const)dst;
1849
+ BYTE* op = ostart;
1850
+ BYTE* const oend = ostart + maxDstSize;
1851
+ size_t errorCode, dumpsLength;
1852
+ const BYTE* litPtr = litStart;
1853
+ const BYTE* const litEnd = litStart + litSize;
1854
+ int nbSeq;
1855
+ const BYTE* dumps;
1856
+ U32* DTableLL = dctx->LLTable;
1857
+ U32* DTableML = dctx->MLTable;
1858
+ U32* DTableOffb = dctx->OffTable;
1859
+ BYTE* const base = (BYTE*) (dctx->base);
1860
+
1861
+ /* Build Decoding Tables */
1862
+ errorCode = ZSTDv01_decodeSeqHeaders(&nbSeq, &dumps, &dumpsLength,
1863
+ DTableLL, DTableML, DTableOffb,
1864
+ ip, iend-ip);
1865
+ if (ZSTDv01_isError(errorCode)) return errorCode;
1866
+ ip += errorCode;
1867
+
1868
+ /* Regen sequences */
1869
+ {
1870
+ seq_t sequence;
1871
+ seqState_t seqState;
1872
+
1873
+ memset(&sequence, 0, sizeof(sequence));
1874
+ seqState.dumps = dumps;
1875
+ seqState.dumpsEnd = dumps + dumpsLength;
1876
+ seqState.prevOffset = 1;
1877
+ errorCode = FSE_initDStream(&(seqState.DStream), ip, iend-ip);
1878
+ if (FSE_isError(errorCode)) return ERROR(corruption_detected);
1879
+ FSE_initDState(&(seqState.stateLL), &(seqState.DStream), DTableLL);
1880
+ FSE_initDState(&(seqState.stateOffb), &(seqState.DStream), DTableOffb);
1881
+ FSE_initDState(&(seqState.stateML), &(seqState.DStream), DTableML);
1882
+
1883
+ for ( ; (FSE_reloadDStream(&(seqState.DStream)) <= FSE_DStream_completed) && (nbSeq>0) ; )
1884
+ {
1885
+ size_t oneSeqSize;
1886
+ nbSeq--;
1887
+ ZSTD_decodeSequence(&sequence, &seqState);
1888
+ oneSeqSize = ZSTD_execSequence(op, sequence, &litPtr, litEnd, base, oend);
1889
+ if (ZSTDv01_isError(oneSeqSize)) return oneSeqSize;
1890
+ op += oneSeqSize;
1891
+ }
1892
+
1893
+ /* check if reached exact end */
1894
+ if ( !FSE_endOfDStream(&(seqState.DStream)) ) return ERROR(corruption_detected); /* requested too much : data is corrupted */
1895
+ if (nbSeq<0) return ERROR(corruption_detected); /* requested too many sequences : data is corrupted */
1896
+
1897
+ /* last literal segment */
1898
+ {
1899
+ size_t lastLLSize = litEnd - litPtr;
1900
+ if (op+lastLLSize > oend) return ERROR(dstSize_tooSmall);
1901
+ if (op != litPtr) memmove(op, litPtr, lastLLSize);
1902
+ op += lastLLSize;
1903
+ }
1904
+ }
1905
+
1906
+ return op-ostart;
1907
+ }
1908
+
1909
+
1910
+ static size_t ZSTD_decompressBlock(
1911
+ void* ctx,
1912
+ void* dst, size_t maxDstSize,
1913
+ const void* src, size_t srcSize)
1914
+ {
1915
+ /* blockType == blockCompressed, srcSize is trusted */
1916
+ const BYTE* ip = (const BYTE*)src;
1917
+ const BYTE* litPtr = NULL;
1918
+ size_t litSize = 0;
1919
+ size_t errorCode;
1920
+
1921
+ /* Decode literals sub-block */
1922
+ errorCode = ZSTDv01_decodeLiteralsBlock(ctx, dst, maxDstSize, &litPtr, &litSize, src, srcSize);
1923
+ if (ZSTDv01_isError(errorCode)) return errorCode;
1924
+ ip += errorCode;
1925
+ srcSize -= errorCode;
1926
+
1927
+ return ZSTD_decompressSequences(ctx, dst, maxDstSize, ip, srcSize, litPtr, litSize);
1928
+ }
1929
+
1930
+
1931
+ size_t ZSTDv01_decompressDCtx(void* ctx, void* dst, size_t maxDstSize, const void* src, size_t srcSize)
1932
+ {
1933
+ const BYTE* ip = (const BYTE*)src;
1934
+ const BYTE* iend = ip + srcSize;
1935
+ BYTE* const ostart = (BYTE* const)dst;
1936
+ BYTE* op = ostart;
1937
+ BYTE* const oend = ostart + maxDstSize;
1938
+ size_t remainingSize = srcSize;
1939
+ U32 magicNumber;
1940
+ size_t errorCode=0;
1941
+ blockProperties_t blockProperties;
1942
+
1943
+ /* Frame Header */
1944
+ if (srcSize < ZSTD_frameHeaderSize+ZSTD_blockHeaderSize) return ERROR(srcSize_wrong);
1945
+ magicNumber = ZSTD_readBE32(src);
1946
+ if (magicNumber != ZSTD_magicNumber) return ERROR(prefix_unknown);
1947
+ ip += ZSTD_frameHeaderSize; remainingSize -= ZSTD_frameHeaderSize;
1948
+
1949
+ /* Loop on each block */
1950
+ while (1)
1951
+ {
1952
+ size_t blockSize = ZSTDv01_getcBlockSize(ip, iend-ip, &blockProperties);
1953
+ if (ZSTDv01_isError(blockSize)) return blockSize;
1954
+
1955
+ ip += ZSTD_blockHeaderSize;
1956
+ remainingSize -= ZSTD_blockHeaderSize;
1957
+ if (blockSize > remainingSize) return ERROR(srcSize_wrong);
1958
+
1959
+ switch(blockProperties.blockType)
1960
+ {
1961
+ case bt_compressed:
1962
+ errorCode = ZSTD_decompressBlock(ctx, op, oend-op, ip, blockSize);
1963
+ break;
1964
+ case bt_raw :
1965
+ errorCode = ZSTD_copyUncompressedBlock(op, oend-op, ip, blockSize);
1966
+ break;
1967
+ case bt_rle :
1968
+ return ERROR(GENERIC); /* not yet supported */
1969
+ break;
1970
+ case bt_end :
1971
+ /* end of frame */
1972
+ if (remainingSize) return ERROR(srcSize_wrong);
1973
+ break;
1974
+ default:
1975
+ return ERROR(GENERIC);
1976
+ }
1977
+ if (blockSize == 0) break; /* bt_end */
1978
+
1979
+ if (ZSTDv01_isError(errorCode)) return errorCode;
1980
+ op += errorCode;
1981
+ ip += blockSize;
1982
+ remainingSize -= blockSize;
1983
+ }
1984
+
1985
+ return op-ostart;
1986
+ }
1987
+
1988
+ size_t ZSTDv01_decompress(void* dst, size_t maxDstSize, const void* src, size_t srcSize)
1989
+ {
1990
+ dctx_t ctx;
1991
+ ctx.base = dst;
1992
+ return ZSTDv01_decompressDCtx(&ctx, dst, maxDstSize, src, srcSize);
1993
+ }
1994
+
1995
+
1996
+ /*******************************
1997
+ * Streaming Decompression API
1998
+ *******************************/
1999
+
2000
+ size_t ZSTDv01_resetDCtx(ZSTDv01_Dctx* dctx)
2001
+ {
2002
+ dctx->expected = ZSTD_frameHeaderSize;
2003
+ dctx->phase = 0;
2004
+ dctx->previousDstEnd = NULL;
2005
+ dctx->base = NULL;
2006
+ return 0;
2007
+ }
2008
+
2009
+ ZSTDv01_Dctx* ZSTDv01_createDCtx(void)
2010
+ {
2011
+ ZSTDv01_Dctx* dctx = (ZSTDv01_Dctx*)malloc(sizeof(ZSTDv01_Dctx));
2012
+ if (dctx==NULL) return NULL;
2013
+ ZSTDv01_resetDCtx(dctx);
2014
+ return dctx;
2015
+ }
2016
+
2017
+ size_t ZSTDv01_freeDCtx(ZSTDv01_Dctx* dctx)
2018
+ {
2019
+ free(dctx);
2020
+ return 0;
2021
+ }
2022
+
2023
+ size_t ZSTDv01_nextSrcSizeToDecompress(ZSTDv01_Dctx* dctx)
2024
+ {
2025
+ return ((dctx_t*)dctx)->expected;
2026
+ }
2027
+
2028
+ size_t ZSTDv01_decompressContinue(ZSTDv01_Dctx* dctx, void* dst, size_t maxDstSize, const void* src, size_t srcSize)
2029
+ {
2030
+ dctx_t* ctx = (dctx_t*)dctx;
2031
+
2032
+ /* Sanity check */
2033
+ if (srcSize != ctx->expected) return ERROR(srcSize_wrong);
2034
+ if (dst != ctx->previousDstEnd) /* not contiguous */
2035
+ ctx->base = dst;
2036
+
2037
+ /* Decompress : frame header */
2038
+ if (ctx->phase == 0)
2039
+ {
2040
+ /* Check frame magic header */
2041
+ U32 magicNumber = ZSTD_readBE32(src);
2042
+ if (magicNumber != ZSTD_magicNumber) return ERROR(prefix_unknown);
2043
+ ctx->phase = 1;
2044
+ ctx->expected = ZSTD_blockHeaderSize;
2045
+ return 0;
2046
+ }
2047
+
2048
+ /* Decompress : block header */
2049
+ if (ctx->phase == 1)
2050
+ {
2051
+ blockProperties_t bp;
2052
+ size_t blockSize = ZSTDv01_getcBlockSize(src, ZSTD_blockHeaderSize, &bp);
2053
+ if (ZSTDv01_isError(blockSize)) return blockSize;
2054
+ if (bp.blockType == bt_end)
2055
+ {
2056
+ ctx->expected = 0;
2057
+ ctx->phase = 0;
2058
+ }
2059
+ else
2060
+ {
2061
+ ctx->expected = blockSize;
2062
+ ctx->bType = bp.blockType;
2063
+ ctx->phase = 2;
2064
+ }
2065
+
2066
+ return 0;
2067
+ }
2068
+
2069
+ /* Decompress : block content */
2070
+ {
2071
+ size_t rSize;
2072
+ switch(ctx->bType)
2073
+ {
2074
+ case bt_compressed:
2075
+ rSize = ZSTD_decompressBlock(ctx, dst, maxDstSize, src, srcSize);
2076
+ break;
2077
+ case bt_raw :
2078
+ rSize = ZSTD_copyUncompressedBlock(dst, maxDstSize, src, srcSize);
2079
+ break;
2080
+ case bt_rle :
2081
+ return ERROR(GENERIC); /* not yet handled */
2082
+ break;
2083
+ case bt_end : /* should never happen (filtered at phase 1) */
2084
+ rSize = 0;
2085
+ break;
2086
+ default:
2087
+ return ERROR(GENERIC);
2088
+ }
2089
+ ctx->phase = 1;
2090
+ ctx->expected = ZSTD_blockHeaderSize;
2091
+ ctx->previousDstEnd = (void*)( ((char*)dst) + rSize);
2092
+ return rSize;
2093
+ }
2094
+
2095
+ }