zstd-ruby 1.4.5.0 → 1.5.1.1

Sign up to get free protection for your applications and to get access to all the features.
Files changed (101) hide show
  1. checksums.yaml +4 -4
  2. data/.github/dependabot.yml +8 -0
  3. data/.github/workflows/ruby.yml +35 -0
  4. data/README.md +2 -2
  5. data/ext/zstdruby/extconf.rb +2 -1
  6. data/ext/zstdruby/libzstd/BUCK +5 -7
  7. data/ext/zstdruby/libzstd/Makefile +225 -222
  8. data/ext/zstdruby/libzstd/README.md +43 -5
  9. data/ext/zstdruby/libzstd/common/bitstream.h +46 -22
  10. data/ext/zstdruby/libzstd/common/compiler.h +182 -22
  11. data/ext/zstdruby/libzstd/common/cpu.h +1 -3
  12. data/ext/zstdruby/libzstd/common/debug.c +1 -1
  13. data/ext/zstdruby/libzstd/common/debug.h +12 -19
  14. data/ext/zstdruby/libzstd/common/entropy_common.c +196 -44
  15. data/ext/zstdruby/libzstd/common/error_private.c +2 -1
  16. data/ext/zstdruby/libzstd/common/error_private.h +82 -3
  17. data/ext/zstdruby/libzstd/common/fse.h +41 -12
  18. data/ext/zstdruby/libzstd/common/fse_decompress.c +139 -22
  19. data/ext/zstdruby/libzstd/common/huf.h +47 -23
  20. data/ext/zstdruby/libzstd/common/mem.h +87 -98
  21. data/ext/zstdruby/libzstd/common/pool.c +23 -17
  22. data/ext/zstdruby/libzstd/common/pool.h +2 -2
  23. data/ext/zstdruby/libzstd/common/portability_macros.h +131 -0
  24. data/ext/zstdruby/libzstd/common/threading.c +6 -5
  25. data/ext/zstdruby/libzstd/common/xxhash.c +6 -846
  26. data/ext/zstdruby/libzstd/common/xxhash.h +5568 -167
  27. data/ext/zstdruby/libzstd/common/zstd_common.c +10 -10
  28. data/ext/zstdruby/libzstd/common/zstd_deps.h +111 -0
  29. data/ext/zstdruby/libzstd/common/zstd_internal.h +189 -142
  30. data/ext/zstdruby/libzstd/common/zstd_trace.h +163 -0
  31. data/ext/zstdruby/libzstd/compress/clevels.h +134 -0
  32. data/ext/zstdruby/libzstd/compress/fse_compress.c +89 -46
  33. data/ext/zstdruby/libzstd/compress/hist.c +27 -29
  34. data/ext/zstdruby/libzstd/compress/hist.h +2 -2
  35. data/ext/zstdruby/libzstd/compress/huf_compress.c +770 -198
  36. data/ext/zstdruby/libzstd/compress/zstd_compress.c +2894 -863
  37. data/ext/zstdruby/libzstd/compress/zstd_compress_internal.h +390 -90
  38. data/ext/zstdruby/libzstd/compress/zstd_compress_literals.c +12 -11
  39. data/ext/zstdruby/libzstd/compress/zstd_compress_literals.h +4 -2
  40. data/ext/zstdruby/libzstd/compress/zstd_compress_sequences.c +31 -8
  41. data/ext/zstdruby/libzstd/compress/zstd_compress_sequences.h +1 -1
  42. data/ext/zstdruby/libzstd/compress/zstd_compress_superblock.c +25 -297
  43. data/ext/zstdruby/libzstd/compress/zstd_compress_superblock.h +1 -1
  44. data/ext/zstdruby/libzstd/compress/zstd_cwksp.h +206 -69
  45. data/ext/zstdruby/libzstd/compress/zstd_double_fast.c +307 -132
  46. data/ext/zstdruby/libzstd/compress/zstd_double_fast.h +1 -1
  47. data/ext/zstdruby/libzstd/compress/zstd_fast.c +322 -143
  48. data/ext/zstdruby/libzstd/compress/zstd_fast.h +1 -1
  49. data/ext/zstdruby/libzstd/compress/zstd_lazy.c +1136 -174
  50. data/ext/zstdruby/libzstd/compress/zstd_lazy.h +59 -1
  51. data/ext/zstdruby/libzstd/compress/zstd_ldm.c +316 -213
  52. data/ext/zstdruby/libzstd/compress/zstd_ldm.h +9 -2
  53. data/ext/zstdruby/libzstd/compress/zstd_ldm_geartab.h +106 -0
  54. data/ext/zstdruby/libzstd/compress/zstd_opt.c +373 -150
  55. data/ext/zstdruby/libzstd/compress/zstd_opt.h +1 -1
  56. data/ext/zstdruby/libzstd/compress/zstdmt_compress.c +152 -444
  57. data/ext/zstdruby/libzstd/compress/zstdmt_compress.h +31 -113
  58. data/ext/zstdruby/libzstd/decompress/huf_decompress.c +1044 -403
  59. data/ext/zstdruby/libzstd/decompress/huf_decompress_amd64.S +571 -0
  60. data/ext/zstdruby/libzstd/decompress/zstd_ddict.c +9 -9
  61. data/ext/zstdruby/libzstd/decompress/zstd_ddict.h +2 -2
  62. data/ext/zstdruby/libzstd/decompress/zstd_decompress.c +450 -105
  63. data/ext/zstdruby/libzstd/decompress/zstd_decompress_block.c +913 -273
  64. data/ext/zstdruby/libzstd/decompress/zstd_decompress_block.h +14 -5
  65. data/ext/zstdruby/libzstd/decompress/zstd_decompress_internal.h +59 -12
  66. data/ext/zstdruby/libzstd/deprecated/zbuff.h +1 -1
  67. data/ext/zstdruby/libzstd/deprecated/zbuff_common.c +1 -1
  68. data/ext/zstdruby/libzstd/deprecated/zbuff_compress.c +24 -4
  69. data/ext/zstdruby/libzstd/deprecated/zbuff_decompress.c +1 -1
  70. data/ext/zstdruby/libzstd/dictBuilder/cover.c +55 -38
  71. data/ext/zstdruby/libzstd/dictBuilder/cover.h +7 -6
  72. data/ext/zstdruby/libzstd/dictBuilder/divsufsort.c +1 -1
  73. data/ext/zstdruby/libzstd/dictBuilder/fastcover.c +43 -34
  74. data/ext/zstdruby/libzstd/dictBuilder/zdict.c +128 -58
  75. data/ext/zstdruby/libzstd/dll/example/Makefile +1 -1
  76. data/ext/zstdruby/libzstd/dll/example/README.md +16 -22
  77. data/ext/zstdruby/libzstd/legacy/zstd_legacy.h +1 -1
  78. data/ext/zstdruby/libzstd/legacy/zstd_v01.c +8 -8
  79. data/ext/zstdruby/libzstd/legacy/zstd_v01.h +1 -1
  80. data/ext/zstdruby/libzstd/legacy/zstd_v02.c +9 -9
  81. data/ext/zstdruby/libzstd/legacy/zstd_v02.h +1 -1
  82. data/ext/zstdruby/libzstd/legacy/zstd_v03.c +9 -9
  83. data/ext/zstdruby/libzstd/legacy/zstd_v03.h +1 -1
  84. data/ext/zstdruby/libzstd/legacy/zstd_v04.c +10 -10
  85. data/ext/zstdruby/libzstd/legacy/zstd_v04.h +1 -1
  86. data/ext/zstdruby/libzstd/legacy/zstd_v05.c +13 -13
  87. data/ext/zstdruby/libzstd/legacy/zstd_v05.h +1 -1
  88. data/ext/zstdruby/libzstd/legacy/zstd_v06.c +13 -13
  89. data/ext/zstdruby/libzstd/legacy/zstd_v06.h +1 -1
  90. data/ext/zstdruby/libzstd/legacy/zstd_v07.c +13 -13
  91. data/ext/zstdruby/libzstd/legacy/zstd_v07.h +1 -1
  92. data/ext/zstdruby/libzstd/libzstd.mk +185 -0
  93. data/ext/zstdruby/libzstd/libzstd.pc.in +4 -3
  94. data/ext/zstdruby/libzstd/modulemap/module.modulemap +4 -0
  95. data/ext/zstdruby/libzstd/{dictBuilder/zdict.h → zdict.h} +154 -7
  96. data/ext/zstdruby/libzstd/zstd.h +699 -214
  97. data/ext/zstdruby/libzstd/{common/zstd_errors.h → zstd_errors.h} +2 -1
  98. data/ext/zstdruby/zstdruby.c +2 -2
  99. data/lib/zstd-ruby/version.rb +1 -1
  100. metadata +15 -6
  101. data/.travis.yml +0 -14
@@ -1,5 +1,5 @@
1
1
  /*
2
- * Copyright (c) 2016-2020, Yann Collet, Facebook, Inc.
2
+ * Copyright (c) Yann Collet, Facebook, Inc.
3
3
  * All rights reserved.
4
4
  *
5
5
  * This source code is licensed under both the BSD-style license (found in the
@@ -13,8 +13,8 @@
13
13
  /*-*************************************
14
14
  * Dependencies
15
15
  ***************************************/
16
- #include <stdlib.h> /* malloc, calloc, free */
17
- #include <string.h> /* memset */
16
+ #define ZSTD_DEPS_NEED_MALLOC
17
+ #include "zstd_deps.h" /* ZSTD_malloc, ZSTD_calloc, ZSTD_free, ZSTD_memset */
18
18
  #include "error_private.h"
19
19
  #include "zstd_internal.h"
20
20
 
@@ -53,31 +53,31 @@ const char* ZSTD_getErrorString(ZSTD_ErrorCode code) { return ERR_getErrorString
53
53
  /*=**************************************************************
54
54
  * Custom allocator
55
55
  ****************************************************************/
56
- void* ZSTD_malloc(size_t size, ZSTD_customMem customMem)
56
+ void* ZSTD_customMalloc(size_t size, ZSTD_customMem customMem)
57
57
  {
58
58
  if (customMem.customAlloc)
59
59
  return customMem.customAlloc(customMem.opaque, size);
60
- return malloc(size);
60
+ return ZSTD_malloc(size);
61
61
  }
62
62
 
63
- void* ZSTD_calloc(size_t size, ZSTD_customMem customMem)
63
+ void* ZSTD_customCalloc(size_t size, ZSTD_customMem customMem)
64
64
  {
65
65
  if (customMem.customAlloc) {
66
66
  /* calloc implemented as malloc+memset;
67
67
  * not as efficient as calloc, but next best guess for custom malloc */
68
68
  void* const ptr = customMem.customAlloc(customMem.opaque, size);
69
- memset(ptr, 0, size);
69
+ ZSTD_memset(ptr, 0, size);
70
70
  return ptr;
71
71
  }
72
- return calloc(1, size);
72
+ return ZSTD_calloc(1, size);
73
73
  }
74
74
 
75
- void ZSTD_free(void* ptr, ZSTD_customMem customMem)
75
+ void ZSTD_customFree(void* ptr, ZSTD_customMem customMem)
76
76
  {
77
77
  if (ptr!=NULL) {
78
78
  if (customMem.customFree)
79
79
  customMem.customFree(customMem.opaque, ptr);
80
80
  else
81
- free(ptr);
81
+ ZSTD_free(ptr);
82
82
  }
83
83
  }
@@ -0,0 +1,111 @@
1
+ /*
2
+ * Copyright (c) Facebook, Inc.
3
+ * All rights reserved.
4
+ *
5
+ * This source code is licensed under both the BSD-style license (found in the
6
+ * LICENSE file in the root directory of this source tree) and the GPLv2 (found
7
+ * in the COPYING file in the root directory of this source tree).
8
+ * You may select, at your option, one of the above-listed licenses.
9
+ */
10
+
11
+ /* This file provides common libc dependencies that zstd requires.
12
+ * The purpose is to allow replacing this file with a custom implementation
13
+ * to compile zstd without libc support.
14
+ */
15
+
16
+ /* Need:
17
+ * NULL
18
+ * INT_MAX
19
+ * UINT_MAX
20
+ * ZSTD_memcpy()
21
+ * ZSTD_memset()
22
+ * ZSTD_memmove()
23
+ */
24
+ #ifndef ZSTD_DEPS_COMMON
25
+ #define ZSTD_DEPS_COMMON
26
+
27
+ #include <limits.h>
28
+ #include <stddef.h>
29
+ #include <string.h>
30
+
31
+ #if defined(__GNUC__) && __GNUC__ >= 4
32
+ # define ZSTD_memcpy(d,s,l) __builtin_memcpy((d),(s),(l))
33
+ # define ZSTD_memmove(d,s,l) __builtin_memmove((d),(s),(l))
34
+ # define ZSTD_memset(p,v,l) __builtin_memset((p),(v),(l))
35
+ #else
36
+ # define ZSTD_memcpy(d,s,l) memcpy((d),(s),(l))
37
+ # define ZSTD_memmove(d,s,l) memmove((d),(s),(l))
38
+ # define ZSTD_memset(p,v,l) memset((p),(v),(l))
39
+ #endif
40
+
41
+ #endif /* ZSTD_DEPS_COMMON */
42
+
43
+ /* Need:
44
+ * ZSTD_malloc()
45
+ * ZSTD_free()
46
+ * ZSTD_calloc()
47
+ */
48
+ #ifdef ZSTD_DEPS_NEED_MALLOC
49
+ #ifndef ZSTD_DEPS_MALLOC
50
+ #define ZSTD_DEPS_MALLOC
51
+
52
+ #include <stdlib.h>
53
+
54
+ #define ZSTD_malloc(s) malloc(s)
55
+ #define ZSTD_calloc(n,s) calloc((n), (s))
56
+ #define ZSTD_free(p) free((p))
57
+
58
+ #endif /* ZSTD_DEPS_MALLOC */
59
+ #endif /* ZSTD_DEPS_NEED_MALLOC */
60
+
61
+ /*
62
+ * Provides 64-bit math support.
63
+ * Need:
64
+ * U64 ZSTD_div64(U64 dividend, U32 divisor)
65
+ */
66
+ #ifdef ZSTD_DEPS_NEED_MATH64
67
+ #ifndef ZSTD_DEPS_MATH64
68
+ #define ZSTD_DEPS_MATH64
69
+
70
+ #define ZSTD_div64(dividend, divisor) ((dividend) / (divisor))
71
+
72
+ #endif /* ZSTD_DEPS_MATH64 */
73
+ #endif /* ZSTD_DEPS_NEED_MATH64 */
74
+
75
+ /* Need:
76
+ * assert()
77
+ */
78
+ #ifdef ZSTD_DEPS_NEED_ASSERT
79
+ #ifndef ZSTD_DEPS_ASSERT
80
+ #define ZSTD_DEPS_ASSERT
81
+
82
+ #include <assert.h>
83
+
84
+ #endif /* ZSTD_DEPS_ASSERT */
85
+ #endif /* ZSTD_DEPS_NEED_ASSERT */
86
+
87
+ /* Need:
88
+ * ZSTD_DEBUG_PRINT()
89
+ */
90
+ #ifdef ZSTD_DEPS_NEED_IO
91
+ #ifndef ZSTD_DEPS_IO
92
+ #define ZSTD_DEPS_IO
93
+
94
+ #include <stdio.h>
95
+ #define ZSTD_DEBUG_PRINT(...) fprintf(stderr, __VA_ARGS__)
96
+
97
+ #endif /* ZSTD_DEPS_IO */
98
+ #endif /* ZSTD_DEPS_NEED_IO */
99
+
100
+ /* Only requested when <stdint.h> is known to be present.
101
+ * Need:
102
+ * intptr_t
103
+ */
104
+ #ifdef ZSTD_DEPS_NEED_STDINT
105
+ #ifndef ZSTD_DEPS_STDINT
106
+ #define ZSTD_DEPS_STDINT
107
+
108
+ #include <stdint.h>
109
+
110
+ #endif /* ZSTD_DEPS_STDINT */
111
+ #endif /* ZSTD_DEPS_NEED_STDINT */
@@ -1,5 +1,5 @@
1
1
  /*
2
- * Copyright (c) 2016-2020, Yann Collet, Facebook, Inc.
2
+ * Copyright (c) Yann Collet, Facebook, Inc.
3
3
  * All rights reserved.
4
4
  *
5
5
  * This source code is licensed under both the BSD-style license (found in the
@@ -19,10 +19,8 @@
19
19
  /*-*************************************
20
20
  * Dependencies
21
21
  ***************************************/
22
- #ifdef __aarch64__
23
- #include <arm_neon.h>
24
- #endif
25
22
  #include "compiler.h"
23
+ #include "cpu.h"
26
24
  #include "mem.h"
27
25
  #include "debug.h" /* assert, DEBUGLOG, RAWLOG, g_debuglevel */
28
26
  #include "error_private.h"
@@ -36,6 +34,11 @@
36
34
  # define XXH_STATIC_LINKING_ONLY /* XXH64_state_t */
37
35
  #endif
38
36
  #include "xxhash.h" /* XXH_reset, update, digest */
37
+ #ifndef ZSTD_NO_TRACE
38
+ # include "zstd_trace.h"
39
+ #else
40
+ # define ZSTD_TRACE 0
41
+ #endif
39
42
 
40
43
  #if defined (__cplusplus)
41
44
  extern "C" {
@@ -55,81 +58,7 @@ extern "C" {
55
58
  #undef MAX
56
59
  #define MIN(a,b) ((a)<(b) ? (a) : (b))
57
60
  #define MAX(a,b) ((a)>(b) ? (a) : (b))
58
-
59
- /**
60
- * Ignore: this is an internal helper.
61
- *
62
- * This is a helper function to help force C99-correctness during compilation.
63
- * Under strict compilation modes, variadic macro arguments can't be empty.
64
- * However, variadic function arguments can be. Using a function therefore lets
65
- * us statically check that at least one (string) argument was passed,
66
- * independent of the compilation flags.
67
- */
68
- static INLINE_KEYWORD UNUSED_ATTR
69
- void _force_has_format_string(const char *format, ...) {
70
- (void)format;
71
- }
72
-
73
- /**
74
- * Ignore: this is an internal helper.
75
- *
76
- * We want to force this function invocation to be syntactically correct, but
77
- * we don't want to force runtime evaluation of its arguments.
78
- */
79
- #define _FORCE_HAS_FORMAT_STRING(...) \
80
- if (0) { \
81
- _force_has_format_string(__VA_ARGS__); \
82
- }
83
-
84
- /**
85
- * Return the specified error if the condition evaluates to true.
86
- *
87
- * In debug modes, prints additional information.
88
- * In order to do that (particularly, printing the conditional that failed),
89
- * this can't just wrap RETURN_ERROR().
90
- */
91
- #define RETURN_ERROR_IF(cond, err, ...) \
92
- if (cond) { \
93
- RAWLOG(3, "%s:%d: ERROR!: check %s failed, returning %s", \
94
- __FILE__, __LINE__, ZSTD_QUOTE(cond), ZSTD_QUOTE(ERROR(err))); \
95
- _FORCE_HAS_FORMAT_STRING(__VA_ARGS__); \
96
- RAWLOG(3, ": " __VA_ARGS__); \
97
- RAWLOG(3, "\n"); \
98
- return ERROR(err); \
99
- }
100
-
101
- /**
102
- * Unconditionally return the specified error.
103
- *
104
- * In debug modes, prints additional information.
105
- */
106
- #define RETURN_ERROR(err, ...) \
107
- do { \
108
- RAWLOG(3, "%s:%d: ERROR!: unconditional check failed, returning %s", \
109
- __FILE__, __LINE__, ZSTD_QUOTE(ERROR(err))); \
110
- _FORCE_HAS_FORMAT_STRING(__VA_ARGS__); \
111
- RAWLOG(3, ": " __VA_ARGS__); \
112
- RAWLOG(3, "\n"); \
113
- return ERROR(err); \
114
- } while(0);
115
-
116
- /**
117
- * If the provided expression evaluates to an error code, returns that error code.
118
- *
119
- * In debug modes, prints additional information.
120
- */
121
- #define FORWARD_IF_ERROR(err, ...) \
122
- do { \
123
- size_t const err_code = (err); \
124
- if (ERR_isError(err_code)) { \
125
- RAWLOG(3, "%s:%d: ERROR!: forwarding error in %s: %s", \
126
- __FILE__, __LINE__, ZSTD_QUOTE(err), ERR_getErrorName(err_code)); \
127
- _FORCE_HAS_FORMAT_STRING(__VA_ARGS__); \
128
- RAWLOG(3, ": " __VA_ARGS__); \
129
- RAWLOG(3, "\n"); \
130
- return err_code; \
131
- } \
132
- } while(0);
61
+ #define BOUNDED(min,val,max) (MAX(min,MIN(val,max)))
133
62
 
134
63
 
135
64
  /*-*************************************
@@ -139,7 +68,7 @@ void _force_has_format_string(const char *format, ...) {
139
68
 
140
69
  #define ZSTD_REP_NUM 3 /* number of repcodes */
141
70
  #define ZSTD_REP_MOVE (ZSTD_REP_NUM-1)
142
- static const U32 repStartValue[ZSTD_REP_NUM] = { 1, 4, 8 };
71
+ static UNUSED_ATTR const U32 repStartValue[ZSTD_REP_NUM] = { 1, 4, 8 };
143
72
 
144
73
  #define KB *(1 <<10)
145
74
  #define MB *(1 <<20)
@@ -153,13 +82,13 @@ static const U32 repStartValue[ZSTD_REP_NUM] = { 1, 4, 8 };
153
82
  #define BIT0 1
154
83
 
155
84
  #define ZSTD_WINDOWLOG_ABSOLUTEMIN 10
156
- static const size_t ZSTD_fcs_fieldSize[4] = { 0, 2, 4, 8 };
157
- static const size_t ZSTD_did_fieldSize[4] = { 0, 1, 2, 4 };
85
+ static UNUSED_ATTR const size_t ZSTD_fcs_fieldSize[4] = { 0, 2, 4, 8 };
86
+ static UNUSED_ATTR const size_t ZSTD_did_fieldSize[4] = { 0, 1, 2, 4 };
158
87
 
159
88
  #define ZSTD_FRAMEIDSIZE 4 /* magic number size */
160
89
 
161
90
  #define ZSTD_BLOCKHEADERSIZE 3 /* C standard doesn't allow `static const` variable to be init using another `static const` variable */
162
- static const size_t ZSTD_blockHeaderSize = ZSTD_BLOCKHEADERSIZE;
91
+ static UNUSED_ATTR const size_t ZSTD_blockHeaderSize = ZSTD_BLOCKHEADERSIZE;
163
92
  typedef enum { bt_raw, bt_rle, bt_compressed, bt_reserved } blockType_e;
164
93
 
165
94
  #define ZSTD_FRAMECHECKSUMSIZE 4
@@ -186,61 +115,86 @@ typedef enum { set_basic, set_rle, set_compressed, set_repeat } symbolEncodingTy
186
115
  #define OffFSELog 8
187
116
  #define MaxFSELog MAX(MAX(MLFSELog, LLFSELog), OffFSELog)
188
117
 
189
- static const U32 LL_bits[MaxLL+1] = { 0, 0, 0, 0, 0, 0, 0, 0,
190
- 0, 0, 0, 0, 0, 0, 0, 0,
191
- 1, 1, 1, 1, 2, 2, 3, 3,
192
- 4, 6, 7, 8, 9,10,11,12,
193
- 13,14,15,16 };
194
- static const S16 LL_defaultNorm[MaxLL+1] = { 4, 3, 2, 2, 2, 2, 2, 2,
195
- 2, 2, 2, 2, 2, 1, 1, 1,
196
- 2, 2, 2, 2, 2, 2, 2, 2,
197
- 2, 3, 2, 1, 1, 1, 1, 1,
198
- -1,-1,-1,-1 };
118
+ #define ZSTD_MAX_HUF_HEADER_SIZE 128 /* header + <= 127 byte tree description */
119
+ /* Each table cannot take more than #symbols * FSELog bits */
120
+ #define ZSTD_MAX_FSE_HEADERS_SIZE (((MaxML + 1) * MLFSELog + (MaxLL + 1) * LLFSELog + (MaxOff + 1) * OffFSELog + 7) / 8)
121
+
122
+ static UNUSED_ATTR const U8 LL_bits[MaxLL+1] = {
123
+ 0, 0, 0, 0, 0, 0, 0, 0,
124
+ 0, 0, 0, 0, 0, 0, 0, 0,
125
+ 1, 1, 1, 1, 2, 2, 3, 3,
126
+ 4, 6, 7, 8, 9,10,11,12,
127
+ 13,14,15,16
128
+ };
129
+ static UNUSED_ATTR const S16 LL_defaultNorm[MaxLL+1] = {
130
+ 4, 3, 2, 2, 2, 2, 2, 2,
131
+ 2, 2, 2, 2, 2, 1, 1, 1,
132
+ 2, 2, 2, 2, 2, 2, 2, 2,
133
+ 2, 3, 2, 1, 1, 1, 1, 1,
134
+ -1,-1,-1,-1
135
+ };
199
136
  #define LL_DEFAULTNORMLOG 6 /* for static allocation */
200
- static const U32 LL_defaultNormLog = LL_DEFAULTNORMLOG;
201
-
202
- static const U32 ML_bits[MaxML+1] = { 0, 0, 0, 0, 0, 0, 0, 0,
203
- 0, 0, 0, 0, 0, 0, 0, 0,
204
- 0, 0, 0, 0, 0, 0, 0, 0,
205
- 0, 0, 0, 0, 0, 0, 0, 0,
206
- 1, 1, 1, 1, 2, 2, 3, 3,
207
- 4, 4, 5, 7, 8, 9,10,11,
208
- 12,13,14,15,16 };
209
- static const S16 ML_defaultNorm[MaxML+1] = { 1, 4, 3, 2, 2, 2, 2, 2,
210
- 2, 1, 1, 1, 1, 1, 1, 1,
211
- 1, 1, 1, 1, 1, 1, 1, 1,
212
- 1, 1, 1, 1, 1, 1, 1, 1,
213
- 1, 1, 1, 1, 1, 1, 1, 1,
214
- 1, 1, 1, 1, 1, 1,-1,-1,
215
- -1,-1,-1,-1,-1 };
137
+ static UNUSED_ATTR const U32 LL_defaultNormLog = LL_DEFAULTNORMLOG;
138
+
139
+ static UNUSED_ATTR const U8 ML_bits[MaxML+1] = {
140
+ 0, 0, 0, 0, 0, 0, 0, 0,
141
+ 0, 0, 0, 0, 0, 0, 0, 0,
142
+ 0, 0, 0, 0, 0, 0, 0, 0,
143
+ 0, 0, 0, 0, 0, 0, 0, 0,
144
+ 1, 1, 1, 1, 2, 2, 3, 3,
145
+ 4, 4, 5, 7, 8, 9,10,11,
146
+ 12,13,14,15,16
147
+ };
148
+ static UNUSED_ATTR const S16 ML_defaultNorm[MaxML+1] = {
149
+ 1, 4, 3, 2, 2, 2, 2, 2,
150
+ 2, 1, 1, 1, 1, 1, 1, 1,
151
+ 1, 1, 1, 1, 1, 1, 1, 1,
152
+ 1, 1, 1, 1, 1, 1, 1, 1,
153
+ 1, 1, 1, 1, 1, 1, 1, 1,
154
+ 1, 1, 1, 1, 1, 1,-1,-1,
155
+ -1,-1,-1,-1,-1
156
+ };
216
157
  #define ML_DEFAULTNORMLOG 6 /* for static allocation */
217
- static const U32 ML_defaultNormLog = ML_DEFAULTNORMLOG;
218
-
219
- static const S16 OF_defaultNorm[DefaultMaxOff+1] = { 1, 1, 1, 1, 1, 1, 2, 2,
220
- 2, 1, 1, 1, 1, 1, 1, 1,
221
- 1, 1, 1, 1, 1, 1, 1, 1,
222
- -1,-1,-1,-1,-1 };
158
+ static UNUSED_ATTR const U32 ML_defaultNormLog = ML_DEFAULTNORMLOG;
159
+
160
+ static UNUSED_ATTR const S16 OF_defaultNorm[DefaultMaxOff+1] = {
161
+ 1, 1, 1, 1, 1, 1, 2, 2,
162
+ 2, 1, 1, 1, 1, 1, 1, 1,
163
+ 1, 1, 1, 1, 1, 1, 1, 1,
164
+ -1,-1,-1,-1,-1
165
+ };
223
166
  #define OF_DEFAULTNORMLOG 5 /* for static allocation */
224
- static const U32 OF_defaultNormLog = OF_DEFAULTNORMLOG;
167
+ static UNUSED_ATTR const U32 OF_defaultNormLog = OF_DEFAULTNORMLOG;
225
168
 
226
169
 
227
170
  /*-*******************************************
228
171
  * Shared functions to include for inlining
229
172
  *********************************************/
230
173
  static void ZSTD_copy8(void* dst, const void* src) {
231
- #ifdef __aarch64__
174
+ #if defined(ZSTD_ARCH_ARM_NEON)
232
175
  vst1_u8((uint8_t*)dst, vld1_u8((const uint8_t*)src));
233
176
  #else
234
- memcpy(dst, src, 8);
177
+ ZSTD_memcpy(dst, src, 8);
235
178
  #endif
236
179
  }
237
-
238
180
  #define COPY8(d,s) { ZSTD_copy8(d,s); d+=8; s+=8; }
181
+
182
+ /* Need to use memmove here since the literal buffer can now be located within
183
+ the dst buffer. In circumstances where the op "catches up" to where the
184
+ literal buffer is, there can be partial overlaps in this call on the final
185
+ copy if the literal is being shifted by less than 16 bytes. */
239
186
  static void ZSTD_copy16(void* dst, const void* src) {
240
- #ifdef __aarch64__
187
+ #if defined(ZSTD_ARCH_ARM_NEON)
241
188
  vst1q_u8((uint8_t*)dst, vld1q_u8((const uint8_t*)src));
189
+ #elif defined(ZSTD_ARCH_X86_SSE2)
190
+ _mm_storeu_si128((__m128i*)dst, _mm_loadu_si128((const __m128i*)src));
191
+ #elif defined(__clang__)
192
+ ZSTD_memmove(dst, src, 16);
242
193
  #else
243
- memcpy(dst, src, 16);
194
+ /* ZSTD_memmove is not inlined properly by gcc */
195
+ BYTE copy16_buf[16];
196
+ ZSTD_memcpy(copy16_buf, src, 16);
197
+ ZSTD_memcpy(dst, copy16_buf, 16);
244
198
  #endif
245
199
  }
246
200
  #define COPY16(d,s) { ZSTD_copy16(d,s); d+=16; s+=16; }
@@ -255,13 +209,13 @@ typedef enum {
255
209
  } ZSTD_overlap_e;
256
210
 
257
211
  /*! ZSTD_wildcopy() :
258
- * Custom version of memcpy(), can over read/write up to WILDCOPY_OVERLENGTH bytes (if length==0)
212
+ * Custom version of ZSTD_memcpy(), can over read/write up to WILDCOPY_OVERLENGTH bytes (if length==0)
259
213
  * @param ovtype controls the overlap detection
260
214
  * - ZSTD_no_overlap: The source and destination are guaranteed to be at least WILDCOPY_VECLEN bytes apart.
261
215
  * - ZSTD_overlap_src_before_dst: The src and dst may overlap, but they MUST be at least 8 bytes apart.
262
216
  * The src buffer must be before the dst buffer.
263
217
  */
264
- MEM_STATIC FORCE_INLINE_ATTR
218
+ MEM_STATIC FORCE_INLINE_ATTR
265
219
  void ZSTD_wildcopy(void* dst, const void* src, ptrdiff_t length, ZSTD_overlap_e const ovtype)
266
220
  {
267
221
  ptrdiff_t diff = (BYTE*)dst - (const BYTE*)src;
@@ -269,8 +223,6 @@ void ZSTD_wildcopy(void* dst, const void* src, ptrdiff_t length, ZSTD_overlap_e
269
223
  BYTE* op = (BYTE*)dst;
270
224
  BYTE* const oend = op + length;
271
225
 
272
- assert(diff >= 8 || (ovtype == ZSTD_no_overlap && diff <= -WILDCOPY_VECLEN));
273
-
274
226
  if (ovtype == ZSTD_overlap_src_before_dst && diff < WILDCOPY_VECLEN) {
275
227
  /* Handle short offset copies. */
276
228
  do {
@@ -284,14 +236,16 @@ void ZSTD_wildcopy(void* dst, const void* src, ptrdiff_t length, ZSTD_overlap_e
284
236
  * one COPY16() in the first call. Then, do two calls per loop since
285
237
  * at that point it is more likely to have a high trip count.
286
238
  */
287
- #ifndef __aarch64__
239
+ #ifdef __aarch64__
288
240
  do {
289
241
  COPY16(op, ip);
290
242
  }
291
243
  while (op < oend);
292
244
  #else
293
- COPY16(op, ip);
294
- if (op >= oend) return;
245
+ ZSTD_copy16(op, ip);
246
+ if (16 >= length) return;
247
+ op += 16;
248
+ ip += 16;
295
249
  do {
296
250
  COPY16(op, ip);
297
251
  COPY16(op, ip);
@@ -305,7 +259,7 @@ MEM_STATIC size_t ZSTD_limitCopy(void* dst, size_t dstCapacity, const void* src,
305
259
  {
306
260
  size_t const length = MIN(dstCapacity, srcSize);
307
261
  if (length > 0) {
308
- memcpy(dst, src, length);
262
+ ZSTD_memcpy(dst, src, length);
309
263
  }
310
264
  return length;
311
265
  }
@@ -320,28 +274,46 @@ MEM_STATIC size_t ZSTD_limitCopy(void* dst, size_t dstCapacity, const void* src,
320
274
  * In which case, resize it down to free some memory */
321
275
  #define ZSTD_WORKSPACETOOLARGE_MAXDURATION 128
322
276
 
277
+ /* Controls whether the input/output buffer is buffered or stable. */
278
+ typedef enum {
279
+ ZSTD_bm_buffered = 0, /* Buffer the input/output */
280
+ ZSTD_bm_stable = 1 /* ZSTD_inBuffer/ZSTD_outBuffer is stable */
281
+ } ZSTD_bufferMode_e;
282
+
323
283
 
324
284
  /*-*******************************************
325
285
  * Private declarations
326
286
  *********************************************/
327
287
  typedef struct seqDef_s {
328
- U32 offset;
288
+ U32 offset; /* offset == rawOffset + ZSTD_REP_NUM, or equivalently, offCode + 1 */
329
289
  U16 litLength;
330
290
  U16 matchLength;
331
291
  } seqDef;
332
292
 
293
+ /* Controls whether seqStore has a single "long" litLength or matchLength. See seqStore_t. */
294
+ typedef enum {
295
+ ZSTD_llt_none = 0, /* no longLengthType */
296
+ ZSTD_llt_literalLength = 1, /* represents a long literal */
297
+ ZSTD_llt_matchLength = 2 /* represents a long match */
298
+ } ZSTD_longLengthType_e;
299
+
333
300
  typedef struct {
334
301
  seqDef* sequencesStart;
335
- seqDef* sequences;
302
+ seqDef* sequences; /* ptr to end of sequences */
336
303
  BYTE* litStart;
337
- BYTE* lit;
304
+ BYTE* lit; /* ptr to end of literals */
338
305
  BYTE* llCode;
339
306
  BYTE* mlCode;
340
307
  BYTE* ofCode;
341
308
  size_t maxNbSeq;
342
309
  size_t maxNbLit;
343
- U32 longLengthID; /* 0 == no longLength; 1 == Lit.longLength; 2 == Match.longLength; */
344
- U32 longLengthPos;
310
+
311
+ /* longLengthPos and longLengthType to allow us to represent either a single litLength or matchLength
312
+ * in the seqStore that has a value larger than U16 (if it exists). To do so, we increment
313
+ * the existing value of the litLength or matchLength by 0x10000.
314
+ */
315
+ ZSTD_longLengthType_e longLengthType;
316
+ U32 longLengthPos; /* Index of the sequence to apply long length modification to */
345
317
  } seqStore_t;
346
318
 
347
319
  typedef struct {
@@ -351,7 +323,7 @@ typedef struct {
351
323
 
352
324
  /**
353
325
  * Returns the ZSTD_sequenceLength for the given sequences. It handles the decoding of long sequences
354
- * indicated by longLengthPos and longLengthID, and adds MINMATCH back to matchLength.
326
+ * indicated by longLengthPos and longLengthType, and adds MINMATCH back to matchLength.
355
327
  */
356
328
  MEM_STATIC ZSTD_sequenceLength ZSTD_getSequenceLength(seqStore_t const* seqStore, seqDef const* seq)
357
329
  {
@@ -359,10 +331,10 @@ MEM_STATIC ZSTD_sequenceLength ZSTD_getSequenceLength(seqStore_t const* seqStore
359
331
  seqLen.litLength = seq->litLength;
360
332
  seqLen.matchLength = seq->matchLength + MINMATCH;
361
333
  if (seqStore->longLengthPos == (U32)(seq - seqStore->sequencesStart)) {
362
- if (seqStore->longLengthID == 1) {
334
+ if (seqStore->longLengthType == ZSTD_llt_literalLength) {
363
335
  seqLen.litLength += 0xFFFF;
364
336
  }
365
- if (seqStore->longLengthID == 2) {
337
+ if (seqStore->longLengthType == ZSTD_llt_matchLength) {
366
338
  seqLen.matchLength += 0xFFFF;
367
339
  }
368
340
  }
@@ -384,9 +356,9 @@ const seqStore_t* ZSTD_getSeqStore(const ZSTD_CCtx* ctx); /* compress & dictBu
384
356
  void ZSTD_seqToCodes(const seqStore_t* seqStorePtr); /* compress, dictBuilder, decodeCorpus (shouldn't get its definition from here) */
385
357
 
386
358
  /* custom memory allocation functions */
387
- void* ZSTD_malloc(size_t size, ZSTD_customMem customMem);
388
- void* ZSTD_calloc(size_t size, ZSTD_customMem customMem);
389
- void ZSTD_free(void* ptr, ZSTD_customMem customMem);
359
+ void* ZSTD_customMalloc(size_t size, ZSTD_customMem customMem);
360
+ void* ZSTD_customCalloc(size_t size, ZSTD_customMem customMem);
361
+ void ZSTD_customFree(void* ptr, ZSTD_customMem customMem);
390
362
 
391
363
 
392
364
  MEM_STATIC U32 ZSTD_highbit32(U32 val) /* compress, dictBuilder, decodeCorpus */
@@ -394,8 +366,18 @@ MEM_STATIC U32 ZSTD_highbit32(U32 val) /* compress, dictBuilder, decodeCorpus
394
366
  assert(val != 0);
395
367
  {
396
368
  # if defined(_MSC_VER) /* Visual */
397
- unsigned long r=0;
398
- return _BitScanReverse(&r, val) ? (unsigned)r : 0;
369
+ # if STATIC_BMI2 == 1
370
+ return _lzcnt_u32(val)^31;
371
+ # else
372
+ if (val != 0) {
373
+ unsigned long r;
374
+ _BitScanReverse(&r, val);
375
+ return (unsigned)r;
376
+ } else {
377
+ /* Should not reach this code path */
378
+ __assume(0);
379
+ }
380
+ # endif
399
381
  # elif defined(__GNUC__) && (__GNUC__ >= 3) /* GCC Intrinsic */
400
382
  return __builtin_clz (val) ^ 31;
401
383
  # elif defined(__ICCARM__) /* IAR Intrinsic */
@@ -413,6 +395,63 @@ MEM_STATIC U32 ZSTD_highbit32(U32 val) /* compress, dictBuilder, decodeCorpus
413
395
  }
414
396
  }
415
397
 
398
+ /**
399
+ * Counts the number of trailing zeros of a `size_t`.
400
+ * Most compilers should support CTZ as a builtin. A backup
401
+ * implementation is provided if the builtin isn't supported, but
402
+ * it may not be terribly efficient.
403
+ */
404
+ MEM_STATIC unsigned ZSTD_countTrailingZeros(size_t val)
405
+ {
406
+ if (MEM_64bits()) {
407
+ # if defined(_MSC_VER) && defined(_WIN64)
408
+ # if STATIC_BMI2
409
+ return _tzcnt_u64(val);
410
+ # else
411
+ if (val != 0) {
412
+ unsigned long r;
413
+ _BitScanForward64(&r, (U64)val);
414
+ return (unsigned)r;
415
+ } else {
416
+ /* Should not reach this code path */
417
+ __assume(0);
418
+ }
419
+ # endif
420
+ # elif defined(__GNUC__) && (__GNUC__ >= 4)
421
+ return __builtin_ctzll((U64)val);
422
+ # else
423
+ static const int DeBruijnBytePos[64] = { 0, 1, 2, 7, 3, 13, 8, 19,
424
+ 4, 25, 14, 28, 9, 34, 20, 56,
425
+ 5, 17, 26, 54, 15, 41, 29, 43,
426
+ 10, 31, 38, 35, 21, 45, 49, 57,
427
+ 63, 6, 12, 18, 24, 27, 33, 55,
428
+ 16, 53, 40, 42, 30, 37, 44, 48,
429
+ 62, 11, 23, 32, 52, 39, 36, 47,
430
+ 61, 22, 51, 46, 60, 50, 59, 58 };
431
+ return DeBruijnBytePos[((U64)((val & -(long long)val) * 0x0218A392CDABBD3FULL)) >> 58];
432
+ # endif
433
+ } else { /* 32 bits */
434
+ # if defined(_MSC_VER)
435
+ if (val != 0) {
436
+ unsigned long r;
437
+ _BitScanForward(&r, (U32)val);
438
+ return (unsigned)r;
439
+ } else {
440
+ /* Should not reach this code path */
441
+ __assume(0);
442
+ }
443
+ # elif defined(__GNUC__) && (__GNUC__ >= 3)
444
+ return __builtin_ctz((U32)val);
445
+ # else
446
+ static const int DeBruijnBytePos[32] = { 0, 1, 28, 2, 29, 14, 24, 3,
447
+ 30, 22, 20, 15, 25, 17, 4, 8,
448
+ 31, 27, 13, 23, 21, 19, 16, 7,
449
+ 26, 12, 18, 6, 11, 5, 10, 9 };
450
+ return DeBruijnBytePos[((U32)((val & -(S32)val) * 0x077CB531U)) >> 27];
451
+ # endif
452
+ }
453
+ }
454
+
416
455
 
417
456
  /* ZSTD_invalidateRepCodes() :
418
457
  * ensures next compression will not use repcodes from previous block.
@@ -439,6 +478,14 @@ size_t ZSTD_getcBlockSize(const void* src, size_t srcSize,
439
478
  size_t ZSTD_decodeSeqHeaders(ZSTD_DCtx* dctx, int* nbSeqPtr,
440
479
  const void* src, size_t srcSize);
441
480
 
481
+ /**
482
+ * @returns true iff the CPU supports dynamic BMI2 dispatch.
483
+ */
484
+ MEM_STATIC int ZSTD_cpuSupportsBmi2(void)
485
+ {
486
+ ZSTD_cpuid_t cpuid = ZSTD_cpuid();
487
+ return ZSTD_cpuid_bmi1(cpuid) && ZSTD_cpuid_bmi2(cpuid);
488
+ }
442
489
 
443
490
  #if defined (__cplusplus)
444
491
  }