oj 3.13.2 → 3.13.6
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/CHANGELOG.md +1280 -0
- data/README.md +1 -1
- data/RELEASE_NOTES.md +55 -0
- data/ext/oj/cache.c +240 -107
- data/ext/oj/cache.h +2 -1
- data/ext/oj/compat.c +1 -2
- data/ext/oj/custom.c +4 -7
- data/ext/oj/extconf.rb +1 -0
- data/ext/oj/intern.c +106 -216
- data/ext/oj/intern.h +0 -1
- data/ext/oj/mimic_json.c +2 -2
- data/ext/oj/object.c +10 -39
- data/ext/oj/oj.c +8 -5
- data/ext/oj/parser.c +167 -148
- data/ext/oj/saj2.c +3 -3
- data/ext/oj/strict.c +2 -3
- data/ext/oj/usual.c +55 -31
- data/ext/oj/wab.c +6 -3
- data/lib/oj/state.rb +8 -7
- data/lib/oj/version.rb +1 -1
- data/pages/Options.md +12 -2
- data/test/bar.rb +9 -28
- data/test/foo.rb +8 -8
- data/test/mem.rb +33 -0
- data/test/perf_once.rb +58 -0
- data/test/perf_parser.rb +6 -1
- data/test/test_parser_usual.rb +9 -5
- metadata +11 -2
data/README.md
CHANGED
data/RELEASE_NOTES.md
ADDED
@@ -0,0 +1,55 @@
|
|
1
|
+
# RELEASE NOTES
|
2
|
+
|
3
|
+
The release notes here are organized by release. For a list of changes
|
4
|
+
see the See [{file:CHANGELOG.md}](CHANGELOG.md) file. In this file are
|
5
|
+
the steps to take to aid in keeping things rolling after updating to
|
6
|
+
the latest version.
|
7
|
+
|
8
|
+
## 3.13.x
|
9
|
+
|
10
|
+
This release included a new cache that performs better than the
|
11
|
+
earlier cache and a new high performance parser.
|
12
|
+
|
13
|
+
### Cache
|
14
|
+
|
15
|
+
The new cache includes a least recently used expiration to reduce
|
16
|
+
memory use. The cache is also self adjusting and will expand as needed
|
17
|
+
for better performance. It also handles Hash keys and string values
|
18
|
+
with two options, `:cache_keys`, a boolean and `:cache_str` an
|
19
|
+
integer. The `:cache_str` if set to more than zero is the limit for
|
20
|
+
the length of string values to cache. The maximum value is 35 which
|
21
|
+
allows strings up to 34 bytes to be cached.
|
22
|
+
|
23
|
+
One interesting aspect of the cache is not so much the string caching
|
24
|
+
which performs similar to the Ruby intern functions but the caching of
|
25
|
+
symbols and object attribute names. There is a significant gain for
|
26
|
+
symbols and object attributes.
|
27
|
+
|
28
|
+
If the cache is not desired then setting the default options to turn
|
29
|
+
it off can be done with this line:
|
30
|
+
|
31
|
+
``` ruby
|
32
|
+
Oj.default_options = { cache_keys: false, cache_str: 0 }
|
33
|
+
```
|
34
|
+
|
35
|
+
### Oj::Parser
|
36
|
+
|
37
|
+
The new parser uses a different core that follows the approach taken
|
38
|
+
by [OjC](https://github.com/ohler55/ojc) and
|
39
|
+
[OjG](https://github.com/ohler55/ojg). It also takes advantage of the
|
40
|
+
bulk Array and Hash functions. Another issue the new parser addresses
|
41
|
+
is option management. Instead of a single global default_options each
|
42
|
+
parser instance maintains it's own options.
|
43
|
+
|
44
|
+
There is a price to be paid when using the Oj::Parser. The API is not
|
45
|
+
the same the older parser. A single parser can only be used in a
|
46
|
+
single thread. This allows reuse of internal buffers for additional
|
47
|
+
improvements in performance.
|
48
|
+
|
49
|
+
The performane advantage of the Oj::Parse is that it is more than 3
|
50
|
+
times faster than the Oj::compat_load call and 6 times faster than the
|
51
|
+
JSON gem.
|
52
|
+
|
53
|
+
### Dump Performance
|
54
|
+
|
55
|
+
Thanks to Watson1978 Oj.dump also received a speed boost.
|
data/ext/oj/cache.c
CHANGED
@@ -1,65 +1,74 @@
|
|
1
1
|
// Copyright (c) 2011, 2021 Peter Ohler. All rights reserved.
|
2
2
|
// Licensed under the MIT License. See LICENSE file in the project root for license details.
|
3
3
|
|
4
|
+
#if HAVE_PTHREAD_MUTEX_INIT
|
5
|
+
#include <pthread.h>
|
6
|
+
#endif
|
7
|
+
#include <stdlib.h>
|
8
|
+
|
4
9
|
#include "cache.h"
|
5
10
|
|
6
|
-
|
11
|
+
// The stdlib calloc, realloc, and free are used instead of the Ruby ALLOC,
|
12
|
+
// ALLOC_N, REALLOC, and xfree since the later could trigger a GC which will
|
13
|
+
// either corrupt memory or if the mark function locks will deadlock.
|
14
|
+
|
15
|
+
#define REHASH_LIMIT 4
|
7
16
|
#define MIN_SHIFT 8
|
17
|
+
#define REUSE_MAX 8192
|
18
|
+
|
19
|
+
#if HAVE_PTHREAD_MUTEX_INIT
|
20
|
+
#define CACHE_LOCK(c) pthread_mutex_lock(&((c)->mutex))
|
21
|
+
#define CACHE_UNLOCK(c) pthread_mutex_unlock(&((c)->mutex))
|
22
|
+
#else
|
23
|
+
#define CACHE_LOCK(c) rb_mutex_lock((c)->mutex)
|
24
|
+
#define CACHE_UNLOCK(c) rb_mutex_unlock((c)->mutex)
|
25
|
+
#endif
|
26
|
+
|
27
|
+
// almost the Murmur hash algorithm
|
28
|
+
#define M 0x5bd1e995
|
8
29
|
|
9
30
|
typedef struct _slot {
|
10
|
-
struct _slot *next;
|
11
|
-
VALUE
|
12
|
-
|
13
|
-
|
14
|
-
|
31
|
+
struct _slot * next;
|
32
|
+
VALUE val;
|
33
|
+
uint64_t hash;
|
34
|
+
volatile uint32_t use_cnt;
|
35
|
+
uint8_t klen;
|
36
|
+
char key[CACHE_MAX_KEY];
|
15
37
|
} * Slot;
|
16
38
|
|
17
39
|
typedef struct _cache {
|
18
|
-
Slot * slots;
|
19
|
-
size_t cnt;
|
40
|
+
volatile Slot * slots;
|
41
|
+
volatile size_t cnt;
|
20
42
|
VALUE (*form)(const char *str, size_t len);
|
21
|
-
|
22
|
-
|
23
|
-
|
43
|
+
uint64_t size;
|
44
|
+
uint64_t mask;
|
45
|
+
VALUE (*intern)(struct _cache *c, const char *key, size_t len);
|
46
|
+
volatile Slot reuse;
|
47
|
+
size_t rcnt;
|
48
|
+
#if HAVE_PTHREAD_MUTEX_INIT
|
49
|
+
pthread_mutex_t mutex;
|
50
|
+
#else
|
51
|
+
VALUE mutex;
|
52
|
+
#endif
|
53
|
+
uint8_t xrate;
|
54
|
+
bool mark;
|
24
55
|
} * Cache;
|
25
56
|
|
26
|
-
// almost the Murmur hash algorithm
|
27
|
-
#define M 0x5bd1e995
|
28
|
-
#define C1 0xCC9E2D51
|
29
|
-
#define C2 0x1B873593
|
30
|
-
#define N 0xE6546B64
|
31
|
-
|
32
57
|
void cache_set_form(Cache c, VALUE (*form)(const char *str, size_t len)) {
|
33
58
|
c->form = form;
|
34
59
|
}
|
35
60
|
|
36
|
-
|
37
|
-
// For debugging only.
|
38
|
-
static void cache_print(Cache c) {
|
39
|
-
for (uint32_t i = 0; i < c->size; i++) {
|
40
|
-
printf("%4d:", i);
|
41
|
-
for (Slot s = c->slots[i]; NULL != s; s = s->next) {
|
42
|
-
char buf[40];
|
43
|
-
strncpy(buf, s->key, s->klen);
|
44
|
-
buf[s->klen] = '\0';
|
45
|
-
printf(" %s", buf);
|
46
|
-
}
|
47
|
-
printf("\n");
|
48
|
-
}
|
49
|
-
}
|
50
|
-
#endif
|
51
|
-
|
52
|
-
static uint32_t hash_calc(const uint8_t *key, size_t len) {
|
61
|
+
static uint64_t hash_calc(const uint8_t *key, size_t len) {
|
53
62
|
const uint8_t *end = key + len;
|
54
63
|
const uint8_t *endless = key + (len & 0xFFFFFFFC);
|
55
|
-
|
56
|
-
|
64
|
+
uint64_t h = (uint64_t)len;
|
65
|
+
uint64_t k;
|
57
66
|
|
58
67
|
while (key < endless) {
|
59
|
-
k = (
|
60
|
-
k |= (
|
61
|
-
k |= (
|
62
|
-
k |= (
|
68
|
+
k = (uint64_t)*key++;
|
69
|
+
k |= (uint64_t)*key++ << 8;
|
70
|
+
k |= (uint64_t)*key++ << 16;
|
71
|
+
k |= (uint64_t)*key++ << 24;
|
63
72
|
|
64
73
|
k *= M;
|
65
74
|
k ^= k >> 24;
|
@@ -83,43 +92,25 @@ static uint32_t hash_calc(const uint8_t *key, size_t len) {
|
|
83
92
|
return h;
|
84
93
|
}
|
85
94
|
|
86
|
-
Cache cache_create(size_t size, VALUE (*form)(const char *str, size_t len), bool mark) {
|
87
|
-
Cache c = ALLOC(struct _cache);
|
88
|
-
int shift = 0;
|
89
|
-
|
90
|
-
for (; REHASH_LIMIT < size; size /= 2, shift++) {
|
91
|
-
}
|
92
|
-
if (shift < MIN_SHIFT) {
|
93
|
-
shift = MIN_SHIFT;
|
94
|
-
}
|
95
|
-
c->size = 1 << shift;
|
96
|
-
c->mask = c->size - 1;
|
97
|
-
c->slots = ALLOC_N(Slot, c->size);
|
98
|
-
memset(c->slots, 0, sizeof(Slot) * c->size);
|
99
|
-
c->form = form;
|
100
|
-
c->cnt = 0;
|
101
|
-
c->mark = mark;
|
102
|
-
|
103
|
-
return c;
|
104
|
-
}
|
105
|
-
|
106
95
|
static void rehash(Cache c) {
|
107
|
-
|
108
|
-
Slot * end
|
96
|
+
uint64_t osize;
|
97
|
+
Slot * end;
|
109
98
|
Slot * sp;
|
110
99
|
|
111
|
-
c->size
|
112
|
-
c->
|
113
|
-
|
114
|
-
|
115
|
-
|
100
|
+
osize = c->size;
|
101
|
+
c->size = osize * 4;
|
102
|
+
c->mask = c->size - 1;
|
103
|
+
c->slots = realloc((void *)c->slots, sizeof(Slot) * c->size);
|
104
|
+
memset((Slot *)c->slots + osize, 0, sizeof(Slot) * osize * 3);
|
105
|
+
end = (Slot *)c->slots + osize;
|
106
|
+
for (sp = (Slot *)c->slots; sp < end; sp++) {
|
116
107
|
Slot s = *sp;
|
117
108
|
Slot next = NULL;
|
118
109
|
|
119
110
|
*sp = NULL;
|
120
111
|
for (; NULL != s; s = next) {
|
121
|
-
|
122
|
-
Slot * bucket = c->slots + h;
|
112
|
+
uint64_t h = s->hash & c->mask;
|
113
|
+
Slot * bucket = (Slot *)c->slots + h;
|
123
114
|
|
124
115
|
next = s->next;
|
125
116
|
s->next = *bucket;
|
@@ -128,8 +119,148 @@ static void rehash(Cache c) {
|
|
128
119
|
}
|
129
120
|
}
|
130
121
|
|
122
|
+
static VALUE lockless_intern(Cache c, const char *key, size_t len) {
|
123
|
+
uint64_t h = hash_calc((const uint8_t *)key, len);
|
124
|
+
Slot * bucket = (Slot *)c->slots + (h & c->mask);
|
125
|
+
Slot b;
|
126
|
+
volatile VALUE rkey;
|
127
|
+
|
128
|
+
while (REUSE_MAX < c->rcnt) {
|
129
|
+
if (NULL != (b = c->reuse)) {
|
130
|
+
c->reuse = b->next;
|
131
|
+
free(b);
|
132
|
+
c->rcnt--;
|
133
|
+
} else {
|
134
|
+
// An accounting error occured somewhere so correct it.
|
135
|
+
c->rcnt = 0;
|
136
|
+
}
|
137
|
+
}
|
138
|
+
for (b = *bucket; NULL != b; b = b->next) {
|
139
|
+
if ((uint8_t)len == b->klen && 0 == strncmp(b->key, key, len)) {
|
140
|
+
b->use_cnt += 16;
|
141
|
+
return b->val;
|
142
|
+
}
|
143
|
+
}
|
144
|
+
rkey = c->form(key, len);
|
145
|
+
if (NULL == (b = c->reuse)) {
|
146
|
+
b = calloc(1, sizeof(struct _slot));
|
147
|
+
} else {
|
148
|
+
c->reuse = b->next;
|
149
|
+
c->rcnt--;
|
150
|
+
}
|
151
|
+
b->hash = h;
|
152
|
+
memcpy(b->key, key, len);
|
153
|
+
b->klen = (uint8_t)len;
|
154
|
+
b->key[len] = '\0';
|
155
|
+
b->val = rkey;
|
156
|
+
b->use_cnt = 4;
|
157
|
+
b->next = *bucket;
|
158
|
+
*bucket = b;
|
159
|
+
c->cnt++; // Don't worry about wrapping. Worse case is the entry is removed and recreated.
|
160
|
+
if (REHASH_LIMIT < c->cnt / c->size) {
|
161
|
+
rehash(c);
|
162
|
+
}
|
163
|
+
return rkey;
|
164
|
+
}
|
165
|
+
|
166
|
+
static VALUE locking_intern(Cache c, const char *key, size_t len) {
|
167
|
+
uint64_t h;
|
168
|
+
Slot * bucket;
|
169
|
+
Slot b;
|
170
|
+
uint64_t old_size;
|
171
|
+
volatile VALUE rkey;
|
172
|
+
|
173
|
+
CACHE_LOCK(c);
|
174
|
+
while (REUSE_MAX < c->rcnt) {
|
175
|
+
if (NULL != (b = c->reuse)) {
|
176
|
+
c->reuse = b->next;
|
177
|
+
free(b);
|
178
|
+
c->rcnt--;
|
179
|
+
} else {
|
180
|
+
// An accounting error occured somewhere so correct it.
|
181
|
+
c->rcnt = 0;
|
182
|
+
}
|
183
|
+
}
|
184
|
+
h = hash_calc((const uint8_t *)key, len);
|
185
|
+
bucket = (Slot *)c->slots + (h & c->mask);
|
186
|
+
for (b = *bucket; NULL != b; b = b->next) {
|
187
|
+
if ((uint8_t)len == b->klen && 0 == strncmp(b->key, key, len)) {
|
188
|
+
b->use_cnt += 4;
|
189
|
+
CACHE_UNLOCK(c);
|
190
|
+
|
191
|
+
return b->val;
|
192
|
+
}
|
193
|
+
}
|
194
|
+
old_size = c->size;
|
195
|
+
// The creation of a new value may trigger a GC which be a problem if the
|
196
|
+
// cache is locked so make sure it is unlocked for the key value creation.
|
197
|
+
if (NULL != (b = c->reuse)) {
|
198
|
+
c->reuse = b->next;
|
199
|
+
c->rcnt--;
|
200
|
+
}
|
201
|
+
CACHE_UNLOCK(c);
|
202
|
+
if (NULL == b) {
|
203
|
+
b = calloc(1, sizeof(struct _slot));
|
204
|
+
}
|
205
|
+
rkey = c->form(key, len);
|
206
|
+
b->hash = h;
|
207
|
+
memcpy(b->key, key, len);
|
208
|
+
b->klen = (uint8_t)len;
|
209
|
+
b->key[len] = '\0';
|
210
|
+
b->val = rkey;
|
211
|
+
b->use_cnt = 16;
|
212
|
+
|
213
|
+
// Lock again to add the new entry.
|
214
|
+
CACHE_LOCK(c);
|
215
|
+
if (old_size != c->size) {
|
216
|
+
h = hash_calc((const uint8_t *)key, len);
|
217
|
+
bucket = (Slot *)c->slots + (h & c->mask);
|
218
|
+
}
|
219
|
+
b->next = *bucket;
|
220
|
+
*bucket = b;
|
221
|
+
c->cnt++; // Don't worry about wrapping. Worse case is the entry is removed and recreated.
|
222
|
+
if (REHASH_LIMIT < c->cnt / c->size) {
|
223
|
+
rehash(c);
|
224
|
+
}
|
225
|
+
CACHE_UNLOCK(c);
|
226
|
+
|
227
|
+
return rkey;
|
228
|
+
}
|
229
|
+
|
230
|
+
Cache cache_create(size_t size, VALUE (*form)(const char *str, size_t len), bool mark, bool locking) {
|
231
|
+
Cache c = calloc(1, sizeof(struct _cache));
|
232
|
+
int shift = 0;
|
233
|
+
|
234
|
+
for (; REHASH_LIMIT < size; size /= 2, shift++) {
|
235
|
+
}
|
236
|
+
if (shift < MIN_SHIFT) {
|
237
|
+
shift = MIN_SHIFT;
|
238
|
+
}
|
239
|
+
#if HAVE_PTHREAD_MUTEX_INIT
|
240
|
+
pthread_mutex_init(&c->mutex, NULL);
|
241
|
+
#else
|
242
|
+
c->mutex = rb_mutex_new();
|
243
|
+
#endif
|
244
|
+
c->size = 1 << shift;
|
245
|
+
c->mask = c->size - 1;
|
246
|
+
c->slots = calloc(c->size, sizeof(Slot));
|
247
|
+
c->form = form;
|
248
|
+
c->xrate = 1; // low
|
249
|
+
c->mark = mark;
|
250
|
+
if (locking) {
|
251
|
+
c->intern = locking_intern;
|
252
|
+
} else {
|
253
|
+
c->intern = lockless_intern;
|
254
|
+
}
|
255
|
+
return c;
|
256
|
+
}
|
257
|
+
|
258
|
+
void cache_set_expunge_rate(Cache c, int rate) {
|
259
|
+
c->xrate = (uint8_t)rate;
|
260
|
+
}
|
261
|
+
|
131
262
|
void cache_free(Cache c) {
|
132
|
-
|
263
|
+
uint64_t i;
|
133
264
|
|
134
265
|
for (i = 0; i < c->size; i++) {
|
135
266
|
Slot next;
|
@@ -137,57 +268,59 @@ void cache_free(Cache c) {
|
|
137
268
|
|
138
269
|
for (s = c->slots[i]; NULL != s; s = next) {
|
139
270
|
next = s->next;
|
140
|
-
|
271
|
+
free(s);
|
141
272
|
}
|
142
273
|
}
|
143
|
-
|
144
|
-
|
274
|
+
free((void *)c->slots);
|
275
|
+
free(c);
|
145
276
|
}
|
146
277
|
|
147
278
|
void cache_mark(Cache c) {
|
148
|
-
|
149
|
-
|
279
|
+
uint64_t i;
|
280
|
+
|
281
|
+
#if !HAVE_PTHREAD_MUTEX_INIT
|
282
|
+
rb_gc_mark(c->mutex);
|
283
|
+
#endif
|
284
|
+
if (0 == c->cnt) {
|
285
|
+
return;
|
286
|
+
}
|
287
|
+
for (i = 0; i < c->size; i++) {
|
288
|
+
Slot s;
|
289
|
+
Slot prev = NULL;
|
290
|
+
Slot next;
|
150
291
|
|
151
|
-
for (
|
152
|
-
|
153
|
-
|
292
|
+
for (s = c->slots[i]; NULL != s; s = next) {
|
293
|
+
next = s->next;
|
294
|
+
if (0 == s->use_cnt) {
|
295
|
+
if (NULL == prev) {
|
296
|
+
c->slots[i] = next;
|
297
|
+
} else {
|
298
|
+
prev->next = next;
|
299
|
+
}
|
300
|
+
c->cnt--;
|
301
|
+
s->next = c->reuse;
|
302
|
+
c->reuse = s;
|
303
|
+
c->rcnt++;
|
304
|
+
continue;
|
305
|
+
}
|
306
|
+
switch (c->xrate) {
|
307
|
+
case 0: break;
|
308
|
+
case 2: s->use_cnt -= 2; break;
|
309
|
+
case 3: s->use_cnt /= 2; break;
|
310
|
+
default: s->use_cnt--; break;
|
311
|
+
}
|
312
|
+
if (c->mark) {
|
154
313
|
rb_gc_mark(s->val);
|
155
314
|
}
|
315
|
+
prev = s;
|
156
316
|
}
|
157
317
|
}
|
158
318
|
}
|
159
319
|
|
160
320
|
VALUE
|
161
321
|
cache_intern(Cache c, const char *key, size_t len) {
|
162
|
-
if (CACHE_MAX_KEY
|
322
|
+
if (CACHE_MAX_KEY <= len) {
|
163
323
|
return c->form(key, len);
|
164
324
|
}
|
165
|
-
|
166
|
-
Slot * bucket = c->slots + (h & c->mask);
|
167
|
-
Slot b;
|
168
|
-
Slot tail = NULL;
|
169
|
-
|
170
|
-
for (b = *bucket; NULL != b; b = b->next) {
|
171
|
-
if ((uint8_t)len == b->klen && 0 == strncmp(b->key, key, len)) {
|
172
|
-
return b->val;
|
173
|
-
}
|
174
|
-
tail = b;
|
175
|
-
}
|
176
|
-
b = ALLOC(struct _slot);
|
177
|
-
b->hash = h;
|
178
|
-
b->next = NULL;
|
179
|
-
memcpy(b->key, key, len);
|
180
|
-
b->klen = (uint8_t)len;
|
181
|
-
b->key[len] = '\0';
|
182
|
-
b->val = c->form(key, len);
|
183
|
-
if (NULL == tail) {
|
184
|
-
*bucket = b;
|
185
|
-
} else {
|
186
|
-
tail->next = b;
|
187
|
-
}
|
188
|
-
c->cnt++;
|
189
|
-
if (REHASH_LIMIT < c->cnt / c->size) {
|
190
|
-
rehash(c);
|
191
|
-
}
|
192
|
-
return b->val;
|
325
|
+
return c->intern(c, key, len);
|
193
326
|
}
|