google-protobuf 3.23.4-x64-mingw-ucrt → 3.24.0.rc.2-x64-mingw-ucrt
Sign up to get free protection for your applications and to get access to all the features.
Potentially problematic release.
This version of google-protobuf might be problematic. Click here for more details.
- checksums.yaml +4 -4
- data/ext/google/protobuf_c/convert.h +0 -2
- data/ext/google/protobuf_c/defs.c +11 -10
- data/ext/google/protobuf_c/defs.h +0 -2
- data/ext/google/protobuf_c/map.c +4 -2
- data/ext/google/protobuf_c/map.h +0 -2
- data/ext/google/protobuf_c/message.c +3 -1
- data/ext/google/protobuf_c/message.h +0 -2
- data/ext/google/protobuf_c/protobuf.c +24 -148
- data/ext/google/protobuf_c/protobuf.h +21 -6
- data/ext/google/protobuf_c/repeated_field.c +5 -2
- data/ext/google/protobuf_c/repeated_field.h +0 -2
- data/ext/google/protobuf_c/ruby-upb.c +7292 -7043
- data/ext/google/protobuf_c/ruby-upb.h +3545 -2574
- data/lib/google/3.1/protobuf_c.so +0 -0
- data/lib/google/3.2/protobuf_c.so +0 -0
- data/lib/google/protobuf/descriptor_pb.rb +9 -1
- data/lib/google/protobuf/object_cache.rb +120 -0
- data/lib/google/protobuf/plugin_pb.rb +1 -1
- data/lib/google/protobuf/well_known_types.rb +2 -8
- data/lib/google/protobuf.rb +1 -0
- metadata +6 -5
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: 75b6a09437df43ad9e4c670a7357a6f2643b892c8365e1ba813a53c36685194a
|
4
|
+
data.tar.gz: 8e5ff18868cb5b8b05dba40f68c2d3f44a35669f8bcd9923fd2cf4954f26bb7a
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: 483f75e7a5f3fb767ffc93765ef63a25e731292cc1e9caebcdc6b05878b1e0215125fc7ed02922ba985a2552fc08d7d2c6580799b4b49c813c42d914d54014c5
|
7
|
+
data.tar.gz: a3715a8b18e215c2238bf204fadc058d76eec53807102cc0b641032841833e9ed2e5ee010522bc8ac599cc34cc6e098d8be487b39be560b3115e98f03854de11
|
@@ -73,6 +73,8 @@ static VALUE rb_str_maybe_null(const char* s) {
|
|
73
73
|
// -----------------------------------------------------------------------------
|
74
74
|
|
75
75
|
typedef struct {
|
76
|
+
// IMPORTANT: WB_PROTECTED objects must only use the RB_OBJ_WRITE()
|
77
|
+
// macro to update VALUE references, as to trigger write barriers.
|
76
78
|
VALUE def_to_descriptor; // Hash table of def* -> Ruby descriptor.
|
77
79
|
upb_DefPool* symtab;
|
78
80
|
} DescriptorPool;
|
@@ -97,7 +99,7 @@ static void DescriptorPool_free(void* _self) {
|
|
97
99
|
static const rb_data_type_t DescriptorPool_type = {
|
98
100
|
"Google::Protobuf::DescriptorPool",
|
99
101
|
{DescriptorPool_mark, DescriptorPool_free, NULL},
|
100
|
-
.flags = RUBY_TYPED_FREE_IMMEDIATELY,
|
102
|
+
.flags = RUBY_TYPED_FREE_IMMEDIATELY | RUBY_TYPED_WB_PROTECTED,
|
101
103
|
};
|
102
104
|
|
103
105
|
static DescriptorPool* ruby_to_DescriptorPool(VALUE val) {
|
@@ -125,11 +127,9 @@ static VALUE DescriptorPool_alloc(VALUE klass) {
|
|
125
127
|
self->def_to_descriptor = Qnil;
|
126
128
|
ret = TypedData_Wrap_Struct(klass, &DescriptorPool_type, self);
|
127
129
|
|
128
|
-
self->def_to_descriptor
|
130
|
+
RB_OBJ_WRITE(ret, &self->def_to_descriptor, rb_hash_new());
|
129
131
|
self->symtab = upb_DefPool_New();
|
130
|
-
|
131
|
-
|
132
|
-
return ret;
|
132
|
+
return ObjectCache_TryAdd(self->symtab, ret);
|
133
133
|
}
|
134
134
|
|
135
135
|
/*
|
@@ -601,7 +601,7 @@ upb_CType ruby_to_fieldtype(VALUE type) {
|
|
601
601
|
|
602
602
|
#define CONVERT(upb, ruby) \
|
603
603
|
if (SYM2ID(type) == rb_intern(#ruby)) { \
|
604
|
-
return kUpb_CType_##upb;
|
604
|
+
return kUpb_CType_##upb; \
|
605
605
|
}
|
606
606
|
|
607
607
|
CONVERT(Float, float);
|
@@ -624,7 +624,7 @@ upb_CType ruby_to_fieldtype(VALUE type) {
|
|
624
624
|
|
625
625
|
static VALUE descriptortype_to_ruby(upb_FieldType type) {
|
626
626
|
switch (type) {
|
627
|
-
#define CONVERT(upb, ruby)
|
627
|
+
#define CONVERT(upb, ruby) \
|
628
628
|
case kUpb_FieldType_##upb: \
|
629
629
|
return ID2SYM(rb_intern(#ruby));
|
630
630
|
CONVERT(Float, float);
|
@@ -709,7 +709,7 @@ static VALUE FieldDescriptor_label(VALUE _self) {
|
|
709
709
|
FieldDescriptor* self = ruby_to_FieldDescriptor(_self);
|
710
710
|
switch (upb_FieldDef_Label(self->fielddef)) {
|
711
711
|
#define CONVERT(upb, ruby) \
|
712
|
-
case kUpb_Label_##upb:
|
712
|
+
case kUpb_Label_##upb: \
|
713
713
|
return ID2SYM(rb_intern(#ruby));
|
714
714
|
|
715
715
|
CONVERT(Optional, optional);
|
@@ -1091,7 +1091,7 @@ static VALUE EnumDescriptor_name(VALUE _self) {
|
|
1091
1091
|
static VALUE EnumDescriptor_lookup_name(VALUE _self, VALUE name) {
|
1092
1092
|
EnumDescriptor* self = ruby_to_EnumDescriptor(_self);
|
1093
1093
|
const char* name_str = rb_id2name(SYM2ID(name));
|
1094
|
-
const upb_EnumValueDef
|
1094
|
+
const upb_EnumValueDef* ev =
|
1095
1095
|
upb_EnumDef_FindValueByName(self->enumdef, name_str);
|
1096
1096
|
if (ev) {
|
1097
1097
|
return INT2NUM(upb_EnumValueDef_Number(ev));
|
@@ -1110,7 +1110,8 @@ static VALUE EnumDescriptor_lookup_name(VALUE _self, VALUE name) {
|
|
1110
1110
|
static VALUE EnumDescriptor_lookup_value(VALUE _self, VALUE number) {
|
1111
1111
|
EnumDescriptor* self = ruby_to_EnumDescriptor(_self);
|
1112
1112
|
int32_t val = NUM2INT(number);
|
1113
|
-
const upb_EnumValueDef* ev =
|
1113
|
+
const upb_EnumValueDef* ev =
|
1114
|
+
upb_EnumDef_FindValueByNumber(self->enumdef, val);
|
1114
1115
|
if (ev) {
|
1115
1116
|
return ID2SYM(rb_intern(upb_EnumValueDef_Name(ev)));
|
1116
1117
|
} else {
|
data/ext/google/protobuf_c/map.c
CHANGED
@@ -93,7 +93,6 @@ VALUE Map_GetRubyWrapper(upb_Map* map, upb_CType key_type, TypeInfo value_type,
|
|
93
93
|
if (val == Qnil) {
|
94
94
|
val = Map_alloc(cMap);
|
95
95
|
Map* self;
|
96
|
-
ObjectCache_Add(map, val);
|
97
96
|
TypedData_Get_Struct(val, Map, &Map_type, self);
|
98
97
|
self->map = map;
|
99
98
|
self->arena = arena;
|
@@ -103,6 +102,7 @@ VALUE Map_GetRubyWrapper(upb_Map* map, upb_CType key_type, TypeInfo value_type,
|
|
103
102
|
const upb_MessageDef* val_m = self->value_type_info.def.msgdef;
|
104
103
|
self->value_type_class = Descriptor_DefToClass(val_m);
|
105
104
|
}
|
105
|
+
return ObjectCache_TryAdd(map, val);
|
106
106
|
}
|
107
107
|
|
108
108
|
return val;
|
@@ -319,7 +319,9 @@ static VALUE Map_init(int argc, VALUE* argv, VALUE _self) {
|
|
319
319
|
|
320
320
|
self->map = upb_Map_New(Arena_get(self->arena), self->key_type,
|
321
321
|
self->value_type_info.type);
|
322
|
-
|
322
|
+
VALUE stored = ObjectCache_TryAdd(self->map, _self);
|
323
|
+
(void)stored;
|
324
|
+
PBRUBY_ASSERT(stored == _self);
|
323
325
|
|
324
326
|
if (init_arg != Qnil) {
|
325
327
|
Map_merge_into_self(_self, init_arg);
|
data/ext/google/protobuf_c/map.h
CHANGED
@@ -108,7 +108,9 @@ void Message_InitPtr(VALUE self_, upb_Message* msg, VALUE arena) {
|
|
108
108
|
Message* self = ruby_to_Message(self_);
|
109
109
|
self->msg = msg;
|
110
110
|
RB_OBJ_WRITE(self_, &self->arena, arena);
|
111
|
-
|
111
|
+
VALUE stored = ObjectCache_TryAdd(msg, self_);
|
112
|
+
(void)stored;
|
113
|
+
PBRUBY_ASSERT(stored == self_);
|
112
114
|
}
|
113
115
|
|
114
116
|
VALUE Message_GetArena(VALUE msg_rb) {
|
@@ -195,7 +195,8 @@ const rb_data_type_t Arena_type = {
|
|
195
195
|
.flags = RUBY_TYPED_FREE_IMMEDIATELY | RUBY_TYPED_WB_PROTECTED,
|
196
196
|
};
|
197
197
|
|
198
|
-
static void*
|
198
|
+
static void *ruby_upb_allocfunc(upb_alloc *alloc, void *ptr, size_t oldsize,
|
199
|
+
size_t size) {
|
199
200
|
if (size == 0) {
|
200
201
|
xfree(ptr);
|
201
202
|
return NULL;
|
@@ -252,164 +253,40 @@ void Arena_register(VALUE module) {
|
|
252
253
|
// Object Cache
|
253
254
|
// -----------------------------------------------------------------------------
|
254
255
|
|
255
|
-
// A pointer -> Ruby Object cache that keeps references to Ruby wrapper
|
256
|
-
// objects. This allows us to look up any Ruby wrapper object by the address
|
257
|
-
// of the object it is wrapping. That way we can avoid ever creating two
|
258
|
-
// different wrapper objects for the same C object, which saves memory and
|
259
|
-
// preserves object identity.
|
260
|
-
//
|
261
|
-
// We use WeakMap for the cache. For Ruby <2.7 we also need a secondary Hash
|
262
|
-
// to store WeakMap keys because Ruby <2.7 WeakMap doesn't allow non-finalizable
|
263
|
-
// keys.
|
264
|
-
//
|
265
|
-
// We also need the secondary Hash if sizeof(long) < sizeof(VALUE), because this
|
266
|
-
// means it may not be possible to fit a pointer into a Fixnum. Keys are
|
267
|
-
// pointers, and if they fit into a Fixnum, Ruby doesn't collect them, but if
|
268
|
-
// they overflow and require allocating a Bignum, they could get collected
|
269
|
-
// prematurely, thus removing the cache entry. This happens on 64-bit Windows,
|
270
|
-
// on which pointers are 64 bits but longs are 32 bits. In this case, we enable
|
271
|
-
// the secondary Hash to hold the keys and prevent them from being collected.
|
272
|
-
|
273
|
-
#if RUBY_API_VERSION_CODE >= 20700 && SIZEOF_LONG >= SIZEOF_VALUE
|
274
|
-
#define USE_SECONDARY_MAP 0
|
275
|
-
#else
|
276
|
-
#define USE_SECONDARY_MAP 1
|
277
|
-
#endif
|
278
|
-
|
279
|
-
#if USE_SECONDARY_MAP
|
280
|
-
|
281
|
-
// Maps Numeric -> Object. The object is then used as a key into the WeakMap.
|
282
|
-
// This is needed for Ruby <2.7 where a number cannot be a key to WeakMap.
|
283
|
-
// The object is used only for its identity; it does not contain any data.
|
284
|
-
VALUE secondary_map = Qnil;
|
285
|
-
|
286
|
-
// Mutations to the map are under a mutex, because SeconaryMap_MaybeGC()
|
287
|
-
// iterates over the map which cannot happen in parallel with insertions, or
|
288
|
-
// Ruby will throw:
|
289
|
-
// can't add a new key into hash during iteration (RuntimeError)
|
290
|
-
VALUE secondary_map_mutex = Qnil;
|
291
|
-
|
292
|
-
// Lambda that will GC entries from the secondary map that are no longer present
|
293
|
-
// in the primary map.
|
294
|
-
VALUE gc_secondary_map_lambda = Qnil;
|
295
|
-
ID length;
|
296
|
-
|
297
|
-
extern VALUE weak_obj_cache;
|
298
|
-
|
299
|
-
static void SecondaryMap_Init() {
|
300
|
-
rb_gc_register_address(&secondary_map);
|
301
|
-
rb_gc_register_address(&gc_secondary_map_lambda);
|
302
|
-
rb_gc_register_address(&secondary_map_mutex);
|
303
|
-
secondary_map = rb_hash_new();
|
304
|
-
gc_secondary_map_lambda = rb_eval_string(
|
305
|
-
"->(secondary, weak) {\n"
|
306
|
-
" secondary.delete_if { |k, v| !weak.key?(v) }\n"
|
307
|
-
"}\n");
|
308
|
-
secondary_map_mutex = rb_mutex_new();
|
309
|
-
length = rb_intern("length");
|
310
|
-
}
|
311
|
-
|
312
|
-
// The secondary map is a regular Hash, and will never shrink on its own.
|
313
|
-
// The main object cache is a WeakMap that will automatically remove entries
|
314
|
-
// when the target object is no longer reachable, but unless we manually
|
315
|
-
// remove the corresponding entries from the secondary map, it will grow
|
316
|
-
// without bound.
|
317
|
-
//
|
318
|
-
// To avoid this unbounded growth we periodically remove entries from the
|
319
|
-
// secondary map that are no longer present in the WeakMap. The logic of
|
320
|
-
// how often to perform this GC is an artbirary tuning parameter that
|
321
|
-
// represents a straightforward CPU/memory tradeoff.
|
322
|
-
//
|
323
|
-
// Requires: secondary_map_mutex is held.
|
324
|
-
static void SecondaryMap_MaybeGC() {
|
325
|
-
PBRUBY_ASSERT(rb_mutex_locked_p(secondary_map_mutex) == Qtrue);
|
326
|
-
size_t weak_len = NUM2ULL(rb_funcall(weak_obj_cache, length, 0));
|
327
|
-
size_t secondary_len = RHASH_SIZE(secondary_map);
|
328
|
-
if (secondary_len < weak_len) {
|
329
|
-
// Logically this case should not be possible: a valid entry cannot exist in
|
330
|
-
// the weak table unless there is a corresponding entry in the secondary
|
331
|
-
// table. It should *always* be the case that secondary_len >= weak_len.
|
332
|
-
//
|
333
|
-
// However ObjectSpace::WeakMap#length (and therefore weak_len) is
|
334
|
-
// unreliable: it overreports its true length by including non-live objects.
|
335
|
-
// However these non-live objects are not yielded in iteration, so we may
|
336
|
-
// have previously deleted them from the secondary map in a previous
|
337
|
-
// invocation of SecondaryMap_MaybeGC().
|
338
|
-
//
|
339
|
-
// In this case, we can't measure any waste, so we just return.
|
340
|
-
return;
|
341
|
-
}
|
342
|
-
size_t waste = secondary_len - weak_len;
|
343
|
-
// GC if we could remove at least 2000 entries or 20% of the table size
|
344
|
-
// (whichever is greater). Since the cost of the GC pass is O(N), we
|
345
|
-
// want to make sure that we condition this on overall table size, to
|
346
|
-
// avoid O(N^2) CPU costs.
|
347
|
-
size_t threshold = PBRUBY_MAX(secondary_len * 0.2, 2000);
|
348
|
-
if (waste > threshold) {
|
349
|
-
rb_funcall(gc_secondary_map_lambda, rb_intern("call"), 2, secondary_map,
|
350
|
-
weak_obj_cache);
|
351
|
-
}
|
352
|
-
}
|
353
|
-
|
354
|
-
// Requires: secondary_map_mutex is held by this thread iff create == true.
|
355
|
-
static VALUE SecondaryMap_Get(VALUE key, bool create) {
|
356
|
-
PBRUBY_ASSERT(!create || rb_mutex_locked_p(secondary_map_mutex) == Qtrue);
|
357
|
-
VALUE ret = rb_hash_lookup(secondary_map, key);
|
358
|
-
if (ret == Qnil && create) {
|
359
|
-
SecondaryMap_MaybeGC();
|
360
|
-
ret = rb_class_new_instance(0, NULL, rb_cObject);
|
361
|
-
rb_hash_aset(secondary_map, key, ret);
|
362
|
-
}
|
363
|
-
return ret;
|
364
|
-
}
|
365
|
-
|
366
|
-
#endif
|
367
|
-
|
368
|
-
// Requires: secondary_map_mutex is held by this thread iff create == true.
|
369
|
-
static VALUE ObjectCache_GetKey(const void *key, bool create) {
|
370
|
-
VALUE key_val = (VALUE)key;
|
371
|
-
PBRUBY_ASSERT((key_val & 3) == 0);
|
372
|
-
VALUE ret = LL2NUM(key_val >> 2);
|
373
|
-
#if USE_SECONDARY_MAP
|
374
|
-
ret = SecondaryMap_Get(ret, create);
|
375
|
-
#endif
|
376
|
-
return ret;
|
377
|
-
}
|
378
|
-
|
379
256
|
// Public ObjectCache API.
|
380
257
|
|
381
258
|
VALUE weak_obj_cache = Qnil;
|
382
259
|
ID item_get;
|
383
|
-
ID
|
260
|
+
ID item_try_add;
|
261
|
+
|
262
|
+
static void ObjectCache_Init(VALUE protobuf) {
|
263
|
+
item_get = rb_intern("get");
|
264
|
+
item_try_add = rb_intern("try_add");
|
384
265
|
|
385
|
-
static void ObjectCache_Init() {
|
386
266
|
rb_gc_register_address(&weak_obj_cache);
|
387
|
-
|
388
|
-
|
389
|
-
|
390
|
-
|
391
|
-
#if USE_SECONDARY_MAP
|
392
|
-
SecondaryMap_Init();
|
267
|
+
#if SIZEOF_LONG >= SIZEOF_VALUE
|
268
|
+
VALUE cache_class = rb_const_get(protobuf, rb_intern("ObjectCache"));
|
269
|
+
#else
|
270
|
+
VALUE cache_class = rb_const_get(protobuf, rb_intern("LegacyObjectCache"));
|
393
271
|
#endif
|
272
|
+
|
273
|
+
weak_obj_cache = rb_class_new_instance(0, NULL, cache_class);
|
274
|
+
rb_const_set(protobuf, rb_intern("OBJECT_CACHE"), weak_obj_cache);
|
275
|
+
rb_const_set(protobuf, rb_intern("SIZEOF_LONG"), INT2NUM(SIZEOF_LONG));
|
276
|
+
rb_const_set(protobuf, rb_intern("SIZEOF_VALUE"), INT2NUM(SIZEOF_VALUE));
|
394
277
|
}
|
395
278
|
|
396
|
-
|
397
|
-
|
398
|
-
|
399
|
-
|
400
|
-
#endif
|
401
|
-
VALUE key_rb = ObjectCache_GetKey(key, true);
|
402
|
-
rb_funcall(weak_obj_cache, item_set, 2, key_rb, val);
|
403
|
-
#if USE_SECONDARY_MAP
|
404
|
-
rb_mutex_unlock(secondary_map_mutex);
|
405
|
-
#endif
|
406
|
-
PBRUBY_ASSERT(ObjectCache_Get(key) == val);
|
279
|
+
VALUE ObjectCache_TryAdd(const void *key, VALUE val) {
|
280
|
+
VALUE key_val = (VALUE)key;
|
281
|
+
PBRUBY_ASSERT((key_val & 3) == 0);
|
282
|
+
return rb_funcall(weak_obj_cache, item_try_add, 2, LL2NUM(key_val), val);
|
407
283
|
}
|
408
284
|
|
409
285
|
// Returns the cached object for this key, if any. Otherwise returns Qnil.
|
410
286
|
VALUE ObjectCache_Get(const void *key) {
|
411
|
-
VALUE
|
412
|
-
|
287
|
+
VALUE key_val = (VALUE)key;
|
288
|
+
PBRUBY_ASSERT((key_val & 3) == 0);
|
289
|
+
return rb_funcall(weak_obj_cache, item_get, 1, LL2NUM(key_val));
|
413
290
|
}
|
414
291
|
|
415
292
|
/*
|
@@ -459,11 +336,10 @@ VALUE Google_Protobuf_deep_copy(VALUE self, VALUE obj) {
|
|
459
336
|
// This must be named "Init_protobuf_c" because the Ruby module is named
|
460
337
|
// "protobuf_c" -- the VM looks for this symbol in our .so.
|
461
338
|
__attribute__((visibility("default"))) void Init_protobuf_c() {
|
462
|
-
ObjectCache_Init();
|
463
|
-
|
464
339
|
VALUE google = rb_define_module("Google");
|
465
340
|
VALUE protobuf = rb_define_module_under(google, "Protobuf");
|
466
341
|
|
342
|
+
ObjectCache_Init(protobuf);
|
467
343
|
Arena_register(protobuf);
|
468
344
|
Defs_register(protobuf);
|
469
345
|
RepeatedField_register(protobuf);
|
@@ -31,8 +31,22 @@
|
|
31
31
|
#ifndef __GOOGLE_PROTOBUF_RUBY_PROTOBUF_H__
|
32
32
|
#define __GOOGLE_PROTOBUF_RUBY_PROTOBUF_H__
|
33
33
|
|
34
|
+
// Ruby 3+ defines NDEBUG itself, see: https://bugs.ruby-lang.org/issues/18777
|
35
|
+
#ifdef NDEBUG
|
36
|
+
#include <ruby.h>
|
37
|
+
#else
|
38
|
+
#include <ruby.h>
|
39
|
+
#undef NDEBUG
|
40
|
+
#endif
|
41
|
+
|
42
|
+
#include <ruby/version.h>
|
43
|
+
|
44
|
+
#if RUBY_API_VERSION_CODE < 20700
|
45
|
+
#error Protobuf requires Ruby >= 2.7
|
46
|
+
#endif
|
47
|
+
|
48
|
+
#include <assert.h> // Must be included after the NDEBUG logic above.
|
34
49
|
#include <ruby/encoding.h>
|
35
|
-
#include <ruby/ruby.h>
|
36
50
|
#include <ruby/vm.h>
|
37
51
|
|
38
52
|
#include "defs.h"
|
@@ -76,10 +90,9 @@ void Arena_Pin(VALUE arena, VALUE obj);
|
|
76
90
|
// being collected (though in Ruby <2.7 is it effectively strong, due to
|
77
91
|
// implementation limitations).
|
78
92
|
|
79
|
-
//
|
80
|
-
//
|
81
|
-
|
82
|
-
void ObjectCache_Add(const void* key, VALUE val);
|
93
|
+
// Tries to add a new entry to the cache, returning the newly installed value or
|
94
|
+
// the pre-existing entry.
|
95
|
+
VALUE ObjectCache_TryAdd(const void* key, VALUE val);
|
83
96
|
|
84
97
|
// Returns the cached object for this key, if any. Otherwise returns Qnil.
|
85
98
|
VALUE ObjectCache_Get(const void* key);
|
@@ -110,7 +123,9 @@ extern VALUE cTypeError;
|
|
110
123
|
do { \
|
111
124
|
} while (false && (expr))
|
112
125
|
#else
|
113
|
-
#define PBRUBY_ASSERT(expr)
|
126
|
+
#define PBRUBY_ASSERT(expr) \
|
127
|
+
if (!(expr)) \
|
128
|
+
rb_bug("Assertion failed at %s:%d, expr: %s", __FILE__, __LINE__, #expr)
|
114
129
|
#endif
|
115
130
|
|
116
131
|
#define PBRUBY_MAX(x, y) (((x) > (y)) ? (x) : (y))
|
@@ -87,7 +87,6 @@ VALUE RepeatedField_GetRubyWrapper(upb_Array* array, TypeInfo type_info,
|
|
87
87
|
if (val == Qnil) {
|
88
88
|
val = RepeatedField_alloc(cRepeatedField);
|
89
89
|
RepeatedField* self;
|
90
|
-
ObjectCache_Add(array, val);
|
91
90
|
TypedData_Get_Struct(val, RepeatedField, &RepeatedField_type, self);
|
92
91
|
self->array = array;
|
93
92
|
self->arena = arena;
|
@@ -95,11 +94,14 @@ VALUE RepeatedField_GetRubyWrapper(upb_Array* array, TypeInfo type_info,
|
|
95
94
|
if (self->type_info.type == kUpb_CType_Message) {
|
96
95
|
self->type_class = Descriptor_DefToClass(type_info.def.msgdef);
|
97
96
|
}
|
97
|
+
val = ObjectCache_TryAdd(array, val);
|
98
98
|
}
|
99
99
|
|
100
100
|
PBRUBY_ASSERT(ruby_to_RepeatedField(val)->type_info.type == type_info.type);
|
101
101
|
PBRUBY_ASSERT(ruby_to_RepeatedField(val)->type_info.def.msgdef ==
|
102
102
|
type_info.def.msgdef);
|
103
|
+
PBRUBY_ASSERT(ruby_to_RepeatedField(val)->array == array);
|
104
|
+
|
103
105
|
return val;
|
104
106
|
}
|
105
107
|
|
@@ -613,7 +615,8 @@ VALUE RepeatedField_init(int argc, VALUE* argv, VALUE _self) {
|
|
613
615
|
|
614
616
|
self->type_info = TypeInfo_FromClass(argc, argv, 0, &self->type_class, &ary);
|
615
617
|
self->array = upb_Array_New(arena, self->type_info.type);
|
616
|
-
|
618
|
+
VALUE stored_val = ObjectCache_TryAdd(self->array, _self);
|
619
|
+
PBRUBY_ASSERT(stored_val == _self);
|
617
620
|
|
618
621
|
if (ary != Qnil) {
|
619
622
|
if (!RB_TYPE_P(ary, T_ARRAY)) {
|