debase-ruby_core_source 0.4.0 → 0.5.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (44) hide show
  1. checksums.yaml +4 -4
  2. data/CHANGELOG.md +5 -1
  3. data/README.md +1 -1
  4. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/addr2line.h +21 -0
  5. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/constant.h +36 -0
  6. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/dln.h +50 -0
  7. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/encdb.h +169 -0
  8. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/eval_intern.h +241 -0
  9. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/gc.h +104 -0
  10. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/id.h +135 -0
  11. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/insns.inc +187 -0
  12. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/insns_info.inc +724 -0
  13. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/internal.h +395 -0
  14. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/iseq.h +140 -0
  15. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/known_errors.inc +731 -0
  16. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/method.h +138 -0
  17. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/node.h +541 -0
  18. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/node_name.inc +212 -0
  19. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/opt_sc.inc +702 -0
  20. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/optinsn.inc +83 -0
  21. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/optunifs.inc +120 -0
  22. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/parse.h +181 -0
  23. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/probes_helper.h +67 -0
  24. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/regenc.h +227 -0
  25. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/regint.h +915 -0
  26. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/regparse.h +367 -0
  27. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/revision.h +1 -0
  28. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/ruby_atomic.h +170 -0
  29. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/siphash.h +48 -0
  30. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/thread_pthread.h +56 -0
  31. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/thread_win32.h +45 -0
  32. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/timev.h +21 -0
  33. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/transcode_data.h +127 -0
  34. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/transdb.h +193 -0
  35. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/version.h +52 -0
  36. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/vm.inc +3196 -0
  37. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/vm_core.h +1015 -0
  38. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/vm_debug.h +41 -0
  39. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/vm_exec.h +173 -0
  40. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/vm_insnhelper.h +274 -0
  41. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/vm_opts.h +56 -0
  42. data/lib/debase/ruby_core_source/ruby-2.0.0-p576/vmtc.inc +101 -0
  43. data/lib/debase/ruby_core_source/version.rb +1 -1
  44. metadata +41 -2
@@ -0,0 +1,1015 @@
1
+ /**********************************************************************
2
+
3
+ vm_core.h -
4
+
5
+ $Author: usa $
6
+ created at: 04/01/01 19:41:38 JST
7
+
8
+ Copyright (C) 2004-2007 Koichi Sasada
9
+
10
+ **********************************************************************/
11
+
12
+ #ifndef RUBY_VM_CORE_H
13
+ #define RUBY_VM_CORE_H
14
+
15
+ #define RUBY_VM_THREAD_MODEL 2
16
+
17
+ #include "ruby/ruby.h"
18
+ #include "ruby/st.h"
19
+
20
+ #include "node.h"
21
+ #include "vm_debug.h"
22
+ #include "vm_opts.h"
23
+ #include "id.h"
24
+ #include "method.h"
25
+ #include "ruby_atomic.h"
26
+
27
+ #if defined(_WIN32)
28
+ #include "thread_win32.h"
29
+ #elif defined(HAVE_PTHREAD_H)
30
+ #include "thread_pthread.h"
31
+ #else
32
+ #error "unsupported thread type"
33
+ #endif
34
+
35
+ #ifndef ENABLE_VM_OBJSPACE
36
+ #ifdef _WIN32
37
+ /*
38
+ * TODO: object space independent st_table.
39
+ * socklist needs st_table in rb_w32_sysinit(), before object space
40
+ * initialization.
41
+ * It is too early now to change st_hash_type, since it breaks binary
42
+ * compatibility.
43
+ */
44
+ #define ENABLE_VM_OBJSPACE 0
45
+ #else
46
+ #define ENABLE_VM_OBJSPACE 1
47
+ #endif
48
+ #endif
49
+
50
+ #include <setjmp.h>
51
+ #include <signal.h>
52
+
53
+ #ifndef NSIG
54
+ # define NSIG (_SIGMAX + 1) /* For QNX */
55
+ #endif
56
+
57
+ #define RUBY_NSIG NSIG
58
+
59
+ #ifdef HAVE_STDARG_PROTOTYPES
60
+ #include <stdarg.h>
61
+ #define va_init_list(a,b) va_start((a),(b))
62
+ #else
63
+ #include <varargs.h>
64
+ #define va_init_list(a,b) va_start((a))
65
+ #endif
66
+
67
+ #if defined(SIGSEGV) && defined(HAVE_SIGALTSTACK) && defined(SA_SIGINFO) && !defined(__NetBSD__)
68
+ #define USE_SIGALTSTACK
69
+ #endif
70
+
71
+ /*****************/
72
+ /* configuration */
73
+ /*****************/
74
+
75
+ /* gcc ver. check */
76
+ #if defined(__GNUC__) && __GNUC__ >= 2
77
+
78
+ #if OPT_TOKEN_THREADED_CODE
79
+ #if OPT_DIRECT_THREADED_CODE
80
+ #undef OPT_DIRECT_THREADED_CODE
81
+ #endif
82
+ #endif
83
+
84
+ #else /* defined(__GNUC__) && __GNUC__ >= 2 */
85
+
86
+ /* disable threaded code options */
87
+ #if OPT_DIRECT_THREADED_CODE
88
+ #undef OPT_DIRECT_THREADED_CODE
89
+ #endif
90
+ #if OPT_TOKEN_THREADED_CODE
91
+ #undef OPT_TOKEN_THREADED_CODE
92
+ #endif
93
+ #endif
94
+
95
+ #ifdef __native_client__
96
+ #undef OPT_DIRECT_THREADED_CODE
97
+ #endif
98
+
99
+ /* call threaded code */
100
+ #if OPT_CALL_THREADED_CODE
101
+ #if OPT_DIRECT_THREADED_CODE
102
+ #undef OPT_DIRECT_THREADED_CODE
103
+ #endif /* OPT_DIRECT_THREADED_CODE */
104
+ #if OPT_STACK_CACHING
105
+ #undef OPT_STACK_CACHING
106
+ #endif /* OPT_STACK_CACHING */
107
+ #endif /* OPT_CALL_THREADED_CODE */
108
+
109
+ /* likely */
110
+ #if __GNUC__ >= 3
111
+ #define LIKELY(x) (__builtin_expect((x), 1))
112
+ #define UNLIKELY(x) (__builtin_expect((x), 0))
113
+ #else /* __GNUC__ >= 3 */
114
+ #define LIKELY(x) (x)
115
+ #define UNLIKELY(x) (x)
116
+ #endif /* __GNUC__ >= 3 */
117
+
118
+ #if __GNUC__ >= 3
119
+ #define UNINITIALIZED_VAR(x) x = x
120
+ #else
121
+ #define UNINITIALIZED_VAR(x) x
122
+ #endif
123
+
124
+ typedef unsigned long rb_num_t;
125
+
126
+ /* iseq data type */
127
+
128
+ struct iseq_compile_data_ensure_node_stack;
129
+
130
+ typedef struct rb_compile_option_struct rb_compile_option_t;
131
+
132
+ struct iseq_inline_cache_entry {
133
+ VALUE ic_vmstat;
134
+ VALUE ic_class;
135
+ union {
136
+ VALUE value;
137
+ long index;
138
+ } ic_value;
139
+ };
140
+
141
+ /* to avoid warning */
142
+ struct rb_thread_struct;
143
+ struct rb_control_frame_struct;
144
+
145
+ /* rb_call_info_t contains calling information including inline cache */
146
+ typedef struct rb_call_info_struct {
147
+ /* fixed at compile time */
148
+ ID mid;
149
+ VALUE flag;
150
+ int orig_argc;
151
+ rb_iseq_t *blockiseq;
152
+
153
+ /* inline cache: keys */
154
+ VALUE vmstat;
155
+ VALUE klass;
156
+
157
+ /* inline cache: values */
158
+ const rb_method_entry_t *me;
159
+ VALUE defined_class;
160
+
161
+ /* temporary values for method calling */
162
+ int argc;
163
+ struct rb_block_struct *blockptr;
164
+ VALUE recv;
165
+ union {
166
+ int opt_pc; /* used by iseq */
167
+ long index; /* used by ivar */
168
+ int missing_reason; /* used by method_missing */
169
+ int inc_sp; /* used by cfunc */
170
+ } aux;
171
+
172
+ VALUE (*call)(struct rb_thread_struct *th, struct rb_control_frame_struct *cfp, struct rb_call_info_struct *ci);
173
+ } rb_call_info_t;
174
+
175
+ #if 1
176
+ #define GetCoreDataFromValue(obj, type, ptr) do { \
177
+ (ptr) = (type*)DATA_PTR(obj); \
178
+ } while (0)
179
+ #else
180
+ #define GetCoreDataFromValue(obj, type, ptr) Data_Get_Struct((obj), type, (ptr))
181
+ #endif
182
+
183
+ #define GetISeqPtr(obj, ptr) \
184
+ GetCoreDataFromValue((obj), rb_iseq_t, (ptr))
185
+
186
+ typedef struct rb_iseq_location_struct {
187
+ VALUE path;
188
+ VALUE absolute_path;
189
+ VALUE base_label;
190
+ VALUE label;
191
+ size_t first_lineno;
192
+ } rb_iseq_location_t;
193
+
194
+ struct rb_iseq_struct;
195
+
196
+ struct rb_iseq_struct {
197
+ /***************/
198
+ /* static data */
199
+ /***************/
200
+
201
+ enum iseq_type {
202
+ ISEQ_TYPE_TOP,
203
+ ISEQ_TYPE_METHOD,
204
+ ISEQ_TYPE_BLOCK,
205
+ ISEQ_TYPE_CLASS,
206
+ ISEQ_TYPE_RESCUE,
207
+ ISEQ_TYPE_ENSURE,
208
+ ISEQ_TYPE_EVAL,
209
+ ISEQ_TYPE_MAIN,
210
+ ISEQ_TYPE_DEFINED_GUARD
211
+ } type; /* instruction sequence type */
212
+
213
+ rb_iseq_location_t location;
214
+
215
+ VALUE *iseq; /* iseq (insn number and operands) */
216
+ VALUE *iseq_encoded; /* encoded iseq */
217
+ unsigned long iseq_size;
218
+ VALUE mark_ary; /* Array: includes operands which should be GC marked */
219
+ VALUE coverage; /* coverage array */
220
+
221
+ /* insn info, must be freed */
222
+ struct iseq_line_info_entry *line_info_table;
223
+ size_t line_info_size;
224
+
225
+ ID *local_table; /* must free */
226
+ int local_table_size;
227
+
228
+ /* sizeof(vars) + 1 */
229
+ int local_size;
230
+
231
+ struct iseq_inline_cache_entry *ic_entries;
232
+ int ic_size;
233
+
234
+ rb_call_info_t *callinfo_entries;
235
+ int callinfo_size;
236
+
237
+ /**
238
+ * argument information
239
+ *
240
+ * def m(a1, a2, ..., aM, # mandatory
241
+ * b1=(...), b2=(...), ..., bN=(...), # optional
242
+ * *c, # rest
243
+ * d1, d2, ..., dO, # post
244
+ * e1:(...), e2:(...), ..., eK:(...), # keyword
245
+ * **f, # keyword rest
246
+ * &g) # block
247
+ * =>
248
+ *
249
+ * argc = M // or 0 if no mandatory arg
250
+ * arg_opts = N+1 // or 0 if no optional arg
251
+ * arg_rest = M+N // or -1 if no rest arg
252
+ * arg_opt_table = [ (arg_opts entries) ]
253
+ * arg_post_start = M+N+(*1) // or 0 if no post arguments
254
+ * arg_post_len = O // or 0 if no post arguments
255
+ * arg_keywords = K // or 0 if no keyword arg
256
+ * arg_block = M+N+(*1)+O+K // or -1 if no block arg
257
+ * arg_keyword = M+N+(*1)+O+K+(&1) // or -1 if no keyword arg/rest
258
+ * arg_simple = 0 if not simple arguments.
259
+ * = 1 if no opt, rest, post, block.
260
+ * = 2 if ambiguous block parameter ({|a|}).
261
+ * arg_size = M+N+O+(*1)+K+(&1)+(**1) argument size.
262
+ */
263
+
264
+ int argc;
265
+ int arg_simple;
266
+ int arg_rest;
267
+ int arg_block;
268
+ int arg_opts;
269
+ int arg_post_len;
270
+ int arg_post_start;
271
+ int arg_size;
272
+ VALUE *arg_opt_table;
273
+ int arg_keyword;
274
+ int arg_keyword_check; /* if this is true, raise an ArgumentError when unknown keyword argument is passed */
275
+ int arg_keywords;
276
+ ID *arg_keyword_table;
277
+
278
+ size_t stack_max; /* for stack overflow check */
279
+
280
+ /* catch table */
281
+ struct iseq_catch_table_entry *catch_table;
282
+ int catch_table_size;
283
+
284
+ /* for child iseq */
285
+ struct rb_iseq_struct *parent_iseq;
286
+ struct rb_iseq_struct *local_iseq;
287
+
288
+ /****************/
289
+ /* dynamic data */
290
+ /****************/
291
+
292
+ VALUE self;
293
+ VALUE orig; /* non-NULL if its data have origin */
294
+
295
+ /* block inlining */
296
+ /*
297
+ * NODE *node;
298
+ * void *special_block_builder;
299
+ * void *cached_special_block_builder;
300
+ * VALUE cached_special_block;
301
+ */
302
+
303
+ /* klass/module nest information stack (cref) */
304
+ NODE *cref_stack;
305
+ VALUE klass;
306
+
307
+ /* misc */
308
+ ID defined_method_id; /* for define_method */
309
+ rb_num_t flip_cnt;
310
+
311
+ /* used at compile time */
312
+ struct iseq_compile_data *compile_data;
313
+ };
314
+
315
+ enum ruby_special_exceptions {
316
+ ruby_error_reenter,
317
+ ruby_error_nomemory,
318
+ ruby_error_sysstack,
319
+ ruby_error_closed_stream,
320
+ ruby_special_error_count
321
+ };
322
+
323
+ #define GetVMPtr(obj, ptr) \
324
+ GetCoreDataFromValue((obj), rb_vm_t, (ptr))
325
+
326
+ #if defined(ENABLE_VM_OBJSPACE) && ENABLE_VM_OBJSPACE
327
+ struct rb_objspace;
328
+ void rb_objspace_free(struct rb_objspace *);
329
+ #endif
330
+
331
+ typedef struct rb_hook_list_struct {
332
+ struct rb_event_hook_struct *hooks;
333
+ rb_event_flag_t events;
334
+ int need_clean;
335
+ } rb_hook_list_t;
336
+
337
+ typedef struct rb_vm_struct {
338
+ VALUE self;
339
+
340
+ rb_global_vm_lock_t gvl;
341
+ rb_thread_lock_t thread_destruct_lock;
342
+
343
+ struct rb_thread_struct *main_thread;
344
+ struct rb_thread_struct *running_thread;
345
+
346
+ st_table *living_threads;
347
+ VALUE thgroup_default;
348
+
349
+ int running;
350
+ int thread_abort_on_exception;
351
+ int trace_running;
352
+ volatile int sleeper;
353
+
354
+ /* object management */
355
+ VALUE mark_object_ary;
356
+
357
+ VALUE special_exceptions[ruby_special_error_count];
358
+
359
+ /* load */
360
+ VALUE top_self;
361
+ VALUE load_path;
362
+ VALUE load_path_snapshot;
363
+ VALUE load_path_check_cache;
364
+ VALUE expanded_load_path;
365
+ VALUE loaded_features;
366
+ VALUE loaded_features_snapshot;
367
+ struct st_table *loaded_features_index;
368
+ struct st_table *loading_table;
369
+
370
+ /* signal */
371
+ struct {
372
+ VALUE cmd;
373
+ int safe;
374
+ } trap_list[RUBY_NSIG];
375
+
376
+ /* hook */
377
+ rb_hook_list_t event_hooks;
378
+
379
+ int src_encoding_index;
380
+
381
+ VALUE verbose, debug, progname;
382
+ VALUE coverages;
383
+
384
+ struct unlinked_method_entry_list_entry *unlinked_method_entry_list;
385
+
386
+ #if defined(ENABLE_VM_OBJSPACE) && ENABLE_VM_OBJSPACE
387
+ struct rb_objspace *objspace;
388
+ #endif
389
+
390
+ /*
391
+ * @shyouhei notes that this is not for storing normal Ruby
392
+ * objects so do *NOT* mark this when you GC.
393
+ */
394
+ struct RArray at_exit;
395
+
396
+ VALUE *defined_strings;
397
+
398
+ /* params */
399
+ struct { /* size in byte */
400
+ size_t thread_vm_stack_size;
401
+ size_t thread_machine_stack_size;
402
+ size_t fiber_vm_stack_size;
403
+ size_t fiber_machine_stack_size;
404
+ } default_params;
405
+ } rb_vm_t;
406
+
407
+ /* default values */
408
+
409
+ #define RUBY_VM_SIZE_ALIGN 4096
410
+
411
+ #define RUBY_VM_THREAD_VM_STACK_SIZE ( 128 * 1024 * sizeof(VALUE)) /* 512 KB or 1024 KB */
412
+ #define RUBY_VM_THREAD_VM_STACK_SIZE_MIN ( 2 * 1024 * sizeof(VALUE)) /* 8 KB or 16 KB */
413
+ #define RUBY_VM_THREAD_MACHINE_STACK_SIZE ( 128 * 1024 * sizeof(VALUE)) /* 512 KB or 1024 KB */
414
+ #define RUBY_VM_THREAD_MACHINE_STACK_SIZE_MIN ( 16 * 1024 * sizeof(VALUE)) /* 64 KB or 128 KB */
415
+
416
+ #define RUBY_VM_FIBER_VM_STACK_SIZE ( 16 * 1024 * sizeof(VALUE)) /* 64 KB or 128 KB */
417
+ #define RUBY_VM_FIBER_VM_STACK_SIZE_MIN ( 2 * 1024 * sizeof(VALUE)) /* 8 KB or 16 KB */
418
+ #define RUBY_VM_FIBER_MACHINE_STACK_SIZE ( 64 * 1024 * sizeof(VALUE)) /* 256 KB or 512 KB */
419
+ #define RUBY_VM_FIBER_MACHINE_STACK_SIZE_MIN ( 16 * 1024 * sizeof(VALUE)) /* 64 KB or 128 KB */
420
+
421
+ #ifndef VM_DEBUG_BP_CHECK
422
+ #define VM_DEBUG_BP_CHECK 0
423
+ #endif
424
+
425
+ typedef struct rb_control_frame_struct {
426
+ VALUE *pc; /* cfp[0] */
427
+ VALUE *sp; /* cfp[1] */
428
+ rb_iseq_t *iseq; /* cfp[2] */
429
+ VALUE flag; /* cfp[3] */
430
+ VALUE self; /* cfp[4] / block[0] */
431
+ VALUE klass; /* cfp[5] / block[1] */
432
+ VALUE *ep; /* cfp[6] / block[2] */
433
+ rb_iseq_t *block_iseq; /* cfp[7] / block[3] */
434
+ VALUE proc; /* cfp[8] / block[4] */
435
+ const rb_method_entry_t *me;/* cfp[9] */
436
+
437
+ #if VM_DEBUG_BP_CHECK
438
+ VALUE *bp_check; /* cfp[10] */
439
+ #endif
440
+ } rb_control_frame_t;
441
+
442
+ typedef struct rb_block_struct {
443
+ VALUE self; /* share with method frame if it's only block */
444
+ VALUE klass; /* share with method frame if it's only block */
445
+ VALUE *ep; /* share with method frame if it's only block */
446
+ rb_iseq_t *iseq;
447
+ VALUE proc;
448
+ } rb_block_t;
449
+
450
+ extern const rb_data_type_t ruby_threadptr_data_type;
451
+
452
+ #define GetThreadPtr(obj, ptr) \
453
+ TypedData_Get_Struct((obj), rb_thread_t, &ruby_threadptr_data_type, (ptr))
454
+
455
+ enum rb_thread_status {
456
+ THREAD_RUNNABLE,
457
+ THREAD_STOPPED,
458
+ THREAD_STOPPED_FOREVER,
459
+ THREAD_KILLED
460
+ };
461
+
462
+ typedef RUBY_JMP_BUF rb_jmpbuf_t;
463
+
464
+ /*
465
+ the members which are written in TH_PUSH_TAG() should be placed at
466
+ the beginning and the end, so that entire region is accessible.
467
+ */
468
+ struct rb_vm_tag {
469
+ VALUE tag;
470
+ VALUE retval;
471
+ rb_jmpbuf_t buf;
472
+ struct rb_vm_tag *prev;
473
+ };
474
+
475
+ struct rb_vm_protect_tag {
476
+ struct rb_vm_protect_tag *prev;
477
+ };
478
+
479
+ struct rb_unblock_callback {
480
+ rb_unblock_function_t *func;
481
+ void *arg;
482
+ };
483
+
484
+ struct rb_mutex_struct;
485
+
486
+ struct rb_thread_struct;
487
+ typedef struct rb_thread_list_struct{
488
+ struct rb_thread_list_struct *next;
489
+ struct rb_thread_struct *th;
490
+ } rb_thread_list_t;
491
+
492
+
493
+ typedef struct rb_thread_struct {
494
+ VALUE self;
495
+ rb_vm_t *vm;
496
+
497
+ /* execution information */
498
+ VALUE *stack; /* must free, must mark */
499
+ size_t stack_size; /* size in word (byte size / sizeof(VALUE)) */
500
+ rb_control_frame_t *cfp;
501
+ int safe_level;
502
+ int raised_flag;
503
+ VALUE last_status; /* $? */
504
+
505
+ /* passing state */
506
+ int state;
507
+
508
+ int waiting_fd;
509
+
510
+ /* for rb_iterate */
511
+ const rb_block_t *passed_block;
512
+
513
+ /* for bmethod */
514
+ const rb_method_entry_t *passed_me;
515
+
516
+ /* for cfunc */
517
+ rb_call_info_t *passed_ci;
518
+
519
+ /* for load(true) */
520
+ VALUE top_self;
521
+ VALUE top_wrapper;
522
+
523
+ /* eval env */
524
+ rb_block_t *base_block;
525
+
526
+ VALUE *root_lep;
527
+ VALUE root_svar;
528
+
529
+ /* thread control */
530
+ rb_thread_id_t thread_id;
531
+ enum rb_thread_status status;
532
+ int to_kill;
533
+ int priority;
534
+
535
+ native_thread_data_t native_thread_data;
536
+ void *blocking_region_buffer;
537
+
538
+ VALUE thgroup;
539
+ VALUE value;
540
+
541
+ /* temporary place of errinfo */
542
+ VALUE errinfo;
543
+
544
+ /* temporary place of retval on OPT_CALL_THREADED_CODE */
545
+ #if OPT_CALL_THREADED_CODE
546
+ VALUE retval;
547
+ #endif
548
+
549
+ /* async errinfo queue */
550
+ VALUE pending_interrupt_queue;
551
+ int pending_interrupt_queue_checked;
552
+ VALUE pending_interrupt_mask_stack;
553
+
554
+ rb_atomic_t interrupt_flag;
555
+ unsigned long interrupt_mask;
556
+ rb_thread_lock_t interrupt_lock;
557
+ struct rb_unblock_callback unblock;
558
+ VALUE locking_mutex;
559
+ struct rb_mutex_struct *keeping_mutexes;
560
+
561
+ struct rb_vm_tag *tag;
562
+ struct rb_vm_protect_tag *protect_tag;
563
+
564
+ /*! Thread-local state of evaluation context.
565
+ *
566
+ * If negative, this thread is evaluating the main program.
567
+ * If positive, this thread is evaluating a program under Kernel::eval
568
+ * family.
569
+ */
570
+ int parse_in_eval;
571
+
572
+ /*! Thread-local state of compiling context.
573
+ *
574
+ * If non-zero, the parser does not automatically print error messages to
575
+ * stderr. */
576
+ int mild_compile_error;
577
+
578
+ /* storage */
579
+ st_table *local_storage;
580
+
581
+ rb_thread_list_t *join_list;
582
+
583
+ VALUE first_proc;
584
+ VALUE first_args;
585
+ VALUE (*first_func)(ANYARGS);
586
+
587
+ /* for GC */
588
+ VALUE *machine_stack_start;
589
+ VALUE *machine_stack_end;
590
+ size_t machine_stack_maxsize;
591
+ #ifdef __ia64
592
+ VALUE *machine_register_stack_start;
593
+ VALUE *machine_register_stack_end;
594
+ size_t machine_register_stack_maxsize;
595
+ #endif
596
+ jmp_buf machine_regs;
597
+ int mark_stack_len;
598
+
599
+ /* statistics data for profiler */
600
+ VALUE stat_insn_usage;
601
+
602
+ /* tracer */
603
+ rb_hook_list_t event_hooks;
604
+ struct rb_trace_arg_struct *trace_arg; /* trace information */
605
+
606
+ /* fiber */
607
+ VALUE fiber;
608
+ VALUE root_fiber;
609
+ rb_jmpbuf_t root_jmpbuf;
610
+
611
+ /* misc */
612
+ int method_missing_reason;
613
+ int abort_on_exception;
614
+ #ifdef USE_SIGALTSTACK
615
+ void *altstack;
616
+ #endif
617
+ unsigned long running_time_us;
618
+ } rb_thread_t;
619
+
620
+ typedef enum {
621
+ VM_DEFINECLASS_TYPE_CLASS = 0x00,
622
+ VM_DEFINECLASS_TYPE_SINGLETON_CLASS = 0x01,
623
+ VM_DEFINECLASS_TYPE_MODULE = 0x02,
624
+ /* 0x03..0x06 is reserved */
625
+ VM_DEFINECLASS_TYPE_MASK = 0x07,
626
+ } rb_vm_defineclass_type_t;
627
+
628
+ #define VM_DEFINECLASS_TYPE(x) ((rb_vm_defineclass_type_t)(x) & VM_DEFINECLASS_TYPE_MASK)
629
+ #define VM_DEFINECLASS_FLAG_SCOPED 0x08
630
+ #define VM_DEFINECLASS_FLAG_HAS_SUPERCLASS 0x10
631
+ #define VM_DEFINECLASS_SCOPED_P(x) ((x) & VM_DEFINECLASS_FLAG_SCOPED)
632
+ #define VM_DEFINECLASS_HAS_SUPERCLASS_P(x) \
633
+ ((x) & VM_DEFINECLASS_FLAG_HAS_SUPERCLASS)
634
+
635
+ /* iseq.c */
636
+ #if defined __GNUC__ && __GNUC__ >= 4
637
+ #pragma GCC visibility push(default)
638
+ #endif
639
+
640
+ /* node -> iseq */
641
+ VALUE rb_iseq_new(NODE*, VALUE, VALUE, VALUE, VALUE, enum iseq_type);
642
+ VALUE rb_iseq_new_top(NODE *node, VALUE name, VALUE path, VALUE absolute_path, VALUE parent);
643
+ VALUE rb_iseq_new_main(NODE *node, VALUE path, VALUE absolute_path);
644
+ VALUE rb_iseq_new_with_bopt(NODE*, VALUE, VALUE, VALUE, VALUE, VALUE, enum iseq_type, VALUE);
645
+ VALUE rb_iseq_new_with_opt(NODE*, VALUE, VALUE, VALUE, VALUE, VALUE, enum iseq_type, const rb_compile_option_t*);
646
+
647
+ /* src -> iseq */
648
+ VALUE rb_iseq_compile(VALUE src, VALUE file, VALUE line);
649
+ VALUE rb_iseq_compile_on_base(VALUE src, VALUE file, VALUE line, rb_block_t *base_block);
650
+ VALUE rb_iseq_compile_with_option(VALUE src, VALUE file, VALUE absolute_path, VALUE line, rb_block_t *base_block, VALUE opt);
651
+
652
+ VALUE rb_iseq_disasm(VALUE self);
653
+ int rb_iseq_disasm_insn(VALUE str, VALUE *iseqval, size_t pos, rb_iseq_t *iseq, VALUE child);
654
+ const char *ruby_node_name(int node);
655
+ int rb_iseq_first_lineno(const rb_iseq_t *iseq);
656
+
657
+ RUBY_EXTERN VALUE rb_cISeq;
658
+ RUBY_EXTERN VALUE rb_cRubyVM;
659
+ RUBY_EXTERN VALUE rb_cEnv;
660
+ RUBY_EXTERN VALUE rb_mRubyVMFrozenCore;
661
+ #if defined __GNUC__ && __GNUC__ >= 4
662
+ #pragma GCC visibility pop
663
+ #endif
664
+
665
+ #define GetProcPtr(obj, ptr) \
666
+ GetCoreDataFromValue((obj), rb_proc_t, (ptr))
667
+
668
+ typedef struct {
669
+ rb_block_t block;
670
+
671
+ VALUE envval; /* for GC mark */
672
+ VALUE blockprocval;
673
+ int safe_level;
674
+ int is_from_method;
675
+ int is_lambda;
676
+ } rb_proc_t;
677
+
678
+ #define GetEnvPtr(obj, ptr) \
679
+ GetCoreDataFromValue((obj), rb_env_t, (ptr))
680
+
681
+ typedef struct {
682
+ VALUE *env;
683
+ int env_size;
684
+ int local_size;
685
+ VALUE prev_envval; /* for GC mark */
686
+ rb_block_t block;
687
+ } rb_env_t;
688
+
689
+ #define GetBindingPtr(obj, ptr) \
690
+ GetCoreDataFromValue((obj), rb_binding_t, (ptr))
691
+
692
+ typedef struct {
693
+ VALUE env;
694
+ VALUE path;
695
+ unsigned short first_lineno;
696
+ } rb_binding_t;
697
+
698
+ /* used by compile time and send insn */
699
+
700
+ enum vm_check_match_type {
701
+ VM_CHECKMATCH_TYPE_WHEN = 1,
702
+ VM_CHECKMATCH_TYPE_CASE = 2,
703
+ VM_CHECKMATCH_TYPE_RESCUE = 3
704
+ };
705
+
706
+ #define VM_CHECKMATCH_TYPE_MASK 0x03
707
+ #define VM_CHECKMATCH_ARRAY 0x04
708
+
709
+ #define VM_CALL_ARGS_SPLAT (0x01 << 1) /* m(*args) */
710
+ #define VM_CALL_ARGS_BLOCKARG (0x01 << 2) /* m(&block) */
711
+ #define VM_CALL_FCALL (0x01 << 3) /* m(...) */
712
+ #define VM_CALL_VCALL (0x01 << 4) /* m */
713
+ #define VM_CALL_TAILCALL (0x01 << 5) /* located at tail position */
714
+ #define VM_CALL_SUPER (0x01 << 6) /* super */
715
+ #define VM_CALL_OPT_SEND (0x01 << 7) /* internal flag */
716
+ #define VM_CALL_ARGS_SKIP_SETUP (0x01 << 8) /* (flag & (SPLAT|BLOCKARG)) && blockiseq == 0 */
717
+
718
+ enum vm_special_object_type {
719
+ VM_SPECIAL_OBJECT_VMCORE = 1,
720
+ VM_SPECIAL_OBJECT_CBASE,
721
+ VM_SPECIAL_OBJECT_CONST_BASE
722
+ };
723
+
724
+ #define VM_FRAME_MAGIC_METHOD 0x11
725
+ #define VM_FRAME_MAGIC_BLOCK 0x21
726
+ #define VM_FRAME_MAGIC_CLASS 0x31
727
+ #define VM_FRAME_MAGIC_TOP 0x41
728
+ #define VM_FRAME_MAGIC_CFUNC 0x61
729
+ #define VM_FRAME_MAGIC_PROC 0x71
730
+ #define VM_FRAME_MAGIC_IFUNC 0x81
731
+ #define VM_FRAME_MAGIC_EVAL 0x91
732
+ #define VM_FRAME_MAGIC_LAMBDA 0xa1
733
+ #define VM_FRAME_MAGIC_RESCUE 0xb1
734
+ #define VM_FRAME_MAGIC_MASK_BITS 8
735
+ #define VM_FRAME_MAGIC_MASK (~(~0<<VM_FRAME_MAGIC_MASK_BITS))
736
+
737
+ #define VM_FRAME_TYPE(cfp) ((cfp)->flag & VM_FRAME_MAGIC_MASK)
738
+
739
+ /* other frame flag */
740
+ #define VM_FRAME_FLAG_PASSED 0x0100
741
+ #define VM_FRAME_FLAG_FINISH 0x0200
742
+ #define VM_FRAME_TYPE_FINISH_P(cfp) (((cfp)->flag & VM_FRAME_FLAG_FINISH) != 0)
743
+
744
+ #define RUBYVM_CFUNC_FRAME_P(cfp) \
745
+ (VM_FRAME_TYPE(cfp) == VM_FRAME_MAGIC_CFUNC)
746
+
747
+ /* inline cache */
748
+ typedef struct iseq_inline_cache_entry *IC;
749
+ typedef rb_call_info_t *CALL_INFO;
750
+
751
+ void rb_vm_change_state(void);
752
+
753
+ typedef VALUE CDHASH;
754
+
755
+ #ifndef FUNC_FASTCALL
756
+ #define FUNC_FASTCALL(x) x
757
+ #endif
758
+
759
+ typedef rb_control_frame_t *
760
+ (FUNC_FASTCALL(*rb_insn_func_t))(rb_thread_t *, rb_control_frame_t *);
761
+
762
+ #define GC_GUARDED_PTR(p) ((VALUE)((VALUE)(p) | 0x01))
763
+ #define GC_GUARDED_PTR_REF(p) ((void *)(((VALUE)(p)) & ~0x03))
764
+ #define GC_GUARDED_PTR_P(p) (((VALUE)(p)) & 0x01)
765
+
766
+ /*
767
+ * block frame:
768
+ * ep[ 0]: prev frame
769
+ * ep[-1]: CREF (for *_eval)
770
+ *
771
+ * method frame:
772
+ * ep[ 0]: block pointer (ptr | VM_ENVVAL_BLOCK_PTR_FLAG)
773
+ */
774
+
775
+ #define VM_ENVVAL_BLOCK_PTR_FLAG 0x02
776
+ #define VM_ENVVAL_BLOCK_PTR(v) (GC_GUARDED_PTR(v) | VM_ENVVAL_BLOCK_PTR_FLAG)
777
+ #define VM_ENVVAL_BLOCK_PTR_P(v) ((v) & VM_ENVVAL_BLOCK_PTR_FLAG)
778
+ #define VM_ENVVAL_PREV_EP_PTR(v) ((VALUE)GC_GUARDED_PTR(v))
779
+ #define VM_ENVVAL_PREV_EP_PTR_P(v) (!(VM_ENVVAL_BLOCK_PTR_P(v)))
780
+
781
+ #define VM_EP_PREV_EP(ep) ((VALUE *)GC_GUARDED_PTR_REF((ep)[0]))
782
+ #define VM_EP_BLOCK_PTR(ep) ((rb_block_t *)GC_GUARDED_PTR_REF((ep)[0]))
783
+ #define VM_EP_LEP_P(ep) VM_ENVVAL_BLOCK_PTR_P((ep)[0])
784
+
785
+ VALUE *rb_vm_ep_local_ep(VALUE *ep);
786
+ rb_block_t *rb_vm_control_frame_block_ptr(rb_control_frame_t *cfp);
787
+
788
+ #define RUBY_VM_PREVIOUS_CONTROL_FRAME(cfp) ((cfp)+1)
789
+ #define RUBY_VM_NEXT_CONTROL_FRAME(cfp) ((cfp)-1)
790
+ #define RUBY_VM_END_CONTROL_FRAME(th) \
791
+ ((rb_control_frame_t *)((th)->stack + (th)->stack_size))
792
+ #define RUBY_VM_VALID_CONTROL_FRAME_P(cfp, ecfp) \
793
+ ((void *)(ecfp) > (void *)(cfp))
794
+ #define RUBY_VM_CONTROL_FRAME_STACK_OVERFLOW_P(th, cfp) \
795
+ (!RUBY_VM_VALID_CONTROL_FRAME_P((cfp), RUBY_VM_END_CONTROL_FRAME(th)))
796
+
797
+ #define RUBY_VM_IFUNC_P(ptr) (BUILTIN_TYPE(ptr) == T_NODE)
798
+ #define RUBY_VM_NORMAL_ISEQ_P(ptr) \
799
+ ((ptr) && !RUBY_VM_IFUNC_P(ptr))
800
+
801
+ #define RUBY_VM_GET_BLOCK_PTR_IN_CFP(cfp) ((rb_block_t *)(&(cfp)->self))
802
+ #define RUBY_VM_GET_CFP_FROM_BLOCK_PTR(b) \
803
+ ((rb_control_frame_t *)((VALUE *)(b) - 4))
804
+ /* magic number `4' is depend on rb_control_frame_t layout. */
805
+
806
+ /* VM related object allocate functions */
807
+ VALUE rb_thread_alloc(VALUE klass);
808
+ VALUE rb_proc_alloc(VALUE klass);
809
+
810
+ /* for debug */
811
+ extern void rb_vmdebug_stack_dump_raw(rb_thread_t *, rb_control_frame_t *);
812
+ extern void rb_vmdebug_debug_print_pre(rb_thread_t *th, rb_control_frame_t *cfp);
813
+ extern void rb_vmdebug_debug_print_post(rb_thread_t *th, rb_control_frame_t *cfp);
814
+
815
+ #define SDR() rb_vmdebug_stack_dump_raw(GET_THREAD(), GET_THREAD()->cfp)
816
+ #define SDR2(cfp) rb_vmdebug_stack_dump_raw(GET_THREAD(), (cfp))
817
+ void rb_vm_bugreport(void);
818
+
819
+ /* functions about thread/vm execution */
820
+ #if defined __GNUC__ && __GNUC__ >= 4
821
+ #pragma GCC visibility push(default)
822
+ #endif
823
+ VALUE rb_iseq_eval(VALUE iseqval);
824
+ VALUE rb_iseq_eval_main(VALUE iseqval);
825
+ #if defined __GNUC__ && __GNUC__ >= 4
826
+ #pragma GCC visibility pop
827
+ #endif
828
+ int rb_thread_method_id_and_class(rb_thread_t *th, ID *idp, VALUE *klassp);
829
+
830
+ VALUE rb_vm_invoke_proc(rb_thread_t *th, rb_proc_t *proc,
831
+ int argc, const VALUE *argv, const rb_block_t *blockptr);
832
+ VALUE rb_vm_make_proc(rb_thread_t *th, const rb_block_t *block, VALUE klass);
833
+ VALUE rb_vm_make_env_object(rb_thread_t *th, rb_control_frame_t *cfp);
834
+ VALUE rb_binding_new_with_cfp(rb_thread_t *th, const rb_control_frame_t *src_cfp);
835
+ void rb_vm_inc_const_missing_count(void);
836
+ void rb_vm_gvl_destroy(rb_vm_t *vm);
837
+ VALUE rb_vm_call(rb_thread_t *th, VALUE recv, VALUE id, int argc,
838
+ const VALUE *argv, const rb_method_entry_t *me,
839
+ VALUE defined_class);
840
+ void rb_unlink_method_entry(rb_method_entry_t *me);
841
+ void rb_gc_mark_unlinked_live_method_entries(void *pvm);
842
+
843
+ void rb_thread_start_timer_thread(void);
844
+ void rb_thread_stop_timer_thread(int);
845
+ void rb_thread_reset_timer_thread(void);
846
+ void rb_thread_wakeup_timer_thread(void);
847
+
848
+ int ruby_thread_has_gvl_p(void);
849
+ typedef int rb_backtrace_iter_func(void *, VALUE, int, VALUE);
850
+ rb_control_frame_t *rb_vm_get_ruby_level_next_cfp(rb_thread_t *th, const rb_control_frame_t *cfp);
851
+ rb_control_frame_t *rb_vm_get_binding_creatable_next_cfp(rb_thread_t *th, const rb_control_frame_t *cfp);
852
+ int rb_vm_get_sourceline(const rb_control_frame_t *);
853
+ VALUE rb_name_err_mesg_new(VALUE obj, VALUE mesg, VALUE recv, VALUE method);
854
+ void rb_vm_stack_to_heap(rb_thread_t *th);
855
+ void ruby_thread_init_stack(rb_thread_t *th);
856
+ int rb_vm_control_frame_id_and_class(const rb_control_frame_t *cfp, ID *idp, VALUE *klassp);
857
+ void rb_vm_rewind_cfp(rb_thread_t *th, rb_control_frame_t *cfp);
858
+
859
+ void rb_gc_mark_machine_stack(rb_thread_t *th);
860
+
861
+ int rb_autoloading_value(VALUE mod, ID id, VALUE* value);
862
+
863
+ #define sysstack_error GET_VM()->special_exceptions[ruby_error_sysstack]
864
+
865
+ #define CHECK_VM_STACK_OVERFLOW(cfp, margin) do \
866
+ if ((VALUE *)((char *)(((VALUE *)(cfp)->sp) + (margin)) + sizeof(rb_control_frame_t)) >= ((VALUE *)(cfp))) { \
867
+ vm_stackoverflow(); \
868
+ } \
869
+ while (0)
870
+
871
+ /* for thread */
872
+
873
+ #if RUBY_VM_THREAD_MODEL == 2
874
+ extern rb_thread_t *ruby_current_thread;
875
+ extern rb_vm_t *ruby_current_vm;
876
+ extern rb_event_flag_t ruby_vm_event_flags;
877
+
878
+ #define GET_VM() ruby_current_vm
879
+
880
+ #ifndef OPT_CALL_CFUNC_WITHOUT_FRAME
881
+ #define OPT_CALL_CFUNC_WITHOUT_FRAME 0
882
+ #endif
883
+
884
+ static inline rb_thread_t *
885
+ GET_THREAD(void)
886
+ {
887
+ rb_thread_t *th = ruby_current_thread;
888
+ #if OPT_CALL_CFUNC_WITHOUT_FRAME
889
+ if (UNLIKELY(th->passed_ci != 0)) {
890
+ void vm_call_cfunc_push_frame(rb_thread_t *th);
891
+ vm_call_cfunc_push_frame(th);
892
+ }
893
+ #endif
894
+ return th;
895
+ }
896
+
897
+ #define rb_thread_set_current_raw(th) (void)(ruby_current_thread = (th))
898
+ #define rb_thread_set_current(th) do { \
899
+ if ((th)->vm->running_thread != (th)) { \
900
+ (th)->running_time_us = 0; \
901
+ } \
902
+ rb_thread_set_current_raw(th); \
903
+ (th)->vm->running_thread = (th); \
904
+ } while (0)
905
+
906
+ #else
907
+ #error "unsupported thread model"
908
+ #endif
909
+
910
+ enum {
911
+ TIMER_INTERRUPT_MASK = 0x01,
912
+ PENDING_INTERRUPT_MASK = 0x02,
913
+ FINALIZER_INTERRUPT_MASK = 0x04,
914
+ TRAP_INTERRUPT_MASK = 0x08
915
+ };
916
+
917
+ #define RUBY_VM_SET_TIMER_INTERRUPT(th) ATOMIC_OR((th)->interrupt_flag, TIMER_INTERRUPT_MASK)
918
+ #define RUBY_VM_SET_INTERRUPT(th) ATOMIC_OR((th)->interrupt_flag, PENDING_INTERRUPT_MASK)
919
+ #define RUBY_VM_SET_FINALIZER_INTERRUPT(th) ATOMIC_OR((th)->interrupt_flag, FINALIZER_INTERRUPT_MASK)
920
+ #define RUBY_VM_SET_TRAP_INTERRUPT(th) ATOMIC_OR((th)->interrupt_flag, TRAP_INTERRUPT_MASK)
921
+ #define RUBY_VM_INTERRUPTED(th) ((th)->interrupt_flag & ~(th)->interrupt_mask & (PENDING_INTERRUPT_MASK|TRAP_INTERRUPT_MASK))
922
+ #define RUBY_VM_INTERRUPTED_ANY(th) ((th)->interrupt_flag & ~(th)->interrupt_mask)
923
+
924
+ int rb_signal_buff_size(void);
925
+ void rb_signal_exec(rb_thread_t *th, int sig);
926
+ void rb_threadptr_check_signal(rb_thread_t *mth);
927
+ void rb_threadptr_signal_raise(rb_thread_t *th, int sig);
928
+ void rb_threadptr_signal_exit(rb_thread_t *th);
929
+ void rb_threadptr_execute_interrupts(rb_thread_t *, int);
930
+ void rb_threadptr_interrupt(rb_thread_t *th);
931
+ void rb_threadptr_unlock_all_locking_mutexes(rb_thread_t *th);
932
+ void rb_threadptr_pending_interrupt_clear(rb_thread_t *th);
933
+ void rb_threadptr_pending_interrupt_enque(rb_thread_t *th, VALUE v);
934
+ int rb_threadptr_pending_interrupt_active_p(rb_thread_t *th);
935
+
936
+ void rb_thread_lock_unlock(rb_thread_lock_t *);
937
+ void rb_thread_lock_destroy(rb_thread_lock_t *);
938
+
939
+ #define RUBY_VM_CHECK_INTS_BLOCKING(th) do { \
940
+ if (UNLIKELY(!rb_threadptr_pending_interrupt_empty_p(th))) { \
941
+ th->pending_interrupt_queue_checked = 0; \
942
+ RUBY_VM_SET_INTERRUPT(th); \
943
+ rb_threadptr_execute_interrupts(th, 1); \
944
+ } \
945
+ else if (UNLIKELY(RUBY_VM_INTERRUPTED_ANY(th))) { \
946
+ rb_threadptr_execute_interrupts(th, 1); \
947
+ } \
948
+ } while (0)
949
+
950
+ #define RUBY_VM_CHECK_INTS(th) do { \
951
+ if (UNLIKELY(RUBY_VM_INTERRUPTED_ANY(th))) { \
952
+ rb_threadptr_execute_interrupts(th, 0); \
953
+ } \
954
+ } while (0)
955
+
956
+ /* tracer */
957
+ struct rb_trace_arg_struct {
958
+ rb_event_flag_t event;
959
+ rb_thread_t *th;
960
+ rb_control_frame_t *cfp;
961
+ VALUE self;
962
+ ID id;
963
+ VALUE klass;
964
+ VALUE data;
965
+
966
+ int klass_solved;
967
+
968
+ /* calc from cfp */
969
+ int lineno;
970
+ VALUE path;
971
+ };
972
+
973
+ void rb_threadptr_exec_event_hooks(struct rb_trace_arg_struct *trace_arg);
974
+ void rb_threadptr_exec_event_hooks_and_pop_frame(struct rb_trace_arg_struct *trace_arg);
975
+
976
+ #define EXEC_EVENT_HOOK_ORIG(th_, flag_, self_, id_, klass_, data_, pop_p_) do { \
977
+ if (UNLIKELY(ruby_vm_event_flags & (flag_))) { \
978
+ if (((th)->event_hooks.events | (th)->vm->event_hooks.events) & (flag_)) { \
979
+ struct rb_trace_arg_struct trace_arg; \
980
+ trace_arg.event = (flag_); \
981
+ trace_arg.th = (th_); \
982
+ trace_arg.cfp = (trace_arg.th)->cfp; \
983
+ trace_arg.self = (self_); \
984
+ trace_arg.id = (id_); \
985
+ trace_arg.klass = (klass_); \
986
+ trace_arg.data = (data_); \
987
+ trace_arg.path = Qundef; \
988
+ trace_arg.klass_solved = 0; \
989
+ if (pop_p_) rb_threadptr_exec_event_hooks_and_pop_frame(&trace_arg); \
990
+ else rb_threadptr_exec_event_hooks(&trace_arg); \
991
+ } \
992
+ } \
993
+ } while (0)
994
+
995
+ #define EXEC_EVENT_HOOK(th_, flag_, self_, id_, klass_, data_) \
996
+ EXEC_EVENT_HOOK_ORIG(th_, flag_, self_, id_, klass_, data_, 0)
997
+
998
+ #define EXEC_EVENT_HOOK_AND_POP_FRAME(th_, flag_, self_, id_, klass_, data_) \
999
+ EXEC_EVENT_HOOK_ORIG(th_, flag_, self_, id_, klass_, data_, 1)
1000
+
1001
+ #if defined __GNUC__ && __GNUC__ >= 4
1002
+ #pragma GCC visibility push(default)
1003
+ #endif
1004
+
1005
+ int rb_thread_check_trap_pending(void);
1006
+
1007
+ extern VALUE rb_get_coverages(void);
1008
+ extern void rb_set_coverages(VALUE);
1009
+ extern void rb_reset_coverages(void);
1010
+
1011
+ #if defined __GNUC__ && __GNUC__ >= 4
1012
+ #pragma GCC visibility pop
1013
+ #endif
1014
+
1015
+ #endif /* RUBY_VM_CORE_H */