ruby-prof 1.1.0-x64-mingw32 → 1.4.2-x64-mingw32
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/CHANGES +48 -1
- data/Rakefile +2 -14
- data/bin/ruby-prof +100 -152
- data/ext/ruby_prof/extconf.rb +8 -28
- data/ext/ruby_prof/rp_aggregate_call_tree.c +59 -0
- data/ext/ruby_prof/rp_aggregate_call_tree.h +13 -0
- data/ext/ruby_prof/rp_allocation.c +67 -59
- data/ext/ruby_prof/rp_allocation.h +3 -3
- data/ext/ruby_prof/rp_call_tree.c +369 -0
- data/ext/ruby_prof/rp_call_tree.h +43 -0
- data/ext/ruby_prof/rp_call_trees.c +288 -0
- data/ext/ruby_prof/rp_call_trees.h +28 -0
- data/ext/ruby_prof/rp_measure_allocations.c +12 -14
- data/ext/ruby_prof/rp_measure_process_time.c +12 -14
- data/ext/ruby_prof/rp_measure_wall_time.c +17 -15
- data/ext/ruby_prof/rp_measurement.c +47 -40
- data/ext/ruby_prof/rp_measurement.h +7 -7
- data/ext/ruby_prof/rp_method.c +116 -255
- data/ext/ruby_prof/rp_method.h +31 -39
- data/ext/ruby_prof/rp_profile.c +316 -303
- data/ext/ruby_prof/rp_profile.h +1 -3
- data/ext/ruby_prof/rp_stack.c +122 -106
- data/ext/ruby_prof/rp_stack.h +17 -20
- data/ext/ruby_prof/rp_thread.c +136 -111
- data/ext/ruby_prof/rp_thread.h +12 -9
- data/ext/ruby_prof/ruby_prof.c +27 -23
- data/ext/ruby_prof/ruby_prof.h +9 -0
- data/ext/ruby_prof/vc/ruby_prof.sln +8 -0
- data/ext/ruby_prof/vc/ruby_prof.vcxproj +22 -7
- data/lib/2.7/ruby_prof.so +0 -0
- data/lib/ruby-prof.rb +5 -5
- data/lib/ruby-prof/assets/call_stack_printer.html.erb +4 -7
- data/lib/ruby-prof/assets/graph_printer.html.erb +5 -6
- data/lib/ruby-prof/{call_info.rb → call_tree.rb} +6 -6
- data/lib/ruby-prof/call_tree_visitor.rb +36 -0
- data/lib/ruby-prof/compatibility.rb +0 -10
- data/lib/ruby-prof/measurement.rb +5 -2
- data/lib/ruby-prof/method_info.rb +3 -15
- data/lib/ruby-prof/printers/abstract_printer.rb +12 -2
- data/lib/ruby-prof/printers/call_info_printer.rb +12 -10
- data/lib/ruby-prof/printers/call_stack_printer.rb +20 -22
- data/lib/ruby-prof/printers/call_tree_printer.rb +1 -1
- data/lib/ruby-prof/printers/dot_printer.rb +3 -3
- data/lib/ruby-prof/printers/flat_printer.rb +3 -2
- data/lib/ruby-prof/printers/graph_printer.rb +4 -5
- data/lib/ruby-prof/printers/multi_printer.rb +2 -2
- data/lib/ruby-prof/profile.rb +8 -4
- data/lib/ruby-prof/rack.rb +51 -127
- data/lib/ruby-prof/thread.rb +3 -18
- data/lib/ruby-prof/version.rb +1 -1
- data/ruby-prof.gemspec +7 -0
- data/test/alias_test.rb +42 -45
- data/test/basic_test.rb +0 -86
- data/test/{call_info_visitor_test.rb → call_tree_visitor_test.rb} +6 -5
- data/test/call_trees_test.rb +66 -0
- data/test/exclude_methods_test.rb +17 -12
- data/test/fiber_test.rb +95 -39
- data/test/gc_test.rb +36 -42
- data/test/inverse_call_tree_test.rb +175 -0
- data/test/line_number_test.rb +67 -70
- data/test/marshal_test.rb +7 -13
- data/test/measure_allocations_test.rb +224 -234
- data/test/measure_allocations_trace_test.rb +224 -234
- data/test/measure_memory_trace_test.rb +814 -469
- data/test/measure_process_time_test.rb +0 -64
- data/test/measure_times.rb +2 -0
- data/test/measure_wall_time_test.rb +34 -58
- data/test/pause_resume_test.rb +19 -10
- data/test/prime.rb +1 -3
- data/test/prime_script.rb +6 -0
- data/test/printer_call_stack_test.rb +0 -1
- data/test/printer_call_tree_test.rb +0 -1
- data/test/printer_flat_test.rb +61 -30
- data/test/printer_graph_html_test.rb +0 -1
- data/test/printer_graph_test.rb +3 -4
- data/test/printers_test.rb +2 -2
- data/test/printing_recursive_graph_test.rb +1 -1
- data/test/profile_test.rb +16 -0
- data/test/rack_test.rb +0 -64
- data/test/recursive_test.rb +50 -54
- data/test/start_stop_test.rb +19 -19
- data/test/test_helper.rb +6 -17
- data/test/thread_test.rb +11 -11
- data/test/unique_call_path_test.rb +25 -95
- metadata +22 -11
- data/ext/ruby_prof/rp_call_info.c +0 -271
- data/ext/ruby_prof/rp_call_info.h +0 -35
- data/lib/2.6.5/ruby_prof.so +0 -0
- data/lib/ruby-prof/call_info_visitor.rb +0 -38
- data/test/parser_timings.rb +0 -24
@@ -0,0 +1,59 @@
|
|
1
|
+
/* Copyright (C) 2005-2019 Shugo Maeda <shugo@ruby-lang.org> and Charlie Savage <cfis@savagexi.com>
|
2
|
+
Please see the LICENSE file for copyright and distribution information */
|
3
|
+
|
4
|
+
#include "rp_aggregate_call_tree.h"
|
5
|
+
|
6
|
+
VALUE cRpAggregateCallTree;
|
7
|
+
|
8
|
+
void prof_aggregate_call_tree_mark(void* data)
|
9
|
+
{
|
10
|
+
prof_call_tree_t* call_tree = (prof_call_tree_t*)data;
|
11
|
+
|
12
|
+
if (call_tree->object != Qnil)
|
13
|
+
rb_gc_mark(call_tree->object);
|
14
|
+
|
15
|
+
if (call_tree->source_file != Qnil)
|
16
|
+
rb_gc_mark(call_tree->source_file);
|
17
|
+
|
18
|
+
prof_measurement_mark(call_tree->measurement);
|
19
|
+
}
|
20
|
+
|
21
|
+
static void prof_aggregate_call_tree_ruby_gc_free(void* data)
|
22
|
+
{
|
23
|
+
prof_call_tree_t* call_tree = (prof_call_tree_t*)data;
|
24
|
+
prof_call_tree_free(call_tree);
|
25
|
+
}
|
26
|
+
|
27
|
+
size_t prof_aggregate_call_tree_size(const void* data)
|
28
|
+
{
|
29
|
+
return sizeof(prof_call_tree_t);
|
30
|
+
}
|
31
|
+
|
32
|
+
static const rb_data_type_t aggregate_call_tree_type =
|
33
|
+
{
|
34
|
+
.wrap_struct_name = "Aggregate_CallTree",
|
35
|
+
.function =
|
36
|
+
{
|
37
|
+
.dmark = prof_aggregate_call_tree_mark,
|
38
|
+
.dfree = prof_aggregate_call_tree_ruby_gc_free,
|
39
|
+
.dsize = prof_aggregate_call_tree_size,
|
40
|
+
},
|
41
|
+
.data = NULL,
|
42
|
+
.flags = RUBY_TYPED_FREE_IMMEDIATELY
|
43
|
+
};
|
44
|
+
|
45
|
+
VALUE prof_aggregate_call_tree_wrap(prof_call_tree_t* call_tree)
|
46
|
+
{
|
47
|
+
if (call_tree->object == Qnil)
|
48
|
+
{
|
49
|
+
call_tree->object = TypedData_Wrap_Struct(cRpAggregateCallTree, &aggregate_call_tree_type, call_tree);
|
50
|
+
}
|
51
|
+
return call_tree->object;
|
52
|
+
}
|
53
|
+
|
54
|
+
void rp_init_aggregate_call_tree()
|
55
|
+
{
|
56
|
+
// AggregateCallTree
|
57
|
+
cRpAggregateCallTree = rb_define_class_under(mProf, "AggregateCallTree", cRpCallTree);
|
58
|
+
rb_undef_method(CLASS_OF(cRpAggregateCallTree), "new");
|
59
|
+
}
|
@@ -0,0 +1,13 @@
|
|
1
|
+
/* Copyright (C) 2005-2019 Shugo Maeda <shugo@ruby-lang.org> and Charlie Savage <cfis@savagexi.com>
|
2
|
+
Please see the LICENSE file for copyright and distribution information */
|
3
|
+
|
4
|
+
#ifndef __RP_AGGREGATE_CALL_TREE_H__
|
5
|
+
#define __RP_AGGREGATE_CALL_TREE_H__
|
6
|
+
|
7
|
+
#include "ruby_prof.h"
|
8
|
+
#include "rp_call_tree.h"
|
9
|
+
|
10
|
+
void rp_init_aggregate_call_tree(void);
|
11
|
+
VALUE prof_aggregate_call_tree_wrap(prof_call_tree_t* call_tree);
|
12
|
+
|
13
|
+
#endif //__RP_AGGREGATE_CALL_TREE_H__
|
@@ -5,12 +5,11 @@
|
|
5
5
|
|
6
6
|
VALUE cRpAllocation;
|
7
7
|
|
8
|
-
prof_allocation_t*
|
9
|
-
allocations_table_lookup(st_table *table, st_data_t key)
|
8
|
+
prof_allocation_t* allocations_table_lookup(st_table* table, st_data_t key)
|
10
9
|
{
|
11
10
|
prof_allocation_t* result = NULL;
|
12
11
|
st_data_t value;
|
13
|
-
if (
|
12
|
+
if (rb_st_lookup(table, key, &value))
|
14
13
|
{
|
15
14
|
result = (prof_allocation_t*)value;
|
16
15
|
}
|
@@ -18,23 +17,20 @@ allocations_table_lookup(st_table *table, st_data_t key)
|
|
18
17
|
return result;
|
19
18
|
}
|
20
19
|
|
21
|
-
void
|
22
|
-
allocations_table_insert(st_table *table, st_data_t key, prof_allocation_t * allocation)
|
20
|
+
void allocations_table_insert(st_table* table, st_data_t key, prof_allocation_t* allocation)
|
23
21
|
{
|
24
|
-
|
22
|
+
rb_st_insert(table, (st_data_t)key, (st_data_t)allocation);
|
25
23
|
}
|
26
24
|
|
27
|
-
st_data_t
|
28
|
-
allocations_key(VALUE klass, int source_line)
|
25
|
+
st_data_t allocations_key(VALUE klass, int source_line)
|
29
26
|
{
|
30
27
|
return (klass << 4) + source_line;
|
31
28
|
}
|
32
29
|
|
33
30
|
/* ====== prof_allocation_t ====== */
|
34
|
-
prof_allocation_t*
|
35
|
-
prof_allocation_create(void)
|
31
|
+
prof_allocation_t* prof_allocation_create(void)
|
36
32
|
{
|
37
|
-
prof_allocation_t
|
33
|
+
prof_allocation_t* result = ALLOC(prof_allocation_t);
|
38
34
|
result->count = 0;
|
39
35
|
result->klass = Qnil;
|
40
36
|
result->klass_name = Qnil;
|
@@ -47,8 +43,19 @@ prof_allocation_create(void)
|
|
47
43
|
return result;
|
48
44
|
}
|
49
45
|
|
50
|
-
prof_allocation_t*
|
51
|
-
|
46
|
+
prof_allocation_t* prof_get_allocation(VALUE self)
|
47
|
+
{
|
48
|
+
/* Can't use Data_Get_Struct because that triggers the event hook
|
49
|
+
ending up in endless recursion. */
|
50
|
+
prof_allocation_t* result = RTYPEDDATA_DATA(self);
|
51
|
+
|
52
|
+
if (!result)
|
53
|
+
rb_raise(rb_eRuntimeError, "This RubyProf::Allocation instance has already been freed, likely because its profile has been freed.");
|
54
|
+
|
55
|
+
return result;
|
56
|
+
}
|
57
|
+
|
58
|
+
prof_allocation_t* prof_allocate_increment(prof_method_t* method, rb_trace_arg_t* trace_arg)
|
52
59
|
{
|
53
60
|
VALUE object = rb_tracearg_object(trace_arg);
|
54
61
|
if (BUILTIN_TYPE(object) == T_IMEMO)
|
@@ -78,76 +85,85 @@ prof_allocate_increment(prof_method_t* method, rb_trace_arg_t* trace_arg)
|
|
78
85
|
return allocation;
|
79
86
|
}
|
80
87
|
|
81
|
-
static void
|
82
|
-
prof_allocation_ruby_gc_free(void *data)
|
88
|
+
static void prof_allocation_ruby_gc_free(void* data)
|
83
89
|
{
|
84
|
-
|
90
|
+
if (data)
|
91
|
+
{
|
92
|
+
prof_allocation_t* allocation = (prof_allocation_t*)data;
|
93
|
+
allocation->object = Qnil;
|
94
|
+
}
|
95
|
+
}
|
85
96
|
|
97
|
+
void prof_allocation_free(prof_allocation_t* allocation)
|
98
|
+
{
|
86
99
|
/* Has this allocation object been accessed by Ruby? If
|
87
100
|
yes clean it up so to avoid a segmentation fault. */
|
88
101
|
if (allocation->object != Qnil)
|
89
102
|
{
|
90
|
-
|
91
|
-
RDATA(allocation->object)->dfree = NULL;
|
92
|
-
RDATA(allocation->object)->data = NULL;
|
103
|
+
RTYPEDDATA(allocation->object)->data = NULL;
|
93
104
|
allocation->object = Qnil;
|
94
105
|
}
|
95
|
-
}
|
96
106
|
|
97
|
-
void
|
98
|
-
prof_allocation_free(prof_allocation_t* allocation)
|
99
|
-
{
|
100
|
-
prof_allocation_ruby_gc_free(allocation);
|
101
107
|
xfree(allocation);
|
102
108
|
}
|
103
109
|
|
104
|
-
size_t
|
105
|
-
prof_allocation_size(const void* data)
|
110
|
+
size_t prof_allocation_size(const void* data)
|
106
111
|
{
|
107
112
|
return sizeof(prof_allocation_t);
|
108
113
|
}
|
109
114
|
|
110
|
-
void
|
111
|
-
prof_allocation_mark(void *data)
|
115
|
+
void prof_allocation_mark(void* data)
|
112
116
|
{
|
117
|
+
if (!data) return;
|
118
|
+
|
113
119
|
prof_allocation_t* allocation = (prof_allocation_t*)data;
|
120
|
+
if (allocation->object != Qnil)
|
121
|
+
rb_gc_mark(allocation->object);
|
122
|
+
|
114
123
|
if (allocation->klass != Qnil)
|
115
124
|
rb_gc_mark(allocation->klass);
|
116
|
-
|
125
|
+
|
117
126
|
if (allocation->klass_name != Qnil)
|
118
127
|
rb_gc_mark(allocation->klass_name);
|
119
128
|
|
120
|
-
if (allocation->object != Qnil)
|
121
|
-
rb_gc_mark(allocation->object);
|
122
|
-
|
123
129
|
if (allocation->source_file != Qnil)
|
124
130
|
rb_gc_mark(allocation->source_file);
|
125
131
|
}
|
126
132
|
|
127
|
-
|
128
|
-
|
133
|
+
static const rb_data_type_t allocation_type =
|
134
|
+
{
|
135
|
+
.wrap_struct_name = "Allocation",
|
136
|
+
.function =
|
137
|
+
{
|
138
|
+
.dmark = prof_allocation_mark,
|
139
|
+
.dfree = prof_allocation_ruby_gc_free,
|
140
|
+
.dsize = prof_allocation_size,
|
141
|
+
},
|
142
|
+
.data = NULL,
|
143
|
+
.flags = RUBY_TYPED_FREE_IMMEDIATELY
|
144
|
+
};
|
145
|
+
|
146
|
+
VALUE prof_allocation_wrap(prof_allocation_t* allocation)
|
129
147
|
{
|
130
148
|
if (allocation->object == Qnil)
|
131
149
|
{
|
132
|
-
allocation->object =
|
150
|
+
allocation->object = TypedData_Wrap_Struct(cRpAllocation, &allocation_type, allocation);
|
133
151
|
}
|
134
152
|
return allocation->object;
|
135
153
|
}
|
136
154
|
|
137
|
-
static VALUE
|
138
|
-
prof_allocation_allocate(VALUE klass)
|
155
|
+
static VALUE prof_allocation_allocate(VALUE klass)
|
139
156
|
{
|
140
157
|
prof_allocation_t* allocation = prof_allocation_create();
|
141
158
|
allocation->object = prof_allocation_wrap(allocation);
|
142
159
|
return allocation->object;
|
143
160
|
}
|
144
161
|
|
145
|
-
prof_allocation_t*
|
146
|
-
prof_allocation_get(VALUE self)
|
162
|
+
prof_allocation_t* prof_allocation_get(VALUE self)
|
147
163
|
{
|
148
164
|
/* Can't use Data_Get_Struct because that triggers the event hook
|
149
165
|
ending up in endless recursion. */
|
150
|
-
prof_allocation_t* result =
|
166
|
+
prof_allocation_t* result = RTYPEDDATA_DATA(self);
|
151
167
|
if (!result)
|
152
168
|
rb_raise(rb_eRuntimeError, "This RubyProf::Allocation instance has already been freed, likely because its profile has been freed.");
|
153
169
|
|
@@ -158,8 +174,7 @@ prof_allocation_get(VALUE self)
|
|
158
174
|
klass -> Class
|
159
175
|
|
160
176
|
Returns the type of Class being allocated. */
|
161
|
-
static VALUE
|
162
|
-
prof_allocation_klass_name(VALUE self)
|
177
|
+
static VALUE prof_allocation_klass_name(VALUE self)
|
163
178
|
{
|
164
179
|
prof_allocation_t* allocation = prof_allocation_get(self);
|
165
180
|
|
@@ -174,8 +189,7 @@ prof_allocation_klass_name(VALUE self)
|
|
174
189
|
|
175
190
|
Returns the klass flags */
|
176
191
|
|
177
|
-
static VALUE
|
178
|
-
prof_allocation_klass_flags(VALUE self)
|
192
|
+
static VALUE prof_allocation_klass_flags(VALUE self)
|
179
193
|
{
|
180
194
|
prof_allocation_t* allocation = prof_allocation_get(self);
|
181
195
|
return INT2FIX(allocation->klass_flags);
|
@@ -185,8 +199,7 @@ prof_allocation_klass_flags(VALUE self)
|
|
185
199
|
source_file -> string
|
186
200
|
|
187
201
|
Returns the the line number where objects were allocated. */
|
188
|
-
static VALUE
|
189
|
-
prof_allocation_source_file(VALUE self)
|
202
|
+
static VALUE prof_allocation_source_file(VALUE self)
|
190
203
|
{
|
191
204
|
prof_allocation_t* allocation = prof_allocation_get(self);
|
192
205
|
return allocation->source_file;
|
@@ -196,8 +209,7 @@ prof_allocation_source_file(VALUE self)
|
|
196
209
|
line -> number
|
197
210
|
|
198
211
|
Returns the the line number where objects were allocated. */
|
199
|
-
static VALUE
|
200
|
-
prof_allocation_source_line(VALUE self)
|
212
|
+
static VALUE prof_allocation_source_line(VALUE self)
|
201
213
|
{
|
202
214
|
prof_allocation_t* allocation = prof_allocation_get(self);
|
203
215
|
return INT2FIX(allocation->source_line);
|
@@ -207,8 +219,7 @@ prof_allocation_source_line(VALUE self)
|
|
207
219
|
count -> number
|
208
220
|
|
209
221
|
Returns the number of times this class has been allocated. */
|
210
|
-
static VALUE
|
211
|
-
prof_allocation_count(VALUE self)
|
222
|
+
static VALUE prof_allocation_count(VALUE self)
|
212
223
|
{
|
213
224
|
prof_allocation_t* allocation = prof_allocation_get(self);
|
214
225
|
return INT2FIX(allocation->count);
|
@@ -218,18 +229,16 @@ prof_allocation_count(VALUE self)
|
|
218
229
|
memory -> number
|
219
230
|
|
220
231
|
Returns the amount of memory allocated. */
|
221
|
-
static VALUE
|
222
|
-
prof_allocation_memory(VALUE self)
|
232
|
+
static VALUE prof_allocation_memory(VALUE self)
|
223
233
|
{
|
224
234
|
prof_allocation_t* allocation = prof_allocation_get(self);
|
225
235
|
return ULL2NUM(allocation->memory);
|
226
236
|
}
|
227
237
|
|
228
238
|
/* :nodoc: */
|
229
|
-
static VALUE
|
230
|
-
prof_allocation_dump(VALUE self)
|
239
|
+
static VALUE prof_allocation_dump(VALUE self)
|
231
240
|
{
|
232
|
-
prof_allocation_t* allocation =
|
241
|
+
prof_allocation_t* allocation = prof_get_allocation(self);
|
233
242
|
|
234
243
|
VALUE result = rb_hash_new();
|
235
244
|
|
@@ -245,10 +254,9 @@ prof_allocation_dump(VALUE self)
|
|
245
254
|
}
|
246
255
|
|
247
256
|
/* :nodoc: */
|
248
|
-
static VALUE
|
249
|
-
prof_allocation_load(VALUE self, VALUE data)
|
257
|
+
static VALUE prof_allocation_load(VALUE self, VALUE data)
|
250
258
|
{
|
251
|
-
prof_allocation_t* allocation =
|
259
|
+
prof_allocation_t* allocation = prof_get_allocation(self);
|
252
260
|
allocation->object = self;
|
253
261
|
|
254
262
|
allocation->key = FIX2LONG(rb_hash_aref(data, ID2SYM(rb_intern("key"))));
|
@@ -264,7 +272,7 @@ prof_allocation_load(VALUE self, VALUE data)
|
|
264
272
|
|
265
273
|
void rp_init_allocation(void)
|
266
274
|
{
|
267
|
-
cRpAllocation = rb_define_class_under(mProf, "Allocation",
|
275
|
+
cRpAllocation = rb_define_class_under(mProf, "Allocation", rb_cObject);
|
268
276
|
rb_undef_method(CLASS_OF(cRpAllocation), "new");
|
269
277
|
rb_define_alloc_func(cRpAllocation, prof_allocation_allocate);
|
270
278
|
|
@@ -7,7 +7,7 @@
|
|
7
7
|
#include "ruby_prof.h"
|
8
8
|
#include "rp_method.h"
|
9
9
|
|
10
|
-
typedef struct
|
10
|
+
typedef struct prof_allocation_t
|
11
11
|
{
|
12
12
|
st_data_t key; /* Key in hash table */
|
13
13
|
unsigned int klass_flags; /* Information about the type of class */
|
@@ -22,10 +22,10 @@ typedef struct
|
|
22
22
|
|
23
23
|
void rp_init_allocation(void);
|
24
24
|
void prof_allocation_free(prof_allocation_t* allocation);
|
25
|
-
void prof_allocation_mark(void
|
25
|
+
void prof_allocation_mark(void* data);
|
26
26
|
VALUE prof_allocation_wrap(prof_allocation_t* allocation);
|
27
27
|
prof_allocation_t* prof_allocation_get(VALUE self);
|
28
|
-
prof_allocation_t* prof_allocate_increment(prof_method_t
|
28
|
+
prof_allocation_t* prof_allocate_increment(prof_method_t* method, rb_trace_arg_t* trace_arg);
|
29
29
|
|
30
30
|
|
31
31
|
#endif //_RP_ALLOCATION_
|
@@ -0,0 +1,369 @@
|
|
1
|
+
/* Copyright (C) 2005-2019 Shugo Maeda <shugo@ruby-lang.org> and Charlie Savage <cfis@savagexi.com>
|
2
|
+
Please see the LICENSE file for copyright and distribution information */
|
3
|
+
|
4
|
+
#include "rp_call_tree.h"
|
5
|
+
|
6
|
+
#define INITIAL_CALL_TREES_SIZE 2
|
7
|
+
|
8
|
+
VALUE cRpCallTree;
|
9
|
+
|
10
|
+
/* ======= prof_call_tree_t ========*/
|
11
|
+
prof_call_tree_t* prof_call_tree_create(prof_method_t* method, prof_call_tree_t* parent, VALUE source_file, int source_line)
|
12
|
+
{
|
13
|
+
prof_call_tree_t* result = ALLOC(prof_call_tree_t);
|
14
|
+
result->method = method;
|
15
|
+
result->parent = parent;
|
16
|
+
result->object = Qnil;
|
17
|
+
result->visits = 0;
|
18
|
+
result->source_line = source_line;
|
19
|
+
result->source_file = source_file;
|
20
|
+
result->children = rb_st_init_numtable();
|
21
|
+
result->measurement = prof_measurement_create();
|
22
|
+
|
23
|
+
return result;
|
24
|
+
}
|
25
|
+
|
26
|
+
prof_call_tree_t* prof_call_tree_copy(prof_call_tree_t* other)
|
27
|
+
{
|
28
|
+
prof_call_tree_t* result = ALLOC(prof_call_tree_t);
|
29
|
+
result->children = rb_st_init_numtable();
|
30
|
+
result->object = Qnil;
|
31
|
+
result->visits = 0;
|
32
|
+
|
33
|
+
result->method = other->method;
|
34
|
+
result->parent = other->parent;
|
35
|
+
result->source_line = other->source_line;
|
36
|
+
result->source_file = other->source_file;
|
37
|
+
|
38
|
+
result->measurement = prof_measurement_create();
|
39
|
+
result->measurement->called = other->measurement->called;
|
40
|
+
result->measurement->total_time = other->measurement->total_time;
|
41
|
+
result->measurement->self_time = other->measurement->self_time;
|
42
|
+
result->measurement->wait_time = other->measurement->wait_time;
|
43
|
+
result->measurement->object = Qnil;
|
44
|
+
|
45
|
+
return result;
|
46
|
+
}
|
47
|
+
|
48
|
+
void prof_call_tree_merge(prof_call_tree_t* result, prof_call_tree_t* other)
|
49
|
+
{
|
50
|
+
result->measurement->called += other->measurement->called;
|
51
|
+
result->measurement->total_time += other->measurement->total_time;
|
52
|
+
result->measurement->self_time += other->measurement->self_time;
|
53
|
+
result->measurement->wait_time += other->measurement->wait_time;
|
54
|
+
}
|
55
|
+
|
56
|
+
static int prof_call_tree_collect_children(st_data_t key, st_data_t value, st_data_t result)
|
57
|
+
{
|
58
|
+
prof_call_tree_t* call_tree = (prof_call_tree_t*)value;
|
59
|
+
VALUE arr = (VALUE)result;
|
60
|
+
rb_ary_push(arr, prof_call_tree_wrap(call_tree));
|
61
|
+
return ST_CONTINUE;
|
62
|
+
}
|
63
|
+
|
64
|
+
static int prof_call_tree_mark_children(st_data_t key, st_data_t value, st_data_t data)
|
65
|
+
{
|
66
|
+
prof_call_tree_t* call_tree = (prof_call_tree_t*)value;
|
67
|
+
rb_st_foreach(call_tree->children, prof_call_tree_mark_children, data);
|
68
|
+
prof_call_tree_mark(call_tree);
|
69
|
+
return ST_CONTINUE;
|
70
|
+
}
|
71
|
+
|
72
|
+
void prof_call_tree_mark(void* data)
|
73
|
+
{
|
74
|
+
if (!data)
|
75
|
+
return;
|
76
|
+
|
77
|
+
prof_call_tree_t* call_tree = (prof_call_tree_t*)data;
|
78
|
+
|
79
|
+
if (call_tree->object != Qnil)
|
80
|
+
rb_gc_mark(call_tree->object);
|
81
|
+
|
82
|
+
if (call_tree->source_file != Qnil)
|
83
|
+
rb_gc_mark(call_tree->source_file);
|
84
|
+
|
85
|
+
prof_method_mark(call_tree->method);
|
86
|
+
prof_measurement_mark(call_tree->measurement);
|
87
|
+
|
88
|
+
// Recurse down through the whole call tree but only from the top node
|
89
|
+
// to avoid calling mark over and over and over.
|
90
|
+
if (!call_tree->parent)
|
91
|
+
rb_st_foreach(call_tree->children, prof_call_tree_mark_children, 0);
|
92
|
+
}
|
93
|
+
|
94
|
+
static void prof_call_tree_ruby_gc_free(void* data)
|
95
|
+
{
|
96
|
+
if (data)
|
97
|
+
{
|
98
|
+
prof_call_tree_t* call_tree = (prof_call_tree_t*)data;
|
99
|
+
call_tree->object = Qnil;
|
100
|
+
}
|
101
|
+
}
|
102
|
+
|
103
|
+
static int prof_call_tree_free_children(st_data_t key, st_data_t value, st_data_t data)
|
104
|
+
{
|
105
|
+
prof_call_tree_t* call_tree = (prof_call_tree_t*)value;
|
106
|
+
prof_call_tree_free(call_tree);
|
107
|
+
return ST_CONTINUE;
|
108
|
+
}
|
109
|
+
|
110
|
+
void prof_call_tree_free(prof_call_tree_t* call_tree_data)
|
111
|
+
{
|
112
|
+
/* Has this call info object been accessed by Ruby? If
|
113
|
+
yes clean it up so to avoid a segmentation fault. */
|
114
|
+
if (call_tree_data->object != Qnil)
|
115
|
+
{
|
116
|
+
RTYPEDDATA(call_tree_data->object)->data = NULL;
|
117
|
+
call_tree_data->object = Qnil;
|
118
|
+
}
|
119
|
+
|
120
|
+
// Free children
|
121
|
+
rb_st_foreach(call_tree_data->children, prof_call_tree_free_children, 0);
|
122
|
+
rb_st_free_table(call_tree_data->children);
|
123
|
+
|
124
|
+
// Free measurement
|
125
|
+
prof_measurement_free(call_tree_data->measurement);
|
126
|
+
|
127
|
+
// Finally free self
|
128
|
+
xfree(call_tree_data);
|
129
|
+
}
|
130
|
+
|
131
|
+
size_t prof_call_tree_size(const void* data)
|
132
|
+
{
|
133
|
+
return sizeof(prof_call_tree_t);
|
134
|
+
}
|
135
|
+
|
136
|
+
static const rb_data_type_t call_tree_type =
|
137
|
+
{
|
138
|
+
.wrap_struct_name = "CallTree",
|
139
|
+
.function =
|
140
|
+
{
|
141
|
+
.dmark = prof_call_tree_mark,
|
142
|
+
.dfree = prof_call_tree_ruby_gc_free,
|
143
|
+
.dsize = prof_call_tree_size,
|
144
|
+
},
|
145
|
+
.data = NULL,
|
146
|
+
.flags = RUBY_TYPED_FREE_IMMEDIATELY
|
147
|
+
};
|
148
|
+
|
149
|
+
VALUE prof_call_tree_wrap(prof_call_tree_t* call_tree)
|
150
|
+
{
|
151
|
+
if (call_tree->object == Qnil)
|
152
|
+
{
|
153
|
+
call_tree->object = TypedData_Wrap_Struct(cRpCallTree, &call_tree_type, call_tree);
|
154
|
+
}
|
155
|
+
return call_tree->object;
|
156
|
+
}
|
157
|
+
|
158
|
+
static VALUE prof_call_tree_allocate(VALUE klass)
|
159
|
+
{
|
160
|
+
prof_call_tree_t* call_tree = prof_call_tree_create(NULL, NULL, Qnil, 0);
|
161
|
+
call_tree->object = prof_call_tree_wrap(call_tree);
|
162
|
+
return call_tree->object;
|
163
|
+
}
|
164
|
+
|
165
|
+
prof_call_tree_t* prof_get_call_tree(VALUE self)
|
166
|
+
{
|
167
|
+
/* Can't use Data_Get_Struct because that triggers the event hook
|
168
|
+
ending up in endless recursion. */
|
169
|
+
prof_call_tree_t* result = RTYPEDDATA_DATA(self);
|
170
|
+
|
171
|
+
if (!result)
|
172
|
+
rb_raise(rb_eRuntimeError, "This RubyProf::CallTree instance has already been freed, likely because its profile has been freed.");
|
173
|
+
|
174
|
+
return result;
|
175
|
+
}
|
176
|
+
|
177
|
+
/* ======= Call Tree Table ========*/
|
178
|
+
static size_t call_tree_table_insert(st_table* table, st_data_t key, prof_call_tree_t* val)
|
179
|
+
{
|
180
|
+
return rb_st_insert(table, (st_data_t)key, (st_data_t)val);
|
181
|
+
}
|
182
|
+
|
183
|
+
prof_call_tree_t* call_tree_table_lookup(st_table* table, st_data_t key)
|
184
|
+
{
|
185
|
+
st_data_t val;
|
186
|
+
if (rb_st_lookup(table, (st_data_t)key, &val))
|
187
|
+
{
|
188
|
+
return (prof_call_tree_t*)val;
|
189
|
+
}
|
190
|
+
else
|
191
|
+
{
|
192
|
+
return NULL;
|
193
|
+
}
|
194
|
+
}
|
195
|
+
|
196
|
+
uint32_t prof_call_figure_depth(prof_call_tree_t* call_tree_data)
|
197
|
+
{
|
198
|
+
uint32_t result = 0;
|
199
|
+
|
200
|
+
while (call_tree_data->parent)
|
201
|
+
{
|
202
|
+
result++;
|
203
|
+
call_tree_data = call_tree_data->parent;
|
204
|
+
}
|
205
|
+
|
206
|
+
return result;
|
207
|
+
}
|
208
|
+
|
209
|
+
void prof_call_tree_add_parent(prof_call_tree_t* self, prof_call_tree_t* parent)
|
210
|
+
{
|
211
|
+
prof_call_tree_add_child(parent, self);
|
212
|
+
self->parent = parent;
|
213
|
+
}
|
214
|
+
|
215
|
+
void prof_call_tree_add_child(prof_call_tree_t* self, prof_call_tree_t* child)
|
216
|
+
{
|
217
|
+
call_tree_table_insert(self->children, child->method->key, child);
|
218
|
+
}
|
219
|
+
|
220
|
+
/* ======= RubyProf::CallTree ========*/
|
221
|
+
|
222
|
+
/* call-seq:
|
223
|
+
parent -> call_tree
|
224
|
+
|
225
|
+
Returns the CallTree parent call_tree object (the method that called this method).*/
|
226
|
+
static VALUE prof_call_tree_parent(VALUE self)
|
227
|
+
{
|
228
|
+
prof_call_tree_t* call_tree = prof_get_call_tree(self);
|
229
|
+
if (call_tree->parent)
|
230
|
+
return prof_call_tree_wrap(call_tree->parent);
|
231
|
+
else
|
232
|
+
return Qnil;
|
233
|
+
}
|
234
|
+
|
235
|
+
/* call-seq:
|
236
|
+
callees -> array
|
237
|
+
|
238
|
+
Returns an array of call info objects that this method called (ie, children).*/
|
239
|
+
static VALUE prof_call_tree_children(VALUE self)
|
240
|
+
{
|
241
|
+
prof_call_tree_t* call_tree = prof_get_call_tree(self);
|
242
|
+
VALUE result = rb_ary_new();
|
243
|
+
rb_st_foreach(call_tree->children, prof_call_tree_collect_children, result);
|
244
|
+
return result;
|
245
|
+
}
|
246
|
+
|
247
|
+
/* call-seq:
|
248
|
+
called -> MethodInfo
|
249
|
+
|
250
|
+
Returns the target method. */
|
251
|
+
static VALUE prof_call_tree_target(VALUE self)
|
252
|
+
{
|
253
|
+
prof_call_tree_t* call_tree = prof_get_call_tree(self);
|
254
|
+
return prof_method_wrap(call_tree->method);
|
255
|
+
}
|
256
|
+
|
257
|
+
/* call-seq:
|
258
|
+
called -> Measurement
|
259
|
+
|
260
|
+
Returns the measurement associated with this call_tree. */
|
261
|
+
static VALUE prof_call_tree_measurement(VALUE self)
|
262
|
+
{
|
263
|
+
prof_call_tree_t* call_tree = prof_get_call_tree(self);
|
264
|
+
return prof_measurement_wrap(call_tree->measurement);
|
265
|
+
}
|
266
|
+
|
267
|
+
/* call-seq:
|
268
|
+
depth -> int
|
269
|
+
|
270
|
+
returns the depth of this call info in the call graph */
|
271
|
+
static VALUE prof_call_tree_depth(VALUE self)
|
272
|
+
{
|
273
|
+
prof_call_tree_t* call_tree_data = prof_get_call_tree(self);
|
274
|
+
uint32_t depth = prof_call_figure_depth(call_tree_data);
|
275
|
+
return rb_int_new(depth);
|
276
|
+
}
|
277
|
+
|
278
|
+
/* call-seq:
|
279
|
+
source_file => string
|
280
|
+
|
281
|
+
return the source file of the method
|
282
|
+
*/
|
283
|
+
static VALUE prof_call_tree_source_file(VALUE self)
|
284
|
+
{
|
285
|
+
prof_call_tree_t* result = prof_get_call_tree(self);
|
286
|
+
return result->source_file;
|
287
|
+
}
|
288
|
+
|
289
|
+
/* call-seq:
|
290
|
+
line_no -> int
|
291
|
+
|
292
|
+
returns the line number of the method */
|
293
|
+
static VALUE prof_call_tree_line(VALUE self)
|
294
|
+
{
|
295
|
+
prof_call_tree_t* result = prof_get_call_tree(self);
|
296
|
+
return INT2FIX(result->source_line);
|
297
|
+
}
|
298
|
+
|
299
|
+
/* :nodoc: */
|
300
|
+
static VALUE prof_call_tree_dump(VALUE self)
|
301
|
+
{
|
302
|
+
prof_call_tree_t* call_tree_data = prof_get_call_tree(self);
|
303
|
+
VALUE result = rb_hash_new();
|
304
|
+
|
305
|
+
rb_hash_aset(result, ID2SYM(rb_intern("measurement")), prof_measurement_wrap(call_tree_data->measurement));
|
306
|
+
|
307
|
+
rb_hash_aset(result, ID2SYM(rb_intern("source_file")), call_tree_data->source_file);
|
308
|
+
rb_hash_aset(result, ID2SYM(rb_intern("source_line")), INT2FIX(call_tree_data->source_line));
|
309
|
+
|
310
|
+
rb_hash_aset(result, ID2SYM(rb_intern("parent")), prof_call_tree_parent(self));
|
311
|
+
rb_hash_aset(result, ID2SYM(rb_intern("children")), prof_call_tree_children(self));
|
312
|
+
rb_hash_aset(result, ID2SYM(rb_intern("target")), prof_call_tree_target(self));
|
313
|
+
|
314
|
+
return result;
|
315
|
+
}
|
316
|
+
|
317
|
+
/* :nodoc: */
|
318
|
+
static VALUE prof_call_tree_load(VALUE self, VALUE data)
|
319
|
+
{
|
320
|
+
VALUE target = Qnil;
|
321
|
+
VALUE parent = Qnil;
|
322
|
+
prof_call_tree_t* call_tree = prof_get_call_tree(self);
|
323
|
+
call_tree->object = self;
|
324
|
+
|
325
|
+
VALUE measurement = rb_hash_aref(data, ID2SYM(rb_intern("measurement")));
|
326
|
+
call_tree->measurement = prof_get_measurement(measurement);
|
327
|
+
|
328
|
+
call_tree->source_file = rb_hash_aref(data, ID2SYM(rb_intern("source_file")));
|
329
|
+
call_tree->source_line = FIX2INT(rb_hash_aref(data, ID2SYM(rb_intern("source_line"))));
|
330
|
+
|
331
|
+
parent = rb_hash_aref(data, ID2SYM(rb_intern("parent")));
|
332
|
+
if (parent != Qnil)
|
333
|
+
call_tree->parent = prof_get_call_tree(parent);
|
334
|
+
|
335
|
+
VALUE callees = rb_hash_aref(data, ID2SYM(rb_intern("children")));
|
336
|
+
for (int i = 0; i < rb_array_len(callees); i++)
|
337
|
+
{
|
338
|
+
VALUE call_tree_object = rb_ary_entry(callees, i);
|
339
|
+
prof_call_tree_t* call_tree_data = prof_get_call_tree(call_tree_object);
|
340
|
+
|
341
|
+
st_data_t key = call_tree_data->method ? call_tree_data->method->key : method_key(Qnil, 0);
|
342
|
+
call_tree_table_insert(call_tree->children, key, call_tree_data);
|
343
|
+
}
|
344
|
+
|
345
|
+
target = rb_hash_aref(data, ID2SYM(rb_intern("target")));
|
346
|
+
call_tree->method = prof_get_method(target);
|
347
|
+
|
348
|
+
return data;
|
349
|
+
}
|
350
|
+
|
351
|
+
void rp_init_call_tree()
|
352
|
+
{
|
353
|
+
/* CallTree */
|
354
|
+
cRpCallTree = rb_define_class_under(mProf, "CallTree", rb_cObject);
|
355
|
+
rb_undef_method(CLASS_OF(cRpCallTree), "new");
|
356
|
+
rb_define_alloc_func(cRpCallTree, prof_call_tree_allocate);
|
357
|
+
|
358
|
+
rb_define_method(cRpCallTree, "parent", prof_call_tree_parent, 0);
|
359
|
+
rb_define_method(cRpCallTree, "children", prof_call_tree_children, 0);
|
360
|
+
rb_define_method(cRpCallTree, "target", prof_call_tree_target, 0);
|
361
|
+
rb_define_method(cRpCallTree, "measurement", prof_call_tree_measurement, 0);
|
362
|
+
|
363
|
+
rb_define_method(cRpCallTree, "depth", prof_call_tree_depth, 0);
|
364
|
+
rb_define_method(cRpCallTree, "source_file", prof_call_tree_source_file, 0);
|
365
|
+
rb_define_method(cRpCallTree, "line", prof_call_tree_line, 0);
|
366
|
+
|
367
|
+
rb_define_method(cRpCallTree, "_dump_data", prof_call_tree_dump, 0);
|
368
|
+
rb_define_method(cRpCallTree, "_load_data", prof_call_tree_load, 1);
|
369
|
+
}
|