ruby-prof 1.4.4-x64-mingw-ucrt
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +7 -0
- data/CHANGES +608 -0
- data/LICENSE +25 -0
- data/README.md +5 -0
- data/Rakefile +98 -0
- data/bin/ruby-prof +328 -0
- data/bin/ruby-prof-check-trace +45 -0
- data/ext/ruby_prof/extconf.rb +22 -0
- data/ext/ruby_prof/rp_aggregate_call_tree.c +59 -0
- data/ext/ruby_prof/rp_aggregate_call_tree.h +13 -0
- data/ext/ruby_prof/rp_allocation.c +287 -0
- data/ext/ruby_prof/rp_allocation.h +31 -0
- data/ext/ruby_prof/rp_call_tree.c +367 -0
- data/ext/ruby_prof/rp_call_tree.h +43 -0
- data/ext/ruby_prof/rp_call_trees.c +288 -0
- data/ext/ruby_prof/rp_call_trees.h +28 -0
- data/ext/ruby_prof/rp_measure_allocations.c +47 -0
- data/ext/ruby_prof/rp_measure_memory.c +46 -0
- data/ext/ruby_prof/rp_measure_process_time.c +66 -0
- data/ext/ruby_prof/rp_measure_wall_time.c +64 -0
- data/ext/ruby_prof/rp_measurement.c +237 -0
- data/ext/ruby_prof/rp_measurement.h +50 -0
- data/ext/ruby_prof/rp_method.c +491 -0
- data/ext/ruby_prof/rp_method.h +62 -0
- data/ext/ruby_prof/rp_profile.c +915 -0
- data/ext/ruby_prof/rp_profile.h +35 -0
- data/ext/ruby_prof/rp_stack.c +212 -0
- data/ext/ruby_prof/rp_stack.h +53 -0
- data/ext/ruby_prof/rp_thread.c +362 -0
- data/ext/ruby_prof/rp_thread.h +39 -0
- data/ext/ruby_prof/ruby_prof.c +52 -0
- data/ext/ruby_prof/ruby_prof.h +26 -0
- data/ext/ruby_prof/vc/ruby_prof.sln +39 -0
- data/ext/ruby_prof/vc/ruby_prof.vcxproj +160 -0
- data/lib/3.1/ruby_prof.so +0 -0
- data/lib/ruby-prof/assets/call_stack_printer.html.erb +711 -0
- data/lib/ruby-prof/assets/call_stack_printer.png +0 -0
- data/lib/ruby-prof/assets/graph_printer.html.erb +355 -0
- data/lib/ruby-prof/call_tree.rb +57 -0
- data/lib/ruby-prof/call_tree_visitor.rb +36 -0
- data/lib/ruby-prof/compatibility.rb +99 -0
- data/lib/ruby-prof/exclude_common_methods.rb +198 -0
- data/lib/ruby-prof/measurement.rb +17 -0
- data/lib/ruby-prof/method_info.rb +78 -0
- data/lib/ruby-prof/printers/abstract_printer.rb +137 -0
- data/lib/ruby-prof/printers/call_info_printer.rb +53 -0
- data/lib/ruby-prof/printers/call_stack_printer.rb +180 -0
- data/lib/ruby-prof/printers/call_tree_printer.rb +147 -0
- data/lib/ruby-prof/printers/dot_printer.rb +132 -0
- data/lib/ruby-prof/printers/flat_printer.rb +53 -0
- data/lib/ruby-prof/printers/graph_html_printer.rb +63 -0
- data/lib/ruby-prof/printers/graph_printer.rb +113 -0
- data/lib/ruby-prof/printers/multi_printer.rb +127 -0
- data/lib/ruby-prof/profile.rb +37 -0
- data/lib/ruby-prof/rack.rb +95 -0
- data/lib/ruby-prof/task.rb +147 -0
- data/lib/ruby-prof/thread.rb +20 -0
- data/lib/ruby-prof/version.rb +3 -0
- data/lib/ruby-prof.rb +52 -0
- data/lib/unprof.rb +10 -0
- data/ruby-prof.gemspec +64 -0
- data/test/abstract_printer_test.rb +26 -0
- data/test/alias_test.rb +122 -0
- data/test/basic_test.rb +43 -0
- data/test/call_tree_visitor_test.rb +32 -0
- data/test/call_trees_test.rb +66 -0
- data/test/duplicate_names_test.rb +32 -0
- data/test/dynamic_method_test.rb +67 -0
- data/test/enumerable_test.rb +21 -0
- data/test/exceptions_test.rb +24 -0
- data/test/exclude_methods_test.rb +151 -0
- data/test/exclude_threads_test.rb +53 -0
- data/test/fiber_test.rb +129 -0
- data/test/gc_test.rb +100 -0
- data/test/inverse_call_tree_test.rb +175 -0
- data/test/line_number_test.rb +158 -0
- data/test/marshal_test.rb +145 -0
- data/test/measure_allocations.rb +26 -0
- data/test/measure_allocations_test.rb +333 -0
- data/test/measure_memory_test.rb +688 -0
- data/test/measure_process_time_test.rb +1614 -0
- data/test/measure_times.rb +56 -0
- data/test/measure_wall_time_test.rb +426 -0
- data/test/multi_printer_test.rb +71 -0
- data/test/no_method_class_test.rb +15 -0
- data/test/pause_resume_test.rb +175 -0
- data/test/prime.rb +54 -0
- data/test/prime_script.rb +6 -0
- data/test/printer_call_stack_test.rb +27 -0
- data/test/printer_call_tree_test.rb +30 -0
- data/test/printer_flat_test.rb +99 -0
- data/test/printer_graph_html_test.rb +59 -0
- data/test/printer_graph_test.rb +40 -0
- data/test/printers_test.rb +141 -0
- data/test/printing_recursive_graph_test.rb +81 -0
- data/test/profile_test.rb +16 -0
- data/test/rack_test.rb +93 -0
- data/test/recursive_test.rb +430 -0
- data/test/singleton_test.rb +38 -0
- data/test/stack_printer_test.rb +64 -0
- data/test/start_stop_test.rb +109 -0
- data/test/test_helper.rb +13 -0
- data/test/thread_test.rb +144 -0
- data/test/unique_call_path_test.rb +136 -0
- data/test/yarv_test.rb +60 -0
- metadata +187 -0
@@ -0,0 +1,43 @@
|
|
1
|
+
/* Copyright (C) 2005-2019 Shugo Maeda <shugo@ruby-lang.org> and Charlie Savage <cfis@savagexi.com>
|
2
|
+
Please see the LICENSE file for copyright and distribution information */
|
3
|
+
|
4
|
+
#ifndef __RP_CALL_TREE_H__
|
5
|
+
#define __RP_CALL_TREE_H__
|
6
|
+
|
7
|
+
#include "ruby_prof.h"
|
8
|
+
#include "rp_measurement.h"
|
9
|
+
#include "rp_method.h"
|
10
|
+
|
11
|
+
extern VALUE cRpCallTree;
|
12
|
+
|
13
|
+
/* Callers and callee information for a method. */
|
14
|
+
typedef struct prof_call_tree_t
|
15
|
+
{
|
16
|
+
prof_method_t* method;
|
17
|
+
struct prof_call_tree_t* parent;
|
18
|
+
st_table* children; /* Call infos that this call info calls */
|
19
|
+
prof_measurement_t* measurement;
|
20
|
+
VALUE object;
|
21
|
+
|
22
|
+
int visits; /* Current visits on the stack */
|
23
|
+
|
24
|
+
unsigned int source_line;
|
25
|
+
VALUE source_file;
|
26
|
+
} prof_call_tree_t;
|
27
|
+
|
28
|
+
prof_call_tree_t* prof_call_tree_create(prof_method_t* method, prof_call_tree_t* parent, VALUE source_file, int source_line);
|
29
|
+
prof_call_tree_t* prof_call_tree_copy(prof_call_tree_t* other);
|
30
|
+
void prof_call_tree_merge(prof_call_tree_t* result, prof_call_tree_t* other);
|
31
|
+
void prof_call_tree_mark(void* data);
|
32
|
+
prof_call_tree_t* call_tree_table_lookup(st_table* table, st_data_t key);
|
33
|
+
|
34
|
+
void prof_call_tree_add_parent(prof_call_tree_t* self, prof_call_tree_t* parent);
|
35
|
+
void prof_call_tree_add_child(prof_call_tree_t* self, prof_call_tree_t* child);
|
36
|
+
|
37
|
+
uint32_t prof_call_figure_depth(prof_call_tree_t* call_tree_data);
|
38
|
+
prof_call_tree_t* prof_get_call_tree(VALUE self);
|
39
|
+
VALUE prof_call_tree_wrap(prof_call_tree_t* call_tree);
|
40
|
+
void prof_call_tree_free(prof_call_tree_t* call_tree);
|
41
|
+
void rp_init_call_tree();
|
42
|
+
|
43
|
+
#endif //__RP_CALL_TREE_H__
|
@@ -0,0 +1,288 @@
|
|
1
|
+
/* Copyright (C) 2005-2013 Shugo Maeda <shugo@ruby-lang.org> and Charlie Savage <cfis@savagexi.com>
|
2
|
+
Please see the LICENSE file for copyright and distribution information */
|
3
|
+
|
4
|
+
#include "rp_aggregate_call_tree.h"
|
5
|
+
#include "rp_call_trees.h"
|
6
|
+
#include "rp_measurement.h"
|
7
|
+
|
8
|
+
#define INITIAL_CALL_TREES_SIZE 2
|
9
|
+
|
10
|
+
VALUE cRpCallTrees;
|
11
|
+
|
12
|
+
/* ======= Call Infos ========*/
|
13
|
+
prof_call_trees_t* prof_get_call_trees(VALUE self)
|
14
|
+
{
|
15
|
+
/* Can't use Data_Get_Struct because that triggers the event hook
|
16
|
+
ending up in endless recursion. */
|
17
|
+
prof_call_trees_t* result = RTYPEDDATA_DATA(self);
|
18
|
+
|
19
|
+
if (!result)
|
20
|
+
rb_raise(rb_eRuntimeError, "This RubyProf::CallTrees instance has already been freed, likely because its profile has been freed.");
|
21
|
+
|
22
|
+
return result;
|
23
|
+
}
|
24
|
+
|
25
|
+
prof_call_trees_t* prof_call_trees_create()
|
26
|
+
{
|
27
|
+
prof_call_trees_t* result = ALLOC(prof_call_trees_t);
|
28
|
+
result->start = ALLOC_N(prof_call_tree_t*, INITIAL_CALL_TREES_SIZE);
|
29
|
+
result->end = result->start + INITIAL_CALL_TREES_SIZE;
|
30
|
+
result->ptr = result->start;
|
31
|
+
result->object = Qnil;
|
32
|
+
return result;
|
33
|
+
}
|
34
|
+
|
35
|
+
void prof_call_trees_mark(void* data)
|
36
|
+
{
|
37
|
+
if (!data) return;
|
38
|
+
|
39
|
+
prof_call_trees_t* call_trees = (prof_call_trees_t*)data;
|
40
|
+
prof_call_tree_t** call_tree;
|
41
|
+
for (call_tree = call_trees->start; call_tree < call_trees->ptr; call_tree++)
|
42
|
+
{
|
43
|
+
prof_call_tree_mark(*call_tree);
|
44
|
+
}
|
45
|
+
}
|
46
|
+
|
47
|
+
void prof_call_trees_free(prof_call_trees_t* call_trees)
|
48
|
+
{
|
49
|
+
/* Has this method object been accessed by Ruby? If
|
50
|
+
yes clean it up so to avoid a segmentation fault. */
|
51
|
+
if (call_trees->object != Qnil)
|
52
|
+
{
|
53
|
+
RTYPEDDATA(call_trees->object)->data = NULL;
|
54
|
+
call_trees->object = Qnil;
|
55
|
+
}
|
56
|
+
|
57
|
+
// Note we do not free our call_tree structures - since they have no parents they will free themselves
|
58
|
+
xfree(call_trees);
|
59
|
+
}
|
60
|
+
|
61
|
+
void prof_call_trees_ruby_gc_free(void* data)
|
62
|
+
{
|
63
|
+
if (data)
|
64
|
+
{
|
65
|
+
// This object gets freed by its owning method
|
66
|
+
prof_call_trees_t* call_trees = (prof_call_trees_t*)data;
|
67
|
+
call_trees->object = Qnil;
|
68
|
+
}
|
69
|
+
}
|
70
|
+
|
71
|
+
static int prof_call_trees_collect_aggregates(st_data_t key, st_data_t value, st_data_t data)
|
72
|
+
{
|
73
|
+
VALUE result = (VALUE)data;
|
74
|
+
prof_call_tree_t* call_tree_data = (prof_call_tree_t*)value;
|
75
|
+
VALUE aggregate_call_tree = prof_aggregate_call_tree_wrap(call_tree_data);
|
76
|
+
rb_ary_push(result, aggregate_call_tree);
|
77
|
+
|
78
|
+
return ST_CONTINUE;
|
79
|
+
}
|
80
|
+
|
81
|
+
static int prof_call_trees_collect_callees(st_data_t key, st_data_t value, st_data_t hash)
|
82
|
+
{
|
83
|
+
st_table* callers = (st_table*)hash;
|
84
|
+
prof_call_tree_t* call_tree_data = (prof_call_tree_t*)value;
|
85
|
+
|
86
|
+
prof_call_tree_t* aggregate_call_tree_data = NULL;
|
87
|
+
|
88
|
+
if (rb_st_lookup(callers, call_tree_data->method->key, (st_data_t*)&aggregate_call_tree_data))
|
89
|
+
{
|
90
|
+
prof_call_tree_merge(aggregate_call_tree_data, call_tree_data);
|
91
|
+
}
|
92
|
+
else
|
93
|
+
{
|
94
|
+
aggregate_call_tree_data = prof_call_tree_copy(call_tree_data);
|
95
|
+
rb_st_insert(callers, call_tree_data->method->key, (st_data_t)aggregate_call_tree_data);
|
96
|
+
}
|
97
|
+
|
98
|
+
return ST_CONTINUE;
|
99
|
+
}
|
100
|
+
|
101
|
+
size_t prof_call_trees_size(const void* data)
|
102
|
+
{
|
103
|
+
return sizeof(prof_call_trees_t);
|
104
|
+
}
|
105
|
+
|
106
|
+
static const rb_data_type_t call_trees_type =
|
107
|
+
{
|
108
|
+
.wrap_struct_name = "CallTrees",
|
109
|
+
.function =
|
110
|
+
{
|
111
|
+
.dmark = prof_call_trees_mark,
|
112
|
+
.dfree = prof_call_trees_ruby_gc_free,
|
113
|
+
.dsize = prof_call_trees_size,
|
114
|
+
},
|
115
|
+
.data = NULL,
|
116
|
+
.flags = RUBY_TYPED_FREE_IMMEDIATELY
|
117
|
+
};
|
118
|
+
|
119
|
+
VALUE prof_call_trees_wrap(prof_call_trees_t* call_trees)
|
120
|
+
{
|
121
|
+
if (call_trees->object == Qnil)
|
122
|
+
{
|
123
|
+
call_trees->object = TypedData_Wrap_Struct(cRpCallTrees, &call_trees_type, call_trees);
|
124
|
+
}
|
125
|
+
return call_trees->object;
|
126
|
+
}
|
127
|
+
|
128
|
+
void prof_add_call_tree(prof_call_trees_t* call_trees, prof_call_tree_t* call_tree)
|
129
|
+
{
|
130
|
+
if (call_trees->ptr == call_trees->end)
|
131
|
+
{
|
132
|
+
size_t len = call_trees->ptr - call_trees->start;
|
133
|
+
size_t new_capacity = (call_trees->end - call_trees->start) * 2;
|
134
|
+
REALLOC_N(call_trees->start, prof_call_tree_t*, new_capacity);
|
135
|
+
call_trees->ptr = call_trees->start + len;
|
136
|
+
call_trees->end = call_trees->start + new_capacity;
|
137
|
+
}
|
138
|
+
*call_trees->ptr = call_tree;
|
139
|
+
call_trees->ptr++;
|
140
|
+
}
|
141
|
+
|
142
|
+
/* ================ Call Infos =================*/
|
143
|
+
/* Document-class: RubyProf::CallTrees
|
144
|
+
The RubyProf::MethodInfo class stores profiling data for a method.
|
145
|
+
One instance of the RubyProf::MethodInfo class is created per method
|
146
|
+
called per thread. Thus, if a method is called in two different
|
147
|
+
thread then there will be two RubyProf::MethodInfo objects
|
148
|
+
created. RubyProf::MethodInfo objects can be accessed via
|
149
|
+
the RubyProf::Profile object. */
|
150
|
+
VALUE prof_call_trees_allocate(VALUE klass)
|
151
|
+
{
|
152
|
+
prof_call_trees_t* call_trees_data = prof_call_trees_create();
|
153
|
+
call_trees_data->object = prof_call_trees_wrap(call_trees_data);
|
154
|
+
return call_trees_data->object;
|
155
|
+
}
|
156
|
+
|
157
|
+
|
158
|
+
/* call-seq:
|
159
|
+
min_depth -> Integer
|
160
|
+
|
161
|
+
Returns the minimum depth of this method in any call tree */
|
162
|
+
VALUE prof_call_trees_min_depth(VALUE self)
|
163
|
+
{
|
164
|
+
unsigned int depth = INT_MAX;
|
165
|
+
|
166
|
+
prof_call_trees_t* call_trees = prof_get_call_trees(self);
|
167
|
+
for (prof_call_tree_t** p_call_tree = call_trees->start; p_call_tree < call_trees->ptr; p_call_tree++)
|
168
|
+
{
|
169
|
+
unsigned int call_tree_depth = prof_call_figure_depth(*p_call_tree);
|
170
|
+
if (call_tree_depth < depth)
|
171
|
+
depth = call_tree_depth;
|
172
|
+
}
|
173
|
+
|
174
|
+
return UINT2NUM(depth);
|
175
|
+
}
|
176
|
+
|
177
|
+
/* call-seq:
|
178
|
+
callers -> array
|
179
|
+
|
180
|
+
Returns an array of all CallTree objects that called this method. */
|
181
|
+
VALUE prof_call_trees_call_trees(VALUE self)
|
182
|
+
{
|
183
|
+
VALUE result = rb_ary_new();
|
184
|
+
|
185
|
+
prof_call_trees_t* call_trees = prof_get_call_trees(self);
|
186
|
+
for (prof_call_tree_t** p_call_tree = call_trees->start; p_call_tree < call_trees->ptr; p_call_tree++)
|
187
|
+
{
|
188
|
+
VALUE call_tree = prof_call_tree_wrap(*p_call_tree);
|
189
|
+
rb_ary_push(result, call_tree);
|
190
|
+
}
|
191
|
+
return result;
|
192
|
+
}
|
193
|
+
|
194
|
+
/* call-seq:
|
195
|
+
callers -> array
|
196
|
+
|
197
|
+
Returns an array of aggregated CallTree objects that called this method (ie, parents).*/
|
198
|
+
VALUE prof_call_trees_callers(VALUE self)
|
199
|
+
{
|
200
|
+
st_table* callers = rb_st_init_numtable();
|
201
|
+
|
202
|
+
prof_call_trees_t* call_trees = prof_get_call_trees(self);
|
203
|
+
for (prof_call_tree_t** p_call_tree = call_trees->start; p_call_tree < call_trees->ptr; p_call_tree++)
|
204
|
+
{
|
205
|
+
prof_call_tree_t* parent = (*p_call_tree)->parent;
|
206
|
+
if (parent == NULL)
|
207
|
+
continue;
|
208
|
+
|
209
|
+
prof_call_tree_t* aggregate_call_tree_data = NULL;
|
210
|
+
|
211
|
+
if (rb_st_lookup(callers, parent->method->key, (st_data_t*)&aggregate_call_tree_data))
|
212
|
+
{
|
213
|
+
prof_call_tree_merge(aggregate_call_tree_data, *p_call_tree);
|
214
|
+
}
|
215
|
+
else
|
216
|
+
{
|
217
|
+
aggregate_call_tree_data = prof_call_tree_copy(*p_call_tree);
|
218
|
+
rb_st_insert(callers, parent->method->key, (st_data_t)aggregate_call_tree_data);
|
219
|
+
}
|
220
|
+
}
|
221
|
+
|
222
|
+
VALUE result = rb_ary_new_capa((long)callers->num_entries);
|
223
|
+
rb_st_foreach(callers, prof_call_trees_collect_aggregates, result);
|
224
|
+
rb_st_free_table(callers);
|
225
|
+
return result;
|
226
|
+
}
|
227
|
+
|
228
|
+
/* call-seq:
|
229
|
+
callees -> array
|
230
|
+
|
231
|
+
Returns an array of aggregated CallTree objects that this method called (ie, children).*/
|
232
|
+
VALUE prof_call_trees_callees(VALUE self)
|
233
|
+
{
|
234
|
+
st_table* callees = rb_st_init_numtable();
|
235
|
+
|
236
|
+
prof_call_trees_t* call_trees = prof_get_call_trees(self);
|
237
|
+
for (prof_call_tree_t** call_tree = call_trees->start; call_tree < call_trees->ptr; call_tree++)
|
238
|
+
{
|
239
|
+
rb_st_foreach((*call_tree)->children, prof_call_trees_collect_callees, (st_data_t)callees);
|
240
|
+
}
|
241
|
+
|
242
|
+
VALUE result = rb_ary_new_capa((long)callees->num_entries);
|
243
|
+
rb_st_foreach(callees, prof_call_trees_collect_aggregates, result);
|
244
|
+
rb_st_free_table(callees);
|
245
|
+
return result;
|
246
|
+
}
|
247
|
+
|
248
|
+
/* :nodoc: */
|
249
|
+
VALUE prof_call_trees_dump(VALUE self)
|
250
|
+
{
|
251
|
+
VALUE result = rb_hash_new();
|
252
|
+
rb_hash_aset(result, ID2SYM(rb_intern("call_trees")), prof_call_trees_call_trees(self));
|
253
|
+
|
254
|
+
return result;
|
255
|
+
}
|
256
|
+
|
257
|
+
/* :nodoc: */
|
258
|
+
VALUE prof_call_trees_load(VALUE self, VALUE data)
|
259
|
+
{
|
260
|
+
prof_call_trees_t* call_trees_data = prof_get_call_trees(self);
|
261
|
+
call_trees_data->object = self;
|
262
|
+
|
263
|
+
VALUE call_trees = rb_hash_aref(data, ID2SYM(rb_intern("call_trees")));
|
264
|
+
for (int i = 0; i < rb_array_len(call_trees); i++)
|
265
|
+
{
|
266
|
+
VALUE call_tree = rb_ary_entry(call_trees, i);
|
267
|
+
prof_call_tree_t* call_tree_data = prof_get_call_tree(call_tree);
|
268
|
+
prof_add_call_tree(call_trees_data, call_tree_data);
|
269
|
+
}
|
270
|
+
|
271
|
+
return data;
|
272
|
+
}
|
273
|
+
|
274
|
+
void rp_init_call_trees()
|
275
|
+
{
|
276
|
+
cRpCallTrees = rb_define_class_under(mProf, "CallTrees", rb_cObject);
|
277
|
+
rb_undef_method(CLASS_OF(cRpCallTrees), "new");
|
278
|
+
rb_define_alloc_func(cRpCallTrees, prof_call_trees_allocate);
|
279
|
+
|
280
|
+
rb_define_method(cRpCallTrees, "min_depth", prof_call_trees_min_depth, 0);
|
281
|
+
|
282
|
+
rb_define_method(cRpCallTrees, "call_trees", prof_call_trees_call_trees, 0);
|
283
|
+
rb_define_method(cRpCallTrees, "callers", prof_call_trees_callers, 0);
|
284
|
+
rb_define_method(cRpCallTrees, "callees", prof_call_trees_callees, 0);
|
285
|
+
|
286
|
+
rb_define_method(cRpCallTrees, "_dump_data", prof_call_trees_dump, 0);
|
287
|
+
rb_define_method(cRpCallTrees, "_load_data", prof_call_trees_load, 1);
|
288
|
+
}
|
@@ -0,0 +1,28 @@
|
|
1
|
+
/* Copyright (C) 2005-2013 Shugo Maeda <shugo@ruby-lang.org> and Charlie Savage <cfis@savagexi.com>
|
2
|
+
Please see the LICENSE file for copyright and distribution information */
|
3
|
+
|
4
|
+
#ifndef __RP_CALL_TREES_H__
|
5
|
+
#define __RP_CALL_TREES_H__
|
6
|
+
|
7
|
+
#include "ruby_prof.h"
|
8
|
+
#include "rp_call_tree.h"
|
9
|
+
|
10
|
+
/* Array of call_tree objects */
|
11
|
+
typedef struct prof_call_trees_t
|
12
|
+
{
|
13
|
+
prof_call_tree_t** start;
|
14
|
+
prof_call_tree_t** end;
|
15
|
+
prof_call_tree_t** ptr;
|
16
|
+
|
17
|
+
VALUE object;
|
18
|
+
} prof_call_trees_t;
|
19
|
+
|
20
|
+
|
21
|
+
void rp_init_call_trees();
|
22
|
+
prof_call_trees_t* prof_call_trees_create();
|
23
|
+
void prof_call_trees_free(prof_call_trees_t* call_trees);
|
24
|
+
prof_call_trees_t* prof_get_call_trees(VALUE self);
|
25
|
+
void prof_add_call_tree(prof_call_trees_t* call_trees, prof_call_tree_t* call_tree);
|
26
|
+
VALUE prof_call_trees_wrap(prof_call_trees_t* call_trees);
|
27
|
+
|
28
|
+
#endif //__RP_CALL_TREES_H__
|
@@ -0,0 +1,47 @@
|
|
1
|
+
/* Copyright (C) 2005-2013 Shugo Maeda <shugo@ruby-lang.org> and Charlie Savage <cfis@savagexi.com>
|
2
|
+
Please see the LICENSE file for copyright and distribution information */
|
3
|
+
|
4
|
+
/* :nodoc: */
|
5
|
+
|
6
|
+
#include "rp_measurement.h"
|
7
|
+
|
8
|
+
static VALUE cMeasureAllocations;
|
9
|
+
VALUE total_allocated_objects_key;
|
10
|
+
|
11
|
+
static double measure_allocations(rb_trace_arg_t* trace_arg)
|
12
|
+
{
|
13
|
+
static double result = 0;
|
14
|
+
|
15
|
+
if (trace_arg)
|
16
|
+
{
|
17
|
+
// Only process creation of new objects
|
18
|
+
rb_event_flag_t event = rb_tracearg_event_flag(trace_arg);
|
19
|
+
if (event == RUBY_INTERNAL_EVENT_NEWOBJ) {
|
20
|
+
// Don't count allocations of internal IMemo objects
|
21
|
+
VALUE object = rb_tracearg_object(trace_arg);
|
22
|
+
if (BUILTIN_TYPE(object) != T_IMEMO)
|
23
|
+
result++;
|
24
|
+
}
|
25
|
+
}
|
26
|
+
return result;
|
27
|
+
}
|
28
|
+
|
29
|
+
prof_measurer_t* prof_measurer_allocations(bool track_allocations)
|
30
|
+
{
|
31
|
+
prof_measurer_t* measure = ALLOC(prof_measurer_t);
|
32
|
+
measure->mode = MEASURE_ALLOCATIONS;
|
33
|
+
measure->measure = measure_allocations;
|
34
|
+
measure->multiplier = 1;
|
35
|
+
// Need to track allocations to get RUBY_INTERNAL_EVENT_NEWOBJ event
|
36
|
+
measure->track_allocations = track_allocations;
|
37
|
+
|
38
|
+
return measure;
|
39
|
+
}
|
40
|
+
|
41
|
+
void rp_init_measure_allocations()
|
42
|
+
{
|
43
|
+
total_allocated_objects_key = ID2SYM(rb_intern("total_allocated_objects"));
|
44
|
+
rb_define_const(mProf, "ALLOCATIONS", INT2NUM(MEASURE_ALLOCATIONS));
|
45
|
+
|
46
|
+
cMeasureAllocations = rb_define_class_under(mMeasure, "Allocations", rb_cObject);
|
47
|
+
}
|
@@ -0,0 +1,46 @@
|
|
1
|
+
/* Copyright (C) 2005-2013 Shugo Maeda <shugo@ruby-lang.org> and Charlie Savage <cfis@savagexi.com>
|
2
|
+
Please see the LICENSE file for copyright and distribution information */
|
3
|
+
|
4
|
+
/* :nodoc: */
|
5
|
+
|
6
|
+
#include "rp_measurement.h"
|
7
|
+
|
8
|
+
static VALUE cMeasureMemory;
|
9
|
+
|
10
|
+
static double measure_memory(rb_trace_arg_t* trace_arg)
|
11
|
+
{
|
12
|
+
static double result = 0;
|
13
|
+
|
14
|
+
if (trace_arg)
|
15
|
+
{
|
16
|
+
// Only process creation of new objects
|
17
|
+
rb_event_flag_t event = rb_tracearg_event_flag(trace_arg);
|
18
|
+
if (event == RUBY_INTERNAL_EVENT_NEWOBJ)
|
19
|
+
{
|
20
|
+
// Don't count allocations of internal IMemo objects
|
21
|
+
VALUE object = rb_tracearg_object(trace_arg);
|
22
|
+
if (BUILTIN_TYPE(object) != T_IMEMO)
|
23
|
+
result += rb_obj_memsize_of(object);
|
24
|
+
}
|
25
|
+
}
|
26
|
+
|
27
|
+
return result;
|
28
|
+
}
|
29
|
+
|
30
|
+
prof_measurer_t* prof_measurer_memory(bool track_allocations)
|
31
|
+
{
|
32
|
+
prof_measurer_t* measure = ALLOC(prof_measurer_t);
|
33
|
+
measure->mode = MEASURE_MEMORY;
|
34
|
+
measure->measure = measure_memory;
|
35
|
+
measure->multiplier = 1;
|
36
|
+
// Need to track allocations to get RUBY_INTERNAL_EVENT_NEWOBJ event
|
37
|
+
measure->track_allocations = true;
|
38
|
+
return measure;
|
39
|
+
}
|
40
|
+
|
41
|
+
void rp_init_measure_memory()
|
42
|
+
{
|
43
|
+
rb_define_const(mProf, "MEMORY", INT2NUM(MEASURE_MEMORY));
|
44
|
+
|
45
|
+
cMeasureMemory = rb_define_class_under(mMeasure, "Allocations", rb_cObject);
|
46
|
+
}
|
@@ -0,0 +1,66 @@
|
|
1
|
+
/* Copyright (C) 2005-2019 Shugo Maeda <shugo@ruby-lang.org> and Charlie Savage <cfis@savagexi.com>
|
2
|
+
Please see the LICENSE file for copyright and distribution information */
|
3
|
+
|
4
|
+
#include "rp_measurement.h"
|
5
|
+
#include <time.h>
|
6
|
+
|
7
|
+
static VALUE cMeasureProcessTime;
|
8
|
+
|
9
|
+
static double measure_process_time(rb_trace_arg_t* trace_arg)
|
10
|
+
{
|
11
|
+
#if defined(_WIN32)
|
12
|
+
FILETIME createTime;
|
13
|
+
FILETIME exitTime;
|
14
|
+
FILETIME kernelTime;
|
15
|
+
FILETIME userTime;
|
16
|
+
|
17
|
+
ULARGE_INTEGER kernelTimeInt;
|
18
|
+
ULARGE_INTEGER userTimeInt;
|
19
|
+
|
20
|
+
GetProcessTimes(GetCurrentProcess(), &createTime, &exitTime, &kernelTime, &userTime);
|
21
|
+
|
22
|
+
kernelTimeInt.LowPart = kernelTime.dwLowDateTime;
|
23
|
+
kernelTimeInt.HighPart = kernelTime.dwHighDateTime;
|
24
|
+
userTimeInt.LowPart = userTime.dwLowDateTime;
|
25
|
+
userTimeInt.HighPart = userTime.dwHighDateTime;
|
26
|
+
|
27
|
+
return (double)(kernelTimeInt.QuadPart + userTimeInt.QuadPart);
|
28
|
+
#elif !defined(CLOCK_PROCESS_CPUTIME_ID)
|
29
|
+
#include <sys/resource.h>
|
30
|
+
struct rusage usage;
|
31
|
+
getrusage(RUSAGE_SELF, &usage);
|
32
|
+
return usage.ru_stime.tv_sec + usage.ru_utime.tv_sec + ((usage.ru_stime.tv_usec + usage.ru_utime.tv_usec) / 1000000.0);
|
33
|
+
#else
|
34
|
+
struct timespec clock;
|
35
|
+
clock_gettime(CLOCK_PROCESS_CPUTIME_ID, &clock);
|
36
|
+
return clock.tv_sec + (clock.tv_nsec / 1000000000.0);
|
37
|
+
#endif
|
38
|
+
}
|
39
|
+
|
40
|
+
static double multiplier_process_time(void)
|
41
|
+
{
|
42
|
+
#if defined(_WIN32)
|
43
|
+
// Times are in 100-nanosecond time units. So instead of 10-9 use 10-7
|
44
|
+
return 1.0 / 10000000.0;
|
45
|
+
#else
|
46
|
+
return 1.0;
|
47
|
+
#endif
|
48
|
+
}
|
49
|
+
|
50
|
+
prof_measurer_t* prof_measurer_process_time(bool track_allocations)
|
51
|
+
{
|
52
|
+
prof_measurer_t* measure = ALLOC(prof_measurer_t);
|
53
|
+
measure->mode = MEASURE_PROCESS_TIME;
|
54
|
+
measure->measure = measure_process_time;
|
55
|
+
measure->multiplier = multiplier_process_time();
|
56
|
+
measure->track_allocations = track_allocations;
|
57
|
+
return measure;
|
58
|
+
}
|
59
|
+
|
60
|
+
void rp_init_measure_process_time()
|
61
|
+
{
|
62
|
+
rb_define_const(mProf, "CLOCKS_PER_SEC", INT2NUM(CLOCKS_PER_SEC));
|
63
|
+
rb_define_const(mProf, "PROCESS_TIME", INT2NUM(MEASURE_PROCESS_TIME));
|
64
|
+
|
65
|
+
cMeasureProcessTime = rb_define_class_under(mMeasure, "ProcessTime", rb_cObject);
|
66
|
+
}
|
@@ -0,0 +1,64 @@
|
|
1
|
+
/* Copyright (C) 2005-2019 Shugo Maeda <shugo@ruby-lang.org> and Charlie Savage <cfis@savagexi.com>
|
2
|
+
Please see the LICENSE file for copyright and distribution information */
|
3
|
+
|
4
|
+
/* :nodoc: */
|
5
|
+
#include "rp_measurement.h"
|
6
|
+
|
7
|
+
#if defined(__APPLE__)
|
8
|
+
#include <mach/mach_time.h>
|
9
|
+
#elif !defined(_WIN32)
|
10
|
+
#include <time.h>
|
11
|
+
#endif
|
12
|
+
|
13
|
+
static VALUE cMeasureWallTime;
|
14
|
+
|
15
|
+
static double measure_wall_time(rb_trace_arg_t* trace_arg)
|
16
|
+
{
|
17
|
+
#if defined(_WIN32)
|
18
|
+
LARGE_INTEGER time;
|
19
|
+
QueryPerformanceCounter(&time);
|
20
|
+
return (double)time.QuadPart;
|
21
|
+
#elif defined(__APPLE__)
|
22
|
+
return mach_absolute_time();// * (uint64_t)mach_timebase.numer / (uint64_t)mach_timebase.denom;
|
23
|
+
#elif defined(__linux__)
|
24
|
+
struct timespec tv;
|
25
|
+
clock_gettime(CLOCK_MONOTONIC, &tv);
|
26
|
+
return tv.tv_sec + (tv.tv_nsec / 1000000000.0);
|
27
|
+
#else
|
28
|
+
struct timeval tv;
|
29
|
+
gettimeofday(&tv, NULL);
|
30
|
+
return tv.tv_sec + (tv.tv_usec / 1000000.0);
|
31
|
+
#endif
|
32
|
+
}
|
33
|
+
|
34
|
+
static double multiplier_wall_time(void)
|
35
|
+
{
|
36
|
+
#if defined(_WIN32)
|
37
|
+
LARGE_INTEGER frequency;
|
38
|
+
QueryPerformanceFrequency(&frequency);
|
39
|
+
return 1.0 / frequency.QuadPart;
|
40
|
+
#elif defined(__APPLE__)
|
41
|
+
mach_timebase_info_data_t mach_timebase;
|
42
|
+
mach_timebase_info(&mach_timebase);
|
43
|
+
return (uint64_t)mach_timebase.numer / (uint64_t)mach_timebase.denom / 1000000000.0;
|
44
|
+
#else
|
45
|
+
return 1.0;
|
46
|
+
#endif
|
47
|
+
}
|
48
|
+
|
49
|
+
prof_measurer_t* prof_measurer_wall_time(bool track_allocations)
|
50
|
+
{
|
51
|
+
prof_measurer_t* measure = ALLOC(prof_measurer_t);
|
52
|
+
measure->mode = MEASURE_WALL_TIME;
|
53
|
+
measure->measure = measure_wall_time;
|
54
|
+
measure->multiplier = multiplier_wall_time();
|
55
|
+
measure->track_allocations = track_allocations;
|
56
|
+
return measure;
|
57
|
+
}
|
58
|
+
|
59
|
+
void rp_init_measure_wall_time()
|
60
|
+
{
|
61
|
+
rb_define_const(mProf, "WALL_TIME", INT2NUM(MEASURE_WALL_TIME));
|
62
|
+
|
63
|
+
cMeasureWallTime = rb_define_class_under(mMeasure, "WallTime", rb_cObject);
|
64
|
+
}
|