/********************************************************************** vm_backtrace.c - $Author: ko1 $ created at: Sun Jun 03 00:14:20 2012 Copyright (C) 1993-2012 Yukihiro Matsumoto **********************************************************************/ #include "eval_intern.h" #include "internal.h" #include "internal/error.h" #include "internal/vm.h" #include "iseq.h" #include "ruby/debug.h" #include "ruby/encoding.h" #include "vm_core.h" static VALUE rb_cBacktrace; static VALUE rb_cBacktraceLocation; static VALUE id2str(ID id) { VALUE str = rb_id2str(id); if (!str) return Qnil; return str; } #define rb_id2str(id) id2str(id) #define BACKTRACE_START 0 #define ALL_BACKTRACE_LINES -1 inline static int calc_lineno(const rb_iseq_t *iseq, const VALUE *pc) { VM_ASSERT(iseq); VM_ASSERT(iseq->body); VM_ASSERT(iseq->body->iseq_encoded); VM_ASSERT(iseq->body->iseq_size); if (! pc) { /* This can happen during VM bootup. */ VM_ASSERT(iseq->body->type == ISEQ_TYPE_TOP); VM_ASSERT(! iseq->body->local_table); VM_ASSERT(! iseq->body->local_table_size); return 0; } else { ptrdiff_t n = pc - iseq->body->iseq_encoded; VM_ASSERT(n <= iseq->body->iseq_size); VM_ASSERT(n >= 0); ASSUME(n >= 0); size_t pos = n; /* no overflow */ if (LIKELY(pos)) { /* use pos-1 because PC points next instruction at the beginning of instruction */ pos--; } #if VMDEBUG && defined(HAVE_BUILTIN___BUILTIN_TRAP) else { /* SDR() is not possible; that causes infinite loop. */ rb_print_backtrace(); __builtin_trap(); } #endif return rb_iseq_line_no(iseq, pos); } } int rb_vm_get_sourceline(const rb_control_frame_t *cfp) { if (VM_FRAME_RUBYFRAME_P(cfp) && cfp->iseq) { const rb_iseq_t *iseq = cfp->iseq; int line = calc_lineno(iseq, cfp->pc); if (line != 0) { return line; } else { return FIX2INT(rb_iseq_first_lineno(iseq)); } } else { return 0; } } typedef struct rb_backtrace_location_struct { enum LOCATION_TYPE { LOCATION_TYPE_ISEQ = 1, LOCATION_TYPE_ISEQ_CALCED, LOCATION_TYPE_CFUNC, } type; union { struct { const rb_iseq_t *iseq; union { const VALUE *pc; int lineno; } lineno; } iseq; struct { ID mid; struct rb_backtrace_location_struct *prev_loc; } cfunc; } body; } rb_backtrace_location_t; struct valued_frame_info { rb_backtrace_location_t *loc; VALUE btobj; }; static void location_mark(void *ptr) { struct valued_frame_info *vfi = (struct valued_frame_info *)ptr; rb_gc_mark(vfi->btobj); } static void location_mark_entry(rb_backtrace_location_t *fi) { switch (fi->type) { case LOCATION_TYPE_ISEQ: case LOCATION_TYPE_ISEQ_CALCED: rb_gc_mark_movable((VALUE)fi->body.iseq.iseq); break; case LOCATION_TYPE_CFUNC: default: break; } } static size_t location_memsize(const void *ptr) { /* rb_backtrace_location_t *fi = (rb_backtrace_location_t *)ptr; */ return sizeof(rb_backtrace_location_t); } static const rb_data_type_t location_data_type = { "frame_info", {location_mark, RUBY_TYPED_DEFAULT_FREE, location_memsize,}, 0, 0, RUBY_TYPED_FREE_IMMEDIATELY }; static inline rb_backtrace_location_t * location_ptr(VALUE locobj) { struct valued_frame_info *vloc; GetCoreDataFromValue(locobj, struct valued_frame_info, vloc); return vloc->loc; } static int location_lineno(rb_backtrace_location_t *loc) { switch (loc->type) { case LOCATION_TYPE_ISEQ: loc->type = LOCATION_TYPE_ISEQ_CALCED; return (loc->body.iseq.lineno.lineno = calc_lineno(loc->body.iseq.iseq, loc->body.iseq.lineno.pc)); case LOCATION_TYPE_ISEQ_CALCED: return loc->body.iseq.lineno.lineno; case LOCATION_TYPE_CFUNC: if (loc->body.cfunc.prev_loc) { return location_lineno(loc->body.cfunc.prev_loc); } return 0; default: rb_bug("location_lineno: unreachable"); UNREACHABLE; } } /* * Returns the line number of this frame. * * For example, using +caller_locations.rb+ from Thread::Backtrace::Location * * loc = c(0..1).first * loc.lineno #=> 2 */ static VALUE location_lineno_m(VALUE self) { return INT2FIX(location_lineno(location_ptr(self))); } static VALUE location_label(rb_backtrace_location_t *loc) { switch (loc->type) { case LOCATION_TYPE_ISEQ: case LOCATION_TYPE_ISEQ_CALCED: return loc->body.iseq.iseq->body->location.label; case LOCATION_TYPE_CFUNC: return rb_id2str(loc->body.cfunc.mid); default: rb_bug("location_label: unreachable"); UNREACHABLE; } } /* * Returns the label of this frame. * * Usually consists of method, class, module, etc names with decoration. * * Consider the following example: * * def foo * puts caller_locations(0).first.label * * 1.times do * puts caller_locations(0).first.label * * 1.times do * puts caller_locations(0).first.label * end * * end * end * * The result of calling +foo+ is this: * * label: foo * label: block in foo * label: block (2 levels) in foo * */ static VALUE location_label_m(VALUE self) { return location_label(location_ptr(self)); } static VALUE location_base_label(rb_backtrace_location_t *loc) { switch (loc->type) { case LOCATION_TYPE_ISEQ: case LOCATION_TYPE_ISEQ_CALCED: return loc->body.iseq.iseq->body->location.base_label; case LOCATION_TYPE_CFUNC: return rb_id2str(loc->body.cfunc.mid); default: rb_bug("location_base_label: unreachable"); UNREACHABLE; } } /* * Returns the base label of this frame. * * Usually same as #label, without decoration. */ static VALUE location_base_label_m(VALUE self) { return location_base_label(location_ptr(self)); } static VALUE location_path(rb_backtrace_location_t *loc) { switch (loc->type) { case LOCATION_TYPE_ISEQ: case LOCATION_TYPE_ISEQ_CALCED: return rb_iseq_path(loc->body.iseq.iseq); case LOCATION_TYPE_CFUNC: if (loc->body.cfunc.prev_loc) { return location_path(loc->body.cfunc.prev_loc); } return Qnil; default: rb_bug("location_path: unreachable"); UNREACHABLE; } } /* * Returns the file name of this frame. This will generally be an absolute * path, unless the frame is in the main script, in which case it will be the * script location passed on the command line. * * For example, using +caller_locations.rb+ from Thread::Backtrace::Location * * loc = c(0..1).first * loc.path #=> caller_locations.rb */ static VALUE location_path_m(VALUE self) { return location_path(location_ptr(self)); } static VALUE location_realpath(rb_backtrace_location_t *loc) { switch (loc->type) { case LOCATION_TYPE_ISEQ: case LOCATION_TYPE_ISEQ_CALCED: return rb_iseq_realpath(loc->body.iseq.iseq); case LOCATION_TYPE_CFUNC: if (loc->body.cfunc.prev_loc) { return location_realpath(loc->body.cfunc.prev_loc); } return Qnil; default: rb_bug("location_realpath: unreachable"); UNREACHABLE; } } /* * Returns the full file path of this frame. * * Same as #path, except that it will return absolute path * even if the frame is in the main script. */ static VALUE location_absolute_path_m(VALUE self) { return location_realpath(location_ptr(self)); } static VALUE location_format(VALUE file, int lineno, VALUE name) { VALUE s = rb_enc_sprintf(rb_enc_compatible(file, name), "%s", RSTRING_PTR(file)); if (lineno != 0) { rb_str_catf(s, ":%d", lineno); } rb_str_cat_cstr(s, ":in "); if (NIL_P(name)) { rb_str_cat_cstr(s, "unknown method"); } else { rb_str_catf(s, "`%s'", RSTRING_PTR(name)); } return s; } static VALUE location_to_str(rb_backtrace_location_t *loc) { VALUE file, name; int lineno; switch (loc->type) { case LOCATION_TYPE_ISEQ: file = rb_iseq_path(loc->body.iseq.iseq); name = loc->body.iseq.iseq->body->location.label; lineno = loc->body.iseq.lineno.lineno = calc_lineno(loc->body.iseq.iseq, loc->body.iseq.lineno.pc); loc->type = LOCATION_TYPE_ISEQ_CALCED; break; case LOCATION_TYPE_ISEQ_CALCED: file = rb_iseq_path(loc->body.iseq.iseq); lineno = loc->body.iseq.lineno.lineno; name = loc->body.iseq.iseq->body->location.label; break; case LOCATION_TYPE_CFUNC: if (loc->body.cfunc.prev_loc) { file = rb_iseq_path(loc->body.cfunc.prev_loc->body.iseq.iseq); lineno = location_lineno(loc->body.cfunc.prev_loc); } else { file = GET_VM()->progname; lineno = 0; } name = rb_id2str(loc->body.cfunc.mid); break; default: rb_bug("location_to_str: unreachable"); } return location_format(file, lineno, name); } /* * Returns a Kernel#caller style string representing this frame. */ static VALUE location_to_str_m(VALUE self) { return location_to_str(location_ptr(self)); } /* * Returns the same as calling +inspect+ on the string representation of * #to_str */ static VALUE location_inspect_m(VALUE self) { return rb_str_inspect(location_to_str(location_ptr(self))); } typedef struct rb_backtrace_struct { rb_backtrace_location_t *backtrace; int backtrace_size; VALUE strary; VALUE locary; } rb_backtrace_t; static void backtrace_mark(void *ptr) { rb_backtrace_t *bt = (rb_backtrace_t *)ptr; size_t i, s = bt->backtrace_size; for (i=0; ibacktrace[i]); } rb_gc_mark_movable(bt->strary); rb_gc_mark_movable(bt->locary); } static void backtrace_free(void *ptr) { rb_backtrace_t *bt = (rb_backtrace_t *)ptr; if (bt->backtrace) ruby_xfree(bt->backtrace); ruby_xfree(bt); } static void location_update_entry(rb_backtrace_location_t *fi) { switch (fi->type) { case LOCATION_TYPE_ISEQ: case LOCATION_TYPE_ISEQ_CALCED: fi->body.iseq.iseq = (rb_iseq_t*)rb_gc_location((VALUE)fi->body.iseq.iseq); break; case LOCATION_TYPE_CFUNC: default: break; } } static void backtrace_update(void *ptr) { rb_backtrace_t *bt = (rb_backtrace_t *)ptr; size_t i, s = bt->backtrace_size; for (i=0; ibacktrace[i]); } bt->strary = rb_gc_location(bt->strary); bt->locary = rb_gc_location(bt->locary); } static size_t backtrace_memsize(const void *ptr) { rb_backtrace_t *bt = (rb_backtrace_t *)ptr; return sizeof(rb_backtrace_t) + sizeof(rb_backtrace_location_t) * bt->backtrace_size; } static const rb_data_type_t backtrace_data_type = { "backtrace", {backtrace_mark, backtrace_free, backtrace_memsize, backtrace_update}, 0, 0, RUBY_TYPED_FREE_IMMEDIATELY }; int rb_backtrace_p(VALUE obj) { return rb_typeddata_is_kind_of(obj, &backtrace_data_type); } static VALUE backtrace_alloc(VALUE klass) { rb_backtrace_t *bt; VALUE obj = TypedData_Make_Struct(klass, rb_backtrace_t, &backtrace_data_type, bt); return obj; } static long backtrace_size(const rb_execution_context_t *ec) { const rb_control_frame_t *last_cfp = ec->cfp; const rb_control_frame_t *start_cfp = RUBY_VM_END_CONTROL_FRAME(ec); if (start_cfp == NULL) { return -1; } start_cfp = RUBY_VM_NEXT_CONTROL_FRAME( RUBY_VM_NEXT_CONTROL_FRAME(start_cfp)); /* skip top frames */ if (start_cfp < last_cfp) { return 0; } return start_cfp - last_cfp + 1; } static int backtrace_each(const rb_execution_context_t *ec, ptrdiff_t from_last, long num_frames, void (*init)(void *arg, size_t size), void (*iter_iseq)(void *arg, const rb_control_frame_t *cfp), void (*iter_cfunc)(void *arg, const rb_control_frame_t *cfp, ID mid), void (*iter_skip)(void *arg, const rb_control_frame_t *cfp), void *arg) { const rb_control_frame_t *last_cfp = ec->cfp; const rb_control_frame_t *start_cfp = RUBY_VM_END_CONTROL_FRAME(ec); const rb_control_frame_t *cfp; ptrdiff_t size, real_size, i, j, last, start = 0; int ret = 0; // In the case the thread vm_stack or cfp is not initialized, there is no backtrace. if (start_cfp == NULL) { init(arg, 0); return ret; } /* <- start_cfp (end control frame) * top frame (dummy) * top frame (dummy) * top frame <- start_cfp * top frame * ... * 2nd frame <- lev:0 * current frame <- ec->cfp */ start_cfp = RUBY_VM_NEXT_CONTROL_FRAME( RUBY_VM_NEXT_CONTROL_FRAME(start_cfp)); /* skip top frames */ if (start_cfp < last_cfp) { real_size = size = last = 0; } else { /* Ensure we don't look at frames beyond the ones requested */ for (; from_last > 0 && start_cfp >= last_cfp; from_last--) { last_cfp = RUBY_VM_PREVIOUS_CONTROL_FRAME(last_cfp); } real_size = size = start_cfp - last_cfp + 1; if (from_last > size) { size = last = 0; ret = 1; } else if (num_frames >= 0 && num_frames < size) { if (from_last + num_frames > size) { size -= from_last; last = size; } else { start = size - from_last - num_frames; size = num_frames; last = start + size; } } else { size -= from_last; last = size; } } init(arg, size); /* If a limited number of frames is requested, scan the VM stack for * ignored frames (iseq without pc). Then adjust the start for the * backtrace to account for skipped frames. */ if (start > 0 && num_frames >= 0 && num_frames < real_size) { ptrdiff_t ignored_frames; bool ignored_frames_before_start = false; for (i=0, j=0, cfp = start_cfp; iiseq && !cfp->pc) { if (j < start) ignored_frames_before_start = true; else i--; } } ignored_frames = j - i; if (ignored_frames) { if (ignored_frames_before_start) { /* There were ignored frames before start. So just decrementing * start for ignored frames could still result in not all desired * frames being captured. * * First, scan to the CFP of the desired start frame. * * Then scan backwards to previous frames, skipping the number of * frames ignored after start and any additional ones before start, * so the number of desired frames will be correctly captured. */ for (i=0, j=0, cfp = start_cfp; i 0 && ignored_frames > 0 && j > 0; j--, ignored_frames--, start--, cfp = RUBY_VM_PREVIOUS_CONTROL_FRAME(cfp)) { if (cfp->iseq && !cfp->pc) { ignored_frames++; } } } else { /* No ignored frames before start frame, just decrement start */ start -= ignored_frames; } } } for (i=0, j=0, cfp = start_cfp; ivm_stack + ec->vm_stack_size) - cfp); */ if (cfp->iseq) { if (cfp->pc) { iter_iseq(arg, cfp); } else { i--; } } else if (RUBYVM_CFUNC_FRAME_P(cfp)) { const rb_callable_method_entry_t *me = rb_vm_frame_method_entry(cfp); ID mid = me->def->original_id; iter_cfunc(arg, cfp, mid); } } return ret; } struct bt_iter_arg { rb_backtrace_t *bt; VALUE btobj; rb_backtrace_location_t *prev_loc; const rb_control_frame_t *prev_cfp; rb_backtrace_location_t *init_loc; }; static bool is_internal_location(const rb_control_frame_t *cfp) { static const char prefix[] = "iseq); return strncmp(prefix, RSTRING_PTR(file), prefix_len) == 0; } static void bt_init(void *ptr, size_t size) { struct bt_iter_arg *arg = (struct bt_iter_arg *)ptr; arg->btobj = backtrace_alloc(rb_cBacktrace); GetCoreDataFromValue(arg->btobj, rb_backtrace_t, arg->bt); arg->bt->backtrace = ZALLOC_N(rb_backtrace_location_t, size+1); arg->bt->backtrace_size = 1; arg->prev_cfp = NULL; arg->init_loc = &arg->bt->backtrace[size]; } static void bt_iter_iseq(void *ptr, const rb_control_frame_t *cfp) { const rb_iseq_t *iseq = cfp->iseq; const VALUE *pc = cfp->pc; struct bt_iter_arg *arg = (struct bt_iter_arg *)ptr; rb_backtrace_location_t *loc = &arg->bt->backtrace[arg->bt->backtrace_size++-1]; loc->type = LOCATION_TYPE_ISEQ; loc->body.iseq.iseq = iseq; loc->body.iseq.lineno.pc = pc; arg->prev_loc = loc; } static void bt_iter_iseq_skip_internal(void *ptr, const rb_control_frame_t *cfp) { struct bt_iter_arg *arg = (struct bt_iter_arg *)ptr; rb_backtrace_location_t *loc = &arg->bt->backtrace[arg->bt->backtrace_size++-1]; if (!is_internal_location(cfp)) { loc->type = LOCATION_TYPE_ISEQ; loc->body.iseq.iseq = cfp->iseq; loc->body.iseq.lineno.pc = cfp->pc; arg->prev_loc = loc; } else if (arg->prev_cfp) { loc->type = LOCATION_TYPE_ISEQ; loc->body.iseq.iseq = arg->prev_cfp->iseq; loc->body.iseq.lineno.pc = arg->prev_cfp->pc; arg->prev_loc = loc; } else { rb_bug("No non-internal backtrace entry before an bt->backtrace[arg->bt->backtrace_size++-1]; loc->type = LOCATION_TYPE_CFUNC; loc->body.cfunc.mid = mid; if (arg->prev_loc) { loc->body.cfunc.prev_loc = arg->prev_loc; } else if (arg->prev_cfp) { const rb_iseq_t *iseq = arg->prev_cfp->iseq; const VALUE *pc = arg->prev_cfp->pc; arg->init_loc->type = LOCATION_TYPE_ISEQ; arg->init_loc->body.iseq.iseq = iseq; arg->init_loc->body.iseq.lineno.pc = pc; loc->body.cfunc.prev_loc = arg->prev_loc = arg->init_loc; } else { loc->body.cfunc.prev_loc = NULL; } } static void bt_iter_skip(void *ptr, const rb_control_frame_t *cfp) { if (cfp->iseq && cfp->pc) { ((struct bt_iter_arg *)ptr)->prev_cfp = cfp; } } static void bt_iter_skip_skip_internal(void *ptr, const rb_control_frame_t *cfp) { if (cfp->iseq && cfp->pc) { if (!is_internal_location(cfp)) { ((struct bt_iter_arg *) ptr)->prev_cfp = cfp; } } } static VALUE rb_ec_partial_backtrace_object(const rb_execution_context_t *ec, long lev, long n, int* level_too_large, bool skip_internal) { struct bt_iter_arg arg; int too_large; arg.prev_loc = 0; too_large = backtrace_each(ec, lev, n, bt_init, skip_internal ? bt_iter_iseq_skip_internal : bt_iter_iseq, bt_iter_cfunc, skip_internal ? bt_iter_skip_skip_internal : bt_iter_skip, &arg); if (level_too_large) *level_too_large = too_large; return arg.btobj; } MJIT_FUNC_EXPORTED VALUE rb_ec_backtrace_object(const rb_execution_context_t *ec) { return rb_ec_partial_backtrace_object(ec, BACKTRACE_START, ALL_BACKTRACE_LINES, NULL, FALSE); } static VALUE backtrace_collect(rb_backtrace_t *bt, VALUE (*func)(rb_backtrace_location_t *, void *arg), void *arg) { VALUE btary; int i; btary = rb_ary_new2(bt->backtrace_size-1); for (i=0; ibacktrace_size-1; i++) { rb_backtrace_location_t *loc = &bt->backtrace[bt->backtrace_size - 2 - i]; rb_ary_push(btary, func(loc, arg)); } return btary; } static VALUE location_to_str_dmyarg(rb_backtrace_location_t *loc, void *dmy) { return location_to_str(loc); } static VALUE backtrace_to_str_ary(VALUE self) { VALUE r; rb_backtrace_t *bt; GetCoreDataFromValue(self, rb_backtrace_t, bt); r = backtrace_collect(bt, location_to_str_dmyarg, 0); RB_GC_GUARD(self); return r; } VALUE rb_backtrace_to_str_ary(VALUE self) { rb_backtrace_t *bt; GetCoreDataFromValue(self, rb_backtrace_t, bt); if (!bt->strary) { bt->strary = backtrace_to_str_ary(self); } return bt->strary; } MJIT_FUNC_EXPORTED void rb_backtrace_use_iseq_first_lineno_for_last_location(VALUE self) { const rb_backtrace_t *bt; const rb_iseq_t *iseq; rb_backtrace_location_t *loc; GetCoreDataFromValue(self, rb_backtrace_t, bt); VM_ASSERT(bt->backtrace_size > 1); loc = &bt->backtrace[bt->backtrace_size - 2]; iseq = loc->body.iseq.iseq; VM_ASSERT(loc->type == LOCATION_TYPE_ISEQ); loc->body.iseq.lineno.lineno = FIX2INT(iseq->body->location.first_lineno); loc->type = LOCATION_TYPE_ISEQ_CALCED; } static VALUE location_create(rb_backtrace_location_t *srcloc, void *btobj) { VALUE obj; struct valued_frame_info *vloc; obj = TypedData_Make_Struct(rb_cBacktraceLocation, struct valued_frame_info, &location_data_type, vloc); vloc->loc = srcloc; vloc->btobj = (VALUE)btobj; return obj; } static VALUE backtrace_to_location_ary(VALUE self) { VALUE r; rb_backtrace_t *bt; GetCoreDataFromValue(self, rb_backtrace_t, bt); r = backtrace_collect(bt, location_create, (void *)self); RB_GC_GUARD(self); return r; } VALUE rb_backtrace_to_location_ary(VALUE self) { rb_backtrace_t *bt; GetCoreDataFromValue(self, rb_backtrace_t, bt); if (!bt->locary) { bt->locary = backtrace_to_location_ary(self); } return bt->locary; } static VALUE backtrace_dump_data(VALUE self) { VALUE str = rb_backtrace_to_str_ary(self); return str; } static VALUE backtrace_load_data(VALUE self, VALUE str) { rb_backtrace_t *bt; GetCoreDataFromValue(self, rb_backtrace_t, bt); bt->strary = str; return self; } static VALUE backtrace_limit(VALUE self) { return LONG2NUM(rb_backtrace_length_limit); } VALUE rb_ec_backtrace_str_ary(const rb_execution_context_t *ec, long lev, long n) { return backtrace_to_str_ary(rb_ec_partial_backtrace_object(ec, lev, n, NULL, FALSE)); } VALUE rb_ec_backtrace_location_ary(const rb_execution_context_t *ec, long lev, long n, bool skip_internal) { return backtrace_to_location_ary(rb_ec_partial_backtrace_object(ec, lev, n, NULL, skip_internal)); } /* make old style backtrace directly */ struct oldbt_arg { VALUE filename; int lineno; void (*func)(void *data, VALUE file, int lineno, VALUE name); void *data; /* result */ }; static void oldbt_init(void *ptr, size_t dmy) { struct oldbt_arg *arg = (struct oldbt_arg *)ptr; arg->filename = GET_VM()->progname; arg->lineno = 0; } static void oldbt_iter_iseq(void *ptr, const rb_control_frame_t *cfp) { const rb_iseq_t *iseq = cfp->iseq; const VALUE *pc = cfp->pc; struct oldbt_arg *arg = (struct oldbt_arg *)ptr; VALUE file = arg->filename = rb_iseq_path(iseq); VALUE name = iseq->body->location.label; int lineno = arg->lineno = calc_lineno(iseq, pc); (arg->func)(arg->data, file, lineno, name); } static void oldbt_iter_cfunc(void *ptr, const rb_control_frame_t *cfp, ID mid) { struct oldbt_arg *arg = (struct oldbt_arg *)ptr; VALUE file = arg->filename; VALUE name = rb_id2str(mid); int lineno = arg->lineno; (arg->func)(arg->data, file, lineno, name); } static void oldbt_print(void *data, VALUE file, int lineno, VALUE name) { FILE *fp = (FILE *)data; if (NIL_P(name)) { fprintf(fp, "\tfrom %s:%d:in unknown method\n", RSTRING_PTR(file), lineno); } else { fprintf(fp, "\tfrom %s:%d:in `%s'\n", RSTRING_PTR(file), lineno, RSTRING_PTR(name)); } } static void vm_backtrace_print(FILE *fp) { struct oldbt_arg arg; arg.func = oldbt_print; arg.data = (void *)fp; backtrace_each(GET_EC(), BACKTRACE_START, ALL_BACKTRACE_LINES, oldbt_init, oldbt_iter_iseq, oldbt_iter_cfunc, NULL, &arg); } static void oldbt_bugreport(void *arg, VALUE file, int line, VALUE method) { const char *filename = NIL_P(file) ? "ruby" : RSTRING_PTR(file); if (!*(int *)arg) { fprintf(stderr, "-- Ruby level backtrace information " "----------------------------------------\n"); *(int *)arg = 1; } if (NIL_P(method)) { fprintf(stderr, "%s:%d:in unknown method\n", filename, line); } else { fprintf(stderr, "%s:%d:in `%s'\n", filename, line, RSTRING_PTR(method)); } } void rb_backtrace_print_as_bugreport(void) { struct oldbt_arg arg; int i = 0; arg.func = oldbt_bugreport; arg.data = (int *)&i; backtrace_each(GET_EC(), BACKTRACE_START, ALL_BACKTRACE_LINES, oldbt_init, oldbt_iter_iseq, oldbt_iter_cfunc, NULL, &arg); } void rb_backtrace(void) { vm_backtrace_print(stderr); } struct print_to_arg { VALUE (*iter)(VALUE recv, VALUE str); VALUE output; }; static void oldbt_print_to(void *data, VALUE file, int lineno, VALUE name) { const struct print_to_arg *arg = data; VALUE str = rb_sprintf("\tfrom %"PRIsVALUE":%d:in ", file, lineno); if (NIL_P(name)) { rb_str_cat2(str, "unknown method\n"); } else { rb_str_catf(str, " `%"PRIsVALUE"'\n", name); } (*arg->iter)(arg->output, str); } void rb_backtrace_each(VALUE (*iter)(VALUE recv, VALUE str), VALUE output) { struct oldbt_arg arg; struct print_to_arg parg; parg.iter = iter; parg.output = output; arg.func = oldbt_print_to; arg.data = &parg; backtrace_each(GET_EC(), BACKTRACE_START, ALL_BACKTRACE_LINES, oldbt_init, oldbt_iter_iseq, oldbt_iter_cfunc, NULL, &arg); } VALUE rb_make_backtrace(void) { return rb_ec_backtrace_str_ary(GET_EC(), BACKTRACE_START, ALL_BACKTRACE_LINES); } static VALUE ec_backtrace_to_ary(const rb_execution_context_t *ec, int argc, const VALUE *argv, int lev_default, int lev_plus, int to_str) { VALUE level, vn; long lev, n; VALUE btval; VALUE r; int too_large; rb_scan_args(argc, argv, "02", &level, &vn); if (argc == 2 && NIL_P(vn)) argc--; switch (argc) { case 0: lev = lev_default + lev_plus; n = ALL_BACKTRACE_LINES; break; case 1: { long beg, len, bt_size = backtrace_size(ec); switch (rb_range_beg_len(level, &beg, &len, bt_size - lev_plus, 0)) { case Qfalse: lev = NUM2LONG(level); if (lev < 0) { rb_raise(rb_eArgError, "negative level (%ld)", lev); } lev += lev_plus; n = ALL_BACKTRACE_LINES; break; case Qnil: return Qnil; default: lev = beg + lev_plus; n = len; break; } break; } case 2: lev = NUM2LONG(level); n = NUM2LONG(vn); if (lev < 0) { rb_raise(rb_eArgError, "negative level (%ld)", lev); } if (n < 0) { rb_raise(rb_eArgError, "negative size (%ld)", n); } lev += lev_plus; break; default: lev = n = 0; /* to avoid warning */ break; } if (n == 0) { return rb_ary_new(); } btval = rb_ec_partial_backtrace_object(ec, lev, n, &too_large, FALSE); if (too_large) { return Qnil; } if (to_str) { r = backtrace_to_str_ary(btval); } else { r = backtrace_to_location_ary(btval); } RB_GC_GUARD(btval); return r; } static VALUE thread_backtrace_to_ary(int argc, const VALUE *argv, VALUE thval, int to_str) { rb_thread_t *target_th = rb_thread_ptr(thval); if (target_th->to_kill || target_th->status == THREAD_KILLED) return Qnil; return ec_backtrace_to_ary(target_th->ec, argc, argv, 0, 0, to_str); } VALUE rb_vm_thread_backtrace(int argc, const VALUE *argv, VALUE thval) { return thread_backtrace_to_ary(argc, argv, thval, 1); } VALUE rb_vm_thread_backtrace_locations(int argc, const VALUE *argv, VALUE thval) { return thread_backtrace_to_ary(argc, argv, thval, 0); } VALUE rb_vm_backtrace(int argc, const VALUE * argv, struct rb_execution_context_struct * ec) { return ec_backtrace_to_ary(ec, argc, argv, 0, 0, 1); } VALUE rb_vm_backtrace_locations(int argc, const VALUE * argv, struct rb_execution_context_struct * ec) { return ec_backtrace_to_ary(ec, argc, argv, 0, 0, 0); } /* * call-seq: * caller(start=1, length=nil) -> array or nil * caller(range) -> array or nil * * Returns the current execution stack---an array containing strings in * the form file:line or file:line: in * `method'. * * The optional _start_ parameter determines the number of initial stack * entries to omit from the top of the stack. * * A second optional +length+ parameter can be used to limit how many entries * are returned from the stack. * * Returns +nil+ if _start_ is greater than the size of * current execution stack. * * Optionally you can pass a range, which will return an array containing the * entries within the specified range. * * def a(skip) * caller(skip) * end * def b(skip) * a(skip) * end * def c(skip) * b(skip) * end * c(0) #=> ["prog:2:in `a'", "prog:5:in `b'", "prog:8:in `c'", "prog:10:in `
'"] * c(1) #=> ["prog:5:in `b'", "prog:8:in `c'", "prog:11:in `
'"] * c(2) #=> ["prog:8:in `c'", "prog:12:in `
'"] * c(3) #=> ["prog:13:in `
'"] * c(4) #=> [] * c(5) #=> nil */ static VALUE rb_f_caller(int argc, VALUE *argv, VALUE _) { return ec_backtrace_to_ary(GET_EC(), argc, argv, 1, 1, 1); } /* * call-seq: * caller_locations(start=1, length=nil) -> array or nil * caller_locations(range) -> array or nil * * Returns the current execution stack---an array containing * backtrace location objects. * * See Thread::Backtrace::Location for more information. * * The optional _start_ parameter determines the number of initial stack * entries to omit from the top of the stack. * * A second optional +length+ parameter can be used to limit how many entries * are returned from the stack. * * Returns +nil+ if _start_ is greater than the size of * current execution stack. * * Optionally you can pass a range, which will return an array containing the * entries within the specified range. */ static VALUE rb_f_caller_locations(int argc, VALUE *argv, VALUE _) { return ec_backtrace_to_ary(GET_EC(), argc, argv, 1, 1, 0); } /* called from Init_vm() in vm.c */ void Init_vm_backtrace(void) { /* :nodoc: */ rb_cBacktrace = rb_define_class_under(rb_cThread, "Backtrace", rb_cObject); rb_define_alloc_func(rb_cBacktrace, backtrace_alloc); rb_undef_method(CLASS_OF(rb_cBacktrace), "new"); rb_marshal_define_compat(rb_cBacktrace, rb_cArray, backtrace_dump_data, backtrace_load_data); rb_define_singleton_method(rb_cBacktrace, "limit", backtrace_limit, 0); /* * An object representation of a stack frame, initialized by * Kernel#caller_locations. * * For example: * * # caller_locations.rb * def a(skip) * caller_locations(skip) * end * def b(skip) * a(skip) * end * def c(skip) * b(skip) * end * * c(0..2).map do |call| * puts call.to_s * end * * Running ruby caller_locations.rb will produce: * * caller_locations.rb:2:in `a' * caller_locations.rb:5:in `b' * caller_locations.rb:8:in `c' * * Here's another example with a slightly different result: * * # foo.rb * class Foo * attr_accessor :locations * def initialize(skip) * @locations = caller_locations(skip) * end * end * * Foo.new(0..2).locations.map do |call| * puts call.to_s * end * * Now run ruby foo.rb and you should see: * * init.rb:4:in `initialize' * init.rb:8:in `new' * init.rb:8:in `
' */ rb_cBacktraceLocation = rb_define_class_under(rb_cBacktrace, "Location", rb_cObject); rb_undef_alloc_func(rb_cBacktraceLocation); rb_undef_method(CLASS_OF(rb_cBacktraceLocation), "new"); rb_define_method(rb_cBacktraceLocation, "lineno", location_lineno_m, 0); rb_define_method(rb_cBacktraceLocation, "label", location_label_m, 0); rb_define_method(rb_cBacktraceLocation, "base_label", location_base_label_m, 0); rb_define_method(rb_cBacktraceLocation, "path", location_path_m, 0); rb_define_method(rb_cBacktraceLocation, "absolute_path", location_absolute_path_m, 0); rb_define_method(rb_cBacktraceLocation, "to_s", location_to_str_m, 0); rb_define_method(rb_cBacktraceLocation, "inspect", location_inspect_m, 0); rb_define_global_function("caller", rb_f_caller, -1); rb_define_global_function("caller_locations", rb_f_caller_locations, -1); } /* debugger API */ RUBY_SYMBOL_EXPORT_BEGIN RUBY_SYMBOL_EXPORT_END struct rb_debug_inspector_struct { rb_execution_context_t *ec; rb_control_frame_t *cfp; VALUE backtrace; VALUE contexts; /* [[klass, binding, iseq, cfp], ...] */ long backtrace_size; }; enum { CALLER_BINDING_SELF, CALLER_BINDING_CLASS, CALLER_BINDING_BINDING, CALLER_BINDING_ISEQ, CALLER_BINDING_CFP }; struct collect_caller_bindings_data { VALUE ary; }; static void collect_caller_bindings_init(void *arg, size_t size) { /* */ } static VALUE get_klass(const rb_control_frame_t *cfp) { VALUE klass; if (rb_vm_control_frame_id_and_class(cfp, 0, 0, &klass)) { if (RB_TYPE_P(klass, T_ICLASS)) { return RBASIC(klass)->klass; } else { return klass; } } else { return Qnil; } } static void collect_caller_bindings_iseq(void *arg, const rb_control_frame_t *cfp) { struct collect_caller_bindings_data *data = (struct collect_caller_bindings_data *)arg; VALUE frame = rb_ary_new2(5); rb_ary_store(frame, CALLER_BINDING_SELF, cfp->self); rb_ary_store(frame, CALLER_BINDING_CLASS, get_klass(cfp)); rb_ary_store(frame, CALLER_BINDING_BINDING, GC_GUARDED_PTR(cfp)); /* create later */ rb_ary_store(frame, CALLER_BINDING_ISEQ, cfp->iseq ? (VALUE)cfp->iseq : Qnil); rb_ary_store(frame, CALLER_BINDING_CFP, GC_GUARDED_PTR(cfp)); rb_ary_push(data->ary, frame); } static void collect_caller_bindings_cfunc(void *arg, const rb_control_frame_t *cfp, ID mid) { struct collect_caller_bindings_data *data = (struct collect_caller_bindings_data *)arg; VALUE frame = rb_ary_new2(5); rb_ary_store(frame, CALLER_BINDING_SELF, cfp->self); rb_ary_store(frame, CALLER_BINDING_CLASS, get_klass(cfp)); rb_ary_store(frame, CALLER_BINDING_BINDING, Qnil); /* not available */ rb_ary_store(frame, CALLER_BINDING_ISEQ, Qnil); /* not available */ rb_ary_store(frame, CALLER_BINDING_CFP, GC_GUARDED_PTR(cfp)); rb_ary_push(data->ary, frame); } static VALUE collect_caller_bindings(const rb_execution_context_t *ec) { struct collect_caller_bindings_data data; VALUE result; int i; data.ary = rb_ary_new(); backtrace_each(ec, BACKTRACE_START, ALL_BACKTRACE_LINES, collect_caller_bindings_init, collect_caller_bindings_iseq, collect_caller_bindings_cfunc, NULL, &data); result = rb_ary_reverse(data.ary); /* bindings should be created from top of frame */ for (i=0; icfp; dbg_context.backtrace = rb_ec_backtrace_location_ary(ec, BACKTRACE_START, ALL_BACKTRACE_LINES, FALSE); dbg_context.backtrace_size = RARRAY_LEN(dbg_context.backtrace); dbg_context.contexts = collect_caller_bindings(ec); EC_PUSH_TAG(ec); if ((state = EC_EXEC_TAG()) == TAG_NONE) { result = (*func)(&dbg_context, data); } EC_POP_TAG(); /* invalidate bindings? */ if (state) { EC_JUMP_TAG(ec, state); } return result; } static VALUE frame_get(const rb_debug_inspector_t *dc, long index) { if (index < 0 || index >= dc->backtrace_size) { rb_raise(rb_eArgError, "no such frame"); } return rb_ary_entry(dc->contexts, index); } VALUE rb_debug_inspector_frame_self_get(const rb_debug_inspector_t *dc, long index) { VALUE frame = frame_get(dc, index); return rb_ary_entry(frame, CALLER_BINDING_SELF); } VALUE rb_debug_inspector_frame_class_get(const rb_debug_inspector_t *dc, long index) { VALUE frame = frame_get(dc, index); return rb_ary_entry(frame, CALLER_BINDING_CLASS); } VALUE rb_debug_inspector_frame_binding_get(const rb_debug_inspector_t *dc, long index) { VALUE frame = frame_get(dc, index); return rb_ary_entry(frame, CALLER_BINDING_BINDING); } VALUE rb_debug_inspector_frame_iseq_get(const rb_debug_inspector_t *dc, long index) { VALUE frame = frame_get(dc, index); VALUE iseq = rb_ary_entry(frame, CALLER_BINDING_ISEQ); return RTEST(iseq) ? rb_iseqw_new((rb_iseq_t *)iseq) : Qnil; } VALUE rb_debug_inspector_backtrace_locations(const rb_debug_inspector_t *dc) { return dc->backtrace; } int rb_profile_frames(int start, int limit, VALUE *buff, int *lines) { int i; const rb_execution_context_t *ec = GET_EC(); const rb_control_frame_t *cfp = ec->cfp, *end_cfp = RUBY_VM_END_CONTROL_FRAME(ec); const rb_callable_method_entry_t *cme; for (i=0; i 0) { start--; continue; } /* record frame info */ cme = rb_vm_frame_method_entry(cfp); if (cme && cme->def->type == VM_METHOD_TYPE_ISEQ) { buff[i] = (VALUE)cme; } else { buff[i] = (VALUE)cfp->iseq; } if (lines) lines[i] = calc_lineno(cfp->iseq, cfp->pc); i++; } else { cme = rb_vm_frame_method_entry(cfp); if (cme && cme->def->type == VM_METHOD_TYPE_CFUNC) { buff[i] = (VALUE)cme; if (lines) lines[i] = 0; i++; } } cfp = RUBY_VM_PREVIOUS_CONTROL_FRAME(cfp); } return i; } static const rb_iseq_t * frame2iseq(VALUE frame) { if (frame == Qnil) return NULL; if (RB_TYPE_P(frame, T_IMEMO)) { switch (imemo_type(frame)) { case imemo_iseq: return (const rb_iseq_t *)frame; case imemo_ment: { const rb_callable_method_entry_t *cme = (rb_callable_method_entry_t *)frame; switch (cme->def->type) { case VM_METHOD_TYPE_ISEQ: return cme->def->body.iseq.iseqptr; default: return NULL; } } default: break; } } rb_bug("frame2iseq: unreachable"); } VALUE rb_profile_frame_path(VALUE frame) { const rb_iseq_t *iseq = frame2iseq(frame); return iseq ? rb_iseq_path(iseq) : Qnil; } static const rb_callable_method_entry_t * cframe(VALUE frame) { if (frame == Qnil) return NULL; if (RB_TYPE_P(frame, T_IMEMO)) { switch (imemo_type(frame)) { case imemo_ment: { const rb_callable_method_entry_t *cme = (rb_callable_method_entry_t *)frame; switch (cme->def->type) { case VM_METHOD_TYPE_CFUNC: return cme; default: return NULL; } } default: return NULL; } } return NULL; } VALUE rb_profile_frame_absolute_path(VALUE frame) { if (cframe(frame)) { static VALUE cfunc_str = Qfalse; if (!cfunc_str) { cfunc_str = rb_str_new_literal(""); rb_gc_register_mark_object(cfunc_str); } return cfunc_str; } const rb_iseq_t *iseq = frame2iseq(frame); return iseq ? rb_iseq_realpath(iseq) : Qnil; } VALUE rb_profile_frame_label(VALUE frame) { const rb_iseq_t *iseq = frame2iseq(frame); return iseq ? rb_iseq_label(iseq) : Qnil; } VALUE rb_profile_frame_base_label(VALUE frame) { const rb_iseq_t *iseq = frame2iseq(frame); return iseq ? rb_iseq_base_label(iseq) : Qnil; } VALUE rb_profile_frame_first_lineno(VALUE frame) { const rb_iseq_t *iseq = frame2iseq(frame); return iseq ? rb_iseq_first_lineno(iseq) : Qnil; } static VALUE frame2klass(VALUE frame) { if (frame == Qnil) return Qnil; if (RB_TYPE_P(frame, T_IMEMO)) { const rb_callable_method_entry_t *cme = (rb_callable_method_entry_t *)frame; if (imemo_type(frame) == imemo_ment) { return cme->defined_class; } } return Qnil; } VALUE rb_profile_frame_classpath(VALUE frame) { VALUE klass = frame2klass(frame); if (klass && !NIL_P(klass)) { if (RB_TYPE_P(klass, T_ICLASS)) { klass = RBASIC(klass)->klass; } else if (FL_TEST(klass, FL_SINGLETON)) { klass = rb_ivar_get(klass, id__attached__); if (!RB_TYPE_P(klass, T_CLASS) && !RB_TYPE_P(klass, T_MODULE)) return rb_sprintf("#<%s:%p>", rb_class2name(rb_obj_class(klass)), (void*)klass); } return rb_class_path(klass); } else { return Qnil; } } VALUE rb_profile_frame_singleton_method_p(VALUE frame) { VALUE klass = frame2klass(frame); if (klass && !NIL_P(klass) && FL_TEST(klass, FL_SINGLETON)) { return Qtrue; } else { return Qfalse; } } VALUE rb_profile_frame_method_name(VALUE frame) { const rb_callable_method_entry_t *cme = cframe(frame); if (cme) { ID mid = cme->def->original_id; return id2str(mid); } const rb_iseq_t *iseq = frame2iseq(frame); return iseq ? rb_iseq_method_name(iseq) : Qnil; } static VALUE qualified_method_name(VALUE frame, VALUE method_name) { if (method_name != Qnil) { VALUE classpath = rb_profile_frame_classpath(frame); VALUE singleton_p = rb_profile_frame_singleton_method_p(frame); if (classpath != Qnil) { return rb_sprintf("%"PRIsVALUE"%s%"PRIsVALUE, classpath, singleton_p == Qtrue ? "." : "#", method_name); } else { return method_name; } } else { return Qnil; } } VALUE rb_profile_frame_qualified_method_name(VALUE frame) { VALUE method_name = rb_profile_frame_method_name(frame); return qualified_method_name(frame, method_name); } VALUE rb_profile_frame_full_label(VALUE frame) { const rb_callable_method_entry_t *cme = cframe(frame); if (cme) { ID mid = cme->def->original_id; VALUE method_name = id2str(mid); return qualified_method_name(frame, method_name); } VALUE label = rb_profile_frame_label(frame); VALUE base_label = rb_profile_frame_base_label(frame); VALUE qualified_method_name = rb_profile_frame_qualified_method_name(frame); if (NIL_P(qualified_method_name) || base_label == qualified_method_name) { return label; } else { long label_length = RSTRING_LEN(label); long base_label_length = RSTRING_LEN(base_label); int prefix_len = rb_long2int(label_length - base_label_length); return rb_sprintf("%.*s%"PRIsVALUE, prefix_len, RSTRING_PTR(label), qualified_method_name); } }