/********************************************************************** vm_trace.c - $Author: ko1 $ created at: Tue Aug 14 19:37:09 2012 Copyright (C) 1993-2012 Yukihiro Matsumoto **********************************************************************/ /* * This file include two parts: * * (1) set_trace_func internal mechanisms * and C level API * * (2) Ruby level API * (2-1) set_trace_func API * (2-2) TracePoint API (not yet) * */ #include "eval_intern.h" #include "internal.h" #include "internal/bits.h" #include "internal/class.h" #include "internal/gc.h" #include "internal/hash.h" #include "internal/symbol.h" #include "internal/thread.h" #include "iseq.h" #include "rjit.h" #include "ruby/atomic.h" #include "ruby/debug.h" #include "vm_core.h" #include "ruby/ractor.h" #include "yjit.h" #include "builtin.h" static VALUE sym_default; /* (1) trace mechanisms */ typedef struct rb_event_hook_struct { rb_event_hook_flag_t hook_flags; rb_event_flag_t events; rb_event_hook_func_t func; VALUE data; struct rb_event_hook_struct *next; struct { rb_thread_t *th; unsigned int target_line; } filter; } rb_event_hook_t; typedef void (*rb_event_hook_raw_arg_func_t)(VALUE data, const rb_trace_arg_t *arg); #define MAX_EVENT_NUM 32 void rb_hook_list_mark(rb_hook_list_t *hooks) { rb_event_hook_t *hook = hooks->hooks; while (hook) { rb_gc_mark(hook->data); hook = hook->next; } } void rb_hook_list_mark_and_update(rb_hook_list_t *hooks) { rb_event_hook_t *hook = hooks->hooks; while (hook) { rb_gc_mark_and_move(&hook->data); hook = hook->next; } } static void clean_hooks(rb_hook_list_t *list); void rb_hook_list_free(rb_hook_list_t *hooks) { hooks->need_clean = true; if (hooks->running == 0) { clean_hooks(hooks); } } /* ruby_vm_event_flags management */ void rb_clear_attr_ccs(void); void rb_clear_bf_ccs(void); static void update_global_event_hook(rb_event_flag_t prev_events, rb_event_flag_t new_events) { rb_event_flag_t new_iseq_events = new_events & ISEQ_TRACE_EVENTS; rb_event_flag_t enabled_iseq_events = ruby_vm_event_enabled_global_flags & ISEQ_TRACE_EVENTS; bool first_time_iseq_events_p = new_iseq_events & ~enabled_iseq_events; bool enable_c_call = (prev_events & RUBY_EVENT_C_CALL) == 0 && (new_events & RUBY_EVENT_C_CALL); bool enable_c_return = (prev_events & RUBY_EVENT_C_RETURN) == 0 && (new_events & RUBY_EVENT_C_RETURN); bool enable_call = (prev_events & RUBY_EVENT_CALL) == 0 && (new_events & RUBY_EVENT_CALL); bool enable_return = (prev_events & RUBY_EVENT_RETURN) == 0 && (new_events & RUBY_EVENT_RETURN); // Modify ISEQs or CCs to enable tracing if (first_time_iseq_events_p) { // write all ISeqs only when new events are added for the first time rb_iseq_trace_set_all(new_iseq_events | enabled_iseq_events); } // if c_call or c_return is activated else if (enable_c_call || enable_c_return) { rb_clear_attr_ccs(); } else if (enable_call || enable_return) { rb_clear_bf_ccs(); } ruby_vm_event_flags = new_events; ruby_vm_event_enabled_global_flags |= new_events; rb_objspace_set_event_hook(new_events); // Invalidate JIT code as needed if (first_time_iseq_events_p || enable_c_call || enable_c_return) { // Invalidate all code when ISEQs are modified to use trace_* insns above. // Also invalidate when enabling c_call or c_return because generated code // never fires these events. // Internal events fire inside C routines so don't need special handling. // Do this after event flags updates so other ractors see updated vm events // when they wake up. rb_yjit_tracing_invalidate_all(); rb_rjit_tracing_invalidate_all(new_iseq_events); } } /* add/remove hooks */ static rb_event_hook_t * alloc_event_hook(rb_event_hook_func_t func, rb_event_flag_t events, VALUE data, rb_event_hook_flag_t hook_flags) { rb_event_hook_t *hook; if ((events & RUBY_INTERNAL_EVENT_MASK) && (events & ~RUBY_INTERNAL_EVENT_MASK)) { rb_raise(rb_eTypeError, "Can not specify normal event and internal event simultaneously."); } hook = ALLOC(rb_event_hook_t); hook->hook_flags = hook_flags; hook->events = events; hook->func = func; hook->data = data; /* no filters */ hook->filter.th = NULL; hook->filter.target_line = 0; return hook; } static void hook_list_connect(VALUE list_owner, rb_hook_list_t *list, rb_event_hook_t *hook, int global_p) { rb_event_flag_t prev_events = list->events; hook->next = list->hooks; list->hooks = hook; list->events |= hook->events; if (global_p) { /* global hooks are root objects at GC mark. */ update_global_event_hook(prev_events, list->events); } else { RB_OBJ_WRITTEN(list_owner, Qundef, hook->data); } } static void connect_event_hook(const rb_execution_context_t *ec, rb_event_hook_t *hook) { rb_hook_list_t *list = rb_ec_ractor_hooks(ec); hook_list_connect(Qundef, list, hook, TRUE); } static void rb_threadptr_add_event_hook(const rb_execution_context_t *ec, rb_thread_t *th, rb_event_hook_func_t func, rb_event_flag_t events, VALUE data, rb_event_hook_flag_t hook_flags) { rb_event_hook_t *hook = alloc_event_hook(func, events, data, hook_flags); hook->filter.th = th; connect_event_hook(ec, hook); } void rb_thread_add_event_hook(VALUE thval, rb_event_hook_func_t func, rb_event_flag_t events, VALUE data) { rb_threadptr_add_event_hook(GET_EC(), rb_thread_ptr(thval), func, events, data, RUBY_EVENT_HOOK_FLAG_SAFE); } void rb_add_event_hook(rb_event_hook_func_t func, rb_event_flag_t events, VALUE data) { rb_add_event_hook2(func, events, data, RUBY_EVENT_HOOK_FLAG_SAFE); } void rb_thread_add_event_hook2(VALUE thval, rb_event_hook_func_t func, rb_event_flag_t events, VALUE data, rb_event_hook_flag_t hook_flags) { rb_threadptr_add_event_hook(GET_EC(), rb_thread_ptr(thval), func, events, data, hook_flags); } void rb_add_event_hook2(rb_event_hook_func_t func, rb_event_flag_t events, VALUE data, rb_event_hook_flag_t hook_flags) { rb_event_hook_t *hook = alloc_event_hook(func, events, data, hook_flags); connect_event_hook(GET_EC(), hook); } static void clean_hooks(rb_hook_list_t *list) { rb_event_hook_t *hook, **nextp = &list->hooks; rb_event_flag_t prev_events = list->events; VM_ASSERT(list->running == 0); VM_ASSERT(list->need_clean == true); list->events = 0; list->need_clean = false; while ((hook = *nextp) != 0) { if (hook->hook_flags & RUBY_EVENT_HOOK_FLAG_DELETED) { *nextp = hook->next; xfree(hook); } else { list->events |= hook->events; /* update active events */ nextp = &hook->next; } } if (list->is_local) { if (list->events == 0) { /* local events */ ruby_xfree(list); } } else { update_global_event_hook(prev_events, list->events); } } static void clean_hooks_check(rb_hook_list_t *list) { if (UNLIKELY(list->need_clean)) { if (list->running == 0) { clean_hooks(list); } } } #define MATCH_ANY_FILTER_TH ((rb_thread_t *)1) /* if func is 0, then clear all funcs */ static int remove_event_hook(const rb_execution_context_t *ec, const rb_thread_t *filter_th, rb_event_hook_func_t func, VALUE data) { rb_hook_list_t *list = rb_ec_ractor_hooks(ec); int ret = 0; rb_event_hook_t *hook = list->hooks; while (hook) { if (func == 0 || hook->func == func) { if (hook->filter.th == filter_th || filter_th == MATCH_ANY_FILTER_TH) { if (UNDEF_P(data) || hook->data == data) { hook->hook_flags |= RUBY_EVENT_HOOK_FLAG_DELETED; ret+=1; list->need_clean = true; } } } hook = hook->next; } clean_hooks_check(list); return ret; } static int rb_threadptr_remove_event_hook(const rb_execution_context_t *ec, const rb_thread_t *filter_th, rb_event_hook_func_t func, VALUE data) { return remove_event_hook(ec, filter_th, func, data); } int rb_thread_remove_event_hook(VALUE thval, rb_event_hook_func_t func) { return rb_threadptr_remove_event_hook(GET_EC(), rb_thread_ptr(thval), func, Qundef); } int rb_thread_remove_event_hook_with_data(VALUE thval, rb_event_hook_func_t func, VALUE data) { return rb_threadptr_remove_event_hook(GET_EC(), rb_thread_ptr(thval), func, data); } int rb_remove_event_hook(rb_event_hook_func_t func) { return remove_event_hook(GET_EC(), NULL, func, Qundef); } int rb_remove_event_hook_with_data(rb_event_hook_func_t func, VALUE data) { return remove_event_hook(GET_EC(), NULL, func, data); } void rb_ec_clear_current_thread_trace_func(const rb_execution_context_t *ec) { rb_threadptr_remove_event_hook(ec, rb_ec_thread_ptr(ec), 0, Qundef); } void rb_ec_clear_all_trace_func(const rb_execution_context_t *ec) { rb_threadptr_remove_event_hook(ec, MATCH_ANY_FILTER_TH, 0, Qundef); } /* invoke hooks */ static void exec_hooks_body(const rb_execution_context_t *ec, rb_hook_list_t *list, const rb_trace_arg_t *trace_arg) { rb_event_hook_t *hook; for (hook = list->hooks; hook; hook = hook->next) { if (!(hook->hook_flags & RUBY_EVENT_HOOK_FLAG_DELETED) && (trace_arg->event & hook->events) && (LIKELY(hook->filter.th == 0) || hook->filter.th == rb_ec_thread_ptr(ec)) && (LIKELY(hook->filter.target_line == 0) || (hook->filter.target_line == (unsigned int)rb_vm_get_sourceline(ec->cfp)))) { if (!(hook->hook_flags & RUBY_EVENT_HOOK_FLAG_RAW_ARG)) { (*hook->func)(trace_arg->event, hook->data, trace_arg->self, trace_arg->id, trace_arg->klass); } else { (*((rb_event_hook_raw_arg_func_t)hook->func))(hook->data, trace_arg); } } } } static int exec_hooks_precheck(const rb_execution_context_t *ec, rb_hook_list_t *list, const rb_trace_arg_t *trace_arg) { if (list->events & trace_arg->event) { list->running++; return TRUE; } else { return FALSE; } } static void exec_hooks_postcheck(const rb_execution_context_t *ec, rb_hook_list_t *list) { list->running--; clean_hooks_check(list); } static void exec_hooks_unprotected(const rb_execution_context_t *ec, rb_hook_list_t *list, const rb_trace_arg_t *trace_arg) { if (exec_hooks_precheck(ec, list, trace_arg) == 0) return; exec_hooks_body(ec, list, trace_arg); exec_hooks_postcheck(ec, list); } static int exec_hooks_protected(rb_execution_context_t *ec, rb_hook_list_t *list, const rb_trace_arg_t *trace_arg) { enum ruby_tag_type state; volatile int raised; if (exec_hooks_precheck(ec, list, trace_arg) == 0) return 0; raised = rb_ec_reset_raised(ec); /* TODO: Support !RUBY_EVENT_HOOK_FLAG_SAFE hooks */ EC_PUSH_TAG(ec); if ((state = EC_EXEC_TAG()) == TAG_NONE) { exec_hooks_body(ec, list, trace_arg); } EC_POP_TAG(); exec_hooks_postcheck(ec, list); if (raised) { rb_ec_set_raised(ec); } return state; } // pop_p: Whether to pop the frame for the TracePoint when it throws. void rb_exec_event_hooks(rb_trace_arg_t *trace_arg, rb_hook_list_t *hooks, int pop_p) { rb_execution_context_t *ec = trace_arg->ec; if (UNLIKELY(trace_arg->event & RUBY_INTERNAL_EVENT_MASK)) { if (ec->trace_arg && (ec->trace_arg->event & RUBY_INTERNAL_EVENT_MASK)) { /* skip hooks because this thread doing INTERNAL_EVENT */ } else { rb_trace_arg_t *prev_trace_arg = ec->trace_arg; ec->trace_arg = trace_arg; /* only global hooks */ exec_hooks_unprotected(ec, rb_ec_ractor_hooks(ec), trace_arg); ec->trace_arg = prev_trace_arg; } } else { if (ec->trace_arg == NULL && /* check reentrant */ trace_arg->self != rb_mRubyVMFrozenCore /* skip special methods. TODO: remove it. */) { const VALUE errinfo = ec->errinfo; const VALUE old_recursive = ec->local_storage_recursive_hash; int state = 0; /* setup */ ec->local_storage_recursive_hash = ec->local_storage_recursive_hash_for_trace; ec->errinfo = Qnil; ec->trace_arg = trace_arg; /* kick hooks */ if ((state = exec_hooks_protected(ec, hooks, trace_arg)) == TAG_NONE) { ec->errinfo = errinfo; } /* cleanup */ ec->trace_arg = NULL; ec->local_storage_recursive_hash_for_trace = ec->local_storage_recursive_hash; ec->local_storage_recursive_hash = old_recursive; if (state) { if (pop_p) { if (VM_FRAME_FINISHED_P(ec->cfp)) { ec->tag = ec->tag->prev; } rb_vm_pop_frame(ec); } EC_JUMP_TAG(ec, state); } } } } VALUE rb_suppress_tracing(VALUE (*func)(VALUE), VALUE arg) { volatile int raised; volatile VALUE result = Qnil; rb_execution_context_t *const ec = GET_EC(); rb_vm_t *const vm = rb_ec_vm_ptr(ec); enum ruby_tag_type state; rb_trace_arg_t dummy_trace_arg; dummy_trace_arg.event = 0; if (!ec->trace_arg) { ec->trace_arg = &dummy_trace_arg; } raised = rb_ec_reset_raised(ec); EC_PUSH_TAG(ec); if (LIKELY((state = EC_EXEC_TAG()) == TAG_NONE)) { result = (*func)(arg); } else { (void)*&vm; /* suppress "clobbered" warning */ } EC_POP_TAG(); if (raised) { rb_ec_reset_raised(ec); } if (ec->trace_arg == &dummy_trace_arg) { ec->trace_arg = NULL; } if (state) { #if defined RUBY_USE_SETJMPEX && RUBY_USE_SETJMPEX RB_GC_GUARD(result); #endif EC_JUMP_TAG(ec, state); } return result; } static void call_trace_func(rb_event_flag_t, VALUE data, VALUE self, ID id, VALUE klass); /* (2-1) set_trace_func (old API) */ /* * call-seq: * set_trace_func(proc) -> proc * set_trace_func(nil) -> nil * * Establishes _proc_ as the handler for tracing, or disables * tracing if the parameter is +nil+. * * *Note:* this method is obsolete, please use TracePoint instead. * * _proc_ takes up to six parameters: * * * an event name string * * a filename string * * a line number * * a method name symbol, or nil * * a binding, or nil * * the class, module, or nil * * _proc_ is invoked whenever an event occurs. * * Events are: * * "c-call":: call a C-language routine * "c-return":: return from a C-language routine * "call":: call a Ruby method * "class":: start a class or module definition * "end":: finish a class or module definition * "line":: execute code on a new line * "raise":: raise an exception * "return":: return from a Ruby method * * Tracing is disabled within the context of _proc_. * * class Test * def test * a = 1 * b = 2 * end * end * * set_trace_func proc { |event, file, line, id, binding, class_or_module| * printf "%8s %s:%-2d %16p %14p\n", event, file, line, id, class_or_module * } * t = Test.new * t.test * * Produces: * * c-return prog.rb:8 :set_trace_func Kernel * line prog.rb:11 nil nil * c-call prog.rb:11 :new Class * c-call prog.rb:11 :initialize BasicObject * c-return prog.rb:11 :initialize BasicObject * c-return prog.rb:11 :new Class * line prog.rb:12 nil nil * call prog.rb:2 :test Test * line prog.rb:3 :test Test * line prog.rb:4 :test Test * return prog.rb:5 :test Test */ static VALUE set_trace_func(VALUE obj, VALUE trace) { rb_remove_event_hook(call_trace_func); if (NIL_P(trace)) { return Qnil; } if (!rb_obj_is_proc(trace)) { rb_raise(rb_eTypeError, "trace_func needs to be Proc"); } rb_add_event_hook(call_trace_func, RUBY_EVENT_ALL, trace); return trace; } static void thread_add_trace_func(rb_execution_context_t *ec, rb_thread_t *filter_th, VALUE trace) { if (!rb_obj_is_proc(trace)) { rb_raise(rb_eTypeError, "trace_func needs to be Proc"); } rb_threadptr_add_event_hook(ec, filter_th, call_trace_func, RUBY_EVENT_ALL, trace, RUBY_EVENT_HOOK_FLAG_SAFE); } /* * call-seq: * thr.add_trace_func(proc) -> proc * * Adds _proc_ as a handler for tracing. * * See Thread#set_trace_func and Kernel#set_trace_func. */ static VALUE thread_add_trace_func_m(VALUE obj, VALUE trace) { thread_add_trace_func(GET_EC(), rb_thread_ptr(obj), trace); return trace; } /* * call-seq: * thr.set_trace_func(proc) -> proc * thr.set_trace_func(nil) -> nil * * Establishes _proc_ on _thr_ as the handler for tracing, or * disables tracing if the parameter is +nil+. * * See Kernel#set_trace_func. */ static VALUE thread_set_trace_func_m(VALUE target_thread, VALUE trace) { rb_execution_context_t *ec = GET_EC(); rb_thread_t *target_th = rb_thread_ptr(target_thread); rb_threadptr_remove_event_hook(ec, target_th, call_trace_func, Qundef); if (NIL_P(trace)) { return Qnil; } else { thread_add_trace_func(ec, target_th, trace); return trace; } } static const char * get_event_name(rb_event_flag_t event) { switch (event) { case RUBY_EVENT_LINE: return "line"; case RUBY_EVENT_CLASS: return "class"; case RUBY_EVENT_END: return "end"; case RUBY_EVENT_CALL: return "call"; case RUBY_EVENT_RETURN: return "return"; case RUBY_EVENT_C_CALL: return "c-call"; case RUBY_EVENT_C_RETURN: return "c-return"; case RUBY_EVENT_RAISE: return "raise"; default: return "unknown"; } } static ID get_event_id(rb_event_flag_t event) { ID id; switch (event) { #define C(name, NAME) case RUBY_EVENT_##NAME: CONST_ID(id, #name); return id; C(line, LINE); C(class, CLASS); C(end, END); C(call, CALL); C(return, RETURN); C(c_call, C_CALL); C(c_return, C_RETURN); C(raise, RAISE); C(b_call, B_CALL); C(b_return, B_RETURN); C(thread_begin, THREAD_BEGIN); C(thread_end, THREAD_END); C(fiber_switch, FIBER_SWITCH); C(script_compiled, SCRIPT_COMPILED); C(rescue, RESCUE); #undef C default: return 0; } } static void get_path_and_lineno(const rb_execution_context_t *ec, const rb_control_frame_t *cfp, rb_event_flag_t event, VALUE *pathp, int *linep) { cfp = rb_vm_get_ruby_level_next_cfp(ec, cfp); if (cfp) { const rb_iseq_t *iseq = cfp->iseq; *pathp = rb_iseq_path(iseq); if (event & (RUBY_EVENT_CLASS | RUBY_EVENT_CALL | RUBY_EVENT_B_CALL)) { *linep = FIX2INT(rb_iseq_first_lineno(iseq)); } else { *linep = rb_vm_get_sourceline(cfp); } } else { *pathp = Qnil; *linep = 0; } } static void call_trace_func(rb_event_flag_t event, VALUE proc, VALUE self, ID id, VALUE klass) { int line; VALUE filename; VALUE eventname = rb_str_new2(get_event_name(event)); VALUE argv[6]; const rb_execution_context_t *ec = GET_EC(); get_path_and_lineno(ec, ec->cfp, event, &filename, &line); if (!klass) { rb_ec_frame_method_id_and_class(ec, &id, 0, &klass); } if (klass) { if (RB_TYPE_P(klass, T_ICLASS)) { klass = RBASIC(klass)->klass; } else if (FL_TEST(klass, FL_SINGLETON)) { klass = RCLASS_ATTACHED_OBJECT(klass); } } argv[0] = eventname; argv[1] = filename; argv[2] = INT2FIX(line); argv[3] = id ? ID2SYM(id) : Qnil; argv[4] = Qnil; if (self && (filename != Qnil) && event != RUBY_EVENT_C_CALL && event != RUBY_EVENT_C_RETURN && (VM_FRAME_RUBYFRAME_P(ec->cfp) && imemo_type_p((VALUE)ec->cfp->iseq, imemo_iseq))) { argv[4] = rb_binding_new(); } argv[5] = klass ? klass : Qnil; rb_proc_call_with_block(proc, 6, argv, Qnil); } /* (2-2) TracePoint API */ static VALUE rb_cTracePoint; typedef struct rb_tp_struct { rb_event_flag_t events; int tracing; /* bool */ rb_thread_t *target_th; VALUE local_target_set; /* Hash: target -> * Qtrue (if target is iseq) or * Qfalse (if target is bmethod) */ void (*func)(VALUE tpval, void *data); void *data; VALUE proc; rb_ractor_t *ractor; VALUE self; } rb_tp_t; static void tp_mark(void *ptr) { rb_tp_t *tp = ptr; rb_gc_mark(tp->proc); rb_gc_mark(tp->local_target_set); if (tp->target_th) rb_gc_mark(tp->target_th->self); } static const rb_data_type_t tp_data_type = { "tracepoint", { tp_mark, RUBY_TYPED_DEFAULT_FREE, NULL, // Nothing allocated externally, so don't need a memsize function }, 0, 0, RUBY_TYPED_FREE_IMMEDIATELY | RUBY_TYPED_WB_PROTECTED | RUBY_TYPED_EMBEDDABLE }; static VALUE tp_alloc(VALUE klass) { rb_tp_t *tp; return TypedData_Make_Struct(klass, rb_tp_t, &tp_data_type, tp); } static rb_event_flag_t symbol2event_flag(VALUE v) { ID id; VALUE sym = rb_to_symbol_type(v); const rb_event_flag_t RUBY_EVENT_A_CALL = RUBY_EVENT_CALL | RUBY_EVENT_B_CALL | RUBY_EVENT_C_CALL; const rb_event_flag_t RUBY_EVENT_A_RETURN = RUBY_EVENT_RETURN | RUBY_EVENT_B_RETURN | RUBY_EVENT_C_RETURN; #define C(name, NAME) CONST_ID(id, #name); if (sym == ID2SYM(id)) return RUBY_EVENT_##NAME C(line, LINE); C(class, CLASS); C(end, END); C(call, CALL); C(return, RETURN); C(c_call, C_CALL); C(c_return, C_RETURN); C(raise, RAISE); C(b_call, B_CALL); C(b_return, B_RETURN); C(thread_begin, THREAD_BEGIN); C(thread_end, THREAD_END); C(fiber_switch, FIBER_SWITCH); C(script_compiled, SCRIPT_COMPILED); C(rescue, RESCUE); /* joke */ C(a_call, A_CALL); C(a_return, A_RETURN); #undef C rb_raise(rb_eArgError, "unknown event: %"PRIsVALUE, rb_sym2str(sym)); } static rb_tp_t * tpptr(VALUE tpval) { rb_tp_t *tp; TypedData_Get_Struct(tpval, rb_tp_t, &tp_data_type, tp); return tp; } static rb_trace_arg_t * get_trace_arg(void) { rb_trace_arg_t *trace_arg = GET_EC()->trace_arg; if (trace_arg == 0) { rb_raise(rb_eRuntimeError, "access from outside"); } return trace_arg; } struct rb_trace_arg_struct * rb_tracearg_from_tracepoint(VALUE tpval) { return get_trace_arg(); } rb_event_flag_t rb_tracearg_event_flag(rb_trace_arg_t *trace_arg) { return trace_arg->event; } VALUE rb_tracearg_event(rb_trace_arg_t *trace_arg) { return ID2SYM(get_event_id(trace_arg->event)); } static void fill_path_and_lineno(rb_trace_arg_t *trace_arg) { if (UNDEF_P(trace_arg->path)) { get_path_and_lineno(trace_arg->ec, trace_arg->cfp, trace_arg->event, &trace_arg->path, &trace_arg->lineno); } } VALUE rb_tracearg_lineno(rb_trace_arg_t *trace_arg) { fill_path_and_lineno(trace_arg); return INT2FIX(trace_arg->lineno); } VALUE rb_tracearg_path(rb_trace_arg_t *trace_arg) { fill_path_and_lineno(trace_arg); return trace_arg->path; } static void fill_id_and_klass(rb_trace_arg_t *trace_arg) { if (!trace_arg->klass_solved) { if (!trace_arg->klass) { rb_vm_control_frame_id_and_class(trace_arg->cfp, &trace_arg->id, &trace_arg->called_id, &trace_arg->klass); } if (trace_arg->klass) { if (RB_TYPE_P(trace_arg->klass, T_ICLASS)) { trace_arg->klass = RBASIC(trace_arg->klass)->klass; } } else { trace_arg->klass = Qnil; } trace_arg->klass_solved = 1; } } VALUE rb_tracearg_parameters(rb_trace_arg_t *trace_arg) { switch (trace_arg->event) { case RUBY_EVENT_CALL: case RUBY_EVENT_RETURN: case RUBY_EVENT_B_CALL: case RUBY_EVENT_B_RETURN: { const rb_control_frame_t *cfp = rb_vm_get_ruby_level_next_cfp(trace_arg->ec, trace_arg->cfp); if (cfp) { int is_proc = 0; if (VM_FRAME_TYPE(cfp) == VM_FRAME_MAGIC_BLOCK && !VM_FRAME_LAMBDA_P(cfp)) { is_proc = 1; } return rb_iseq_parameters(cfp->iseq, is_proc); } break; } case RUBY_EVENT_C_CALL: case RUBY_EVENT_C_RETURN: { fill_id_and_klass(trace_arg); if (trace_arg->klass && trace_arg->id) { const rb_method_entry_t *me; VALUE iclass = Qnil; me = rb_method_entry_without_refinements(trace_arg->klass, trace_arg->called_id, &iclass); return rb_unnamed_parameters(rb_method_entry_arity(me)); } break; } case RUBY_EVENT_RAISE: case RUBY_EVENT_LINE: case RUBY_EVENT_CLASS: case RUBY_EVENT_END: case RUBY_EVENT_SCRIPT_COMPILED: case RUBY_EVENT_RESCUE: rb_raise(rb_eRuntimeError, "not supported by this event"); break; } return Qnil; } VALUE rb_tracearg_method_id(rb_trace_arg_t *trace_arg) { fill_id_and_klass(trace_arg); return trace_arg->id ? ID2SYM(trace_arg->id) : Qnil; } VALUE rb_tracearg_callee_id(rb_trace_arg_t *trace_arg) { fill_id_and_klass(trace_arg); return trace_arg->called_id ? ID2SYM(trace_arg->called_id) : Qnil; } VALUE rb_tracearg_defined_class(rb_trace_arg_t *trace_arg) { fill_id_and_klass(trace_arg); return trace_arg->klass; } VALUE rb_tracearg_binding(rb_trace_arg_t *trace_arg) { rb_control_frame_t *cfp; switch (trace_arg->event) { case RUBY_EVENT_C_CALL: case RUBY_EVENT_C_RETURN: return Qnil; } cfp = rb_vm_get_binding_creatable_next_cfp(trace_arg->ec, trace_arg->cfp); if (cfp && imemo_type_p((VALUE)cfp->iseq, imemo_iseq)) { return rb_vm_make_binding(trace_arg->ec, cfp); } else { return Qnil; } } VALUE rb_tracearg_self(rb_trace_arg_t *trace_arg) { return trace_arg->self; } VALUE rb_tracearg_return_value(rb_trace_arg_t *trace_arg) { if (trace_arg->event & (RUBY_EVENT_RETURN | RUBY_EVENT_C_RETURN | RUBY_EVENT_B_RETURN)) { /* ok */ } else { rb_raise(rb_eRuntimeError, "not supported by this event"); } if (UNDEF_P(trace_arg->data)) { rb_bug("rb_tracearg_return_value: unreachable"); } return trace_arg->data; } VALUE rb_tracearg_raised_exception(rb_trace_arg_t *trace_arg) { if (trace_arg->event & (RUBY_EVENT_RAISE | RUBY_EVENT_RESCUE)) { /* ok */ } else { rb_raise(rb_eRuntimeError, "not supported by this event"); } if (UNDEF_P(trace_arg->data)) { rb_bug("rb_tracearg_raised_exception: unreachable"); } return trace_arg->data; } VALUE rb_tracearg_eval_script(rb_trace_arg_t *trace_arg) { VALUE data = trace_arg->data; if (trace_arg->event & (RUBY_EVENT_SCRIPT_COMPILED)) { /* ok */ } else { rb_raise(rb_eRuntimeError, "not supported by this event"); } if (UNDEF_P(data)) { rb_bug("rb_tracearg_raised_exception: unreachable"); } if (rb_obj_is_iseq(data)) { return Qnil; } else { VM_ASSERT(RB_TYPE_P(data, T_ARRAY)); /* [src, iseq] */ return RARRAY_AREF(data, 0); } } VALUE rb_tracearg_instruction_sequence(rb_trace_arg_t *trace_arg) { VALUE data = trace_arg->data; if (trace_arg->event & (RUBY_EVENT_SCRIPT_COMPILED)) { /* ok */ } else { rb_raise(rb_eRuntimeError, "not supported by this event"); } if (UNDEF_P(data)) { rb_bug("rb_tracearg_raised_exception: unreachable"); } if (rb_obj_is_iseq(data)) { return rb_iseqw_new((const rb_iseq_t *)data); } else { VM_ASSERT(RB_TYPE_P(data, T_ARRAY)); VM_ASSERT(rb_obj_is_iseq(RARRAY_AREF(data, 1))); /* [src, iseq] */ return rb_iseqw_new((const rb_iseq_t *)RARRAY_AREF(data, 1)); } } VALUE rb_tracearg_object(rb_trace_arg_t *trace_arg) { if (trace_arg->event & (RUBY_INTERNAL_EVENT_NEWOBJ | RUBY_INTERNAL_EVENT_FREEOBJ)) { /* ok */ } else { rb_raise(rb_eRuntimeError, "not supported by this event"); } if (UNDEF_P(trace_arg->data)) { rb_bug("rb_tracearg_object: unreachable"); } return trace_arg->data; } static VALUE tracepoint_attr_event(rb_execution_context_t *ec, VALUE tpval) { return rb_tracearg_event(get_trace_arg()); } static VALUE tracepoint_attr_lineno(rb_execution_context_t *ec, VALUE tpval) { return rb_tracearg_lineno(get_trace_arg()); } static VALUE tracepoint_attr_path(rb_execution_context_t *ec, VALUE tpval) { return rb_tracearg_path(get_trace_arg()); } static VALUE tracepoint_attr_parameters(rb_execution_context_t *ec, VALUE tpval) { return rb_tracearg_parameters(get_trace_arg()); } static VALUE tracepoint_attr_method_id(rb_execution_context_t *ec, VALUE tpval) { return rb_tracearg_method_id(get_trace_arg()); } static VALUE tracepoint_attr_callee_id(rb_execution_context_t *ec, VALUE tpval) { return rb_tracearg_callee_id(get_trace_arg()); } static VALUE tracepoint_attr_defined_class(rb_execution_context_t *ec, VALUE tpval) { return rb_tracearg_defined_class(get_trace_arg()); } static VALUE tracepoint_attr_binding(rb_execution_context_t *ec, VALUE tpval) { return rb_tracearg_binding(get_trace_arg()); } static VALUE tracepoint_attr_self(rb_execution_context_t *ec, VALUE tpval) { return rb_tracearg_self(get_trace_arg()); } static VALUE tracepoint_attr_return_value(rb_execution_context_t *ec, VALUE tpval) { return rb_tracearg_return_value(get_trace_arg()); } static VALUE tracepoint_attr_raised_exception(rb_execution_context_t *ec, VALUE tpval) { return rb_tracearg_raised_exception(get_trace_arg()); } static VALUE tracepoint_attr_eval_script(rb_execution_context_t *ec, VALUE tpval) { return rb_tracearg_eval_script(get_trace_arg()); } static VALUE tracepoint_attr_instruction_sequence(rb_execution_context_t *ec, VALUE tpval) { return rb_tracearg_instruction_sequence(get_trace_arg()); } static void tp_call_trace(VALUE tpval, rb_trace_arg_t *trace_arg) { rb_tp_t *tp = tpptr(tpval); if (tp->func) { (*tp->func)(tpval, tp->data); } else { if (tp->ractor == NULL || tp->ractor == GET_RACTOR()) { rb_proc_call_with_block((VALUE)tp->proc, 1, &tpval, Qnil); } } } VALUE rb_tracepoint_enable(VALUE tpval) { rb_tp_t *tp; tp = tpptr(tpval); if (tp->local_target_set != Qfalse) { rb_raise(rb_eArgError, "can't nest-enable a targeting TracePoint"); } if (tp->tracing) { return Qundef; } if (tp->target_th) { rb_thread_add_event_hook2(tp->target_th->self, (rb_event_hook_func_t)tp_call_trace, tp->events, tpval, RUBY_EVENT_HOOK_FLAG_SAFE | RUBY_EVENT_HOOK_FLAG_RAW_ARG); } else { rb_add_event_hook2((rb_event_hook_func_t)tp_call_trace, tp->events, tpval, RUBY_EVENT_HOOK_FLAG_SAFE | RUBY_EVENT_HOOK_FLAG_RAW_ARG); } tp->tracing = 1; return Qundef; } static const rb_iseq_t * iseq_of(VALUE target) { VALUE iseqv = rb_funcall(rb_cISeq, rb_intern("of"), 1, target); if (NIL_P(iseqv)) { rb_raise(rb_eArgError, "specified target is not supported"); } else { return rb_iseqw_to_iseq(iseqv); } } const rb_method_definition_t *rb_method_def(VALUE method); /* proc.c */ static VALUE rb_tracepoint_enable_for_target(VALUE tpval, VALUE target, VALUE target_line) { rb_tp_t *tp = tpptr(tpval); const rb_iseq_t *iseq = iseq_of(target); int n = 0; unsigned int line = 0; bool target_bmethod = false; if (tp->tracing > 0) { rb_raise(rb_eArgError, "can't nest-enable a targeting TracePoint"); } if (!NIL_P(target_line)) { if ((tp->events & RUBY_EVENT_LINE) == 0) { rb_raise(rb_eArgError, "target_line is specified, but line event is not specified"); } else { line = NUM2UINT(target_line); } } VM_ASSERT(tp->local_target_set == Qfalse); RB_OBJ_WRITE(tpval, &tp->local_target_set, rb_obj_hide(rb_ident_hash_new())); /* bmethod */ if (rb_obj_is_method(target)) { rb_method_definition_t *def = (rb_method_definition_t *)rb_method_def(target); if (def->type == VM_METHOD_TYPE_BMETHOD && (tp->events & (RUBY_EVENT_CALL | RUBY_EVENT_RETURN))) { if (def->body.bmethod.hooks == NULL) { def->body.bmethod.hooks = ZALLOC(rb_hook_list_t); } rb_hook_list_connect_tracepoint(target, def->body.bmethod.hooks, tpval, 0); rb_hash_aset(tp->local_target_set, target, Qfalse); target_bmethod = true; n++; } } /* iseq */ n += rb_iseq_add_local_tracepoint_recursively(iseq, tp->events, tpval, line, target_bmethod); rb_hash_aset(tp->local_target_set, (VALUE)iseq, Qtrue); if ((tp->events & (RUBY_EVENT_CALL | RUBY_EVENT_RETURN)) && iseq->body->builtin_attrs & BUILTIN_ATTR_SINGLE_NOARG_LEAF) { rb_clear_bf_ccs(); } if (n == 0) { rb_raise(rb_eArgError, "can not enable any hooks"); } rb_yjit_tracing_invalidate_all(); rb_rjit_tracing_invalidate_all(tp->events); ruby_vm_event_local_num++; tp->tracing = 1; return Qnil; } static int disable_local_event_iseq_i(VALUE target, VALUE iseq_p, VALUE tpval) { if (iseq_p) { rb_iseq_remove_local_tracepoint_recursively((rb_iseq_t *)target, tpval); } else { /* bmethod */ rb_method_definition_t *def = (rb_method_definition_t *)rb_method_def(target); rb_hook_list_t *hooks = def->body.bmethod.hooks; VM_ASSERT(hooks != NULL); rb_hook_list_remove_tracepoint(hooks, tpval); if (hooks->events == 0) { rb_hook_list_free(def->body.bmethod.hooks); def->body.bmethod.hooks = NULL; } } return ST_CONTINUE; } VALUE rb_tracepoint_disable(VALUE tpval) { rb_tp_t *tp; tp = tpptr(tpval); if (tp->local_target_set) { rb_hash_foreach(tp->local_target_set, disable_local_event_iseq_i, tpval); RB_OBJ_WRITE(tpval, &tp->local_target_set, Qfalse); ruby_vm_event_local_num--; } else { if (tp->target_th) { rb_thread_remove_event_hook_with_data(tp->target_th->self, (rb_event_hook_func_t)tp_call_trace, tpval); } else { rb_remove_event_hook_with_data((rb_event_hook_func_t)tp_call_trace, tpval); } } tp->tracing = 0; tp->target_th = NULL; return Qundef; } void rb_hook_list_connect_tracepoint(VALUE target, rb_hook_list_t *list, VALUE tpval, unsigned int target_line) { rb_tp_t *tp = tpptr(tpval); rb_event_hook_t *hook = alloc_event_hook((rb_event_hook_func_t)tp_call_trace, tp->events, tpval, RUBY_EVENT_HOOK_FLAG_SAFE | RUBY_EVENT_HOOK_FLAG_RAW_ARG); hook->filter.target_line = target_line; hook_list_connect(target, list, hook, FALSE); } void rb_hook_list_remove_tracepoint(rb_hook_list_t *list, VALUE tpval) { rb_event_hook_t *hook = list->hooks; rb_event_flag_t events = 0; while (hook) { if (hook->data == tpval) { hook->hook_flags |= RUBY_EVENT_HOOK_FLAG_DELETED; list->need_clean = true; } else if ((hook->hook_flags & RUBY_EVENT_HOOK_FLAG_DELETED) == 0) { events |= hook->events; } hook = hook->next; } list->events = events; } static VALUE tracepoint_enable_m(rb_execution_context_t *ec, VALUE tpval, VALUE target, VALUE target_line, VALUE target_thread) { rb_tp_t *tp = tpptr(tpval); int previous_tracing = tp->tracing; if (target_thread == sym_default) { if (rb_block_given_p() && NIL_P(target) && NIL_P(target_line)) { target_thread = rb_thread_current(); } else { target_thread = Qnil; } } /* check target_thread */ if (RTEST(target_thread)) { if (tp->target_th) { rb_raise(rb_eArgError, "can not override target_thread filter"); } tp->target_th = rb_thread_ptr(target_thread); RUBY_ASSERT(tp->target_th->self == target_thread); RB_OBJ_WRITTEN(tpval, Qundef, target_thread); } else { tp->target_th = NULL; } if (NIL_P(target)) { if (!NIL_P(target_line)) { rb_raise(rb_eArgError, "only target_line is specified"); } rb_tracepoint_enable(tpval); } else { rb_tracepoint_enable_for_target(tpval, target, target_line); } if (rb_block_given_p()) { return rb_ensure(rb_yield, Qundef, previous_tracing ? rb_tracepoint_enable : rb_tracepoint_disable, tpval); } else { return RBOOL(previous_tracing); } } static VALUE tracepoint_disable_m(rb_execution_context_t *ec, VALUE tpval) { rb_tp_t *tp = tpptr(tpval); int previous_tracing = tp->tracing; if (rb_block_given_p()) { if (tp->local_target_set != Qfalse) { rb_raise(rb_eArgError, "can't disable a targeting TracePoint in a block"); } rb_tracepoint_disable(tpval); return rb_ensure(rb_yield, Qundef, previous_tracing ? rb_tracepoint_enable : rb_tracepoint_disable, tpval); } else { rb_tracepoint_disable(tpval); return RBOOL(previous_tracing); } } VALUE rb_tracepoint_enabled_p(VALUE tpval) { rb_tp_t *tp = tpptr(tpval); return RBOOL(tp->tracing); } static VALUE tracepoint_enabled_p(rb_execution_context_t *ec, VALUE tpval) { return rb_tracepoint_enabled_p(tpval); } static VALUE tracepoint_new(VALUE klass, rb_thread_t *target_th, rb_event_flag_t events, void (func)(VALUE, void*), void *data, VALUE proc) { VALUE tpval = tp_alloc(klass); rb_tp_t *tp; TypedData_Get_Struct(tpval, rb_tp_t, &tp_data_type, tp); RB_OBJ_WRITE(tpval, &tp->proc, proc); tp->ractor = rb_ractor_shareable_p(proc) ? NULL : GET_RACTOR(); tp->func = func; tp->data = data; tp->events = events; tp->self = tpval; return tpval; } VALUE rb_tracepoint_new(VALUE target_thval, rb_event_flag_t events, void (*func)(VALUE, void *), void *data) { rb_thread_t *target_th = NULL; if (RTEST(target_thval)) { target_th = rb_thread_ptr(target_thval); /* TODO: Test it! * Warning: This function is not tested. */ } return tracepoint_new(rb_cTracePoint, target_th, events, func, data, Qundef); } static VALUE tracepoint_new_s(rb_execution_context_t *ec, VALUE self, VALUE args) { rb_event_flag_t events = 0; long i; long argc = RARRAY_LEN(args); if (argc > 0) { for (i=0; itrace_arg; if (trace_arg) { switch (trace_arg->event) { case RUBY_EVENT_LINE: { VALUE sym = rb_tracearg_method_id(trace_arg); if (NIL_P(sym)) break; return rb_sprintf("#", rb_tracearg_event(trace_arg), rb_tracearg_path(trace_arg), FIX2INT(rb_tracearg_lineno(trace_arg)), sym); } case RUBY_EVENT_CALL: case RUBY_EVENT_C_CALL: case RUBY_EVENT_RETURN: case RUBY_EVENT_C_RETURN: return rb_sprintf("#", rb_tracearg_event(trace_arg), rb_tracearg_method_id(trace_arg), rb_tracearg_path(trace_arg), FIX2INT(rb_tracearg_lineno(trace_arg))); case RUBY_EVENT_THREAD_BEGIN: case RUBY_EVENT_THREAD_END: return rb_sprintf("#", rb_tracearg_event(trace_arg), rb_tracearg_self(trace_arg)); default: break; } return rb_sprintf("#", rb_tracearg_event(trace_arg), rb_tracearg_path(trace_arg), FIX2INT(rb_tracearg_lineno(trace_arg))); } else { return rb_sprintf("#", tp->tracing ? "enabled" : "disabled"); } } static void tracepoint_stat_event_hooks(VALUE hash, VALUE key, rb_event_hook_t *hook) { int active = 0, deleted = 0; while (hook) { if (hook->hook_flags & RUBY_EVENT_HOOK_FLAG_DELETED) { deleted++; } else { active++; } hook = hook->next; } rb_hash_aset(hash, key, rb_ary_new3(2, INT2FIX(active), INT2FIX(deleted))); } static VALUE tracepoint_stat_s(rb_execution_context_t *ec, VALUE self) { rb_vm_t *vm = GET_VM(); VALUE stat = rb_hash_new(); tracepoint_stat_event_hooks(stat, vm->self, rb_ec_ractor_hooks(ec)->hooks); /* TODO: thread local hooks */ return stat; } static VALUE disallow_reentry(VALUE val) { rb_trace_arg_t *arg = (rb_trace_arg_t *)val; rb_execution_context_t *ec = GET_EC(); if (ec->trace_arg != NULL) rb_bug("should be NULL, but %p", (void *)ec->trace_arg); ec->trace_arg = arg; return Qnil; } static VALUE tracepoint_allow_reentry(rb_execution_context_t *ec, VALUE self) { const rb_trace_arg_t *arg = ec->trace_arg; if (arg == NULL) rb_raise(rb_eRuntimeError, "No need to allow reentrance."); ec->trace_arg = NULL; return rb_ensure(rb_yield, Qnil, disallow_reentry, (VALUE)arg); } #include "trace_point.rbinc" /* This function is called from inits.c */ void Init_vm_trace(void) { sym_default = ID2SYM(rb_intern_const("default")); /* trace_func */ rb_define_global_function("set_trace_func", set_trace_func, 1); rb_define_method(rb_cThread, "set_trace_func", thread_set_trace_func_m, 1); rb_define_method(rb_cThread, "add_trace_func", thread_add_trace_func_m, 1); rb_cTracePoint = rb_define_class("TracePoint", rb_cObject); rb_undef_alloc_func(rb_cTracePoint); } /* * Ruby actually has two separate mechanisms for enqueueing work from contexts * where it is not safe to run Ruby code, to run later on when it is safe. One * is async-signal-safe but more limited, and accessed through the * `rb_postponed_job_preregister` and `rb_postponed_job_trigger` functions. The * other is more flexible but cannot be used in signal handlers, and is accessed * through the `rb_workqueue_register` function. * * The postponed job functions form part of Ruby's extension API, but the * workqueue functions are for internal use only. */ struct rb_workqueue_job { struct ccan_list_node jnode; /* <=> vm->workqueue */ rb_postponed_job_func_t func; void *data; }; // Used for VM memsize reporting. Returns the size of a list of rb_workqueue_job // structs. Defined here because the struct definition lives here as well. size_t rb_vm_memsize_workqueue(struct ccan_list_head *workqueue) { struct rb_workqueue_job *work = 0; size_t size = 0; ccan_list_for_each(workqueue, work, jnode) { size += sizeof(struct rb_workqueue_job); } return size; } /* * thread-safe and called from non-Ruby thread * returns FALSE on failure (ENOMEM), TRUE otherwise */ int rb_workqueue_register(unsigned flags, rb_postponed_job_func_t func, void *data) { struct rb_workqueue_job *wq_job = malloc(sizeof(*wq_job)); rb_vm_t *vm = GET_VM(); if (!wq_job) return FALSE; wq_job->func = func; wq_job->data = data; rb_nativethread_lock_lock(&vm->workqueue_lock); ccan_list_add_tail(&vm->workqueue, &wq_job->jnode); rb_nativethread_lock_unlock(&vm->workqueue_lock); // TODO: current implementation affects only main ractor RUBY_VM_SET_POSTPONED_JOB_INTERRUPT(rb_vm_main_ractor_ec(vm)); return TRUE; } #define PJOB_TABLE_SIZE (sizeof(rb_atomic_t) * CHAR_BIT) /* pre-registered jobs table, for async-safe jobs */ typedef struct rb_postponed_job_queue { struct { rb_postponed_job_func_t func; void *data; } table[PJOB_TABLE_SIZE]; /* Bits in this are set when the corresponding entry in prereg_table has non-zero * triggered_count; i.e. somebody called rb_postponed_job_trigger */ rb_atomic_t triggered_bitset; } rb_postponed_job_queues_t; void rb_vm_postponed_job_queue_init(rb_vm_t *vm) { /* use mimmalloc; postponed job registration is a dependency of objspace, so this gets * called _VERY_ early inside Init_BareVM */ rb_postponed_job_queues_t *pjq = ruby_mimmalloc(sizeof(rb_postponed_job_queues_t)); pjq->triggered_bitset = 0; memset(pjq->table, 0, sizeof(pjq->table)); vm->postponed_job_queue = pjq; } static rb_execution_context_t * get_valid_ec(rb_vm_t *vm) { rb_execution_context_t *ec = rb_current_execution_context(false); if (ec == NULL) ec = rb_vm_main_ractor_ec(vm); return ec; } void rb_vm_postponed_job_atfork(void) { rb_vm_t *vm = GET_VM(); rb_postponed_job_queues_t *pjq = vm->postponed_job_queue; /* make sure we set the interrupt flag on _this_ thread if we carried any pjobs over * from the other side of the fork */ if (pjq->triggered_bitset) { RUBY_VM_SET_POSTPONED_JOB_INTERRUPT(get_valid_ec(vm)); } } /* Frees the memory managed by the postponed job infrastructure at shutdown */ void rb_vm_postponed_job_free(void) { rb_vm_t *vm = GET_VM(); ruby_xfree(vm->postponed_job_queue); vm->postponed_job_queue = NULL; } // Used for VM memsize reporting. Returns the total size of the postponed job // queue infrastructure. size_t rb_vm_memsize_postponed_job_queue(void) { return sizeof(rb_postponed_job_queues_t); } rb_postponed_job_handle_t rb_postponed_job_preregister(unsigned int flags, rb_postponed_job_func_t func, void *data) { /* The doc comments say that this function should be called under the GVL, because * that is actually required to get the guarantee that "if a given (func, data) pair * was already pre-registered, this method will return the same handle instance". * * However, the actual implementation here is called without the GVL, from inside * rb_postponed_job_register, to support that legacy interface. In the presence * of concurrent calls to both _preregister and _register functions on the same * func, however, the data may get mixed up between them. */ rb_postponed_job_queues_t *pjq = GET_VM()->postponed_job_queue; for (unsigned int i = 0; i < PJOB_TABLE_SIZE; i++) { /* Try and set this slot to equal `func` */ rb_postponed_job_func_t existing_func = (rb_postponed_job_func_t)RUBY_ATOMIC_PTR_CAS(pjq->table[i], NULL, (void *)func); if (existing_func == NULL || existing_func == func) { /* Either this slot was NULL, and we set it to func, or, this slot was already equal to func. * In either case, clobber the data with our data. Note that concurrent calls to * rb_postponed_job_register with the same func & different data will result in either of the * datas being written */ RUBY_ATOMIC_PTR_EXCHANGE(pjq->table[i].data, data); return (rb_postponed_job_handle_t)i; } else { /* Try the next slot if this one already has a func in it */ continue; } } /* full */ return POSTPONED_JOB_HANDLE_INVALID; } void rb_postponed_job_trigger(rb_postponed_job_handle_t h) { rb_vm_t *vm = GET_VM(); rb_postponed_job_queues_t *pjq = vm->postponed_job_queue; RUBY_ATOMIC_OR(pjq->triggered_bitset, (((rb_atomic_t)1UL) << h)); RUBY_VM_SET_POSTPONED_JOB_INTERRUPT(get_valid_ec(vm)); } static int pjob_register_legacy_impl(unsigned int flags, rb_postponed_job_func_t func, void *data) { /* We _know_ calling preregister from a signal handler like this is racy; what is * and is not promised is very exhaustively documented in debug.h */ rb_postponed_job_handle_t h = rb_postponed_job_preregister(0, func, data); if (h == POSTPONED_JOB_HANDLE_INVALID) { return 0; } rb_postponed_job_trigger(h); return 1; } int rb_postponed_job_register(unsigned int flags, rb_postponed_job_func_t func, void *data) { return pjob_register_legacy_impl(flags, func, data); } int rb_postponed_job_register_one(unsigned int flags, rb_postponed_job_func_t func, void *data) { return pjob_register_legacy_impl(flags, func, data); } void rb_postponed_job_flush(rb_vm_t *vm) { rb_postponed_job_queues_t *pjq = GET_VM()->postponed_job_queue; rb_execution_context_t *ec = GET_EC(); const rb_atomic_t block_mask = POSTPONED_JOB_INTERRUPT_MASK | TRAP_INTERRUPT_MASK; volatile rb_atomic_t saved_mask = ec->interrupt_mask & block_mask; VALUE volatile saved_errno = ec->errinfo; struct ccan_list_head tmp; ccan_list_head_init(&tmp); rb_nativethread_lock_lock(&vm->workqueue_lock); ccan_list_append_list(&tmp, &vm->workqueue); rb_nativethread_lock_unlock(&vm->workqueue_lock); rb_atomic_t triggered_bits = RUBY_ATOMIC_EXCHANGE(pjq->triggered_bitset, 0); ec->errinfo = Qnil; /* mask POSTPONED_JOB dispatch */ ec->interrupt_mask |= block_mask; { EC_PUSH_TAG(ec); if (EC_EXEC_TAG() == TAG_NONE) { /* execute postponed jobs */ while (triggered_bits) { unsigned int i = bit_length(triggered_bits) - 1; triggered_bits ^= ((1UL) << i); /* toggle ith bit off */ rb_postponed_job_func_t func = pjq->table[i].func; void *data = pjq->table[i].data; (func)(data); } /* execute workqueue jobs */ struct rb_workqueue_job *wq_job; while ((wq_job = ccan_list_pop(&tmp, struct rb_workqueue_job, jnode))) { rb_postponed_job_func_t func = wq_job->func; void *data = wq_job->data; free(wq_job); (func)(data); } } EC_POP_TAG(); } /* restore POSTPONED_JOB mask */ ec->interrupt_mask &= ~(saved_mask ^ block_mask); ec->errinfo = saved_errno; /* If we threw an exception, there might be leftover workqueue items; carry them over * to a subsequent execution of flush */ if (!ccan_list_empty(&tmp)) { rb_nativethread_lock_lock(&vm->workqueue_lock); ccan_list_prepend_list(&vm->workqueue, &tmp); rb_nativethread_lock_unlock(&vm->workqueue_lock); RUBY_VM_SET_POSTPONED_JOB_INTERRUPT(GET_EC()); } /* likewise with any remaining-to-be-executed bits of the preregistered postponed * job table */ if (triggered_bits) { RUBY_ATOMIC_OR(pjq->triggered_bitset, triggered_bits); RUBY_VM_SET_POSTPONED_JOB_INTERRUPT(GET_EC()); } }