2006-12-31 18:02:22 +03:00
|
|
|
/**********************************************************************
|
|
|
|
|
|
|
|
insnhelper.h - helper macros to implement each instructions
|
|
|
|
|
|
|
|
$Author$
|
|
|
|
created at: 04/01/01 15:50:34 JST
|
|
|
|
|
* blockinlining.c, compile.c, compile.h, debug.c, debug.h,
id.c, insnhelper.h, insns.def, thread.c, thread_pthread.ci,
thread_pthread.h, thread_win32.ci, thread_win32.h, vm.h,
vm_dump.c, vm_evalbody.ci, vm_opts.h: fix comments and
copyright year.
git-svn-id: svn+ssh://ci.ruby-lang.org/ruby/trunk@13920 b2dd03c8-39d4-4d8f-98ff-823fe69b080e
2007-11-14 01:13:04 +03:00
|
|
|
Copyright (C) 2004-2007 Koichi Sasada
|
2006-12-31 18:02:22 +03:00
|
|
|
|
|
|
|
**********************************************************************/
|
|
|
|
|
2008-01-18 11:56:11 +03:00
|
|
|
#ifndef RUBY_INSNHELPER_H
|
|
|
|
#define RUBY_INSNHELPER_H
|
2006-12-31 18:02:22 +03:00
|
|
|
|
2017-10-21 09:22:43 +03:00
|
|
|
RUBY_SYMBOL_EXPORT_BEGIN
|
|
|
|
|
2018-08-07 19:27:45 +03:00
|
|
|
RUBY_EXTERN VALUE ruby_vm_const_missing_count;
|
|
|
|
RUBY_EXTERN rb_serial_t ruby_vm_global_method_state;
|
|
|
|
RUBY_EXTERN rb_serial_t ruby_vm_global_constant_state;
|
|
|
|
RUBY_EXTERN rb_serial_t ruby_vm_class_serial;
|
* common.mk: clean up
- remove blockinlining.$(OBJEXT) to built
- make ENCODING_H_INCLDUES variable (include/ruby/encoding.h)
- make VM_CORE_H_INCLUDES variable (vm_core.h)
- simplify rules.
- make depends rule to output depend status using gcc -MM.
* include/ruby/mvm.h, include/ruby/vm.h: rename mvm.h to vm.h.
* include/ruby.h: ditto.
* load.c: add inclusion explicitly.
* enumerator.c, object.c, parse.y, thread.c, vm_dump.c:
remove useless inclusion.
* eval_intern.h: cleanup inclusion.
* vm_core.h: rb_thread_t should be defined in this file.
* vm_evalbody.c, vm_exec.c: rename vm_evalbody.c to vm_exec.c.
* vm.h, vm_exec.h: rename vm.h to vm_exec.h.
* insnhelper.h, vm_insnhelper.h: rename insnhelper.h to vm_insnhelper.h.
* vm.c, vm_insnhelper.c, vm_insnhelper.h:
- rename vm_eval() to vm_exec_core().
- rename vm_eval_body() to vm_exec().
- cleanup include order.
* vm_method.c: fix comment.
git-svn-id: svn+ssh://ci.ruby-lang.org/ruby/trunk@19466 b2dd03c8-39d4-4d8f-98ff-823fe69b080e
2008-09-23 04:20:28 +04:00
|
|
|
|
2017-10-21 09:22:43 +03:00
|
|
|
RUBY_SYMBOL_EXPORT_END
|
|
|
|
|
2012-11-22 10:28:17 +04:00
|
|
|
#if VM_COLLECT_USAGE_DETAILS
|
2012-10-04 16:31:05 +04:00
|
|
|
#define COLLECT_USAGE_INSN(insn) vm_collect_usage_insn(insn)
|
|
|
|
#define COLLECT_USAGE_OPERAND(insn, n, op) vm_collect_usage_operand((insn), (n), ((VALUE)(op)))
|
* probes.d: add DTrace probe declarations. [ruby-core:27448]
* array.c (empty_ary_alloc, ary_new): added array create DTrace probe.
* compile.c (rb_insns_name): allowing DTrace probes to access
instruction sequence name.
* Makefile.in: translate probes.d file to appropriate header file.
* common.mk: declare dependencies on the DTrace header.
* configure.in: add a test for existence of DTrace.
* eval.c (setup_exception): add a probe for when an exception is
raised.
* gc.c: Add DTrace probes for mark begin and end, and sweep begin and
end.
* hash.c (empty_hash_alloc): Add a probe for hash allocation.
* insns.def: Add probes for function entry and return.
* internal.h: function declaration for compile.c change.
* load.c (rb_f_load): add probes for `load` entry and exit, require
entry and exit, and wrapping search_required for load path search.
* object.c (rb_obj_alloc): added a probe for general object creation.
* parse.y (yycompile0): added a probe around parse and compile phase.
* string.c (empty_str_alloc, str_new): DTrace probes for string
allocation.
* test/dtrace/*: tests for DTrace probes.
* vm.c (vm_invoke_proc): add probes for function return on exception
raise, hash create, and instruction sequence execution.
* vm_core.h: add probe declarations for function entry and exit.
* vm_dump.c: add probes header file.
* vm_eval.c (vm_call0_cfunc, vm_call0_cfunc_with_frame): add probe on
function entry and return.
* vm_exec.c: expose instruction number to instruction name function.
* vm_insnshelper.c: add function entry and exit probes for cfunc
methods.
* vm_insnhelper.h: vm usage information is always collected, so
uncomment the functions.
12 19:14:50 2012 Akinori MUSHA <knu@iDaemons.org>
* configure.in (isinf, isnan): isinf() and isnan() are macros on
DragonFly which cannot be found by AC_REPLACE_FUNCS(). This
workaround enforces the fact that they exist on DragonFly.
12 15:59:38 2012 Shugo Maeda <shugo@ruby-lang.org>
* vm_core.h (rb_call_info_t::refinements), compile.c (new_callinfo),
vm_insnhelper.c (vm_search_method): revert r37616 because it's too
slow. [ruby-dev:46477]
* test/ruby/test_refinement.rb (test_inline_method_cache): skip
the test until the bug is fixed efficiently.
git-svn-id: svn+ssh://ci.ruby-lang.org/ruby/trunk@37631 b2dd03c8-39d4-4d8f-98ff-823fe69b080e
2012-11-13 01:52:12 +04:00
|
|
|
|
2012-10-04 16:31:05 +04:00
|
|
|
#define COLLECT_USAGE_REGISTER(reg, s) vm_collect_usage_register((reg), (s))
|
|
|
|
#else
|
2012-11-22 10:28:17 +04:00
|
|
|
#define COLLECT_USAGE_INSN(insn) /* none */
|
|
|
|
#define COLLECT_USAGE_OPERAND(insn, n, op) /* none */
|
2012-10-04 16:31:05 +04:00
|
|
|
#define COLLECT_USAGE_REGISTER(reg, s) /* none */
|
|
|
|
#endif
|
2006-12-31 18:02:22 +03:00
|
|
|
|
|
|
|
/**********************************************************/
|
|
|
|
/* deal with stack */
|
|
|
|
/**********************************************************/
|
|
|
|
|
2018-06-27 13:36:49 +03:00
|
|
|
#define PUSH(x) (SET_SV(x), INC_SP(1))
|
2006-12-31 18:02:22 +03:00
|
|
|
#define TOPN(n) (*(GET_SP()-(n)-1))
|
2007-06-05 21:39:52 +04:00
|
|
|
#define POPN(n) (DEC_SP(n))
|
|
|
|
#define POP() (DEC_SP(1))
|
2007-06-01 08:05:46 +04:00
|
|
|
#define STACK_ADDR_FROM_TOP(n) (GET_SP()-(n))
|
2006-12-31 18:02:22 +03:00
|
|
|
|
|
|
|
/**********************************************************/
|
|
|
|
/* deal with registers */
|
|
|
|
/**********************************************************/
|
|
|
|
|
2016-11-05 19:31:25 +03:00
|
|
|
#define VM_REG_CFP (reg_cfp)
|
|
|
|
#define VM_REG_PC (VM_REG_CFP->pc)
|
|
|
|
#define VM_REG_SP (VM_REG_CFP->sp)
|
|
|
|
#define VM_REG_EP (VM_REG_CFP->ep)
|
2006-12-31 18:02:22 +03:00
|
|
|
|
2007-06-24 14:33:00 +04:00
|
|
|
#define RESTORE_REGS() do { \
|
2017-10-27 09:21:50 +03:00
|
|
|
VM_REG_CFP = ec->cfp; \
|
2007-06-24 14:33:00 +04:00
|
|
|
} while (0)
|
2006-12-31 18:02:22 +03:00
|
|
|
|
2018-12-28 04:06:04 +03:00
|
|
|
#if VM_COLLECT_USAGE_DETAILS
|
2012-06-11 07:14:59 +04:00
|
|
|
enum vm_regan_regtype {
|
|
|
|
VM_REGAN_PC = 0,
|
|
|
|
VM_REGAN_SP = 1,
|
|
|
|
VM_REGAN_EP = 2,
|
|
|
|
VM_REGAN_CFP = 3,
|
|
|
|
VM_REGAN_SELF = 4,
|
2018-01-02 09:41:46 +03:00
|
|
|
VM_REGAN_ISEQ = 5
|
2012-06-11 07:14:59 +04:00
|
|
|
};
|
|
|
|
enum vm_regan_acttype {
|
|
|
|
VM_REGAN_ACT_GET = 0,
|
2018-01-02 09:41:46 +03:00
|
|
|
VM_REGAN_ACT_SET = 1
|
2012-06-11 07:14:59 +04:00
|
|
|
};
|
|
|
|
|
2012-10-04 16:31:05 +04:00
|
|
|
#define COLLECT_USAGE_REGISTER_HELPER(a, b, v) \
|
|
|
|
(COLLECT_USAGE_REGISTER((VM_REGAN_##a), (VM_REGAN_ACT_##b)), (v))
|
2006-12-31 18:02:22 +03:00
|
|
|
#else
|
2012-10-04 16:31:05 +04:00
|
|
|
#define COLLECT_USAGE_REGISTER_HELPER(a, b, v) (v)
|
2006-12-31 18:02:22 +03:00
|
|
|
#endif
|
|
|
|
|
|
|
|
/* PC */
|
2016-11-05 19:31:25 +03:00
|
|
|
#define GET_PC() (COLLECT_USAGE_REGISTER_HELPER(PC, GET, VM_REG_PC))
|
|
|
|
#define SET_PC(x) (VM_REG_PC = (COLLECT_USAGE_REGISTER_HELPER(PC, SET, (x))))
|
2006-12-31 18:02:22 +03:00
|
|
|
#define GET_CURRENT_INSN() (*GET_PC())
|
|
|
|
#define GET_OPERAND(n) (GET_PC()[(n)])
|
2016-11-05 19:31:25 +03:00
|
|
|
#define ADD_PC(n) (SET_PC(VM_REG_PC + (n)))
|
2017-11-14 15:58:36 +03:00
|
|
|
#define JUMP(dst) (SET_PC(VM_REG_PC + (dst)))
|
2006-12-31 18:02:22 +03:00
|
|
|
|
2012-06-11 07:14:59 +04:00
|
|
|
/* frame pointer, environment pointer */
|
2016-11-05 19:31:25 +03:00
|
|
|
#define GET_CFP() (COLLECT_USAGE_REGISTER_HELPER(CFP, GET, VM_REG_CFP))
|
|
|
|
#define GET_EP() (COLLECT_USAGE_REGISTER_HELPER(EP, GET, VM_REG_EP))
|
|
|
|
#define SET_EP(x) (VM_REG_EP = (COLLECT_USAGE_REGISTER_HELPER(EP, SET, (x))))
|
2012-06-11 07:14:59 +04:00
|
|
|
#define GET_LEP() (VM_EP_LEP(GET_EP()))
|
2006-12-31 18:02:22 +03:00
|
|
|
|
|
|
|
/* SP */
|
2016-11-05 19:31:25 +03:00
|
|
|
#define GET_SP() (COLLECT_USAGE_REGISTER_HELPER(SP, GET, VM_REG_SP))
|
|
|
|
#define SET_SP(x) (VM_REG_SP = (COLLECT_USAGE_REGISTER_HELPER(SP, SET, (x))))
|
|
|
|
#define INC_SP(x) (VM_REG_SP += (COLLECT_USAGE_REGISTER_HELPER(SP, SET, (x))))
|
|
|
|
#define DEC_SP(x) (VM_REG_SP -= (COLLECT_USAGE_REGISTER_HELPER(SP, SET, (x))))
|
2006-12-31 18:02:22 +03:00
|
|
|
#define SET_SV(x) (*GET_SP() = (x))
|
2018-07-19 16:25:22 +03:00
|
|
|
/* set current stack value as x */
|
2006-12-31 18:02:22 +03:00
|
|
|
|
|
|
|
/* instruction sequence C struct */
|
|
|
|
#define GET_ISEQ() (GET_CFP()->iseq)
|
|
|
|
|
|
|
|
/**********************************************************/
|
|
|
|
/* deal with variables */
|
|
|
|
/**********************************************************/
|
|
|
|
|
2016-07-28 14:02:30 +03:00
|
|
|
#define GET_PREV_EP(ep) ((VALUE *)((ep)[VM_ENV_DATA_INDEX_SPECVAL] & ~0x03))
|
2006-12-31 18:02:22 +03:00
|
|
|
|
|
|
|
/**********************************************************/
|
|
|
|
/* deal with values */
|
|
|
|
/**********************************************************/
|
|
|
|
|
2012-10-04 16:31:05 +04:00
|
|
|
#define GET_SELF() (COLLECT_USAGE_REGISTER_HELPER(SELF, GET, GET_CFP()->self))
|
2006-12-31 18:02:22 +03:00
|
|
|
|
|
|
|
/**********************************************************/
|
|
|
|
/* deal with control flow 2: method/iterator */
|
|
|
|
/**********************************************************/
|
|
|
|
|
2012-10-15 00:59:21 +04:00
|
|
|
/* set fastpath when cached method is *NOT* protected
|
|
|
|
* because inline method cache does not care about receiver.
|
|
|
|
*/
|
2012-10-15 21:40:50 +04:00
|
|
|
|
2019-12-17 07:22:24 +03:00
|
|
|
static inline void
|
|
|
|
CC_SET_FASTPATH(CALL_CACHE cc, vm_call_handler func, bool enabled)
|
|
|
|
{
|
|
|
|
if (LIKELY(enabled)) {
|
|
|
|
cc->call = func;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void
|
|
|
|
CC_SET_ME(CALL_CACHE cc, const rb_callable_method_entry_t *me)
|
|
|
|
{
|
|
|
|
cc->me = me;
|
2019-12-17 09:49:41 +03:00
|
|
|
cc->method_serial = me ? me->def->method_serial : 0;
|
2019-12-17 07:22:24 +03:00
|
|
|
}
|
2019-12-16 11:38:41 +03:00
|
|
|
|
2016-07-28 14:02:30 +03:00
|
|
|
#define GET_BLOCK_HANDLER() (GET_LEP()[VM_ENV_DATA_INDEX_SPECVAL])
|
2006-12-31 18:02:22 +03:00
|
|
|
|
|
|
|
/**********************************************************/
|
|
|
|
/* deal with control flow 3: exception */
|
|
|
|
/**********************************************************/
|
|
|
|
|
|
|
|
|
2018-09-13 06:46:46 +03:00
|
|
|
/**********************************************************/
|
|
|
|
/* deal with stack canary */
|
|
|
|
/**********************************************************/
|
|
|
|
|
|
|
|
#if VM_CHECK_MODE > 0
|
|
|
|
#define SETUP_CANARY() \
|
2019-02-01 10:26:39 +03:00
|
|
|
VALUE *canary; \
|
2018-09-14 10:44:44 +03:00
|
|
|
if (leaf) { \
|
2018-09-13 06:46:46 +03:00
|
|
|
canary = GET_SP(); \
|
|
|
|
SET_SV(vm_stack_canary); \
|
2019-02-01 10:26:39 +03:00
|
|
|
} \
|
|
|
|
else {\
|
|
|
|
SET_SV(Qfalse); /* cleanup */ \
|
2018-09-13 06:46:46 +03:00
|
|
|
}
|
|
|
|
#define CHECK_CANARY() \
|
2019-02-01 10:26:39 +03:00
|
|
|
if (leaf) { \
|
|
|
|
if (*canary == vm_stack_canary) { \
|
|
|
|
*canary = Qfalse; /* cleanup */ \
|
|
|
|
} \
|
|
|
|
else { \
|
|
|
|
vm_canary_is_found_dead(INSN_ATTR(bin), *canary); \
|
|
|
|
} \
|
2018-09-13 06:46:46 +03:00
|
|
|
}
|
|
|
|
#else
|
|
|
|
#define SETUP_CANARY() /* void */
|
|
|
|
#define CHECK_CANARY() /* void */
|
|
|
|
#endif
|
|
|
|
|
2006-12-31 18:02:22 +03:00
|
|
|
/**********************************************************/
|
|
|
|
/* others */
|
|
|
|
/**********************************************************/
|
|
|
|
|
2018-09-14 10:44:44 +03:00
|
|
|
#ifndef MJIT_HEADER
|
|
|
|
#define CALL_SIMPLE_METHOD() do { \
|
|
|
|
rb_snum_t x = leaf ? INSN_ATTR(width) : 0; \
|
2019-07-31 04:36:05 +03:00
|
|
|
rb_snum_t y = attr_width_opt_send_without_block(0); \
|
2018-09-14 10:44:44 +03:00
|
|
|
rb_snum_t z = x - y; \
|
|
|
|
ADD_PC(z); \
|
|
|
|
DISPATCH_ORIGINAL_INSN(opt_send_without_block); \
|
|
|
|
} while (0)
|
|
|
|
#endif
|
|
|
|
|
2019-12-13 03:42:34 +03:00
|
|
|
#define PREV_CLASS_SERIAL() (ruby_vm_class_serial)
|
2013-11-09 07:34:49 +04:00
|
|
|
#define NEXT_CLASS_SERIAL() (++ruby_vm_class_serial)
|
2013-12-09 14:51:02 +04:00
|
|
|
#define GET_GLOBAL_METHOD_STATE() (ruby_vm_global_method_state)
|
|
|
|
#define INC_GLOBAL_METHOD_STATE() (++ruby_vm_global_method_state)
|
|
|
|
#define GET_GLOBAL_CONSTANT_STATE() (ruby_vm_global_constant_state)
|
|
|
|
#define INC_GLOBAL_CONSTANT_STATE() (++ruby_vm_global_constant_state)
|
2011-06-13 15:25:44 +04:00
|
|
|
|
2015-03-11 15:49:27 +03:00
|
|
|
static inline struct vm_throw_data *
|
2019-07-25 11:15:48 +03:00
|
|
|
THROW_DATA_NEW(VALUE val, const rb_control_frame_t *cf, int st)
|
2015-03-10 21:39:46 +03:00
|
|
|
{
|
2019-07-25 11:15:48 +03:00
|
|
|
struct vm_throw_data *obj = (struct vm_throw_data *)rb_imemo_new(imemo_throw_data, val, (VALUE)cf, 0, 0);
|
|
|
|
obj->throw_state = st;
|
|
|
|
return obj;
|
2015-03-10 21:39:46 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
static inline VALUE
|
2015-03-11 15:49:27 +03:00
|
|
|
THROW_DATA_VAL(const struct vm_throw_data *obj)
|
2015-03-10 21:39:46 +03:00
|
|
|
{
|
2017-04-07 10:50:30 +03:00
|
|
|
VM_ASSERT(THROW_DATA_P(obj));
|
2015-03-10 21:39:46 +03:00
|
|
|
return obj->throw_obj;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline const rb_control_frame_t *
|
2015-03-11 15:49:27 +03:00
|
|
|
THROW_DATA_CATCH_FRAME(const struct vm_throw_data *obj)
|
2015-03-10 21:39:46 +03:00
|
|
|
{
|
2017-04-07 10:50:30 +03:00
|
|
|
VM_ASSERT(THROW_DATA_P(obj));
|
2015-03-10 21:39:46 +03:00
|
|
|
return obj->catch_frame;
|
|
|
|
}
|
|
|
|
|
2017-04-06 05:56:23 +03:00
|
|
|
static inline int
|
2015-03-11 15:49:27 +03:00
|
|
|
THROW_DATA_STATE(const struct vm_throw_data *obj)
|
2015-03-10 21:39:46 +03:00
|
|
|
{
|
2017-04-07 10:50:30 +03:00
|
|
|
VM_ASSERT(THROW_DATA_P(obj));
|
2019-07-22 11:44:58 +03:00
|
|
|
return obj->throw_state;
|
2015-03-10 21:39:46 +03:00
|
|
|
}
|
|
|
|
|
2017-04-06 05:56:23 +03:00
|
|
|
static inline int
|
|
|
|
THROW_DATA_CONSUMED_P(const struct vm_throw_data *obj)
|
|
|
|
{
|
|
|
|
VM_ASSERT(THROW_DATA_P(obj));
|
|
|
|
return obj->flags & THROW_DATA_CONSUMED;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void
|
|
|
|
THROW_DATA_CATCH_FRAME_SET(struct vm_throw_data *obj, const rb_control_frame_t *cfp)
|
|
|
|
{
|
2017-04-07 10:50:30 +03:00
|
|
|
VM_ASSERT(THROW_DATA_P(obj));
|
2017-04-06 05:56:23 +03:00
|
|
|
obj->catch_frame = cfp;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void
|
|
|
|
THROW_DATA_STATE_SET(struct vm_throw_data *obj, int st)
|
|
|
|
{
|
2017-04-07 10:50:30 +03:00
|
|
|
VM_ASSERT(THROW_DATA_P(obj));
|
2019-07-22 11:44:58 +03:00
|
|
|
obj->throw_state = st;
|
2017-04-06 05:56:23 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
static inline void
|
|
|
|
THROW_DATA_CONSUMED_SET(struct vm_throw_data *obj)
|
|
|
|
{
|
|
|
|
if (THROW_DATA_P(obj) &&
|
|
|
|
THROW_DATA_STATE(obj) == TAG_BREAK) {
|
|
|
|
obj->flags |= THROW_DATA_CONSUMED;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
VALUE size packed callinfo (ci).
Now, rb_call_info contains how to call the method with tuple of
(mid, orig_argc, flags, kwarg). Most of cases, kwarg == NULL and
mid+argc+flags only requires 64bits. So this patch packed
rb_call_info to VALUE (1 word) on such cases. If we can not
represent it in VALUE, then use imemo_callinfo which contains
conventional callinfo (rb_callinfo, renamed from rb_call_info).
iseq->body->ci_kw_size is removed because all of callinfo is VALUE
size (packed ci or a pointer to imemo_callinfo).
To access ci information, we need to use these functions:
vm_ci_mid(ci), _flag(ci), _argc(ci), _kwarg(ci).
struct rb_call_info_kw_arg is renamed to rb_callinfo_kwarg.
rb_funcallv_with_cc() and rb_method_basic_definition_p_with_cc()
is temporary removed because cd->ci should be marked.
2020-01-08 02:20:36 +03:00
|
|
|
#define IS_ARGS_SPLAT(ci) (vm_ci_flag(ci) & VM_CALL_ARGS_SPLAT)
|
|
|
|
#define IS_ARGS_KEYWORD(ci) (vm_ci_flag(ci) & VM_CALL_KWARG)
|
|
|
|
#define IS_ARGS_KW_SPLAT(ci) (vm_ci_flag(ci) & VM_CALL_KW_SPLAT)
|
|
|
|
#define IS_ARGS_KW_OR_KW_SPLAT(ci) (vm_ci_flag(ci) & (VM_CALL_KWARG | VM_CALL_KW_SPLAT))
|
2018-02-06 17:07:57 +03:00
|
|
|
|
2019-03-21 09:25:09 +03:00
|
|
|
/* If this returns true, an optimized function returned by `vm_call_iseq_setup_func`
|
|
|
|
can be used as a fastpath. */
|
|
|
|
static bool
|
VALUE size packed callinfo (ci).
Now, rb_call_info contains how to call the method with tuple of
(mid, orig_argc, flags, kwarg). Most of cases, kwarg == NULL and
mid+argc+flags only requires 64bits. So this patch packed
rb_call_info to VALUE (1 word) on such cases. If we can not
represent it in VALUE, then use imemo_callinfo which contains
conventional callinfo (rb_callinfo, renamed from rb_call_info).
iseq->body->ci_kw_size is removed because all of callinfo is VALUE
size (packed ci or a pointer to imemo_callinfo).
To access ci information, we need to use these functions:
vm_ci_mid(ci), _flag(ci), _argc(ci), _kwarg(ci).
struct rb_call_info_kw_arg is renamed to rb_callinfo_kwarg.
rb_funcallv_with_cc() and rb_method_basic_definition_p_with_cc()
is temporary removed because cd->ci should be marked.
2020-01-08 02:20:36 +03:00
|
|
|
vm_call_iseq_optimizable_p(const struct rb_callinfo *ci, const struct rb_call_cache *cc)
|
2019-03-21 09:25:09 +03:00
|
|
|
{
|
|
|
|
return !IS_ARGS_SPLAT(ci) && !IS_ARGS_KEYWORD(ci) &&
|
|
|
|
!(METHOD_ENTRY_VISI(cc->me) == METHOD_VISI_PROTECTED);
|
|
|
|
}
|
|
|
|
|
2008-01-18 11:56:11 +03:00
|
|
|
#endif /* RUBY_INSNHELPER_H */
|