cinfo->ret.storage = ArgInFloatReg;
cinfo->ret.reg = 8;
break;
+ case MONO_TYPE_GENERICINST:
+ if (!mono_type_generic_inst_is_valuetype (sig->ret)) {
+ cinfo->ret.storage = ArgInIReg;
+ cinfo->ret.reg = IA64_R8;
+ break;
+ }
+ /* Fall through */
case MONO_TYPE_VALUETYPE:
case MONO_TYPE_TYPEDBYREF: {
guint32 tmp_gr = 0, tmp_fr = 0, tmp_stacksize = 0;
case MONO_TYPE_ARRAY:
add_general (&gr, &stack_size, ainfo);
break;
+ case MONO_TYPE_GENERICINST:
+ if (!mono_type_generic_inst_is_valuetype (sig->params [i])) {
+ add_general (&gr, &stack_size, ainfo);
+ break;
+ }
+ /* Fall through */
case MONO_TYPE_VALUETYPE:
case MONO_TYPE_TYPEDBYREF:
/* FIXME: */
case MONO_TYPE_SZARRAY:
case MONO_TYPE_ARRAY:
return TRUE;
+ case MONO_TYPE_GENERICINST:
+ if (!mono_type_generic_inst_is_valuetype (t))
+ return TRUE;
+ return FALSE;
case MONO_TYPE_VALUETYPE:
return FALSE;
}
mono_ia64_alloc_stacked_registers (MonoCompile *cfg)
{
CallInfo *cinfo;
- guint32 reserved_regs = 3;
+ guint32 reserved_regs;
+ MonoMethodHeader *header;
if (cfg->arch.reg_local0 > 0)
/* Already done */
cinfo = get_call_info (mono_method_signature (cfg->method), FALSE);
- /* Three registers are reserved for use by the prolog/epilog */
- reserved_regs = 3;
+ header = mono_method_get_header (cfg->method);
+
+ /* Some registers are reserved for use by the prolog/epilog */
+ reserved_regs = header->num_clauses ? 4 : 3;
if ((mono_jit_trace_calls != NULL && mono_trace_eval (cfg->method)) ||
(cfg->prof_options & MONO_PROFILE_ENTER_LEAVE)) {
cfg->arch.reg_saved_ar_pfs = cfg->arch.reg_local0 - 1;
cfg->arch.reg_saved_b0 = cfg->arch.reg_local0 - 2;
- cfg->arch.reg_saved_sp = cfg->arch.reg_local0 - 3;
+ cfg->arch.reg_fp = cfg->arch.reg_local0 - 3;
+
+ /*
+ * Frames without handlers save sp to fp, frames with handlers save it into
+ * a dedicated register.
+ */
+ if (header->num_clauses)
+ cfg->arch.reg_saved_sp = cfg->arch.reg_local0 - 4;
+ else
+ cfg->arch.reg_saved_sp = cfg->arch.reg_fp;
if ((mono_jit_trace_calls != NULL && mono_trace_eval (cfg->method)) ||
(cfg->prof_options & MONO_PROFILE_ENTER_LEAVE)) {
- cfg->arch.reg_saved_return_val = cfg->arch.reg_local0 - 4;
+ cfg->arch.reg_saved_return_val = cfg->arch.reg_local0 - reserved_regs;
}
/*
}
else {
/* Locals are allocated backwards from %fp */
- cfg->frame_reg = cfg->arch.reg_saved_sp;
+ cfg->frame_reg = cfg->arch.reg_fp;
offset = 0;
}
if ((i >= sig->hasthis) && (MONO_TYPE_ISSTRUCT(arg_type))) {
MonoInst *stack_addr;
- gint align;
+ guint32 align;
guint32 size;
if (arg_type->type == MONO_TYPE_TYPEDBYREF) {
switch (ins->opcode) {
case OP_MOVE:
case OP_FMOVE:
+ case OP_SETREG:
/*
* Removes:
*
break;
case OP_VCALL:
case OP_VCALL_REG:
- case OP_VCALL_MEMBASE:
+ case OP_VCALL_MEMBASE: {
+ ArgStorage storage;
+
cinfo = get_call_info (((MonoCallInst*)ins)->signature, FALSE);
- ArgStorage storage = cinfo->ret.storage;
+ storage = cinfo->ret.storage;
if (storage == ArgAggregate) {
MonoInst *local = (MonoInst*)cfg->arch.ret_var_addr_local;
}
g_free (cinfo);
break;
+ }
default:
g_assert_not_reached ();
}
mono_add_patch_info (cfg, code.buf + code.nins - cfg->native_code, patch_type, data); \
} while (0)
+#define emit_cond_system_exception(cfg,code,exc_name,predicate) do { \
+ MonoInst *tins = mono_branch_optimize_exception_target (cfg, bb, exc_name); \
+ if (tins == NULL) \
+ add_patch_info (cfg, code, MONO_PATCH_INFO_EXC, exc_name); \
+ else \
+ add_patch_info (cfg, code, MONO_PATCH_INFO_BB, tins->inst_true_bb); \
+ ia64_br_cond_pred (code, (predicate), 0); \
+} while (0)
+
static Ia64CodegenState
emit_call (MonoCompile *cfg, Ia64CodegenState code, guint32 patch_type, gconstpointer data)
{
if ((patch_type == MONO_PATCH_INFO_ABS) || (patch_type == MONO_PATCH_INFO_INTERNAL_METHOD)) {
/* Indirect call */
+ /* mono_arch_patch_callsite will patch this */
/* mono_arch_nullify_class_init_trampoline will patch this */
ia64_movl (code, GP_SCRATCH_REG, 0);
ia64_ld8_inc_imm (code, GP_SCRATCH_REG2, GP_SCRATCH_REG, 8);
case OP_COND_EXC_IOV:
case OP_COND_EXC_OV:
- add_patch_info (cfg, code,
- MONO_PATCH_INFO_EXC, "OverflowException");
- ia64_br_cond_pred (code, 6, 0);
+ emit_cond_system_exception (cfg, code, "OverflowException", 6);
break;
case OP_COND_EXC_IC:
case OP_COND_EXC_C:
- add_patch_info (cfg, code,
- MONO_PATCH_INFO_EXC, "OverflowException");
- ia64_br_cond_pred (code, 7, 0);
+ emit_cond_system_exception (cfg, code, "OverflowException", 7);
break;
case OP_IA64_COND_EXC:
- add_patch_info (cfg, code,
- MONO_PATCH_INFO_EXC, ins->inst_p1);
- ia64_br_cond_pred (code, 6, 0);
+ emit_cond_system_exception (cfg, code, ins->inst_p1, 6);
break;
case OP_IA64_CSET:
ia64_mov_pred (code, 7, ins->dreg, IA64_R0);
ia64_stfd_hint (code, ins->inst_destbasereg, ins->sreg1, 0);
break;
case OP_STORER4_MEMBASE_REG:
- ia64_stfs_hint (code, ins->inst_destbasereg, ins->sreg1, 0);
+ ia64_fnorm_s_sf (code, FP_SCRATCH_REG, ins->sreg1, 0);
+ ia64_stfs_hint (code, ins->inst_destbasereg, FP_SCRATCH_REG, 0);
break;
case OP_LOADR8_MEMBASE:
ia64_ldfd (code, ins->dreg, ins->inst_basereg);
break;
case OP_LOADR4_MEMBASE:
ia64_ldfs (code, ins->dreg, ins->inst_basereg);
+ ia64_fnorm_d_sf (code, ins->dreg, ins->dreg, 0);
break;
case CEE_CONV_R4:
ia64_setf_sig (code, ins->dreg, ins->sreg1);
case CEE_CKFINITE:
/* Quiet NaN */
ia64_fclass_m (code, 6, 7, ins->sreg1, 0x080);
- add_patch_info (cfg, code,
- MONO_PATCH_INFO_EXC, "ArithmeticException");
- ia64_br_cond_pred (code, 6, 0);
+ emit_cond_system_exception (cfg, code, "ArithmeticException", 6);
/* Signaling NaN */
ia64_fclass_m (code, 6, 7, ins->sreg1, 0x040);
- add_patch_info (cfg, code,
- MONO_PATCH_INFO_EXC, "ArithmeticException");
- ia64_br_cond_pred (code, 6, 0);
+ emit_cond_system_exception (cfg, code, "ArithmeticException", 6);
/* Positive infinity */
ia64_fclass_m (code, 6, 7, ins->sreg1, 0x021);
- add_patch_info (cfg, code,
- MONO_PATCH_INFO_EXC, "ArithmeticException");
- ia64_br_cond_pred (code, 6, 0);
+ emit_cond_system_exception (cfg, code, "ArithmeticException", 6);
/* Negative infinity */
ia64_fclass_m (code, 6, 7, ins->sreg1, 0x022);
- add_patch_info (cfg, code,
- MONO_PATCH_INFO_EXC, "ArithmeticException");
- ia64_br_cond_pred (code, 6, 0);
+ emit_cond_system_exception (cfg, code, "ArithmeticException", 6);
break;
/* Calls */
MonoInst *spvar = mono_find_spvar_for_region (cfg, bb->region);
/*
- * We might be called by call_filter, in which case the
- * the register stack is not set up correctly. So do it now.
- * Allocate a stack frame and set the fp register from the value
- * passed in by the caller.
- * R15 is used since it is writable using libunwind.
+ * R15 determines our caller. It is used since it is writable using
+ * libunwind.
* R15 == 0 means we are called by OP_CALL_HANDLER or via resume_context ()
+ * R15 != 0 means we are called by call_filter ().
*/
ia64_codegen_set_one_ins_per_bundle (code, TRUE);
ia64_cmp_eq (code, 6, 7, IA64_R15, IA64_R0);
- /* Alloc is not predictable so we have to use a branch */
- ia64_br_cond_pred (code, 6, 3);
+
+ ia64_br_cond_pred (code, 6, 6);
+
+ /*
+ * Called by call_filter:
+ * Allocate a new stack frame, and set the fp register from the
+ * value passed in by the caller.
+ * We allocate a similar frame as is done by the prolog, so
+ * if an exception is thrown while executing the filter, the
+ * unwinder can unwind through the filter frame using the unwind
+ * info for the prolog.
+ */
ia64_alloc (code, cfg->arch.reg_saved_ar_pfs, cfg->arch.reg_local0 - cfg->arch.reg_in0, cfg->arch.reg_out0 - cfg->arch.reg_local0, cfg->arch.n_out_regs, 0);
+ ia64_mov_from_br (code, cfg->arch.reg_saved_b0, IA64_B0);
+ ia64_mov (code, cfg->arch.reg_saved_sp, IA64_SP);
ia64_mov (code, cfg->frame_reg, IA64_R15);
+ /* Signal to endfilter that we are called by call_filter */
+ ia64_mov (code, GP_SCRATCH_REG, IA64_R0);
+
/* Save the return address */
ia64_adds_imm (code, GP_SCRATCH_REG2, spvar->inst_offset, cfg->frame_reg);
ia64_st8_hint (code, GP_SCRATCH_REG2, GP_SCRATCH_REG, 0);
break;
}
- case CEE_ENDFINALLY: {
- MonoInst *spvar = mono_find_spvar_for_region (cfg, bb->region);
- /* Return the saved arp_pfs value to call_filter */
- ia64_mov (code, IA64_R9, cfg->arch.reg_saved_ar_pfs);
- ia64_adds_imm (code, GP_SCRATCH_REG, spvar->inst_offset, cfg->frame_reg);
- ia64_ld8_hint (code, GP_SCRATCH_REG, GP_SCRATCH_REG, 0);
- ia64_mov_to_br (code, IA64_B6, GP_SCRATCH_REG);
- ia64_br_cond_reg (code, IA64_B6);
- break;
- }
+ case CEE_ENDFINALLY:
case OP_ENDFILTER: {
- /* FIXME: Return the value */
+ /* FIXME: Return the value in ENDFILTER */
MonoInst *spvar = mono_find_spvar_for_region (cfg, bb->region);
- /* Return the saved arp_pfs value to call_filter */
- ia64_mov (code, IA64_R9, cfg->arch.reg_saved_ar_pfs);
+
+ /* Load the return address */
ia64_adds_imm (code, GP_SCRATCH_REG, spvar->inst_offset, cfg->frame_reg);
ia64_ld8_hint (code, GP_SCRATCH_REG, GP_SCRATCH_REG, 0);
+
+ /* Test caller */
+ ia64_cmp_eq (code, 6, 7, GP_SCRATCH_REG, IA64_R0);
+ ia64_br_cond_pred (code, 7, 4);
+
+ /* Called by call_filter */
+ /* Pop frame */
+ ia64_mov_to_ar_i (code, IA64_PFS, cfg->arch.reg_saved_ar_pfs);
+ ia64_mov_to_br (code, IA64_B0, cfg->arch.reg_saved_b0);
+ ia64_br_ret_reg (code, IA64_B0);
+
+ /* Called by CALL_HANDLER */
ia64_mov_to_br (code, IA64_B6, GP_SCRATCH_REG);
ia64_br_cond_reg (code, IA64_B6);
break;
static void
ia64_real_emit_bundle (Ia64CodegenState *code, int *deps_start, int *stops, int n, guint64 template, guint64 ins1, guint64 ins2, guint64 ins3, guint8 nops)
{
- g_assert (n <= code->nins);
int stop_pos, i, deps_to_shift, dep_shift;
+ g_assert (n <= code->nins);
+
// if (n > 1) printf ("FOUND: %ld.\n", template);
ia64_emit_bundle_template (code, template, ins1, ins2, ins3);
target = mono_resolve_patch_target (method, domain, code, patch_info, run_cctors);
+ if (patch_info->type == MONO_PATCH_INFO_NONE)
+ continue;
if (mono_compile_aot) {
NOT_IMPLEMENTED;
}
ia64_mov_from_br (code, cfg->arch.reg_saved_b0, IA64_B0);
if ((alloc_size || cinfo->stack_usage) && !cfg->arch.omit_fp) {
- ia64_unw_save_reg (code, UNW_IA64_SP, UNW_IA64_GR + cfg->frame_reg);
- ia64_mov (code, cfg->frame_reg, IA64_SP);
+ ia64_unw_save_reg (code, UNW_IA64_SP, UNW_IA64_GR + cfg->arch.reg_saved_sp);
+ ia64_mov (code, cfg->arch.reg_saved_sp, IA64_SP);
+ if (cfg->frame_reg != cfg->arch.reg_saved_sp)
+ ia64_mov (code, cfg->frame_reg, IA64_SP);
}
if (alloc_size) {
mono_arch_emit_exceptions (MonoCompile *cfg)
{
MonoJumpInfo *patch_info;
- int nthrows;
+ int i, nthrows;
Ia64CodegenState code;
gboolean empty = TRUE;
//unw_dyn_region_info_t *r_exceptions;
-
- /*
MonoClass *exc_classes [16];
guint8 *exc_throw_start [16], *exc_throw_end [16];
- */
guint32 code_size = 0;
/* Compute needed space */
MonoClass *exc_class;
guint8* throw_ip;
guint8* buf;
+ guint64 exc_token_index;
exc_class = mono_class_from_name (mono_defaults.corlib, "System", patch_info->data.name);
g_assert (exc_class);
+ exc_token_index = mono_metadata_token_index (exc_class->type_token);
throw_ip = cfg->native_code + patch_info->ip.i;
ia64_begin_bundle (code);
ia64_patch (cfg->native_code + patch_info->ip.i, code.buf);
- ia64_movl (code, cfg->arch.reg_out0 + 0, exc_class->type_token);
+ /* Find a throw sequence for the same exception class */
+ for (i = 0; i < nthrows; ++i)
+ if (exc_classes [i] == exc_class)
+ break;
- ia64_begin_bundle (code);
+ if (i < nthrows) {
+ gint64 offset = exc_throw_end [i] - 16 - throw_ip;
- patch_info->data.name = "mono_arch_throw_corlib_exception";
- patch_info->type = MONO_PATCH_INFO_INTERNAL_METHOD;
- patch_info->ip.i = code.buf - cfg->native_code;
+ if (ia64_is_adds_imm (offset))
+ ia64_adds_imm (code, cfg->arch.reg_out0 + 1, offset, IA64_R0);
+ else
+ ia64_movl (code, cfg->arch.reg_out0 + 1, offset);
- /* Indirect call */
- ia64_movl (code, GP_SCRATCH_REG, 0);
- ia64_ld8_inc_imm (code, GP_SCRATCH_REG2, GP_SCRATCH_REG, 8);
- ia64_mov_to_br (code, IA64_B6, GP_SCRATCH_REG2);
- ia64_ld8 (code, IA64_GP, GP_SCRATCH_REG);
+ buf = code.buf + code.nins;
+ ia64_br_cond_pred (code, 0, 0);
+ ia64_begin_bundle (code);
+ ia64_patch (buf, exc_throw_start [i]);
- /* Compute the offset */
- buf = code.buf + 32;
- ia64_movl (code, cfg->arch.reg_out0 + 1, buf - throw_ip);
+ patch_info->type = MONO_PATCH_INFO_NONE;
+ }
+ else {
+ /* Arg1 */
+ buf = code.buf;
+ ia64_movl (code, cfg->arch.reg_out0 + 1, 0);
- ia64_br_call_reg (code, IA64_B0, IA64_B6);
+ ia64_begin_bundle (code);
+
+ if (nthrows < 16) {
+ exc_classes [nthrows] = exc_class;
+ exc_throw_start [nthrows] = code.buf;
+ }
+
+ /* Arg2 */
+ if (ia64_is_adds_imm (exc_token_index))
+ ia64_adds_imm (code, cfg->arch.reg_out0 + 0, exc_token_index, IA64_R0);
+ else
+ ia64_movl (code, cfg->arch.reg_out0 + 0, exc_token_index);
+
+ patch_info->data.name = "mono_arch_throw_corlib_exception";
+ patch_info->type = MONO_PATCH_INFO_INTERNAL_METHOD;
+ patch_info->ip.i = code.buf + code.nins - cfg->native_code;
+
+ /* Indirect call */
+ ia64_movl (code, GP_SCRATCH_REG, 0);
+ ia64_ld8_inc_imm (code, GP_SCRATCH_REG2, GP_SCRATCH_REG, 8);
+ ia64_mov_to_br (code, IA64_B6, GP_SCRATCH_REG2);
+ ia64_ld8 (code, IA64_GP, GP_SCRATCH_REG);
+
+ ia64_br_call_reg (code, IA64_B0, IA64_B6);
+
+ /* Patch up the throw offset */
+ ia64_begin_bundle (code);
+
+ ia64_patch (buf, (gpointer)(code.buf - 16 - throw_ip));
+
+ if (nthrows < 16) {
+ exc_throw_end [nthrows] = code.buf;
+ nthrows ++;
+ }
+ }
empty = FALSE;
break;
vtarg->dreg = mono_regstate_next_int (cfg->rs);
mono_bblock_add_inst (cfg->cbb, vtarg);
- mono_call_inst_add_outarg_reg (call, vtarg->dreg, out_reg, FALSE);
+ mono_call_inst_add_outarg_reg (cfg, call, vtarg->dreg, out_reg, FALSE);
out_reg ++;
}
this->dreg = mono_regstate_next_int (cfg->rs);
mono_bblock_add_inst (cfg->cbb, this);
- mono_call_inst_add_outarg_reg (call, this->dreg, out_reg, FALSE);
+ mono_call_inst_add_outarg_reg (cfg, call, this->dreg, out_reg, FALSE);
}
}