add_general (&gr, &stack_size, ainfo, FALSE);
break;
case MONO_TYPE_GENERICINST:
- if (!mono_type_generic_inst_is_valuetype (sig->params [i])) {
+ if (!mono_type_generic_inst_is_valuetype (ptype)) {
add_general (&gr, &stack_size, ainfo, FALSE);
break;
}
cinfo->ret.reg = sparc_f0;
break;
case MONO_TYPE_GENERICINST:
- if (!mono_type_generic_inst_is_valuetype (sig->ret)) {
+ if (!mono_type_generic_inst_is_valuetype (ret_type)) {
cinfo->ret.storage = ArgInIReg;
cinfo->ret.reg = sparc_i0;
if (gr < 1)
int i, offset, size, align, curinst;
CallInfo *cinfo;
- header = mono_method_get_header (cfg->method);
+ header = cfg->header;
sig = mono_method_signature (cfg->method);
cfg->ret->opcode = OP_REGVAR;
cfg->ret->inst_c0 = cinfo->ret.reg;
break;
- case ArgInIRegPair:
- if (((sig->ret->type == MONO_TYPE_I8) || (sig->ret->type == MONO_TYPE_U8))) {
+ case ArgInIRegPair: {
+ MonoType *t = mono_type_get_underlying_type (sig->ret);
+ if (((t->type == MONO_TYPE_I8) || (t->type == MONO_TYPE_U8))) {
MonoInst *low = get_vreg_to_inst (cfg, cfg->ret->dreg + 1);
MonoInst *high = get_vreg_to_inst (cfg, cfg->ret->dreg + 2);
cfg->ret->opcode = OP_REGVAR;
cfg->ret->inst_c0 = cinfo->ret.reg;
break;
+ }
case ArgOnStack:
#ifdef SPARCV9
g_assert_not_reached ();
}
}
- /* Add a properly aligned dword for use by int<->float conversion opcodes */
- offset += 8;
- offset = ALIGN_TO (offset, 8);
- cfg->arch.float_spill_slot_offset = offset;
-
/*
* spillvars are stored between the normal locals and the storage reserved
* by the ABI.
low->flags |= MONO_INST_VOLATILE;
high->flags |= MONO_INST_VOLATILE;
}
+
+ /* Add a properly aligned dword for use by int<->float conversion opcodes */
+ cfg->arch.float_spill_slot = mono_compile_create_var (cfg, &mono_defaults.double_class->byval_arg, OP_ARG);
+ ((MonoInst*)cfg->arch.float_spill_slot)->flags |= MONO_INST_VOLATILE;
}
static void
else
arg_type = sig->params [i - sig->hasthis];
+ arg_type = mono_type_get_underlying_type (arg_type);
if ((i >= sig->hasthis) && (MONO_TYPE_ISSTRUCT(sig->params [i - sig->hasthis])))
emit_pass_vtype (cfg, call, cinfo, ainfo, arg_type, in, sig->pinvoke);
else if (!arg_type->byref && ((arg_type->type == MONO_TYPE_I8) || (arg_type->type == MONO_TYPE_U8)))
mono_arch_emit_setret (MonoCompile *cfg, MonoMethod *method, MonoInst *val)
{
CallInfo *cinfo = get_call_info (cfg, mono_method_signature (method), FALSE);
+ MonoType *ret = mini_type_get_underlying_type (cfg->generic_sharing_context, mono_method_signature (method)->ret);
switch (cinfo->ret.storage) {
case ArgInIReg:
MONO_EMIT_NEW_UNALU (cfg, OP_MOVE, cfg->ret->dreg, val->dreg);
break;
case ArgInIRegPair:
- MONO_EMIT_NEW_UNALU (cfg, OP_MOVE, cfg->ret->dreg + 2, val->dreg + 2);
- MONO_EMIT_NEW_UNALU (cfg, OP_MOVE, cfg->ret->dreg + 1, val->dreg + 1);
+ if (ret->type == MONO_TYPE_I8 || ret->type == MONO_TYPE_U8) {
+ MONO_EMIT_NEW_UNALU (cfg, OP_LMOVE, cfg->ret->dreg, val->dreg);
+ } else {
+ MONO_EMIT_NEW_UNALU (cfg, OP_MOVE, cfg->ret->dreg + 2, val->dreg + 2);
+ MONO_EMIT_NEW_UNALU (cfg, OP_MOVE, cfg->ret->dreg + 1, val->dreg + 1);
+ }
break;
case ArgInFReg:
- if (mono_method_signature (method)->ret->type == MONO_TYPE_R4)
+ if (ret->type == MONO_TYPE_R4)
MONO_EMIT_NEW_UNALU (cfg, OP_SETFRET, cfg->ret->dreg, val->dreg);
else
MONO_EMIT_NEW_UNALU (cfg, OP_FMOVE, cfg->ret->dreg, val->dreg);
* Only do this if the method is small since BPr only has a 16bit
* displacement.
*/
- if (v64 && (mono_method_get_header (cfg->method)->code_size < 10000) && last_ins &&
+ if (v64 && (cfg->header->code_size < 10000) && last_ins &&
(last_ins->opcode == OP_COMPARE_IMM) &&
(last_ins->inst_imm == 0)) {
switch (ins->opcode) {
int size = 0;
guint32 *code, *start;
- g_assert (!fail_tramp);
-
for (i = 0; i < count; ++i) {
MonoIMTCheckItem *item = imt_entries [i];
if (item->is_equals) {
item->chunk_size += CMP_SIZE;
item->chunk_size += BR_SMALL_SIZE + JUMP_IMM_SIZE;
} else {
+ if (fail_tramp)
+ item->chunk_size += 16;
item->chunk_size += JUMP_IMM_SIZE;
#if ENABLE_WRONG_METHOD_CHECK
item->chunk_size += CMP_SIZE + BR_SMALL_SIZE + 1;
}
size += item->chunk_size;
}
- code = mono_domain_code_reserve (domain, size * 4);
+ if (fail_tramp)
+ code = mono_method_alloc_generic_virtual_thunk (domain, size * 4);
+ else
+ code = mono_domain_code_reserve (domain, size * 4);
start = code;
-
for (i = 0; i < count; ++i) {
MonoIMTCheckItem *item = imt_entries [i];
item->code_target = (guint8*)code;
if (item->is_equals) {
- if (item->check_target_idx) {
- if (!item->compare_done) {
+ gboolean fail_case = !item->check_target_idx && fail_tramp;
+
+ if (item->check_target_idx || fail_case) {
+ if (!item->compare_done || fail_case) {
sparc_set (code, (guint32)item->key, sparc_g5);
sparc_cmp (code, MONO_ARCH_IMT_REG, sparc_g5);
}
sparc_ld (code, sparc_g5, 0, sparc_g5);
sparc_jmpl (code, sparc_g5, sparc_g0, sparc_g0);
sparc_nop (code);
+
+ if (fail_case) {
+ sparc_patch (item->jmp_code, code);
+ sparc_set (code, fail_tramp, sparc_g5);
+ sparc_jmpl (code, sparc_g5, sparc_g0, sparc_g0);
+ sparc_nop (code);
+ item->jmp_code = NULL;
+ }
} else {
/* enable the commented code to assert on wrong method */
#if ENABLE_WRONG_METHOD_CHECK
return (MonoMethod*)regs [sparc_g1];
}
-MonoObject*
-mono_arch_find_this_argument (mgreg_t *regs, MonoMethod *method, MonoGenericSharingContext *gsctx)
+gpointer
+mono_arch_get_this_arg_from_call (MonoGenericSharingContext *gsctx, MonoMethodSignature *sig, mgreg_t *regs, guint8 *code)
{
mono_sparc_flushw ();
break;
}
case OP_ICONV_TO_R4: {
- gint32 offset = cfg->arch.float_spill_slot_offset;
+ MonoInst *spill = cfg->arch.float_spill_slot;
+ gint32 reg = spill->inst_basereg;
+ gint32 offset = spill->inst_offset;
+
+ g_assert (spill->opcode == OP_REGOFFSET);
#ifdef SPARCV9
if (!sparc_is_imm13 (offset)) {
sparc_set (code, offset, sparc_o7);
- sparc_stx (code, ins->sreg1, sparc_sp, offset);
- sparc_lddf (code, sparc_sp, offset, FP_SCRATCH_REG);
+ sparc_stx (code, ins->sreg1, reg, offset);
+ sparc_lddf (code, reg, offset, FP_SCRATCH_REG);
} else {
- sparc_stx_imm (code, ins->sreg1, sparc_sp, offset);
- sparc_lddf_imm (code, sparc_sp, offset, FP_SCRATCH_REG);
+ sparc_stx_imm (code, ins->sreg1, reg, offset);
+ sparc_lddf_imm (code, reg, offset, FP_SCRATCH_REG);
}
sparc_fxtos (code, FP_SCRATCH_REG, FP_SCRATCH_REG);
#else
if (!sparc_is_imm13 (offset)) {
sparc_set (code, offset, sparc_o7);
- sparc_st (code, ins->sreg1, sparc_sp, sparc_o7);
- sparc_ldf (code, sparc_sp, sparc_o7, FP_SCRATCH_REG);
+ sparc_st (code, ins->sreg1, reg, sparc_o7);
+ sparc_ldf (code, reg, sparc_o7, FP_SCRATCH_REG);
} else {
- sparc_st_imm (code, ins->sreg1, sparc_sp, offset);
- sparc_ldf_imm (code, sparc_sp, offset, FP_SCRATCH_REG);
+ sparc_st_imm (code, ins->sreg1, reg, offset);
+ sparc_ldf_imm (code, reg, offset, FP_SCRATCH_REG);
}
sparc_fitos (code, FP_SCRATCH_REG, FP_SCRATCH_REG);
#endif
break;
}
case OP_ICONV_TO_R8: {
- gint32 offset = cfg->arch.float_spill_slot_offset;
+ MonoInst *spill = cfg->arch.float_spill_slot;
+ gint32 reg = spill->inst_basereg;
+ gint32 offset = spill->inst_offset;
+
+ g_assert (spill->opcode == OP_REGOFFSET);
+
#ifdef SPARCV9
if (!sparc_is_imm13 (offset)) {
sparc_set (code, offset, sparc_o7);
- sparc_stx (code, ins->sreg1, sparc_sp, sparc_o7);
- sparc_lddf (code, sparc_sp, sparc_o7, FP_SCRATCH_REG);
+ sparc_stx (code, ins->sreg1, reg, sparc_o7);
+ sparc_lddf (code, reg, sparc_o7, FP_SCRATCH_REG);
} else {
- sparc_stx_imm (code, ins->sreg1, sparc_sp, offset);
- sparc_lddf_imm (code, sparc_sp, offset, FP_SCRATCH_REG);
+ sparc_stx_imm (code, ins->sreg1, reg, offset);
+ sparc_lddf_imm (code, reg, offset, FP_SCRATCH_REG);
}
sparc_fxtod (code, FP_SCRATCH_REG, ins->dreg);
#else
if (!sparc_is_imm13 (offset)) {
sparc_set (code, offset, sparc_o7);
- sparc_st (code, ins->sreg1, sparc_sp, sparc_o7);
- sparc_ldf (code, sparc_sp, sparc_o7, FP_SCRATCH_REG);
+ sparc_st (code, ins->sreg1, reg, sparc_o7);
+ sparc_ldf (code, reg, sparc_o7, FP_SCRATCH_REG);
} else {
- sparc_st_imm (code, ins->sreg1, sparc_sp, offset);
- sparc_ldf_imm (code, sparc_sp, offset, FP_SCRATCH_REG);
+ sparc_st_imm (code, ins->sreg1, reg, offset);
+ sparc_ldf_imm (code, reg, offset, FP_SCRATCH_REG);
}
sparc_fitod (code, FP_SCRATCH_REG, ins->dreg);
#endif
#endif
case OP_FCONV_TO_I4:
case OP_FCONV_TO_U4: {
- gint32 offset = cfg->arch.float_spill_slot_offset;
+ MonoInst *spill = cfg->arch.float_spill_slot;
+ gint32 reg = spill->inst_basereg;
+ gint32 offset = spill->inst_offset;
+
+ g_assert (spill->opcode == OP_REGOFFSET);
+
sparc_fdtoi (code, ins->sreg1, FP_SCRATCH_REG);
if (!sparc_is_imm13 (offset)) {
sparc_set (code, offset, sparc_o7);
- sparc_stdf (code, FP_SCRATCH_REG, sparc_sp, sparc_o7);
- sparc_ld (code, sparc_sp, sparc_o7, ins->dreg);
+ sparc_stdf (code, FP_SCRATCH_REG, reg, sparc_o7);
+ sparc_ld (code, reg, sparc_o7, ins->dreg);
} else {
- sparc_stdf_imm (code, FP_SCRATCH_REG, sparc_sp, offset);
- sparc_ld_imm (code, sparc_sp, offset, ins->dreg);
+ sparc_stdf_imm (code, FP_SCRATCH_REG, reg, offset);
+ sparc_ld_imm (code, reg, offset, ins->dreg);
}
switch (ins->opcode) {
EMIT_FLOAT_COND_BRANCH (ins, sparc_fbu, 1, 1);
break;
case OP_CKFINITE: {
- gint32 offset = cfg->arch.float_spill_slot_offset;
+ MonoInst *spill = cfg->arch.float_spill_slot;
+ gint32 reg = spill->inst_basereg;
+ gint32 offset = spill->inst_offset;
+
+ g_assert (spill->opcode == OP_REGOFFSET);
+
if (!sparc_is_imm13 (offset)) {
sparc_set (code, offset, sparc_o7);
- sparc_stdf (code, ins->sreg1, sparc_sp, sparc_o7);
- sparc_lduh (code, sparc_sp, sparc_o7, sparc_o7);
+ sparc_stdf (code, ins->sreg1, reg, sparc_o7);
+ sparc_lduh (code, reg, sparc_o7, sparc_o7);
} else {
- sparc_stdf_imm (code, ins->sreg1, sparc_sp, offset);
- sparc_lduh_imm (code, sparc_sp, offset, sparc_o7);
+ sparc_stdf_imm (code, ins->sreg1, reg, offset);
+ sparc_lduh_imm (code, reg, offset, sparc_o7);
}
sparc_srl_imm (code, sparc_o7, 4, sparc_o7);
sparc_and_imm (code, FALSE, sparc_o7, 2047, sparc_o7);