X-Git-Url: http://nsz.repo.hu/git/?a=blobdiff_plain;f=ir%2Fbe%2Fsparc%2Fsparc_transform.c;h=3e70dc90056a5cc876661ddcf4836360d972556d;hb=0e4318adc38a54110c1b1f7f1d5edc0ee2c78170;hp=d1b1c0af2a92efa513ea73c4a2a1ca73d5356f78;hpb=8ec4a1e10fe72948e6076936d408accdb4e90dc0;p=libfirm diff --git a/ir/be/sparc/sparc_transform.c b/ir/be/sparc/sparc_transform.c index d1b1c0af2..3e70dc900 100644 --- a/ir/be/sparc/sparc_transform.c +++ b/ir/be/sparc/sparc_transform.c @@ -26,6 +26,7 @@ #include "config.h" #include +#include #include "irnode_t.h" #include "irgraph_t.h" @@ -60,19 +61,55 @@ DEBUG_ONLY(static firm_dbg_module_t *dbg = NULL;) -static beabi_helper_env_t *abihelper; static const arch_register_t *sp_reg = &sparc_registers[REG_SP]; static const arch_register_t *fp_reg = &sparc_registers[REG_FRAME_POINTER]; -static calling_convention_t *cconv = NULL; +static calling_convention_t *current_cconv = NULL; +static be_stackorder_t *stackorder; static ir_mode *mode_gp; +static ir_mode *mode_flags; static ir_mode *mode_fp; static ir_mode *mode_fp2; //static ir_mode *mode_fp4; static pmap *node_to_stack; +static size_t start_mem_offset; +static ir_node *start_mem; +static size_t start_g0_offset; +static ir_node *start_g0; +static size_t start_g7_offset; +static ir_node *start_g7; +static size_t start_sp_offset; +static ir_node *start_sp; +static size_t start_fp_offset; +static ir_node *start_fp; +static ir_node *frame_base; +static size_t start_params_offset; +static size_t start_callee_saves_offset; + +static const arch_register_t *const omit_fp_callee_saves[] = { + &sparc_registers[REG_L0], + &sparc_registers[REG_L1], + &sparc_registers[REG_L2], + &sparc_registers[REG_L3], + &sparc_registers[REG_L4], + &sparc_registers[REG_L5], + &sparc_registers[REG_L6], + &sparc_registers[REG_L7], + &sparc_registers[REG_I0], + &sparc_registers[REG_I1], + &sparc_registers[REG_I2], + &sparc_registers[REG_I3], + &sparc_registers[REG_I4], + &sparc_registers[REG_I5], +}; -static inline int mode_needs_gp_reg(ir_mode *mode) +static inline bool mode_needs_gp_reg(ir_mode *mode) { - return mode_is_int(mode) || mode_is_reference(mode); + if (mode_is_int(mode) || mode_is_reference(mode)) { + /* we should only see 32bit code */ + assert(get_mode_size_bits(mode) <= 32); + return true; + } + return false; } /** @@ -239,10 +276,13 @@ static ir_node *gen_helper_binop_args(ir_node *node, } mode1 = get_irn_mode(op1); mode2 = get_irn_mode(op2); + /* we shouldn't see 64bit code */ + assert(get_mode_size_bits(mode1) <= 32); + assert(get_mode_size_bits(mode2) <= 32); if (is_imm_encodeable(op2)) { - ir_node *new_op1 = be_transform_node(op1); int32_t immediate = get_tarval_long(get_Const_tarval(op2)); + new_op1 = be_transform_node(op1); if (! (flags & MATCH_MODE_NEUTRAL) && needs_extension(mode1)) { new_op1 = gen_extension(dbgi, block, new_op1, mode1); } @@ -308,28 +348,107 @@ static ir_node *gen_helper_unfpop(ir_node *node, ir_mode *mode, new_unop_fp_func new_func_double, new_unop_fp_func new_func_quad) { - ir_node *block = be_transform_node(get_nodes_block(node)); - ir_node *op1 = get_binop_left(node); - ir_node *new_op1 = be_transform_node(op1); - dbg_info *dbgi = get_irn_dbg_info(node); - unsigned bits = get_mode_size_bits(mode); + ir_node *block = be_transform_node(get_nodes_block(node)); + ir_node *op = get_unop_op(node); + ir_node *new_op = be_transform_node(op); + dbg_info *dbgi = get_irn_dbg_info(node); + unsigned bits = get_mode_size_bits(mode); switch (bits) { case 32: - return new_func_single(dbgi, block, new_op1, mode); + return new_func_single(dbgi, block, new_op, mode); case 64: - return new_func_double(dbgi, block, new_op1, mode); + return new_func_double(dbgi, block, new_op, mode); case 128: - return new_func_quad(dbgi, block, new_op1, mode); + return new_func_quad(dbgi, block, new_op, mode); default: break; } panic("unsupported mode %+F for float op", mode); } -static ir_node *get_g0(void) +typedef ir_node* (*new_binopx_imm_func)(dbg_info *dbgi, ir_node *block, + ir_node *op1, ir_node *flags, + ir_entity *imm_entity, int32_t imm); + +typedef ir_node* (*new_binopx_reg_func)(dbg_info *dbgi, ir_node *block, + ir_node *op1, ir_node *op2, + ir_node *flags); + +static ir_node *gen_helper_binopx(ir_node *node, match_flags_t match_flags, + new_binopx_reg_func new_binopx_reg, + new_binopx_imm_func new_binopx_imm) +{ + dbg_info *dbgi = get_irn_dbg_info(node); + ir_node *block = be_transform_node(get_nodes_block(node)); + ir_node *op1 = get_irn_n(node, 0); + ir_node *op2 = get_irn_n(node, 1); + ir_node *flags = get_irn_n(node, 2); + ir_node *new_flags = be_transform_node(flags); + ir_node *new_op1; + ir_node *new_op2; + + /* only support for mode-neutral implemented so far */ + assert(match_flags & MATCH_MODE_NEUTRAL); + + if (is_imm_encodeable(op2)) { + int32_t immediate = get_tarval_long(get_Const_tarval(op2)); + new_op1 = be_transform_node(op1); + return new_binopx_imm(dbgi, block, new_op1, new_flags, NULL, immediate); + } + new_op2 = be_transform_node(op2); + if ((match_flags & MATCH_COMMUTATIVE) && is_imm_encodeable(op1)) { + int32_t immediate = get_tarval_long(get_Const_tarval(op1)); + return new_binopx_imm(dbgi, block, new_op2, new_flags, NULL, immediate); + } + new_op1 = be_transform_node(op1); + return new_binopx_reg(dbgi, block, new_op1, new_op2, new_flags); + +} + +static ir_node *get_g0(ir_graph *irg) { - return be_prolog_get_reg_value(abihelper, &sparc_registers[REG_G0]); + if (start_g0 == NULL) { + /* this is already the transformed start node */ + ir_node *start = get_irg_start(irg); + assert(is_sparc_Start(start)); + start_g0 = new_r_Proj(start, mode_gp, start_g0_offset); + } + return start_g0; +} + +static ir_node *get_g7(ir_graph *irg) +{ + if (start_g7 == NULL) { + ir_node *start = get_irg_start(irg); + assert(is_sparc_Start(start)); + start_g7 = new_r_Proj(start, mode_gp, start_g7_offset); + } + return start_g7; +} + +static ir_node *make_tls_offset(dbg_info *dbgi, ir_node *block, + ir_entity *entity, int32_t offset) +{ + ir_node *hi = new_bd_sparc_SetHi(dbgi, block, entity, offset); + ir_node *low = new_bd_sparc_Xor_imm(dbgi, block, hi, entity, offset); + return low; +} + +static ir_node *make_address(dbg_info *dbgi, ir_node *block, ir_entity *entity, + int32_t offset) +{ + if (get_entity_owner(entity) == get_tls_type()) { + ir_graph *irg = get_irn_irg(block); + ir_node *g7 = get_g7(irg); + ir_node *offsetn = make_tls_offset(dbgi, block, entity, offset); + ir_node *add = new_bd_sparc_Add_reg(dbgi, block, g7, offsetn); + return add; + } else { + ir_node *hi = new_bd_sparc_SetHi(dbgi, block, entity, offset); + ir_node *low = new_bd_sparc_Or_imm(dbgi, block, hi, entity, offset); + return low; + } } typedef struct address_t { @@ -363,15 +482,28 @@ static void match_address(ir_node *ptr, address_t *address, bool use_ptr2) * won't save anything but produce multiple sethi+or combinations with * just different offsets */ if (is_SymConst(base) && get_irn_n_edges(base) == 1) { - dbg_info *dbgi = get_irn_dbg_info(ptr); - ir_node *block = get_nodes_block(ptr); - ir_node *new_block = be_transform_node(block); - entity = get_SymConst_entity(base); - base = new_bd_sparc_SetHi(dbgi, new_block, entity, offset); - } else if (use_ptr2 && is_Add(base) && entity == NULL && offset == 0) { + ir_entity *sc_entity = get_SymConst_entity(base); + dbg_info *dbgi = get_irn_dbg_info(ptr); + ir_node *block = get_nodes_block(ptr); + ir_node *new_block = be_transform_node(block); + + if (get_entity_owner(sc_entity) == get_tls_type()) { + if (!use_ptr2) { + goto only_offset; + } else { + ptr2 = make_tls_offset(dbgi, new_block, sc_entity, offset); + offset = 0; + base = get_g7(get_irn_irg(base)); + } + } else { + entity = sc_entity; + base = new_bd_sparc_SetHi(dbgi, new_block, entity, offset); + } + } else if (use_ptr2 && is_Add(base) && offset == 0) { ptr2 = be_transform_node(get_Add_right(base)); base = be_transform_node(get_Add_left(base)); } else { +only_offset: if (sparc_is_value_imm_encodeable(offset)) { base = be_transform_node(base); } else { @@ -437,6 +569,34 @@ static ir_node *gen_Add(ir_node *node) new_bd_sparc_Add_reg, new_bd_sparc_Add_imm); } +static ir_node *gen_AddCC_t(ir_node *node) +{ + return gen_helper_binop(node, MATCH_COMMUTATIVE | MATCH_MODE_NEUTRAL, + new_bd_sparc_AddCC_reg, new_bd_sparc_AddCC_imm); +} + +static ir_node *gen_Proj_AddCC_t(ir_node *node) +{ + long pn = get_Proj_proj(node); + ir_node *pred = get_Proj_pred(node); + ir_node *new_pred = be_transform_node(pred); + + switch (pn) { + case pn_sparc_AddCC_t_res: + return new_r_Proj(new_pred, mode_gp, pn_sparc_AddCC_res); + case pn_sparc_AddCC_t_flags: + return new_r_Proj(new_pred, mode_flags, pn_sparc_AddCC_flags); + default: + panic("Invalid AddCC_t proj found"); + } +} + +static ir_node *gen_AddX_t(ir_node *node) +{ + return gen_helper_binopx(node, MATCH_COMMUTATIVE | MATCH_MODE_NEUTRAL, + new_bd_sparc_AddX_reg, new_bd_sparc_AddX_imm); +} + /** * Creates an sparc Sub. * @@ -452,12 +612,41 @@ static ir_node *gen_Sub(ir_node *node) new_bd_sparc_fsub_d, new_bd_sparc_fsub_q); } - return gen_helper_binop(node, MATCH_NONE, new_bd_sparc_Sub_reg, new_bd_sparc_Sub_imm); + return gen_helper_binop(node, MATCH_MODE_NEUTRAL, + new_bd_sparc_Sub_reg, new_bd_sparc_Sub_imm); +} + +static ir_node *gen_SubCC_t(ir_node *node) +{ + return gen_helper_binop(node, MATCH_MODE_NEUTRAL, + new_bd_sparc_SubCC_reg, new_bd_sparc_SubCC_imm); +} + +static ir_node *gen_Proj_SubCC_t(ir_node *node) +{ + long pn = get_Proj_proj(node); + ir_node *pred = get_Proj_pred(node); + ir_node *new_pred = be_transform_node(pred); + + switch (pn) { + case pn_sparc_SubCC_t_res: + return new_r_Proj(new_pred, mode_gp, pn_sparc_SubCC_res); + case pn_sparc_SubCC_t_flags: + return new_r_Proj(new_pred, mode_flags, pn_sparc_SubCC_flags); + default: + panic("Invalid SubCC_t proj found"); + } +} + +static ir_node *gen_SubX_t(ir_node *node) +{ + return gen_helper_binopx(node, MATCH_MODE_NEUTRAL, + new_bd_sparc_SubX_reg, new_bd_sparc_SubX_imm); } -static ir_node *create_ldf(dbg_info *dbgi, ir_node *block, ir_node *ptr, - ir_node *mem, ir_mode *mode, ir_entity *entity, - long offset, bool is_frame_entity) +ir_node *create_ldf(dbg_info *dbgi, ir_node *block, ir_node *ptr, + ir_node *mem, ir_mode *mode, ir_entity *entity, + long offset, bool is_frame_entity) { unsigned bits = get_mode_size_bits(mode); assert(mode_is_float(mode)); @@ -474,10 +663,10 @@ static ir_node *create_ldf(dbg_info *dbgi, ir_node *block, ir_node *ptr, } } -static ir_node *create_stf(dbg_info *dbgi, ir_node *block, ir_node *value, - ir_node *ptr, ir_node *mem, ir_mode *mode, - ir_entity *entity, long offset, - bool is_frame_entity) +ir_node *create_stf(dbg_info *dbgi, ir_node *block, ir_node *value, + ir_node *ptr, ir_node *mem, ir_mode *mode, + ir_entity *entity, long offset, + bool is_frame_entity) { unsigned bits = get_mode_size_bits(mode); assert(mode_is_float(mode)); @@ -511,6 +700,10 @@ static ir_node *gen_Load(ir_node *node) ir_node *new_load = NULL; address_t address; + if (get_Load_unaligned(node) == align_non_aligned) { + panic("sparc: transformation of unaligned Loads not implemented yet"); + } + if (mode_is_float(mode)) { match_address(ptr, &address, false); new_load = create_ldf(dbgi, block, address.ptr, new_mem, mode, @@ -551,12 +744,17 @@ static ir_node *gen_Store(ir_node *node) ir_node *new_store = NULL; address_t address; + if (get_Store_unaligned(node) == align_non_aligned) { + panic("sparc: transformation of unaligned Stores not implemented yet"); + } + if (mode_is_float(mode)) { /* TODO: variants with reg+reg address mode */ match_address(ptr, &address, false); new_store = create_stf(dbgi, block, new_val, address.ptr, new_mem, mode, address.entity, address.offset, false); } else { + assert(get_mode_size_bits(mode) <= 32); match_address(ptr, &address, true); if (address.ptr2 != NULL) { assert(address.entity == NULL && address.offset == 0); @@ -605,8 +803,13 @@ static ir_node *gen_Mulh(ir_node *node) if (mode_is_float(mode)) panic("FP not supported yet"); - mul = gen_helper_binop(node, MATCH_COMMUTATIVE, new_bd_sparc_Mulh_reg, new_bd_sparc_Mulh_imm); - return new_r_Proj(mul, mode_gp, pn_sparc_Mulh_low); + if (mode_is_signed(mode)) { + mul = gen_helper_binop(node, MATCH_COMMUTATIVE, new_bd_sparc_SMulh_reg, new_bd_sparc_SMulh_imm); + return new_r_Proj(mul, mode_gp, pn_sparc_SMulh_low); + } else { + mul = gen_helper_binop(node, MATCH_COMMUTATIVE, new_bd_sparc_UMulh_reg, new_bd_sparc_UMulh_imm); + return new_r_Proj(mul, mode_gp, pn_sparc_UMulh_low); + } } static ir_node *gen_sign_extension_value(ir_node *node) @@ -654,7 +857,8 @@ static ir_node *gen_Div(ir_node *node) new_right); } } else { - ir_node *left_high = get_g0(); + ir_graph *irg = get_irn_irg(node); + ir_node *left_high = get_g0(irg); if (is_imm_encodeable(right)) { int32_t immediate = get_tarval_long(get_Const_tarval(right)); res = new_bd_sparc_UDiv_imm(dbgi, new_block, left_high, left_low, @@ -669,27 +873,6 @@ static ir_node *gen_Div(ir_node *node) return res; } -#if 0 -static ir_node *gen_Abs(ir_node *node) -{ - ir_mode *const mode = get_irn_mode(node); - - if (mode_is_float(mode)) { - return gen_helper_unfpop(node, mode, new_bd_sparc_fabs_s, - new_bd_sparc_fabs_d, new_bd_sparc_fabs_q); - } else { - ir_node *const block = be_transform_node(get_nodes_block(node)); - dbg_info *const dbgi = get_irn_dbg_info(node); - ir_node *const op = get_Abs_op(node); - ir_node *const new_op = be_transform_node(op); - ir_node *const sra = new_bd_sparc_Sra_imm(dbgi, block, new_op, NULL, 31); - ir_node *const xor = new_bd_sparc_Xor_reg(dbgi, block, new_op, sra); - ir_node *const sub = new_bd_sparc_Sub_reg(dbgi, block, xor, sra); - return sub; - } -} -#endif - /** * Transforms a Not node. * @@ -698,7 +881,8 @@ static ir_node *gen_Abs(ir_node *node) static ir_node *gen_Not(ir_node *node) { ir_node *op = get_Not_op(node); - ir_node *zero = get_g0(); + ir_graph *irg = get_irn_irg(node); + ir_node *zero = get_g0(irg); dbg_info *dbgi = get_irn_dbg_info(node); ir_node *block = be_transform_node(get_nodes_block(node)); ir_node *new_op = be_transform_node(op); @@ -714,22 +898,23 @@ static ir_node *gen_helper_bitop(ir_node *node, new_binop_reg_func new_reg, new_binop_imm_func new_imm, new_binop_reg_func new_not_reg, - new_binop_imm_func new_not_imm) + new_binop_imm_func new_not_imm, + match_flags_t flags) { ir_node *op1 = get_binop_left(node); ir_node *op2 = get_binop_right(node); if (is_Not(op1)) { return gen_helper_binop_args(node, op2, get_Not_op(op1), - MATCH_MODE_NEUTRAL, + flags, new_not_reg, new_not_imm); } if (is_Not(op2)) { return gen_helper_binop_args(node, op1, get_Not_op(op2), - MATCH_MODE_NEUTRAL, + flags, new_not_reg, new_not_imm); } return gen_helper_binop_args(node, op1, op2, - MATCH_MODE_NEUTRAL | MATCH_COMMUTATIVE, + flags | MATCH_COMMUTATIVE, new_reg, new_imm); } @@ -739,7 +924,8 @@ static ir_node *gen_And(ir_node *node) new_bd_sparc_And_reg, new_bd_sparc_And_imm, new_bd_sparc_AndN_reg, - new_bd_sparc_AndN_imm); + new_bd_sparc_AndN_imm, + MATCH_MODE_NEUTRAL); } static ir_node *gen_Or(ir_node *node) @@ -748,7 +934,8 @@ static ir_node *gen_Or(ir_node *node) new_bd_sparc_Or_reg, new_bd_sparc_Or_imm, new_bd_sparc_OrN_reg, - new_bd_sparc_OrN_imm); + new_bd_sparc_OrN_imm, + MATCH_MODE_NEUTRAL); } static ir_node *gen_Eor(ir_node *node) @@ -757,21 +944,31 @@ static ir_node *gen_Eor(ir_node *node) new_bd_sparc_Xor_reg, new_bd_sparc_Xor_imm, new_bd_sparc_XNor_reg, - new_bd_sparc_XNor_imm); + new_bd_sparc_XNor_imm, + MATCH_MODE_NEUTRAL); } static ir_node *gen_Shl(ir_node *node) { + ir_mode *mode = get_irn_mode(node); + if (get_mode_modulo_shift(mode) != 32) + panic("modulo_shift!=32 not supported by sparc backend"); return gen_helper_binop(node, MATCH_NONE, new_bd_sparc_Sll_reg, new_bd_sparc_Sll_imm); } static ir_node *gen_Shr(ir_node *node) { + ir_mode *mode = get_irn_mode(node); + if (get_mode_modulo_shift(mode) != 32) + panic("modulo_shift!=32 not supported by sparc backend"); return gen_helper_binop(node, MATCH_NONE, new_bd_sparc_Srl_reg, new_bd_sparc_Srl_imm); } static ir_node *gen_Shrs(ir_node *node) { + ir_mode *mode = get_irn_mode(node); + if (get_mode_modulo_shift(mode) != 32) + panic("modulo_shift!=32 not supported by sparc backend"); return gen_helper_binop(node, MATCH_NONE, new_bd_sparc_Sra_reg, new_bd_sparc_Sra_imm); } @@ -795,7 +992,7 @@ static ir_node *gen_Minus(ir_node *node) dbgi = get_irn_dbg_info(node); op = get_Minus_op(node); new_op = be_transform_node(op); - zero = get_g0(); + zero = get_g0(get_irn_irg(node)); return new_bd_sparc_Sub_reg(dbgi, block, zero, new_op); } @@ -833,7 +1030,7 @@ static ir_node *gen_float_const(dbg_info *dbgi, ir_node *block, ir_tarval *tv) { ir_entity *entity = create_float_const_entity(tv); ir_node *hi = new_bd_sparc_SetHi(dbgi, block, entity, 0); - ir_node *mem = new_r_NoMem(current_ir_graph); + ir_node *mem = get_irg_no_mem(current_ir_graph); ir_mode *mode = get_tarval_mode(tv); ir_node *new_op = create_ldf(dbgi, block, hi, mem, mode, entity, 0, false); @@ -857,9 +1054,10 @@ static ir_node *gen_Const(ir_node *node) value = get_tarval_long(tv); if (value == 0) { - return get_g0(); + return get_g0(get_irn_irg(node)); } else if (sparc_is_value_imm_encodeable(value)) { - return new_bd_sparc_Or_imm(dbgi, block, get_g0(), NULL, value); + ir_graph *irg = get_irn_irg(node); + return new_bd_sparc_Or_imm(dbgi, block, get_g0(irg), NULL, value); } else { ir_node *hi = new_bd_sparc_SetHi(dbgi, block, NULL, value); if ((value & 0x3ff) != 0) { @@ -880,14 +1078,6 @@ static ir_mode *get_cmp_mode(ir_node *b_value) return get_irn_mode(op); } -static ir_node *make_address(dbg_info *dbgi, ir_node *block, ir_entity *entity, - int32_t offset) -{ - ir_node *hi = new_bd_sparc_SetHi(dbgi, block, entity, offset); - ir_node *low = new_bd_sparc_Or_imm(dbgi, block, hi, entity, offset); - return low; -} - static ir_node *gen_SwitchJmp(ir_node *node) { dbg_info *dbgi = get_irn_dbg_info(node); @@ -897,7 +1087,7 @@ static ir_node *gen_SwitchJmp(ir_node *node) long default_pn = get_Cond_default_proj(node); ir_entity *entity; ir_node *table_address; - ir_node *index; + ir_node *idx; ir_node *load; ir_node *address; @@ -908,18 +1098,13 @@ static ir_node *gen_SwitchJmp(ir_node *node) set_entity_visibility(entity, ir_visibility_private); add_entity_linkage(entity, IR_LINKAGE_CONSTANT); - /* TODO: this code does not construct code to check for access - * out-of bounds of the jumptable yet. I think we should put this stuff - * into the switch_lowering phase to get some additional optimisations - * done. */ - /* construct base address */ table_address = make_address(dbgi, block, entity, 0); /* scale index */ - index = new_bd_sparc_Sll_imm(dbgi, block, new_selector, NULL, 2); + idx = new_bd_sparc_Sll_imm(dbgi, block, new_selector, NULL, 2); /* load from jumptable */ - load = new_bd_sparc_Ld_reg(dbgi, block, table_address, index, - new_r_NoMem(current_ir_graph), + load = new_bd_sparc_Ld_reg(dbgi, block, table_address, idx, + get_irg_no_mem(current_ir_graph), mode_gp); address = new_r_Proj(load, mode_gp, pn_sparc_Ld_res); @@ -930,33 +1115,49 @@ static ir_node *gen_Cond(ir_node *node) { ir_node *selector = get_Cond_selector(node); ir_mode *mode = get_irn_mode(selector); - ir_mode *cmp_mode; ir_node *block; ir_node *flag_node; bool is_unsigned; ir_relation relation; dbg_info *dbgi; - // switch/case jumps + /* switch/case jumps */ if (mode != mode_b) { return gen_SwitchJmp(node); } - // regular if/else jumps - assert(is_Cmp(selector)); + block = be_transform_node(get_nodes_block(node)); + dbgi = get_irn_dbg_info(node); - cmp_mode = get_cmp_mode(selector); + /* regular if/else jumps */ + if (is_Cmp(selector)) { + ir_mode *cmp_mode; - block = be_transform_node(get_nodes_block(node)); - dbgi = get_irn_dbg_info(node); - flag_node = be_transform_node(selector); - relation = get_Cmp_relation(selector); - is_unsigned = !mode_is_signed(cmp_mode); - if (mode_is_float(cmp_mode)) { - assert(!is_unsigned); - return new_bd_sparc_fbfcc(dbgi, block, flag_node, relation); + cmp_mode = get_cmp_mode(selector); + flag_node = be_transform_node(selector); + relation = get_Cmp_relation(selector); + is_unsigned = !mode_is_signed(cmp_mode); + if (mode_is_float(cmp_mode)) { + assert(!is_unsigned); + return new_bd_sparc_fbfcc(dbgi, block, flag_node, relation); + } else { + return new_bd_sparc_Bicc(dbgi, block, flag_node, relation, is_unsigned); + } } else { - return new_bd_sparc_Bicc(dbgi, block, flag_node, relation, is_unsigned); + /* in this case, the selector must already deliver a mode_b value. + * this happens, for example, when the Cond is connected to a Conv + * which converts its argument to mode_b. */ + ir_node *new_op; + ir_graph *irg; + assert(mode == mode_b); + + block = be_transform_node(get_nodes_block(node)); + irg = get_irn_irg(block); + dbgi = get_irn_dbg_info(node); + new_op = be_transform_node(selector); + /* follow the SPARC architecture manual and use orcc for tst */ + flag_node = new_bd_sparc_OrCCZero_reg(dbgi, block, new_op, get_g0(irg)); + return new_bd_sparc_Bicc(dbgi, block, flag_node, ir_relation_less_greater, true); } } @@ -996,19 +1197,22 @@ static ir_node *gen_Cmp(ir_node *node) new_bd_sparc_AndCCZero_reg, new_bd_sparc_AndCCZero_imm, new_bd_sparc_AndNCCZero_reg, - new_bd_sparc_AndNCCZero_imm); + new_bd_sparc_AndNCCZero_imm, + MATCH_NONE); } else if (is_Or(op1)) { return gen_helper_bitop(op1, new_bd_sparc_OrCCZero_reg, new_bd_sparc_OrCCZero_imm, new_bd_sparc_OrNCCZero_reg, - new_bd_sparc_OrNCCZero_imm); + new_bd_sparc_OrNCCZero_imm, + MATCH_NONE); } else if (is_Eor(op1)) { return gen_helper_bitop(op1, new_bd_sparc_XorCCZero_reg, new_bd_sparc_XorCCZero_imm, new_bd_sparc_XNorCCZero_reg, - new_bd_sparc_XNorCCZero_imm); + new_bd_sparc_XNorCCZero_imm, + MATCH_NONE); } } @@ -1076,7 +1280,7 @@ static ir_node *create_ftoi(dbg_info *dbgi, ir_node *block, ir_node *op, { ir_graph *irg = get_irn_irg(block); ir_node *sp = get_irg_frame(irg); - ir_node *nomem = new_r_NoMem(irg); + ir_node *nomem = get_irg_no_mem(irg); ir_node *stf = create_stf(dbgi, block, ftoi, sp, nomem, src_mode, NULL, 0, true); ir_node *ld = new_bd_sparc_Ld_imm(dbgi, block, sp, stf, mode_gp, @@ -1093,7 +1297,7 @@ static ir_node *create_itof(dbg_info *dbgi, ir_node *block, ir_node *op, { ir_graph *irg = get_irn_irg(block); ir_node *sp = get_irg_frame(irg); - ir_node *nomem = new_r_NoMem(irg); + ir_node *nomem = get_irg_no_mem(irg); ir_node *st = new_bd_sparc_St_imm(dbgi, block, op, sp, nomem, mode_gp, NULL, 0, true); ir_node *ldf = new_bd_sparc_Ldf_s(dbgi, block, sp, st, mode_fp, @@ -1119,7 +1323,7 @@ static ir_node *gen_Conv(ir_node *node) ir_node *op = get_Conv_op(node); ir_mode *src_mode = get_irn_mode(op); ir_mode *dst_mode = get_irn_mode(node); - dbg_info *dbg = get_irn_dbg_info(node); + dbg_info *dbgi = get_irn_dbg_info(node); ir_node *new_op; int src_bits = get_mode_size_bits(src_mode); @@ -1138,21 +1342,21 @@ static ir_node *gen_Conv(ir_node *node) if (mode_is_float(src_mode)) { if (mode_is_float(dst_mode)) { /* float -> float conv */ - return create_fftof(dbg, block, new_op, src_mode, dst_mode); + return create_fftof(dbgi, block, new_op, src_mode, dst_mode); } else { /* float -> int conv */ if (!mode_is_signed(dst_mode)) panic("float to unsigned not implemented yet"); - return create_ftoi(dbg, block, new_op, src_mode); + return create_ftoi(dbgi, block, new_op, src_mode); } } else { /* int -> float conv */ if (src_bits < 32) { - new_op = gen_extension(dbg, block, new_op, src_mode); + new_op = gen_extension(dbgi, block, new_op, src_mode); } else if (src_bits == 32 && !mode_is_signed(src_mode)) { panic("unsigned to float not lowered!"); } - return create_itof(dbg, block, new_op, dst_mode); + return create_itof(dbgi, block, new_op, dst_mode); } } else if (src_mode == mode_b) { panic("ConvB not lowered %+F", node); @@ -1165,6 +1369,11 @@ static ir_node *gen_Conv(ir_node *node) return new_op; } + if (dst_mode == mode_b) { + /* mode_b lowering already took care that we only have 0/1 values */ + return new_op; + } + if (src_bits < dst_bits) { min_bits = src_bits; min_mode = src_mode; @@ -1178,9 +1387,9 @@ static ir_node *gen_Conv(ir_node *node) } if (mode_is_signed(min_mode)) { - return gen_sign_extension(dbg, block, new_op, min_bits); + return gen_sign_extension(dbgi, block, new_op, min_bits); } else { - return gen_zero_extension(dbg, block, new_op, min_bits); + return gen_zero_extension(dbgi, block, new_op, min_bits); } } } @@ -1193,85 +1402,15 @@ static ir_node *gen_Unknown(ir_node *node) ir_node *block = be_transform_node(get_nodes_block(node)); return gen_float_const(NULL, block, get_mode_null(mode)); } else if (mode_needs_gp_reg(mode)) { - return get_g0(); + ir_graph *irg = get_irn_irg(node); + return get_g0(irg); } panic("Unexpected Unknown mode"); } /** - * Produces the type which sits between the stack args and the locals on the - * stack. - */ -static ir_type *sparc_get_between_type(void) -{ - static ir_type *between_type = NULL; - static ir_type *between_type0 = NULL; - - if (cconv->omit_fp) { - if (between_type0 == NULL) { - between_type0 - = new_type_class(new_id_from_str("sparc_between_type")); - set_type_size_bytes(between_type0, 0); - } - return between_type0; - } - - if (between_type == NULL) { - between_type = new_type_class(new_id_from_str("sparc_between_type")); - set_type_size_bytes(between_type, SPARC_MIN_STACKSIZE); - } - - return between_type; -} - -static void create_stacklayout(ir_graph *irg) -{ - ir_entity *entity = get_irg_entity(irg); - ir_type *function_type = get_entity_type(entity); - be_stack_layout_t *layout = be_get_irg_stack_layout(irg); - ir_type *arg_type; - int p; - int n_params; - - /* calling conventions must be decided by now */ - assert(cconv != NULL); - - /* construct argument type */ - arg_type = new_type_struct(id_mangle_u(get_entity_ident(entity), new_id_from_chars("arg_type", 8))); - n_params = get_method_n_params(function_type); - for (p = 0; p < n_params; ++p) { - reg_or_stackslot_t *param = &cconv->parameters[p]; - char buf[128]; - ident *id; - - if (param->type == NULL) - continue; - - snprintf(buf, sizeof(buf), "param_%d", p); - id = new_id_from_str(buf); - param->entity = new_entity(arg_type, id, param->type); - set_entity_offset(param->entity, param->offset); - } - - memset(layout, 0, sizeof(*layout)); - - layout->frame_type = get_irg_frame_type(irg); - layout->between_type = sparc_get_between_type(); - layout->arg_type = arg_type; - layout->initial_offset = 0; - layout->initial_bias = 0; - layout->stack_dir = -1; - layout->sp_relative = cconv->omit_fp; - - assert(N_FRAME_TYPES == 3); - layout->order[0] = layout->frame_type; - layout->order[1] = layout->between_type; - layout->order[2] = layout->arg_type; -} - -/** - * transform the start node to the prolog code + initial barrier + * transform the start node to the prolog code */ static ir_node *gen_Start(ir_node *node) { @@ -1281,77 +1420,145 @@ static ir_node *gen_Start(ir_node *node) ir_node *block = get_nodes_block(node); ir_node *new_block = be_transform_node(block); dbg_info *dbgi = get_irn_dbg_info(node); - ir_node *mem; + struct obstack *obst = be_get_be_obst(irg); + const arch_register_req_t *req; + size_t n_outs; ir_node *start; - ir_node *sp; size_t i; + size_t o; + + /* start building list of start constraints */ + assert(obstack_object_size(obst) == 0); + + /* calculate number of outputs */ + n_outs = 4; /* memory, g0, g7, sp */ + if (!current_cconv->omit_fp) + ++n_outs; /* framepointer */ + /* function parameters */ + n_outs += current_cconv->n_param_regs; + /* callee saves */ + if (current_cconv->omit_fp) { + n_outs += ARRAY_SIZE(omit_fp_callee_saves); + } + + start = new_bd_sparc_Start(dbgi, new_block, n_outs); + + o = 0; - /* stackpointer is important at function prolog */ - be_prolog_add_reg(abihelper, sp_reg, + /* first output is memory */ + start_mem_offset = o; + arch_set_irn_register_req_out(start, o, arch_no_register_req); + ++o; + + /* the zero register */ + start_g0_offset = o; + req = be_create_reg_req(obst, &sparc_registers[REG_G0], + arch_register_req_type_ignore); + arch_set_irn_register_req_out(start, o, req); + arch_set_irn_register_out(start, o, &sparc_registers[REG_G0]); + ++o; + + /* g7 is used for TLS data */ + start_g7_offset = o; + req = be_create_reg_req(obst, &sparc_registers[REG_G7], + arch_register_req_type_ignore); + arch_set_irn_register_req_out(start, o, req); + arch_set_irn_register_out(start, o, &sparc_registers[REG_G7]); + ++o; + + /* we need an output for the stackpointer */ + start_sp_offset = o; + req = be_create_reg_req(obst, sp_reg, arch_register_req_type_produces_sp | arch_register_req_type_ignore); - be_prolog_add_reg(abihelper, &sparc_registers[REG_G0], - arch_register_req_type_ignore); + arch_set_irn_register_req_out(start, o, req); + arch_set_irn_register_out(start, o, sp_reg); + ++o; + + if (!current_cconv->omit_fp) { + start_fp_offset = o; + req = be_create_reg_req(obst, fp_reg, arch_register_req_type_ignore); + arch_set_irn_register_req_out(start, o, req); + arch_set_irn_register_out(start, o, fp_reg); + ++o; + } + /* function parameters in registers */ + start_params_offset = o; for (i = 0; i < get_method_n_params(function_type); ++i) { - const reg_or_stackslot_t *param = &cconv->parameters[i]; - if (param->reg0 != NULL) { - be_prolog_add_reg(abihelper, param->reg0, - arch_register_req_type_none); + const reg_or_stackslot_t *param = ¤t_cconv->parameters[i]; + const arch_register_t *reg0 = param->reg0; + const arch_register_t *reg1 = param->reg1; + if (reg0 != NULL) { + arch_set_irn_register_req_out(start, o, reg0->single_req); + arch_set_irn_register_out(start, o, reg0); + ++o; } - if (param->reg1 != NULL) { - be_prolog_add_reg(abihelper, param->reg1, - arch_register_req_type_none); + if (reg1 != NULL) { + arch_set_irn_register_req_out(start, o, reg1->single_req); + arch_set_irn_register_out(start, o, reg1); + ++o; } } /* we need the values of the callee saves (Note: non omit-fp mode has no * callee saves) */ - if (cconv->omit_fp) { + start_callee_saves_offset = o; + if (current_cconv->omit_fp) { size_t n_callee_saves = ARRAY_SIZE(omit_fp_callee_saves); size_t c; for (c = 0; c < n_callee_saves; ++c) { - be_prolog_add_reg(abihelper, omit_fp_callee_saves[c], - arch_register_req_type_none); + const arch_register_t *reg = omit_fp_callee_saves[c]; + arch_set_irn_register_req_out(start, o, reg->single_req); + arch_set_irn_register_out(start, o, reg); + ++o; } - } else { - be_prolog_add_reg(abihelper, fp_reg, arch_register_req_type_ignore); } + assert(n_outs == o); - start = be_prolog_create_start(abihelper, dbgi, new_block); - mem = be_prolog_get_memory(abihelper); - sp = be_prolog_get_reg_value(abihelper, sp_reg); + return start; +} - if (!cconv->omit_fp) { - ir_node *save = new_bd_sparc_Save_imm(NULL, block, sp, NULL, - -SPARC_MIN_STACKSIZE); - arch_irn_add_flags(save, arch_irn_flags_prolog); - arch_set_irn_register(save, sp_reg); - sp = save; +static ir_node *get_initial_sp(ir_graph *irg) +{ + if (start_sp == NULL) { + ir_node *start = get_irg_start(irg); + start_sp = new_r_Proj(start, mode_gp, start_sp_offset); } + return start_sp; +} - sp = be_new_IncSP(sp_reg, new_block, sp, BE_STACK_FRAME_SIZE_EXPAND, 0); - arch_irn_add_flags(sp, arch_irn_flags_prolog); - be_prolog_set_reg_value(abihelper, sp_reg, sp); - be_prolog_set_memory(abihelper, mem); +static ir_node *get_initial_fp(ir_graph *irg) +{ + if (start_fp == NULL) { + ir_node *start = get_irg_start(irg); + start_fp = new_r_Proj(start, mode_gp, start_fp_offset); + } + return start_fp; +} - return start; +static ir_node *get_initial_mem(ir_graph *irg) +{ + if (start_mem == NULL) { + ir_node *start = get_irg_start(irg); + start_mem = new_r_Proj(start, mode_M, start_mem_offset); + } + return start_mem; } static ir_node *get_stack_pointer_for(ir_node *node) { /* get predecessor in stack_order list */ - ir_node *stack_pred = be_get_stack_pred(abihelper, node); - ir_node *stack_pred_transformed; + ir_node *stack_pred = be_get_stack_pred(stackorder, node); ir_node *stack; if (stack_pred == NULL) { /* first stack user in the current block. We can simply use the * initial sp_proj for it */ - ir_node *sp_proj = be_prolog_get_reg_value(abihelper, sp_reg); - return sp_proj; + ir_graph *irg = get_irn_irg(node); + return get_initial_sp(irg); } - stack_pred_transformed = be_transform_node(stack_pred); - stack = (ir_node*)pmap_get(node_to_stack, stack_pred); + be_transform_node(stack_pred); + stack = (ir_node*)pmap_get(node_to_stack, stack_pred); if (stack == NULL) { return get_stack_pointer_for(stack_pred); } @@ -1365,64 +1572,66 @@ static ir_node *get_stack_pointer_for(ir_node *node) static ir_node *gen_Return(ir_node *node) { ir_node *block = get_nodes_block(node); + ir_graph *irg = get_irn_irg(node); ir_node *new_block = be_transform_node(block); dbg_info *dbgi = get_irn_dbg_info(node); ir_node *mem = get_Return_mem(node); ir_node *new_mem = be_transform_node(mem); ir_node *sp = get_stack_pointer_for(node); size_t n_res = get_Return_n_ress(node); - ir_node *barrier; + struct obstack *be_obst = be_get_be_obst(irg); ir_node *bereturn; + ir_node **in; + const arch_register_req_t **reqs; size_t i; + size_t p; + size_t n_ins; + + /* estimate number of return values */ + n_ins = 2 + n_res; /* memory + stackpointer, return values */ + if (current_cconv->omit_fp) + n_ins += ARRAY_SIZE(omit_fp_callee_saves); + + in = ALLOCAN(ir_node*, n_ins); + reqs = OALLOCN(be_obst, const arch_register_req_t*, n_ins); + p = 0; + + in[p] = new_mem; + reqs[p] = arch_no_register_req; + ++p; - be_epilog_begin(abihelper); - be_epilog_set_memory(abihelper, new_mem); - /* connect stack pointer with initial stack pointer. fix_stack phase - will later serialize all stack pointer adjusting nodes */ - be_epilog_add_reg(abihelper, sp_reg, - arch_register_req_type_produces_sp | arch_register_req_type_ignore, - sp); + in[p] = sp; + reqs[p] = sp_reg->single_req; + ++p; /* result values */ for (i = 0; i < n_res; ++i) { ir_node *res_value = get_Return_res(node, i); ir_node *new_res_value = be_transform_node(res_value); - const reg_or_stackslot_t *slot = &cconv->results[i]; - const arch_register_t *reg = slot->reg0; - assert(slot->reg1 == NULL); - be_epilog_add_reg(abihelper, reg, arch_register_req_type_none, - new_res_value); + const reg_or_stackslot_t *slot = ¤t_cconv->results[i]; + assert(slot->req1 == NULL); + in[p] = new_res_value; + reqs[p] = slot->req0; + ++p; } /* callee saves */ - if (cconv->omit_fp) { - size_t n_callee_saves = ARRAY_SIZE(omit_fp_callee_saves); - size_t i; + if (current_cconv->omit_fp) { + ir_node *start = get_irg_start(irg); + size_t n_callee_saves = ARRAY_SIZE(omit_fp_callee_saves); for (i = 0; i < n_callee_saves; ++i) { const arch_register_t *reg = omit_fp_callee_saves[i]; + ir_mode *mode = reg->reg_class->mode; ir_node *value - = be_prolog_get_reg_value(abihelper, reg); - be_epilog_add_reg(abihelper, reg, arch_register_req_type_none, - value); + = new_r_Proj(start, mode, i + start_callee_saves_offset); + in[p] = value; + reqs[p] = reg->single_req; + ++p; } } + assert(p == n_ins); - /* we need a restore instruction */ - if (!cconv->omit_fp) { - ir_node *restore = new_bd_sparc_RestoreZero(NULL, block); - arch_irn_add_flags(restore, arch_irn_flags_epilog); - add_irn_dep(restore, barrier); - arch_set_irn_register(restore, sp_reg); - be_epilog_set_reg_value(abihelper, sp_reg, restore); - } else { - /* epilog code: an incsp */ - sp = be_epilog_get_reg_value(abihelper, sp_reg); - sp = be_new_IncSP(sp_reg, new_block, sp, - BE_STACK_FRAME_SIZE_SHRINK, 0); - be_epilog_set_reg_value(abihelper, sp_reg, sp); - } - - bereturn = be_epilog_create_return(abihelper, dbgi, new_block); - arch_irn_add_flags(bereturn, arch_irn_flags_epilog); + bereturn = new_bd_sparc_Return_reg(dbgi, new_block, n_ins, in); + arch_set_irn_register_reqs_in(bereturn, reqs); return bereturn; } @@ -1432,7 +1641,7 @@ static ir_node *bitcast_int_to_float(dbg_info *dbgi, ir_node *block, { ir_graph *irg = current_ir_graph; ir_node *sp = get_irg_frame(irg); - ir_node *nomem = new_r_NoMem(irg); + ir_node *nomem = get_irg_no_mem(irg); ir_node *st = new_bd_sparc_St_imm(dbgi, block, value0, sp, nomem, mode_gp, NULL, 0, true); ir_mode *mode; @@ -1465,7 +1674,7 @@ static void bitcast_float_to_int(dbg_info *dbgi, ir_node *block, { ir_graph *irg = current_ir_graph; ir_node *stack = get_irg_frame(irg); - ir_node *nomem = new_r_NoMem(irg); + ir_node *nomem = get_irg_no_mem(irg); ir_node *stf = create_stf(dbgi, block, node, stack, nomem, float_mode, NULL, 0, true); int bits = get_mode_size_bits(float_mode); @@ -1482,8 +1691,8 @@ static void bitcast_float_to_int(dbg_info *dbgi, ir_node *block, set_irn_pinned(ld, op_pin_state_floats); result[1] = new_r_Proj(ld2, mode_gp, pn_sparc_Ld_res); - arch_irn_add_flags(ld, (arch_irn_flags_t)sparc_arch_irn_flag_needs_64bit_spillslot); - arch_irn_add_flags(ld2, (arch_irn_flags_t)sparc_arch_irn_flag_needs_64bit_spillslot); + arch_add_irn_flags(ld, (arch_irn_flags_t)sparc_arch_irn_flag_needs_64bit_spillslot); + arch_add_irn_flags(ld2, (arch_irn_flags_t)sparc_arch_irn_flag_needs_64bit_spillslot); } else { assert(bits == 32); result[1] = NULL; @@ -1501,26 +1710,31 @@ static ir_node *gen_Call(ir_node *node) dbg_info *dbgi = get_irn_dbg_info(node); ir_type *type = get_Call_type(node); size_t n_params = get_Call_n_params(node); - size_t n_param_regs = sizeof(param_regs)/sizeof(param_regs[0]); + size_t n_ress = get_method_n_ress(type); /* max inputs: memory, callee, register arguments */ - int max_inputs = 2 + n_param_regs; - ir_node **in = ALLOCAN(ir_node*, max_inputs); ir_node **sync_ins = ALLOCAN(ir_node*, n_params); struct obstack *obst = be_get_be_obst(irg); - const arch_register_req_t **in_req - = OALLOCNZ(obst, const arch_register_req_t*, max_inputs); calling_convention_t *cconv = sparc_decide_calling_convention(type, NULL); + size_t n_param_regs = cconv->n_param_regs; + /* param-regs + mem + stackpointer + callee */ + unsigned max_inputs = 3 + n_param_regs; + ir_node **in = ALLOCAN(ir_node*, max_inputs); + const arch_register_req_t **in_req + = OALLOCNZ(obst, const arch_register_req_t*, max_inputs); int in_arity = 0; int sync_arity = 0; int n_caller_saves - = sizeof(caller_saves)/sizeof(caller_saves[0]); + = rbitset_popcount(cconv->caller_saves, N_SPARC_REGISTERS); ir_entity *entity = NULL; ir_node *new_frame = get_stack_pointer_for(node); + bool aggregate_return + = get_method_calling_convention(type) & cc_compound_ret; ir_node *incsp; int mem_pos; ir_node *res; size_t p; + size_t r; int i; int o; int out_arity; @@ -1552,6 +1766,7 @@ static ir_node *gen_Call(ir_node *node) ir_mode *mode = get_type_mode(param_type); ir_node *new_values[2]; ir_node *str; + int offset; if (mode_is_float(mode) && param->reg0 != NULL) { unsigned size_bits = get_mode_size_bits(mode); @@ -1584,18 +1799,20 @@ static ir_node *gen_Call(ir_node *node) mode = mode_gp; } - /* create a parameter frame if necessary */ + /* we need to skip over our save area when constructing the call + * arguments on stack */ + offset = param->offset + SPARC_MIN_STACKSIZE; + if (mode_is_float(mode)) { str = create_stf(dbgi, new_block, new_value, incsp, new_mem, - mode, NULL, param->offset, true); + mode, NULL, offset, true); } else { str = new_bd_sparc_St_imm(dbgi, new_block, new_value, incsp, - new_mem, mode, NULL, param->offset, true); + new_mem, mode, NULL, offset, true); } set_irn_pinned(str, op_pin_state_floats); sync_ins[sync_arity++] = str; } - assert(in_arity <= max_inputs); /* construct memory input */ if (sync_arity == 0) { @@ -1613,28 +1830,43 @@ static ir_node *gen_Call(ir_node *node) in_req[in_arity] = sparc_reg_classes[CLASS_sparc_gp].class_req; ++in_arity; } + assert(in_arity <= (int)max_inputs); /* outputs: * - memory + * - results * - caller saves */ - out_arity = 1 + n_caller_saves; + out_arity = 1 + cconv->n_reg_results + n_caller_saves; /* create call node */ if (entity != NULL) { res = new_bd_sparc_Call_imm(dbgi, new_block, in_arity, in, out_arity, - entity, 0); + entity, 0, aggregate_return); } else { - res = new_bd_sparc_Call_reg(dbgi, new_block, in_arity, in, out_arity); + res = new_bd_sparc_Call_reg(dbgi, new_block, in_arity, in, out_arity, + aggregate_return); } - arch_set_in_register_reqs(res, in_req); + arch_set_irn_register_reqs_in(res, in_req); /* create output register reqs */ o = 0; - arch_set_out_register_req(res, o++, arch_no_register_req); - for (i = 0; i < n_caller_saves; ++i) { - const arch_register_t *reg = caller_saves[i]; - arch_set_out_register_req(res, o++, reg->single_req); + arch_set_irn_register_req_out(res, o++, arch_no_register_req); + /* add register requirements for the result regs */ + for (r = 0; r < n_ress; ++r) { + const reg_or_stackslot_t *result_info = &cconv->results[r]; + const arch_register_req_t *req = result_info->req0; + if (req != NULL) { + arch_set_irn_register_req_out(res, o++, req); + } + assert(result_info->req1 == NULL); + } + for (i = 0; i < N_SPARC_REGISTERS; ++i) { + const arch_register_t *reg; + if (!rbitset_is_set(cconv->caller_saves, i)) + continue; + reg = &sparc_registers[i]; + arch_set_irn_register_req_out(res, o++, reg->single_req); } assert(o == out_arity); @@ -1667,14 +1899,106 @@ static ir_node *gen_Sel(ir_node *node) /* must be the frame pointer all other sels must have been lowered * already */ assert(is_Proj(ptr) && is_Start(get_Proj_pred(ptr))); - /* we should not have value types from parameters anymore - they should be - lowered */ - assert(get_entity_owner(entity) != - get_method_value_param_type(get_entity_type(get_irg_entity(get_irn_irg(node))))); return new_bd_sparc_FrameAddr(dbgi, new_block, new_ptr, entity, 0); } +static ir_node *gen_Alloc(ir_node *node) +{ + dbg_info *dbgi = get_irn_dbg_info(node); + ir_node *block = get_nodes_block(node); + ir_node *new_block = be_transform_node(block); + ir_type *type = get_Alloc_type(node); + ir_node *size = get_Alloc_count(node); + ir_node *stack_pred = get_stack_pointer_for(node); + ir_node *subsp; + if (get_Alloc_where(node) != stack_alloc) + panic("only stack-alloc supported in sparc backend (at %+F)", node); + /* lowerer should have transformed all allocas to byte size */ + if (type != get_unknown_type() && get_type_size_bytes(type) != 1) + panic("Found non-byte alloc in sparc backend (at %+F)", node); + + if (is_Const(size)) { + ir_tarval *tv = get_Const_tarval(size); + long sizel = get_tarval_long(tv); + subsp = be_new_IncSP(sp_reg, new_block, stack_pred, sizel, 0); + set_irn_dbg_info(subsp, dbgi); + } else { + ir_node *new_size = be_transform_node(size); + subsp = new_bd_sparc_SubSP(dbgi, new_block, stack_pred, new_size); + arch_set_irn_register(subsp, sp_reg); + } + + /* if we are the last IncSP producer in a block then we have to keep + * the stack value. + * Note: This here keeps all producers which is more than necessary */ + keep_alive(subsp); + + pmap_insert(node_to_stack, node, subsp); + /* the "result" is the unmodified sp value */ + return stack_pred; +} + +static ir_node *gen_Proj_Alloc(ir_node *node) +{ + ir_node *alloc = get_Proj_pred(node); + long pn = get_Proj_proj(node); + + switch ((pn_Alloc)pn) { + case pn_Alloc_M: { + ir_node *alloc_mem = get_Alloc_mem(alloc); + return be_transform_node(alloc_mem); + } + case pn_Alloc_res: { + ir_node *new_alloc = be_transform_node(alloc); + return new_alloc; + } + case pn_Alloc_X_regular: + case pn_Alloc_X_except: + panic("sparc backend: exception output of alloc not supported (at %+F)", + node); + } + panic("sparc backend: invalid Proj->Alloc"); +} + +static ir_node *gen_Free(ir_node *node) +{ + dbg_info *dbgi = get_irn_dbg_info(node); + ir_node *block = get_nodes_block(node); + ir_node *new_block = be_transform_node(block); + ir_type *type = get_Free_type(node); + ir_node *size = get_Free_count(node); + ir_node *mem = get_Free_mem(node); + ir_node *new_mem = be_transform_node(mem); + ir_node *stack_pred = get_stack_pointer_for(node); + ir_node *addsp; + if (get_Alloc_where(node) != stack_alloc) + panic("only stack-alloc supported in sparc backend (at %+F)", node); + /* lowerer should have transformed all allocas to byte size */ + if (type != get_unknown_type() && get_type_size_bytes(type) != 1) + panic("Found non-byte alloc in sparc backend (at %+F)", node); + + if (is_Const(size)) { + ir_tarval *tv = get_Const_tarval(size); + long sizel = get_tarval_long(tv); + addsp = be_new_IncSP(sp_reg, new_block, stack_pred, -sizel, 0); + set_irn_dbg_info(addsp, dbgi); + } else { + ir_node *new_size = be_transform_node(size); + addsp = new_bd_sparc_AddSP(dbgi, new_block, stack_pred, new_size); + arch_set_irn_register(addsp, sp_reg); + } + + /* if we are the last IncSP producer in a block then we have to keep + * the stack value. + * Note: This here keeps all producers which is more than necessary */ + keep_alive(addsp); + + pmap_insert(node_to_stack, node, addsp); + /* the "result" is the unmodified sp value */ + return new_mem; +} + static const arch_register_req_t float1_req = { arch_register_req_type_normal, &sparc_reg_classes[CLASS_sparc_fp], @@ -1746,7 +2070,7 @@ static ir_node *gen_Phi(ir_node *node) phi = new_ir_node(dbgi, irg, block, op_Phi, mode, get_irn_arity(node), get_irn_in(node) + 1); copy_node_attr(irg, node, phi); be_duplicate_deps(node, phi); - arch_set_out_register_req(phi, 0, req); + arch_set_irn_register_req_out(phi, 0, req); be_enqueue_preds(node); return phi; } @@ -1773,7 +2097,10 @@ static ir_node *gen_Proj_Load(ir_node *node) break; case iro_sparc_Ldf: if (pn == pn_Load_res) { - return new_rd_Proj(dbgi, new_load, mode_fp, pn_sparc_Ldf_res); + const sparc_load_store_attr_t *attr + = get_sparc_load_store_attr_const(new_load); + ir_mode *mode = attr->load_store_mode; + return new_rd_Proj(dbgi, new_load, mode, pn_sparc_Ldf_res); } else if (pn == pn_Load_M) { return new_rd_Proj(dbgi, new_load, mode_M, pn_sparc_Ld_M); } @@ -1825,16 +2152,23 @@ static ir_node *gen_Proj_Div(ir_node *node) ir_node *pred = get_Proj_pred(node); ir_node *new_pred = be_transform_node(pred); long pn = get_Proj_proj(node); + ir_mode *res_mode; - assert(is_sparc_SDiv(new_pred) || is_sparc_UDiv(new_pred) - || is_sparc_fdiv(new_pred)); + if (is_sparc_SDiv(new_pred) || is_sparc_UDiv(new_pred)) { + res_mode = mode_gp; + } else if (is_sparc_fdiv(new_pred)) { + res_mode = get_Div_resmode(pred); + } else { + panic("sparc backend: Div transformed to something unexpected: %+F", + new_pred); + } assert((int)pn_sparc_SDiv_res == (int)pn_sparc_UDiv_res); assert((int)pn_sparc_SDiv_M == (int)pn_sparc_UDiv_M); assert((int)pn_sparc_SDiv_res == (int)pn_sparc_fdiv_res); assert((int)pn_sparc_SDiv_M == (int)pn_sparc_fdiv_M); switch (pn) { case pn_Div_res: - return new_r_Proj(new_pred, mode_gp, pn_sparc_SDiv_res); + return new_r_Proj(new_pred, res_mode, pn_sparc_SDiv_res); case pn_Div_M: return new_r_Proj(new_pred, mode_gp, pn_sparc_SDiv_M); default: @@ -1843,10 +2177,16 @@ static ir_node *gen_Proj_Div(ir_node *node) panic("Unsupported Proj from Div"); } -static ir_node *get_frame_base(void) +static ir_node *get_frame_base(ir_graph *irg) { - const arch_register_t *reg = cconv->omit_fp ? sp_reg : fp_reg; - return be_prolog_get_reg_value(abihelper, reg); + if (frame_base == NULL) { + if (current_cconv->omit_fp) { + frame_base = get_initial_sp(irg); + } else { + frame_base = get_initial_fp(irg); + } + } + return frame_base; } static ir_node *gen_Proj_Start(ir_node *node) @@ -1861,50 +2201,62 @@ static ir_node *gen_Proj_Start(ir_node *node) case pn_Start_X_initial_exec: /* exchange ProjX with a jump */ return new_bd_sparc_Ba(NULL, new_block); - case pn_Start_M: - return be_prolog_get_memory(abihelper); + case pn_Start_M: { + ir_graph *irg = get_irn_irg(node); + return get_initial_mem(irg); + } case pn_Start_T_args: - /* we should never need this explicitely */ - return new_r_Bad(get_irn_irg(block)); + return new_r_Bad(get_irn_irg(block), mode_T); case pn_Start_P_frame_base: - return get_frame_base(); - case pn_Start_P_tls: - return new_r_Bad(current_ir_graph); - case pn_Start_max: - break; + return get_frame_base(get_irn_irg(block)); } panic("Unexpected start proj: %ld\n", pn); } static ir_node *gen_Proj_Proj_Start(ir_node *node) { - long pn = get_Proj_proj(node); - ir_node *block = get_nodes_block(node); - ir_node *new_block = be_transform_node(block); - ir_entity *entity = get_irg_entity(current_ir_graph); - ir_type *method_type = get_entity_type(entity); - ir_type *param_type = get_method_param_type(method_type, pn); + long pn = get_Proj_proj(node); + ir_node *block = get_nodes_block(node); + ir_graph *irg = get_irn_irg(node); + ir_node *new_block = be_transform_node(block); + ir_node *args = get_Proj_pred(node); + ir_node *start = get_Proj_pred(args); + ir_node *new_start = be_transform_node(start); const reg_or_stackslot_t *param; /* Proj->Proj->Start must be a method argument */ assert(get_Proj_proj(get_Proj_pred(node)) == pn_Start_T_args); - param = &cconv->parameters[pn]; + param = ¤t_cconv->parameters[pn]; if (param->reg0 != NULL) { /* argument transmitted in register */ - ir_mode *mode = get_type_mode(param_type); - const arch_register_t *reg = param->reg0; - ir_node *value = be_prolog_get_reg_value(abihelper, reg); + const arch_register_t *reg = param->reg0; + ir_mode *reg_mode = reg->reg_class->mode; + long new_pn = param->reg_offset + start_params_offset; + ir_node *value = new_r_Proj(new_start, reg_mode, new_pn); + bool is_float = false; + + { + ir_entity *entity = get_irg_entity(irg); + ir_type *method_type = get_entity_type(entity); + if (pn < (long)get_method_n_params(method_type)) { + ir_type *param_type = get_method_param_type(method_type, pn); + ir_mode *mode = get_type_mode(param_type); + is_float = mode_is_float(mode); + } + } - if (mode_is_float(mode)) { + if (is_float) { + const arch_register_t *reg1 = param->reg1; ir_node *value1 = NULL; - if (param->reg1 != NULL) { - value1 = be_prolog_get_reg_value(abihelper, param->reg1); + if (reg1 != NULL) { + ir_mode *reg1_mode = reg1->reg_class->mode; + value1 = new_r_Proj(new_start, reg1_mode, new_pn+1); } else if (param->entity != NULL) { - ir_node *fp = be_prolog_get_reg_value(abihelper, fp_reg); - ir_node *mem = be_prolog_get_memory(abihelper); + ir_node *fp = get_initial_fp(irg); + ir_node *mem = get_initial_mem(irg); ir_node *ld = new_bd_sparc_Ld_imm(NULL, new_block, fp, mem, mode_gp, param->entity, 0, true); @@ -1917,11 +2269,11 @@ static ir_node *gen_Proj_Proj_Start(ir_node *node) return value; } else { /* argument transmitted on stack */ - ir_node *mem = be_prolog_get_memory(abihelper); - ir_mode *mode = get_type_mode(param->type); - ir_node *base = get_frame_base(); - ir_node *load; - ir_node *value; + ir_node *mem = get_initial_mem(irg); + ir_mode *mode = get_type_mode(param->type); + ir_node *base = get_frame_base(irg); + ir_node *load; + ir_node *value; if (mode_is_float(mode)) { load = create_ldf(NULL, new_block, base, mem, mode, @@ -1950,30 +2302,11 @@ static ir_node *gen_Proj_Call(ir_node *node) case pn_Call_X_regular: case pn_Call_X_except: case pn_Call_T_result: - case pn_Call_P_value_res_base: - case pn_Call_max: break; } panic("Unexpected Call proj %ld\n", pn); } -/** - * Finds number of output value of a mode_T node which is constrained to - * a single specific register. - */ -static int find_out_for_reg(ir_node *node, const arch_register_t *reg) -{ - int n_outs = arch_irn_get_n_outs(node); - int o; - - for (o = 0; o < n_outs; ++o) { - const arch_register_req_t *req = arch_get_out_register_req(node, o); - if (req == reg->single_req) - return o; - } - return -1; -} - static ir_node *gen_Proj_Proj_Call(ir_node *node) { long pn = get_Proj_proj(node); @@ -1982,21 +2315,17 @@ static ir_node *gen_Proj_Proj_Call(ir_node *node) ir_type *function_type = get_Call_type(call); calling_convention_t *cconv = sparc_decide_calling_convention(function_type, NULL); - const reg_or_stackslot_t *res = &cconv->results[pn]; - const arch_register_t *reg = res->reg0; - ir_mode *mode; - int regn; + const reg_or_stackslot_t *res = &cconv->results[pn]; + ir_mode *mode = get_irn_mode(node); + long new_pn = 1 + res->reg_offset; - assert(res->reg0 != NULL && res->reg1 == NULL); - regn = find_out_for_reg(new_call, reg); - if (regn < 0) { - panic("Internal error in calling convention for return %+F", node); + assert(res->req0 != NULL && res->req1 == NULL); + if (mode_needs_gp_reg(mode)) { + mode = mode_gp; } - mode = res->reg0->reg_class->mode; - sparc_free_calling_convention(cconv); - return new_r_Proj(new_call, mode, regn); + return new_r_Proj(new_call, mode, new_pn); } /** @@ -2007,6 +2336,8 @@ static ir_node *gen_Proj(ir_node *node) ir_node *pred = get_Proj_pred(node); switch (get_irn_opcode(pred)) { + case iro_Alloc: + return gen_Proj_Alloc(node); case iro_Store: return gen_Proj_Store(node); case iro_Load: @@ -2031,6 +2362,11 @@ static ir_node *gen_Proj(ir_node *node) /* FALLTHROUGH */ } default: + if (is_sparc_AddCC_t(pred)) { + return gen_Proj_AddCC_t(node); + } else if (is_sparc_SubCC_t(pred)) { + return gen_Proj_SubCC_t(node); + } panic("code selection didn't expect Proj after %+F\n", pred); } } @@ -2055,6 +2391,7 @@ static void sparc_register_transformers(void) be_start_transform_setup(); be_set_transform_function(op_Add, gen_Add); + be_set_transform_function(op_Alloc, gen_Alloc); be_set_transform_function(op_And, gen_And); be_set_transform_function(op_Call, gen_Call); be_set_transform_function(op_Cmp, gen_Cmp); @@ -2063,6 +2400,7 @@ static void sparc_register_transformers(void) be_set_transform_function(op_Conv, gen_Conv); be_set_transform_function(op_Div, gen_Div); be_set_transform_function(op_Eor, gen_Eor); + be_set_transform_function(op_Free, gen_Free); be_set_transform_function(op_Jmp, gen_Jmp); be_set_transform_function(op_Load, gen_Load); be_set_transform_function(op_Minus, gen_Minus); @@ -2083,28 +2421,11 @@ static void sparc_register_transformers(void) be_set_transform_function(op_SymConst, gen_SymConst); be_set_transform_function(op_Unknown, gen_Unknown); + be_set_transform_function(op_sparc_AddX_t, gen_AddX_t); + be_set_transform_function(op_sparc_AddCC_t,gen_AddCC_t); be_set_transform_function(op_sparc_Save, be_duplicate_node); -} - -/* hack to avoid unused fp proj at start barrier */ -static void assure_fp_keep(void) -{ - unsigned n_users = 0; - const ir_edge_t *edge; - ir_node *fp_proj = be_prolog_get_reg_value(abihelper, fp_reg); - - foreach_out_edge(fp_proj, edge) { - ir_node *succ = get_edge_src_irn(edge); - if (is_End(succ) || is_Anchor(succ)) - continue; - ++n_users; - } - - if (n_users == 0) { - ir_node *block = get_nodes_block(fp_proj); - ir_node *in[1] = { fp_proj }; - be_new_Keep(block, 1, in); - } + be_set_transform_function(op_sparc_SubX_t, gen_SubX_t); + be_set_transform_function(op_sparc_SubCC_t,gen_SubCC_t); } /** @@ -2119,22 +2440,34 @@ void sparc_transform_graph(ir_graph *irg) node_to_stack = pmap_create(); - mode_gp = mode_Iu; - mode_fp = mode_F; - mode_fp2 = mode_D; + mode_gp = mode_Iu; + mode_fp = mode_F; + mode_fp2 = mode_D; + mode_flags = mode_Bu; //mode_fp4 = ? - abihelper = be_abihelper_prepare(irg); - be_collect_stacknodes(abihelper); - cconv = sparc_decide_calling_convention(get_entity_type(entity), irg); - create_stacklayout(irg); + start_mem = NULL; + start_g0 = NULL; + start_g7 = NULL; + start_sp = NULL; + start_fp = NULL; + frame_base = NULL; + + stackorder = be_collect_stacknodes(irg); + current_cconv + = sparc_decide_calling_convention(get_entity_type(entity), irg); + if (sparc_variadic_fixups(irg, current_cconv)) { + sparc_free_calling_convention(current_cconv); + current_cconv + = sparc_decide_calling_convention(get_entity_type(entity), irg); + } + sparc_create_stacklayout(irg, current_cconv); + be_add_parameter_entity_stores(irg); be_transform_graph(irg, NULL); - if (!cconv->omit_fp) - assure_fp_keep(); - be_abihelper_finish(abihelper); - sparc_free_calling_convention(cconv); + be_free_stackorder(stackorder); + sparc_free_calling_convention(current_cconv); frame_type = get_irg_frame_type(irg); if (get_type_state(frame_type) == layout_undefined)