X-Git-Url: http://nsz.repo.hu/git/?a=blobdiff_plain;f=ir%2Fbe%2Fbenode.c;h=9d1004cdf785fed828f4cb5f0adaffe4696d438a;hb=bd019d8cd785c1a433c19dee44bf3f9e82397286;hp=87c0b84734233ed31881fa6abe2077f879cfb812;hpb=f8d7cadcac7404b56a67f0c6b6cd71a76cda9bd6;p=libfirm diff --git a/ir/be/benode.c b/ir/be/benode.c index 87c0b8473..9d1004cdf 100644 --- a/ir/be/benode.c +++ b/ir/be/benode.c @@ -1,5 +1,5 @@ /* - * Copyright (C) 1995-2008 University of Karlsruhe. All right reserved. + * Copyright (C) 1995-2011 University of Karlsruhe. All right reserved. * * This file is part of libFirm. * @@ -22,7 +22,6 @@ * @brief Backend node support for generic backend nodes. * @author Sebastian Hack * @date 17.05.2005 - * @version $Id$ * * Backend node support for generic backend nodes. * This file provides Perm, Copy, Spill and Reload nodes. @@ -58,8 +57,13 @@ #include "beirgmod.h" +typedef struct be_node_attr_t { + except_attr exc; +} be_node_attr_t; + /** The be_Return nodes attribute type. */ typedef struct { + be_node_attr_t base; int num_ret_vals; /**< number of return values */ unsigned pop; /**< number of bytes that should be popped */ int emit_pop; /**< if set, emit pop bytes, even if pop = 0 */ @@ -67,6 +71,7 @@ typedef struct { /** The be_IncSP attribute type. */ typedef struct { + be_node_attr_t base; int offset; /**< The offset by which the stack shall be expanded/shrinked. */ int align; /**< whether stack should be aligned after the @@ -75,20 +80,23 @@ typedef struct { /** The be_Frame attribute type. */ typedef struct { + be_node_attr_t base; ir_entity *ent; int offset; } be_frame_attr_t; /** The be_Call attribute type. */ typedef struct { + be_node_attr_t base; ir_entity *ent; /**< called entity if this is a static call. */ unsigned pop; ir_type *call_tp; /**< call type, copied from the original Call */ } be_call_attr_t; typedef struct { - ir_entity **in_entities; - ir_entity **out_entities; + be_node_attr_t base; + ir_entity **in_entities; + ir_entity **out_entities; } be_memperm_attr_t; ir_op *op_be_Spill; @@ -105,19 +113,16 @@ ir_op *op_be_AddSP; ir_op *op_be_SubSP; ir_op *op_be_Start; ir_op *op_be_FrameAddr; -ir_op *op_be_Barrier; - -static const ir_op_ops be_node_op_ops; /** * Compare the attributes of two be_FrameAddr nodes. * * @return zero if both nodes have identically attributes */ -static int FrameAddr_cmp_attr(ir_node *a, ir_node *b) +static int FrameAddr_cmp_attr(const ir_node *a, const ir_node *b) { - const be_frame_attr_t *a_attr = get_irn_generic_attr_const(a); - const be_frame_attr_t *b_attr = get_irn_generic_attr_const(b); + const be_frame_attr_t *a_attr = (const be_frame_attr_t*)get_irn_generic_attr_const(a); + const be_frame_attr_t *b_attr = (const be_frame_attr_t*)get_irn_generic_attr_const(b); if (a_attr->ent != b_attr->ent || a_attr->offset != b_attr->offset) return 1; @@ -130,10 +135,10 @@ static int FrameAddr_cmp_attr(ir_node *a, ir_node *b) * * @return zero if both nodes have identically attributes */ -static int Return_cmp_attr(ir_node *a, ir_node *b) +static int Return_cmp_attr(const ir_node *a, const ir_node *b) { - const be_return_attr_t *a_attr = get_irn_generic_attr_const(a); - const be_return_attr_t *b_attr = get_irn_generic_attr_const(b); + const be_return_attr_t *a_attr = (const be_return_attr_t*)get_irn_generic_attr_const(a); + const be_return_attr_t *b_attr = (const be_return_attr_t*)get_irn_generic_attr_const(b); if (a_attr->num_ret_vals != b_attr->num_ret_vals) return 1; @@ -150,10 +155,10 @@ static int Return_cmp_attr(ir_node *a, ir_node *b) * * @return zero if both nodes have identically attributes */ -static int IncSP_cmp_attr(ir_node *a, ir_node *b) +static int IncSP_cmp_attr(const ir_node *a, const ir_node *b) { - const be_incsp_attr_t *a_attr = get_irn_generic_attr_const(a); - const be_incsp_attr_t *b_attr = get_irn_generic_attr_const(b); + const be_incsp_attr_t *a_attr = (const be_incsp_attr_t*)get_irn_generic_attr_const(a); + const be_incsp_attr_t *b_attr = (const be_incsp_attr_t*)get_irn_generic_attr_const(b); if (a_attr->offset != b_attr->offset) return 1; @@ -166,10 +171,10 @@ static int IncSP_cmp_attr(ir_node *a, ir_node *b) * * @return zero if both nodes have identically attributes */ -static int Call_cmp_attr(ir_node *a, ir_node *b) +static int Call_cmp_attr(const ir_node *a, const ir_node *b) { - const be_call_attr_t *a_attr = get_irn_generic_attr_const(a); - const be_call_attr_t *b_attr = get_irn_generic_attr_const(b); + const be_call_attr_t *a_attr = (const be_call_attr_t*)get_irn_generic_attr_const(a); + const be_call_attr_t *b_attr = (const be_call_attr_t*)get_irn_generic_attr_const(b); if (a_attr->ent != b_attr->ent || a_attr->call_tp != b_attr->call_tp) @@ -183,8 +188,7 @@ static arch_register_req_t *allocate_reg_req(const ir_node *node) ir_graph *irg = get_irn_irg(node); struct obstack *obst = be_get_be_obst(irg); - arch_register_req_t *req = obstack_alloc(obst, sizeof(*req)); - memset(req, 0, sizeof(*req)); + arch_register_req_t *req = OALLOCZ(obst, arch_register_req_t); return req; } @@ -204,7 +208,7 @@ void be_set_constr_out(ir_node *node, int pos, const arch_register_req_t *req) /** * Initializes the generic attribute of all be nodes and return it. */ -static void *init_node_attr(ir_node *node, int n_inputs, int n_outputs) +static void init_node_attr(ir_node *node, int n_inputs, int n_outputs) { ir_graph *irg = get_irn_irg(node); struct obstack *obst = be_get_be_obst(irg); @@ -233,8 +237,6 @@ static void *init_node_attr(ir_node *node, int n_inputs, int n_outputs) } else { info->out_infos = NEW_ARR_F(reg_out_info_t, 0); } - - return get_irn_generic_attr(node); } static void add_register_req_out(ir_node *node) @@ -264,21 +266,23 @@ ir_node *be_new_Spill(const arch_register_class_t *cls, in[0] = frame; in[1] = to_spill; res = new_ir_node(NULL, irg, bl, op_be_Spill, mode_M, 2, in); - a = init_node_attr(res, 2, 1); + init_node_attr(res, 2, 1); + a = (be_frame_attr_t*) get_irn_generic_attr(res); a->ent = NULL; a->offset = 0; + a->base.exc.pin_state = op_pin_state_pinned; - be_node_set_reg_class_in(res, be_pos_Spill_frame, cls_frame); - be_node_set_reg_class_in(res, be_pos_Spill_val, cls); + be_node_set_reg_class_in(res, n_be_Spill_frame, cls_frame); + be_node_set_reg_class_in(res, n_be_Spill_val, cls); /* * For spills and reloads, we return "none" as requirement for frame * pointer, so every input is ok. Some backends need this (STA). * Matze: we should investigate if this is really needed, this solution * looks very hacky to me */ - be_set_constr_in(res, be_pos_Spill_frame, arch_no_register_req); + be_set_constr_in(res, n_be_Spill_frame, arch_no_register_req); - arch_set_out_register_req(res, 0, arch_no_register_req); + arch_set_irn_register_req_out(res, 0, arch_no_register_req); return res; } @@ -290,6 +294,7 @@ ir_node *be_new_Reload(const arch_register_class_t *cls, ir_node *in[2]; ir_node *res; ir_graph *irg = get_Block_irg(block); + be_frame_attr_t *a; in[0] = frame; in[1] = mem; @@ -298,8 +303,13 @@ ir_node *be_new_Reload(const arch_register_class_t *cls, init_node_attr(res, 2, 1); be_node_set_reg_class_out(res, 0, cls); - be_node_set_reg_class_in(res, be_pos_Reload_frame, cls_frame); - arch_irn_set_flags(res, arch_irn_flags_rematerializable); + be_node_set_reg_class_in(res, n_be_Reload_frame, cls_frame); + arch_set_irn_flags(res, arch_irn_flags_rematerializable); + + a = (be_frame_attr_t*) get_irn_generic_attr(res); + a->ent = NULL; + a->offset = 0; + a->base.exc.pin_state = op_pin_state_pinned; /* * For spills and reloads, we return "none" as requirement for frame @@ -307,7 +317,7 @@ ir_node *be_new_Reload(const arch_register_class_t *cls, * Matze: we should investigate if this is really needed, this solution * looks very hacky to me */ - be_set_constr_in(res, be_pos_Reload_frame, arch_no_register_req); + be_set_constr_in(res, n_be_Reload_frame, arch_no_register_req); return res; } @@ -315,35 +325,38 @@ ir_node *be_new_Reload(const arch_register_class_t *cls, ir_node *be_get_Reload_mem(const ir_node *irn) { assert(be_is_Reload(irn)); - return get_irn_n(irn, be_pos_Reload_mem); + return get_irn_n(irn, n_be_Reload_mem); } ir_node *be_get_Reload_frame(const ir_node *irn) { assert(be_is_Reload(irn)); - return get_irn_n(irn, be_pos_Reload_frame); + return get_irn_n(irn, n_be_Reload_frame); } ir_node *be_get_Spill_val(const ir_node *irn) { assert(be_is_Spill(irn)); - return get_irn_n(irn, be_pos_Spill_val); + return get_irn_n(irn, n_be_Spill_val); } ir_node *be_get_Spill_frame(const ir_node *irn) { assert(be_is_Spill(irn)); - return get_irn_n(irn, be_pos_Spill_frame); + return get_irn_n(irn, n_be_Spill_frame); } ir_node *be_new_Perm(const arch_register_class_t *cls, ir_node *block, int n, ir_node *in[]) { - int i; - ir_graph *irg = get_Block_irg(block); + int i; + ir_graph *irg = get_Block_irg(block); + be_node_attr_t *attr; ir_node *irn = new_ir_node(NULL, irg, block, op_be_Perm, mode_T, n, in); init_node_attr(irn, n, n); + attr = (be_node_attr_t*) get_irn_generic_attr(irn); + attr->exc.pin_state = op_pin_state_pinned; for (i = 0; i < n; ++i) { be_node_set_reg_class_in(irn, i, cls); be_node_set_reg_class_out(irn, i, cls); @@ -365,7 +378,7 @@ void be_Perm_reduce(ir_node *perm, int new_size, int *map) assert(be_is_Perm(perm)); assert(new_size <= arity); - new_in = alloca(new_size * sizeof(*new_in)); + new_in = ALLOCAN(ir_node*, new_size); /* save the old register data */ memcpy(old_in_reqs, info->in_reqs, arity * sizeof(old_in_reqs[0])); @@ -401,31 +414,37 @@ ir_node *be_new_MemPerm(ir_node *block, int n, ir_node *in[]) init_node_attr(irn, n + 1, n); be_node_set_reg_class_in(irn, 0, sp->reg_class); - attr = get_irn_generic_attr(irn); + attr = (be_memperm_attr_t*)get_irn_generic_attr(irn); attr->in_entities = OALLOCNZ(irg->obst, ir_entity*, n); attr->out_entities = OALLOCNZ(irg->obst, ir_entity*, n); return irn; } -ir_node *be_new_Copy(const arch_register_class_t *cls, ir_node *bl, ir_node *op) +ir_node *be_new_Copy(ir_node *bl, ir_node *op) { ir_node *in[1]; ir_node *res; arch_register_req_t *req; + be_node_attr_t *attr; ir_graph *irg = get_Block_irg(bl); + const arch_register_req_t *in_req = arch_get_irn_register_req(op); + const arch_register_class_t *cls = in_req->cls; in[0] = op; res = new_ir_node(NULL, irg, bl, op_be_Copy, get_irn_mode(op), 1, in); init_node_attr(res, 1, 1); + attr = (be_node_attr_t*) get_irn_generic_attr(res); + attr->exc.pin_state = op_pin_state_floats; be_node_set_reg_class_in(res, 0, cls); be_node_set_reg_class_out(res, 0, cls); req = allocate_reg_req(res); req->cls = cls; - req->type = arch_register_req_type_should_be_same; + req->type = arch_register_req_type_should_be_same + | (in_req->type & arch_register_req_type_aligned); req->other_same = 1U << 0; - req->width = 1; + req->width = in_req->width; be_set_constr_out(res, 0, req); return res; @@ -433,12 +452,12 @@ ir_node *be_new_Copy(const arch_register_class_t *cls, ir_node *bl, ir_node *op) ir_node *be_get_Copy_op(const ir_node *cpy) { - return get_irn_n(cpy, be_pos_Copy_op); + return get_irn_n(cpy, n_be_Copy_op); } void be_set_Copy_op(ir_node *cpy, ir_node *op) { - set_irn_n(cpy, be_pos_Copy_op, op); + set_irn_n(cpy, n_be_Copy_op, op); } ir_node *be_new_Keep(ir_node *block, int n, ir_node *in[]) @@ -446,9 +465,12 @@ ir_node *be_new_Keep(ir_node *block, int n, ir_node *in[]) int i; ir_node *res; ir_graph *irg = get_Block_irg(block); + be_node_attr_t *attr; res = new_ir_node(NULL, irg, block, op_be_Keep, mode_ANY, -1, NULL); init_node_attr(res, -1, 1); + attr = (be_node_attr_t*) get_irn_generic_attr(res); + attr->exc.pin_state = op_pin_state_pinned; for (i = 0; i < n; ++i) { add_irn_n(res, in[i]); @@ -474,61 +496,63 @@ ir_node *be_new_Call(dbg_info *dbg, ir_graph *irg, ir_node *bl, ir_node *mem, ir_type *call_tp) { be_call_attr_t *a; - int real_n = be_pos_Call_first_arg + n; + int real_n = n_be_Call_first_arg + n; ir_node *irn; ir_node **real_in; NEW_ARR_A(ir_node *, real_in, real_n); - real_in[be_pos_Call_mem] = mem; - real_in[be_pos_Call_sp] = sp; - real_in[be_pos_Call_ptr] = ptr; - memcpy(&real_in[be_pos_Call_first_arg], in, n * sizeof(in[0])); + real_in[n_be_Call_mem] = mem; + real_in[n_be_Call_sp] = sp; + real_in[n_be_Call_ptr] = ptr; + memcpy(&real_in[n_be_Call_first_arg], in, n * sizeof(in[0])); irn = new_ir_node(dbg, irg, bl, op_be_Call, mode_T, real_n, real_in); - a = init_node_attr(irn, real_n, n_outs); - a->ent = NULL; - a->call_tp = call_tp; - a->pop = 0; + init_node_attr(irn, real_n, n_outs); + a = (be_call_attr_t*)get_irn_generic_attr(irn); + a->ent = NULL; + a->call_tp = call_tp; + a->pop = 0; + a->base.exc.pin_state = op_pin_state_pinned; return irn; } ir_entity *be_Call_get_entity(const ir_node *call) { - const be_call_attr_t *a = get_irn_generic_attr_const(call); + const be_call_attr_t *a = (const be_call_attr_t*)get_irn_generic_attr_const(call); assert(be_is_Call(call)); return a->ent; } void be_Call_set_entity(ir_node *call, ir_entity *ent) { - be_call_attr_t *a = get_irn_generic_attr(call); + be_call_attr_t *a = (be_call_attr_t*)get_irn_generic_attr(call); assert(be_is_Call(call)); a->ent = ent; } ir_type *be_Call_get_type(ir_node *call) { - const be_call_attr_t *a = get_irn_generic_attr_const(call); + const be_call_attr_t *a = (const be_call_attr_t*)get_irn_generic_attr_const(call); assert(be_is_Call(call)); return a->call_tp; } void be_Call_set_type(ir_node *call, ir_type *call_tp) { - be_call_attr_t *a = get_irn_generic_attr(call); + be_call_attr_t *a = (be_call_attr_t*)get_irn_generic_attr(call); assert(be_is_Call(call)); a->call_tp = call_tp; } void be_Call_set_pop(ir_node *call, unsigned pop) { - be_call_attr_t *a = get_irn_generic_attr(call); + be_call_attr_t *a = (be_call_attr_t*)get_irn_generic_attr(call); a->pop = pop; } unsigned be_Call_get_pop(const ir_node *call) { - const be_call_attr_t *a = get_irn_generic_attr_const(call); + const be_call_attr_t *a = (const be_call_attr_t*)get_irn_generic_attr_const(call); return a->pop; } @@ -547,35 +571,36 @@ ir_node *be_new_Return(dbg_info *dbg, ir_graph *irg, ir_node *block, int n_res, } be_set_constr_out(res, 0, arch_no_register_req); - a = get_irn_generic_attr(res); - a->num_ret_vals = n_res; - a->pop = pop; - a->emit_pop = 0; + a = (be_return_attr_t*)get_irn_generic_attr(res); + a->num_ret_vals = n_res; + a->pop = pop; + a->emit_pop = 0; + a->base.exc.pin_state = op_pin_state_pinned; return res; } int be_Return_get_n_rets(const ir_node *ret) { - const be_return_attr_t *a = get_irn_generic_attr_const(ret); + const be_return_attr_t *a = (const be_return_attr_t*)get_irn_generic_attr_const(ret); return a->num_ret_vals; } unsigned be_Return_get_pop(const ir_node *ret) { - const be_return_attr_t *a = get_irn_generic_attr_const(ret); + const be_return_attr_t *a = (const be_return_attr_t*)get_irn_generic_attr_const(ret); return a->pop; } int be_Return_get_emit_pop(const ir_node *ret) { - const be_return_attr_t *a = get_irn_generic_attr_const(ret); + const be_return_attr_t *a = (const be_return_attr_t*)get_irn_generic_attr_const(ret); return a->emit_pop; } void be_Return_set_emit_pop(ir_node *ret, int emit_pop) { - be_return_attr_t *a = get_irn_generic_attr(ret); + be_return_attr_t *a = (be_return_attr_t*)get_irn_generic_attr(ret); a->emit_pop = emit_pop; } @@ -600,9 +625,11 @@ ir_node *be_new_IncSP(const arch_register_t *sp, ir_node *bl, in[0] = old_sp; irn = new_ir_node(NULL, irg, bl, op_be_IncSP, sp->reg_class->mode, sizeof(in) / sizeof(in[0]), in); - a = init_node_attr(irn, 1, 1); - a->offset = offset; - a->align = align; + init_node_attr(irn, 1, 1); + a = (be_incsp_attr_t*)get_irn_generic_attr(irn); + a->offset = offset; + a->align = align; + a->base.exc.pin_state = op_pin_state_pinned; /* Set output constraint to stack register. */ be_node_set_reg_class_in(irn, 0, sp->reg_class); @@ -615,44 +642,49 @@ ir_node *be_new_AddSP(const arch_register_t *sp, ir_node *bl, ir_node *old_sp, ir_node *sz) { ir_node *irn; - ir_node *in[be_pos_AddSP_last]; - const arch_register_class_t *cls; + ir_node *in[n_be_AddSP_last]; ir_graph *irg; + be_node_attr_t *attr; - in[be_pos_AddSP_old_sp] = old_sp; - in[be_pos_AddSP_size] = sz; + in[n_be_AddSP_old_sp] = old_sp; + in[n_be_AddSP_size] = sz; irg = get_Block_irg(bl); - irn = new_ir_node(NULL, irg, bl, op_be_AddSP, mode_T, be_pos_AddSP_last, in); - init_node_attr(irn, be_pos_AddSP_last, pn_be_AddSP_last); + irn = new_ir_node(NULL, irg, bl, op_be_AddSP, mode_T, n_be_AddSP_last, in); + init_node_attr(irn, n_be_AddSP_last, pn_be_AddSP_last); + attr = (be_node_attr_t*) get_irn_generic_attr(irn); + attr->exc.pin_state = op_pin_state_pinned; /* Set output constraint to stack register. */ - be_set_constr_single_reg_in(irn, be_pos_AddSP_old_sp, sp, 0); - be_node_set_reg_class_in(irn, be_pos_AddSP_size, arch_register_get_class(sp)); + be_set_constr_single_reg_in(irn, n_be_AddSP_old_sp, sp, + arch_register_req_type_none); + be_node_set_reg_class_in(irn, n_be_AddSP_size, sp->reg_class); be_set_constr_single_reg_out(irn, pn_be_AddSP_sp, sp, arch_register_req_type_produces_sp); - cls = arch_register_get_class(sp); - return irn; } ir_node *be_new_SubSP(const arch_register_t *sp, ir_node *bl, ir_node *old_sp, ir_node *sz) { ir_node *irn; - ir_node *in[be_pos_SubSP_last]; + ir_node *in[n_be_SubSP_last]; ir_graph *irg; + be_node_attr_t *attr; - in[be_pos_SubSP_old_sp] = old_sp; - in[be_pos_SubSP_size] = sz; + in[n_be_SubSP_old_sp] = old_sp; + in[n_be_SubSP_size] = sz; irg = get_Block_irg(bl); - irn = new_ir_node(NULL, irg, bl, op_be_SubSP, mode_T, be_pos_SubSP_last, in); - init_node_attr(irn, be_pos_SubSP_last, pn_be_SubSP_last); + irn = new_ir_node(NULL, irg, bl, op_be_SubSP, mode_T, n_be_SubSP_last, in); + init_node_attr(irn, n_be_SubSP_last, pn_be_SubSP_last); + attr = (be_node_attr_t*) get_irn_generic_attr(irn); + attr->exc.pin_state = op_pin_state_pinned; /* Set output constraint to stack register. */ - be_set_constr_single_reg_in(irn, be_pos_SubSP_old_sp, sp, 0); - be_node_set_reg_class_in(irn, be_pos_SubSP_size, arch_register_get_class(sp)); + be_set_constr_single_reg_in(irn, n_be_SubSP_old_sp, sp, + arch_register_req_type_none); + be_node_set_reg_class_in(irn, n_be_SubSP_size, sp->reg_class); be_set_constr_single_reg_out(irn, pn_be_SubSP_sp, sp, arch_register_req_type_produces_sp); return irn; @@ -663,9 +695,12 @@ ir_node *be_new_Start(dbg_info *dbgi, ir_node *bl, int n_outs) ir_node *res; int i; ir_graph *irg = get_Block_irg(bl); + be_node_attr_t *attr; res = new_ir_node(dbgi, irg, bl, op_be_Start, mode_T, 0, NULL); init_node_attr(res, 0, -1); + attr = (be_node_attr_t*) get_irn_generic_attr(res); + attr->exc.pin_state = op_pin_state_pinned; for (i = 0; i < n_outs; ++i) { add_register_req_out(res); } @@ -682,9 +717,11 @@ ir_node *be_new_FrameAddr(const arch_register_class_t *cls_frame, ir_node *bl, i in[0] = frame; irn = new_ir_node(NULL, irg, bl, op_be_FrameAddr, get_irn_mode(frame), 1, in); - a = init_node_attr(irn, 1, 1); - a->ent = ent; - a->offset = 0; + init_node_attr(irn, 1, 1); + a = (be_frame_attr_t*)get_irn_generic_attr(irn); + a->ent = ent; + a->offset = 0; + a->base.exc.pin_state = op_pin_state_floats; be_node_set_reg_class_in(irn, 0, cls_frame); be_node_set_reg_class_out(irn, 0, cls_frame); @@ -694,73 +731,50 @@ ir_node *be_new_FrameAddr(const arch_register_class_t *cls_frame, ir_node *bl, i ir_node *be_get_FrameAddr_frame(const ir_node *node) { assert(be_is_FrameAddr(node)); - return get_irn_n(node, be_pos_FrameAddr_ptr); + return get_irn_n(node, n_be_FrameAddr_ptr); } ir_entity *be_get_FrameAddr_entity(const ir_node *node) { - const be_frame_attr_t *attr = get_irn_generic_attr_const(node); + const be_frame_attr_t *attr = (const be_frame_attr_t*)get_irn_generic_attr_const(node); return attr->ent; } -ir_node *be_new_CopyKeep(const arch_register_class_t *cls, ir_node *bl, ir_node *src, int n, ir_node *in_keep[], ir_mode *mode) +ir_node *be_new_CopyKeep(ir_node *bl, ir_node *src, int n, ir_node *in_keep[]) { ir_node *irn; ir_node **in = ALLOCAN(ir_node*, n + 1); ir_graph *irg = get_Block_irg(bl); + const arch_register_req_t *req = arch_get_irn_register_req(src); + const arch_register_class_t *cls = req->cls; + ir_mode *mode = get_irn_mode(src); + be_node_attr_t *attr; in[0] = src; memcpy(&in[1], in_keep, n * sizeof(in[0])); irn = new_ir_node(NULL, irg, bl, op_be_CopyKeep, mode, n + 1, in); init_node_attr(irn, n + 1, 1); + attr = (be_node_attr_t*) get_irn_generic_attr(irn); + attr->exc.pin_state = op_pin_state_floats; be_node_set_reg_class_in(irn, 0, cls); be_node_set_reg_class_out(irn, 0, cls); return irn; } -ir_node *be_new_CopyKeep_single(const arch_register_class_t *cls, ir_node *bl, ir_node *src, ir_node *keep, ir_mode *mode) +ir_node *be_new_CopyKeep_single(ir_node *bl, ir_node *src, ir_node *keep) { - return be_new_CopyKeep(cls, bl, src, 1, &keep, mode); + return be_new_CopyKeep(bl, src, 1, &keep); } ir_node *be_get_CopyKeep_op(const ir_node *cpy) { - return get_irn_n(cpy, be_pos_CopyKeep_op); + return get_irn_n(cpy, n_be_CopyKeep_op); } void be_set_CopyKeep_op(ir_node *cpy, ir_node *op) { - set_irn_n(cpy, be_pos_CopyKeep_op, op); -} - -ir_node *be_new_Barrier(ir_node *bl, int n, ir_node *in[]) -{ - ir_node *res; - int i; - ir_graph *irg = get_Block_irg(bl); - - res = new_ir_node(NULL, irg, bl, op_be_Barrier, mode_T, -1, NULL); - init_node_attr(res, -1, -1); - for (i = 0; i < n; ++i) { - add_irn_n(res, in[i]); - add_register_req_in(res); - add_register_req_out(res); - } - - return res; -} - -ir_node *be_Barrier_append_node(ir_node *barrier, ir_node *node) -{ - ir_mode *mode = get_irn_mode(node); - int n = add_irn_n(barrier, node); - - ir_node *proj = new_r_Proj(barrier, mode, n); - add_register_req_in(barrier); - add_register_req_out(barrier); - - return proj; + set_irn_n(cpy, n_be_CopyKeep_op, op); } static bool be_has_frame_entity(const ir_node *irn) @@ -778,7 +792,7 @@ static bool be_has_frame_entity(const ir_node *irn) ir_entity *be_get_frame_entity(const ir_node *irn) { if (be_has_frame_entity(irn)) { - const be_frame_attr_t *a = get_irn_generic_attr_const(irn); + const be_frame_attr_t *a = (const be_frame_attr_t*)get_irn_generic_attr_const(irn); return a->ent; } return NULL; @@ -788,7 +802,7 @@ int be_get_frame_offset(const ir_node *irn) { assert(is_be_node(irn)); if (be_has_frame_entity(irn)) { - const be_frame_attr_t *a = get_irn_generic_attr_const(irn); + const be_frame_attr_t *a = (const be_frame_attr_t*)get_irn_generic_attr_const(irn); return a->offset; } return 0; @@ -796,7 +810,7 @@ int be_get_frame_offset(const ir_node *irn) void be_set_MemPerm_in_entity(const ir_node *irn, int n, ir_entity *ent) { - const be_memperm_attr_t *attr = get_irn_generic_attr_const(irn); + const be_memperm_attr_t *attr = (const be_memperm_attr_t*)get_irn_generic_attr_const(irn); assert(be_is_MemPerm(irn)); assert(n < be_get_MemPerm_entity_arity(irn)); @@ -806,7 +820,7 @@ void be_set_MemPerm_in_entity(const ir_node *irn, int n, ir_entity *ent) ir_entity* be_get_MemPerm_in_entity(const ir_node* irn, int n) { - const be_memperm_attr_t *attr = get_irn_generic_attr_const(irn); + const be_memperm_attr_t *attr = (const be_memperm_attr_t*)get_irn_generic_attr_const(irn); assert(be_is_MemPerm(irn)); assert(n < be_get_MemPerm_entity_arity(irn)); @@ -816,7 +830,7 @@ ir_entity* be_get_MemPerm_in_entity(const ir_node* irn, int n) void be_set_MemPerm_out_entity(const ir_node *irn, int n, ir_entity *ent) { - const be_memperm_attr_t *attr = get_irn_generic_attr_const(irn); + const be_memperm_attr_t *attr = (const be_memperm_attr_t*)get_irn_generic_attr_const(irn); assert(be_is_MemPerm(irn)); assert(n < be_get_MemPerm_entity_arity(irn)); @@ -826,7 +840,7 @@ void be_set_MemPerm_out_entity(const ir_node *irn, int n, ir_entity *ent) ir_entity* be_get_MemPerm_out_entity(const ir_node* irn, int n) { - const be_memperm_attr_t *attr = get_irn_generic_attr_const(irn); + const be_memperm_attr_t *attr = (const be_memperm_attr_t*)get_irn_generic_attr_const(irn); assert(be_is_MemPerm(irn)); assert(n < be_get_MemPerm_entity_arity(irn)); @@ -842,7 +856,7 @@ int be_get_MemPerm_entity_arity(const ir_node *irn) const arch_register_req_t *be_create_reg_req(struct obstack *obst, const arch_register_t *reg, arch_register_req_type_t additional_types) { - arch_register_req_t *req = obstack_alloc(obst, sizeof(*req)); + arch_register_req_t *req = OALLOC(obst, arch_register_req_t); const arch_register_class_t *cls = arch_register_get_class(reg); unsigned *limited_bitset; @@ -874,22 +888,23 @@ void be_set_constr_single_reg_in(ir_node *node, int pos, void be_set_constr_single_reg_out(ir_node *node, int pos, const arch_register_t *reg, arch_register_req_type_t additional_types) { + ir_graph *irg = get_irn_irg(node); + be_irg_t *birg = be_birg_from_irg(irg); const arch_register_req_t *req; /* if we have an ignore register, add ignore flag and just assign it */ - if (reg->type & arch_register_type_ignore) { + if (!rbitset_is_set(birg->allocatable_regs, reg->global_index)) { additional_types |= arch_register_req_type_ignore; } if (additional_types == 0) { req = reg->single_req; } else { - ir_graph *irg = get_irn_irg(node); struct obstack *obst = be_get_be_obst(irg); req = be_create_reg_req(obst, reg, additional_types); } - arch_irn_set_register(node, pos, reg); + arch_set_irn_register_out(node, pos, reg); be_set_constr_out(node, pos, req); } @@ -919,71 +934,25 @@ void be_set_IncSP_pred(ir_node *incsp, ir_node *pred) void be_set_IncSP_offset(ir_node *irn, int offset) { - be_incsp_attr_t *a = get_irn_generic_attr(irn); + be_incsp_attr_t *a = (be_incsp_attr_t*)get_irn_generic_attr(irn); assert(be_is_IncSP(irn)); a->offset = offset; } int be_get_IncSP_offset(const ir_node *irn) { - const be_incsp_attr_t *a = get_irn_generic_attr_const(irn); + const be_incsp_attr_t *a = (const be_incsp_attr_t*)get_irn_generic_attr_const(irn); assert(be_is_IncSP(irn)); return a->offset; } int be_get_IncSP_align(const ir_node *irn) { - const be_incsp_attr_t *a = get_irn_generic_attr_const(irn); + const be_incsp_attr_t *a = (const be_incsp_attr_t*)get_irn_generic_attr_const(irn); assert(be_is_IncSP(irn)); return a->align; } -ir_node *be_spill(ir_node *block, ir_node *irn) -{ - ir_graph *irg = get_Block_irg(block); - ir_node *frame = get_irg_frame(irg); - const arch_register_class_t *cls = arch_get_irn_reg_class_out(irn); - const arch_register_class_t *cls_frame = arch_get_irn_reg_class_out(frame); - ir_node *spill; - - spill = be_new_Spill(cls, cls_frame, block, frame, irn); - return spill; -} - -ir_node *be_reload(const arch_register_class_t *cls, ir_node *insert, ir_mode *mode, ir_node *spill) -{ - ir_node *reload; - ir_node *bl = is_Block(insert) ? insert : get_nodes_block(insert); - ir_graph *irg = get_Block_irg(bl); - ir_node *frame = get_irg_frame(irg); - const arch_register_class_t *cls_frame = arch_get_irn_reg_class_out(frame); - - assert(be_is_Spill(spill) || (is_Phi(spill) && get_irn_mode(spill) == mode_M)); - - reload = be_new_Reload(cls, cls_frame, bl, frame, spill, mode); - - if (is_Block(insert)) { - insert = sched_skip(insert, 0, sched_skip_cf_predicator, NULL); - sched_add_after(insert, reload); - } else { - sched_add_before(insert, reload); - } - - return reload; -} - - -static arch_irn_class_t be_node_classify(const ir_node *irn) -{ - switch (get_irn_opcode(irn)) { - case beo_Spill: return arch_irn_class_spill; - case beo_Reload: return arch_irn_class_reload; - case beo_Perm: return arch_irn_class_perm; - case beo_Copy: return arch_irn_class_copy; - default: return 0; - } -} - static ir_entity *be_node_get_frame_entity(const ir_node *irn) { return be_get_frame_entity(irn); @@ -995,7 +964,7 @@ void be_node_set_frame_entity(ir_node *irn, ir_entity *ent) assert(be_has_frame_entity(irn)); - a = get_irn_generic_attr(irn); + a = (be_frame_attr_t*)get_irn_generic_attr(irn); a->ent = ent; } @@ -1006,7 +975,7 @@ static void be_node_set_frame_offset(ir_node *irn, int offset) if (!be_has_frame_entity(irn)) return; - a = get_irn_generic_attr(irn); + a = (be_frame_attr_t*)get_irn_generic_attr(irn); a->offset = offset; } @@ -1024,7 +993,6 @@ static int be_node_get_sp_bias(const ir_node *irn) /* for be nodes */ static const arch_irn_ops_t be_node_irn_ops = { - be_node_classify, be_node_get_frame_entity, be_node_set_frame_offset, be_node_get_sp_bias, @@ -1034,10 +1002,61 @@ static const arch_irn_ops_t be_node_irn_ops = { NULL, /* perform_memory_operand */ }; -static arch_irn_class_t dummy_classify(const ir_node *node) +static int get_start_reg_index(ir_graph *irg, const arch_register_t *reg) { - (void) node; - return 0; + ir_node *start = get_irg_start(irg); + unsigned n_outs = arch_get_irn_n_outs(start); + int i; + + /* do a naive linear search... */ + for (i = 0; i < (int)n_outs; ++i) { + const arch_register_req_t *out_req + = arch_get_irn_register_req_out(start, i); + if (! (out_req->type & arch_register_req_type_limited)) + continue; + if (out_req->cls != arch_register_get_class(reg)) + continue; + if (!rbitset_is_set(out_req->limited, reg->index)) + continue; + return i; + } + panic("Tried querying undefined register '%s' at Start", reg->name); +} + +ir_node *be_get_initial_reg_value(ir_graph *irg, const arch_register_t *reg) +{ + int i = get_start_reg_index(irg, reg); + ir_node *start = get_irg_start(irg); + ir_mode *mode = arch_register_class_mode(arch_register_get_class(reg)); + const ir_edge_t *edge; + + foreach_out_edge(start, edge) { + ir_node *proj = get_edge_src_irn(edge); + if (!is_Proj(proj)) // maybe End/Anchor + continue; + if (get_Proj_proj(proj) == i) { + return proj; + } + } + return new_r_Proj(start, mode, i); +} + +int be_find_return_reg_input(ir_node *ret, const arch_register_t *reg) +{ + int arity = get_irn_arity(ret); + int i; + /* do a naive linear search... */ + for (i = 0; i < arity; ++i) { + const arch_register_req_t *req = arch_get_irn_register_req_in(ret, i); + if (! (req->type & arch_register_req_type_limited)) + continue; + if (req->cls != arch_register_get_class(reg)) + continue; + if (!rbitset_is_set(req->limited, reg->index)) + continue; + return i; + } + panic("Tried querying undefined register '%s' at Return", reg->name); } static ir_entity* dummy_get_frame_entity(const ir_node *node) @@ -1061,7 +1080,6 @@ static int dummy_get_sp_bias(const ir_node *node) /* for "middleend" nodes */ static const arch_irn_ops_t dummy_be_irn_ops = { - dummy_classify, dummy_get_frame_entity, dummy_set_frame_offset, dummy_get_sp_bias, @@ -1141,7 +1159,6 @@ void be_dump_phi_reg_reqs(FILE *F, ir_node *node, dump_reason_t reason) } static const arch_irn_ops_t phi_irn_ops = { - dummy_classify, dummy_get_frame_entity, dummy_set_frame_offset, dummy_get_sp_bias, @@ -1171,26 +1188,20 @@ static void dump_node(FILE *f, ir_node *irn, dump_reason_t reason) break; case dump_node_nodeattr_txt: if (be_is_Call(irn)) { - const be_call_attr_t *a = get_irn_generic_attr_const(irn); + const be_call_attr_t *a = (const be_call_attr_t*)get_irn_generic_attr_const(irn); if (a->ent) fprintf(f, " [%s] ", get_entity_name(a->ent)); } if (be_is_IncSP(irn)) { - const be_incsp_attr_t *attr = get_irn_generic_attr_const(irn); - if (attr->offset == BE_STACK_FRAME_SIZE_EXPAND) { - fprintf(f, " [Setup Stackframe] "); - } else if (attr->offset == BE_STACK_FRAME_SIZE_SHRINK) { - fprintf(f, " [Destroy Stackframe] "); - } else { - fprintf(f, " [%d] ", attr->offset); - } + const be_incsp_attr_t *attr = (const be_incsp_attr_t*)get_irn_generic_attr_const(irn); + fprintf(f, " [%d] ", attr->offset); } break; case dump_node_info_txt: arch_dump_reqs_and_registers(f, irn); if (be_has_frame_entity(irn)) { - const be_frame_attr_t *a = get_irn_generic_attr_const(irn); + const be_frame_attr_t *a = (const be_frame_attr_t*)get_irn_generic_attr_const(irn); if (a->ent) { unsigned size = get_type_size_bytes(get_entity_type(a->ent)); ir_fprintf(f, "frame entity: %+F, offset 0x%x (%d), size 0x%x (%d) bytes\n", @@ -1201,18 +1212,13 @@ static void dump_node(FILE *f, ir_node *irn, dump_reason_t reason) switch (get_irn_opcode(irn)) { case beo_IncSP: { - const be_incsp_attr_t *a = get_irn_generic_attr_const(irn); + const be_incsp_attr_t *a = (const be_incsp_attr_t*)get_irn_generic_attr_const(irn); fprintf(f, "align: %d\n", a->align); - if (a->offset == BE_STACK_FRAME_SIZE_EXPAND) - fprintf(f, "offset: FRAME_SIZE\n"); - else if (a->offset == BE_STACK_FRAME_SIZE_SHRINK) - fprintf(f, "offset: -FRAME SIZE\n"); - else - fprintf(f, "offset: %d\n", a->offset); + fprintf(f, "offset: %d\n", a->offset); break; } case beo_Call: { - const be_call_attr_t *a = get_irn_generic_attr_const(irn); + const be_call_attr_t *a = (const be_call_attr_t*)get_irn_generic_attr_const(irn); if (a->ent) fprintf(f, "\ncalling: %s\n", get_entity_name(a->ent)); @@ -1257,10 +1263,11 @@ static void copy_attr(ir_graph *irg, const ir_node *old_node, ir_node *new_node) memcpy(new_attr, old_attr, get_op_attr_size(get_irn_op(old_node))); + new_info->flags = old_info->flags; if (old_info->out_infos != NULL) { - unsigned n_outs = ARR_LEN(old_info->out_infos); + size_t n_outs = ARR_LEN(old_info->out_infos); /* need dyanmic out infos? */ - if (be_is_Barrier(new_node) || be_is_Perm(new_node)) { + if (be_is_Perm(new_node)) { new_info->out_infos = NEW_ARR_F(reg_out_info_t, n_outs); } else { new_info->out_infos = NEW_ARR_D(reg_out_info_t, obst, n_outs); @@ -1314,24 +1321,24 @@ int is_be_node(const ir_node *irn) void be_init_op(void) { - ir_opcode opc; + unsigned opc; /* Acquire all needed opcodes. */ - op_be_Spill = new_ir_op(beo_Spill, "be_Spill", op_pin_state_pinned, irop_flag_none, oparity_unary, 0, sizeof(be_frame_attr_t), &be_node_op_ops); - op_be_Reload = new_ir_op(beo_Reload, "be_Reload", op_pin_state_pinned, irop_flag_none, oparity_zero, 0, sizeof(be_frame_attr_t), &be_node_op_ops); - op_be_Perm = new_ir_op(beo_Perm, "be_Perm", op_pin_state_pinned, irop_flag_none, oparity_variable, 0, 0, &be_node_op_ops); - op_be_MemPerm = new_ir_op(beo_MemPerm, "be_MemPerm", op_pin_state_pinned, irop_flag_none, oparity_variable, 0, sizeof(be_memperm_attr_t), &be_node_op_ops); - op_be_Copy = new_ir_op(beo_Copy, "be_Copy", op_pin_state_floats, irop_flag_none, oparity_unary, 0, 0, &be_node_op_ops); - op_be_Keep = new_ir_op(beo_Keep, "be_Keep", op_pin_state_floats, irop_flag_keep, oparity_dynamic, 0, 0, &be_node_op_ops); - op_be_CopyKeep = new_ir_op(beo_CopyKeep, "be_CopyKeep", op_pin_state_floats, irop_flag_keep, oparity_variable, 0, 0, &be_node_op_ops); - op_be_Call = new_ir_op(beo_Call, "be_Call", op_pin_state_pinned, irop_flag_fragile|irop_flag_uses_memory, oparity_variable, 0, sizeof(be_call_attr_t), &be_node_op_ops); - op_be_Return = new_ir_op(beo_Return, "be_Return", op_pin_state_pinned, irop_flag_cfopcode, oparity_dynamic, 0, sizeof(be_return_attr_t), &be_node_op_ops); - op_be_AddSP = new_ir_op(beo_AddSP, "be_AddSP", op_pin_state_pinned, irop_flag_none, oparity_unary, 0, 0, &be_node_op_ops); - op_be_SubSP = new_ir_op(beo_SubSP, "be_SubSP", op_pin_state_pinned, irop_flag_none, oparity_unary, 0, 0, &be_node_op_ops); - op_be_IncSP = new_ir_op(beo_IncSP, "be_IncSP", op_pin_state_pinned, irop_flag_none, oparity_unary, 0, sizeof(be_incsp_attr_t), &be_node_op_ops); - op_be_Start = new_ir_op(beo_Start, "be_Start", op_pin_state_pinned, irop_flag_none, oparity_zero, 0, 0, &be_node_op_ops); - op_be_FrameAddr = new_ir_op(beo_FrameAddr, "be_FrameAddr", op_pin_state_floats, irop_flag_none, oparity_unary, 0, sizeof(be_frame_attr_t), &be_node_op_ops); - op_be_Barrier = new_ir_op(beo_Barrier, "be_Barrier", op_pin_state_pinned, irop_flag_none, oparity_dynamic, 0, 0, &be_node_op_ops); + op_be_Spill = new_ir_op(beo_Spill, "be_Spill", op_pin_state_exc_pinned, irop_flag_none, oparity_unary, 0, sizeof(be_frame_attr_t), &be_node_op_ops); + op_be_Reload = new_ir_op(beo_Reload, "be_Reload", op_pin_state_exc_pinned, irop_flag_none, oparity_zero, 0, sizeof(be_frame_attr_t), &be_node_op_ops); + op_be_Perm = new_ir_op(beo_Perm, "be_Perm", op_pin_state_exc_pinned, irop_flag_none, oparity_variable, 0, sizeof(be_node_attr_t), &be_node_op_ops); + op_be_MemPerm = new_ir_op(beo_MemPerm, "be_MemPerm", op_pin_state_exc_pinned, irop_flag_none, oparity_variable, 0, sizeof(be_memperm_attr_t), &be_node_op_ops); + op_be_Copy = new_ir_op(beo_Copy, "be_Copy", op_pin_state_exc_pinned, irop_flag_none, oparity_unary, 0, sizeof(be_node_attr_t), &be_node_op_ops); + op_be_Keep = new_ir_op(beo_Keep, "be_Keep", op_pin_state_exc_pinned, irop_flag_keep, oparity_dynamic, 0, sizeof(be_node_attr_t), &be_node_op_ops); + op_be_CopyKeep = new_ir_op(beo_CopyKeep, "be_CopyKeep", op_pin_state_exc_pinned, irop_flag_keep, oparity_variable, 0, sizeof(be_node_attr_t), &be_node_op_ops); + op_be_Call = new_ir_op(beo_Call, "be_Call", op_pin_state_exc_pinned, irop_flag_fragile|irop_flag_uses_memory, oparity_variable, 0, sizeof(be_call_attr_t), &be_node_op_ops); + ir_op_set_fragile_indices(op_be_Call, n_be_Call_mem, pn_be_Call_X_regular, pn_be_Call_X_except); + op_be_Return = new_ir_op(beo_Return, "be_Return", op_pin_state_exc_pinned, irop_flag_cfopcode, oparity_dynamic, 0, sizeof(be_return_attr_t), &be_node_op_ops); + op_be_AddSP = new_ir_op(beo_AddSP, "be_AddSP", op_pin_state_exc_pinned, irop_flag_none, oparity_unary, 0, sizeof(be_node_attr_t), &be_node_op_ops); + op_be_SubSP = new_ir_op(beo_SubSP, "be_SubSP", op_pin_state_exc_pinned, irop_flag_none, oparity_unary, 0, sizeof(be_node_attr_t), &be_node_op_ops); + op_be_IncSP = new_ir_op(beo_IncSP, "be_IncSP", op_pin_state_exc_pinned, irop_flag_none, oparity_unary, 0, sizeof(be_incsp_attr_t), &be_node_op_ops); + op_be_Start = new_ir_op(beo_Start, "be_Start", op_pin_state_exc_pinned, irop_flag_none, oparity_zero, 0, sizeof(be_node_attr_t), &be_node_op_ops); + op_be_FrameAddr = new_ir_op(beo_FrameAddr, "be_FrameAddr", op_pin_state_exc_pinned, irop_flag_none, oparity_unary, 0, sizeof(be_frame_attr_t), &be_node_op_ops); op_be_Spill->ops.node_cmp_attr = FrameAddr_cmp_attr; op_be_Reload->ops.node_cmp_attr = FrameAddr_cmp_attr; @@ -1347,7 +1354,6 @@ void be_init_op(void) op_be_IncSP->ops.node_cmp_attr = IncSP_cmp_attr; op_be_Start->ops.node_cmp_attr = be_nodes_equal; op_be_FrameAddr->ops.node_cmp_attr = FrameAddr_cmp_attr; - op_be_Barrier->ops.node_cmp_attr = be_nodes_equal; /* attach out dummy_ops to middle end nodes */ for (opc = iro_First; opc <= iro_Last; ++opc) {