X-Git-Url: http://nsz.repo.hu/git/?a=blobdiff_plain;f=ir%2Fbe%2Fbenode.c;h=3d7c28557d7ea1f411d88e170d2feb03da4203c4;hb=6b45fde8adadc605dfe4c3a181c7fd8f78ac227c;hp=3383f3883cee69c86360cbac49986423f3c6e093;hpb=4764ebb82834c3370640980c9299f0dbb1ac598d;p=libfirm diff --git a/ir/be/benode.c b/ir/be/benode.c index 3383f3883..3d7c28557 100644 --- a/ir/be/benode.c +++ b/ir/be/benode.c @@ -1,5 +1,5 @@ /* - * Copyright (C) 1995-2008 University of Karlsruhe. All right reserved. + * Copyright (C) 1995-2011 University of Karlsruhe. All right reserved. * * This file is part of libFirm. * @@ -58,53 +58,46 @@ #include "beirgmod.h" -#define get_irn_attr(irn) get_irn_generic_attr(irn) -#define get_irn_attr_const(irn) get_irn_generic_attr_const(irn) - -typedef struct { - const arch_register_req_t *in_req; -} be_reg_data_t; - -/** The generic be nodes attribute type. */ -typedef struct { - be_reg_data_t *reg_data; +typedef struct be_node_attr_t { + except_attr exc; } be_node_attr_t; /** The be_Return nodes attribute type. */ typedef struct { - be_node_attr_t node_attr; /**< base attributes of every be node. */ - int num_ret_vals; /**< number of return values */ - unsigned pop; /**< number of bytes that should be popped */ - int emit_pop; /**< if set, emit pop bytes, even if pop = 0 */ + be_node_attr_t base; + int num_ret_vals; /**< number of return values */ + unsigned pop; /**< number of bytes that should be popped */ + int emit_pop; /**< if set, emit pop bytes, even if pop = 0 */ } be_return_attr_t; /** The be_IncSP attribute type. */ typedef struct { - be_node_attr_t node_attr; /**< base attributes of every be node. */ - int offset; /**< The offset by which the stack shall be expanded/shrinked. */ - int align; /**< whether stack should be aligned after the - IncSP */ + be_node_attr_t base; + int offset; /**< The offset by which the stack shall be + expanded/shrinked. */ + int align; /**< whether stack should be aligned after the + IncSP */ } be_incsp_attr_t; /** The be_Frame attribute type. */ typedef struct { - be_node_attr_t node_attr; /**< base attributes of every be node. */ + be_node_attr_t base; ir_entity *ent; int offset; } be_frame_attr_t; /** The be_Call attribute type. */ typedef struct { - be_node_attr_t node_attr; /**< base attributes of every be node. */ - ir_entity *ent; /**< The called entity if this is a static call. */ + be_node_attr_t base; + ir_entity *ent; /**< called entity if this is a static call. */ unsigned pop; - ir_type *call_tp; /**< The call type, copied from the original Call node. */ + ir_type *call_tp; /**< call type, copied from the original Call */ } be_call_attr_t; typedef struct { - be_node_attr_t node_attr; /**< base attributes of every be node. */ - ir_entity **in_entities; - ir_entity **out_entities; + be_node_attr_t base; + ir_entity **in_entities; + ir_entity **out_entities; } be_memperm_attr_t; ir_op *op_be_Spill; @@ -121,62 +114,21 @@ ir_op *op_be_AddSP; ir_op *op_be_SubSP; ir_op *op_be_Start; ir_op *op_be_FrameAddr; -ir_op *op_be_Barrier; - -static const ir_op_ops be_node_op_ops; - -#define N irop_flag_none -#define L irop_flag_labeled -#define C irop_flag_commutative -#define X irop_flag_cfopcode -#define I irop_flag_ip_cfopcode -#define F irop_flag_fragile -#define Y irop_flag_forking -#define H irop_flag_highlevel -#define c irop_flag_constlike -#define K irop_flag_keep -#define M irop_flag_uses_memory - -/** - * Compare two be node attributes. - * - * @return zero if both attributes are identically - */ -static int node_cmp_attr(ir_node *a, ir_node *b) -{ - const be_node_attr_t *a_attr = get_irn_attr_const(a); - const be_node_attr_t *b_attr = get_irn_attr_const(b); - int i, len = ARR_LEN(a_attr->reg_data); - - if (len != ARR_LEN(b_attr->reg_data)) - return 1; - - if (!be_nodes_equal(a, b)) - return 1; - - for (i = len - 1; i >= 0; --i) { - if (!reg_reqs_equal(a_attr->reg_data[i].in_req, - b_attr->reg_data[i].in_req)) - return 1; - } - - return 0; -} /** * Compare the attributes of two be_FrameAddr nodes. * * @return zero if both nodes have identically attributes */ -static int FrameAddr_cmp_attr(ir_node *a, ir_node *b) +static int FrameAddr_cmp_attr(const ir_node *a, const ir_node *b) { - const be_frame_attr_t *a_attr = get_irn_attr_const(a); - const be_frame_attr_t *b_attr = get_irn_attr_const(b); + const be_frame_attr_t *a_attr = (const be_frame_attr_t*)get_irn_generic_attr_const(a); + const be_frame_attr_t *b_attr = (const be_frame_attr_t*)get_irn_generic_attr_const(b); if (a_attr->ent != b_attr->ent || a_attr->offset != b_attr->offset) return 1; - return node_cmp_attr(a, b); + return be_nodes_equal(a, b); } /** @@ -184,10 +136,10 @@ static int FrameAddr_cmp_attr(ir_node *a, ir_node *b) * * @return zero if both nodes have identically attributes */ -static int Return_cmp_attr(ir_node *a, ir_node *b) +static int Return_cmp_attr(const ir_node *a, const ir_node *b) { - const be_return_attr_t *a_attr = get_irn_attr_const(a); - const be_return_attr_t *b_attr = get_irn_attr_const(b); + const be_return_attr_t *a_attr = (const be_return_attr_t*)get_irn_generic_attr_const(a); + const be_return_attr_t *b_attr = (const be_return_attr_t*)get_irn_generic_attr_const(b); if (a_attr->num_ret_vals != b_attr->num_ret_vals) return 1; @@ -196,7 +148,7 @@ static int Return_cmp_attr(ir_node *a, ir_node *b) if (a_attr->emit_pop != b_attr->emit_pop) return 1; - return node_cmp_attr(a, b); + return be_nodes_equal(a, b); } /** @@ -204,15 +156,15 @@ static int Return_cmp_attr(ir_node *a, ir_node *b) * * @return zero if both nodes have identically attributes */ -static int IncSP_cmp_attr(ir_node *a, ir_node *b) +static int IncSP_cmp_attr(const ir_node *a, const ir_node *b) { - const be_incsp_attr_t *a_attr = get_irn_attr_const(a); - const be_incsp_attr_t *b_attr = get_irn_attr_const(b); + const be_incsp_attr_t *a_attr = (const be_incsp_attr_t*)get_irn_generic_attr_const(a); + const be_incsp_attr_t *b_attr = (const be_incsp_attr_t*)get_irn_generic_attr_const(b); if (a_attr->offset != b_attr->offset) return 1; - return node_cmp_attr(a, b); + return be_nodes_equal(a, b); } /** @@ -220,16 +172,16 @@ static int IncSP_cmp_attr(ir_node *a, ir_node *b) * * @return zero if both nodes have identically attributes */ -static int Call_cmp_attr(ir_node *a, ir_node *b) +static int Call_cmp_attr(const ir_node *a, const ir_node *b) { - const be_call_attr_t *a_attr = get_irn_attr_const(a); - const be_call_attr_t *b_attr = get_irn_attr_const(b); + const be_call_attr_t *a_attr = (const be_call_attr_t*)get_irn_generic_attr_const(a); + const be_call_attr_t *b_attr = (const be_call_attr_t*)get_irn_generic_attr_const(b); if (a_attr->ent != b_attr->ent || a_attr->call_tp != b_attr->call_tp) return 1; - return node_cmp_attr(a, b); + return be_nodes_equal(a, b); } static arch_register_req_t *allocate_reg_req(const ir_node *node) @@ -237,17 +189,15 @@ static arch_register_req_t *allocate_reg_req(const ir_node *node) ir_graph *irg = get_irn_irg(node); struct obstack *obst = be_get_be_obst(irg); - arch_register_req_t *req = obstack_alloc(obst, sizeof(*req)); - memset(req, 0, sizeof(*req)); + arch_register_req_t *req = OALLOCZ(obst, arch_register_req_t); return req; } void be_set_constr_in(ir_node *node, int pos, const arch_register_req_t *req) { - const be_node_attr_t *attr = get_irn_attr_const(node); - be_reg_data_t *rd = &attr->reg_data[pos]; - assert(pos < ARR_LEN(attr->reg_data)); - rd->in_req = req; + backend_info_t *info = be_get_info(node); + assert(pos < get_irn_arity(node)); + info->in_reqs[pos] = req; } void be_set_constr_out(ir_node *node, int pos, const arch_register_req_t *req) @@ -259,24 +209,24 @@ void be_set_constr_out(ir_node *node, int pos, const arch_register_req_t *req) /** * Initializes the generic attribute of all be nodes and return it. */ -static void *init_node_attr(ir_node *node, int n_inputs, int n_outputs) +static void init_node_attr(ir_node *node, int n_inputs, int n_outputs) { ir_graph *irg = get_irn_irg(node); struct obstack *obst = be_get_be_obst(irg); - be_node_attr_t *a = get_irn_attr(node); backend_info_t *info = be_get_info(node); - - memset(a, 0, sizeof(get_op_attr_size(get_irn_op(node)))); + const arch_register_req_t **in_reqs; if (n_inputs >= 0) { int i; - a->reg_data = NEW_ARR_D(be_reg_data_t, obst, n_inputs); + assert(n_inputs == get_irn_arity(node)); + in_reqs = OALLOCN(obst, const arch_register_req_t*, n_inputs); for (i = 0; i < n_inputs; ++i) { - a->reg_data[i].in_req = arch_no_register_req; + in_reqs[i] = arch_no_register_req; } } else { - a->reg_data = NEW_ARR_F(be_reg_data_t, 0); + in_reqs = NEW_ARR_F(const arch_register_req_t*, 0); } + info->in_reqs = in_reqs; if (n_outputs >= 0) { int i; @@ -288,8 +238,6 @@ static void *init_node_attr(ir_node *node, int n_inputs, int n_outputs) } else { info->out_infos = NEW_ARR_F(reg_out_info_t, 0); } - - return a; } static void add_register_req_out(ir_node *node) @@ -303,11 +251,8 @@ static void add_register_req_out(ir_node *node) static void add_register_req_in(ir_node *node) { - be_node_attr_t *a = get_irn_attr(node); - be_reg_data_t regreq; - memset(®req, 0, sizeof(regreq)); - regreq.in_req = arch_no_register_req; - ARR_APP1(be_reg_data_t, a->reg_data, regreq); + backend_info_t *info = be_get_info(node); + ARR_APP1(const arch_register_req_t*, info->in_reqs, arch_no_register_req); } ir_node *be_new_Spill(const arch_register_class_t *cls, @@ -322,19 +267,21 @@ ir_node *be_new_Spill(const arch_register_class_t *cls, in[0] = frame; in[1] = to_spill; res = new_ir_node(NULL, irg, bl, op_be_Spill, mode_M, 2, in); - a = init_node_attr(res, 2, 1); + init_node_attr(res, 2, 1); + a = (be_frame_attr_t*) get_irn_generic_attr(res); a->ent = NULL; a->offset = 0; + a->base.exc.pin_state = op_pin_state_pinned; - be_node_set_reg_class_in(res, be_pos_Spill_frame, cls_frame); - be_node_set_reg_class_in(res, be_pos_Spill_val, cls); + be_node_set_reg_class_in(res, n_be_Spill_frame, cls_frame); + be_node_set_reg_class_in(res, n_be_Spill_val, cls); /* * For spills and reloads, we return "none" as requirement for frame * pointer, so every input is ok. Some backends need this (STA). * Matze: we should investigate if this is really needed, this solution * looks very hacky to me */ - be_set_constr_in(res, be_pos_Spill_frame, arch_no_register_req); + be_set_constr_in(res, n_be_Spill_frame, arch_no_register_req); arch_set_out_register_req(res, 0, arch_no_register_req); @@ -348,6 +295,7 @@ ir_node *be_new_Reload(const arch_register_class_t *cls, ir_node *in[2]; ir_node *res; ir_graph *irg = get_Block_irg(block); + be_frame_attr_t *a; in[0] = frame; in[1] = mem; @@ -356,16 +304,21 @@ ir_node *be_new_Reload(const arch_register_class_t *cls, init_node_attr(res, 2, 1); be_node_set_reg_class_out(res, 0, cls); - be_node_set_reg_class_in(res, be_pos_Reload_frame, cls_frame); + be_node_set_reg_class_in(res, n_be_Reload_frame, cls_frame); arch_irn_set_flags(res, arch_irn_flags_rematerializable); + a = (be_frame_attr_t*) get_irn_generic_attr(res); + a->ent = NULL; + a->offset = 0; + a->base.exc.pin_state = op_pin_state_pinned; + /* * For spills and reloads, we return "none" as requirement for frame * pointer, so every input is ok. Some backends need this (e.g. STA). * Matze: we should investigate if this is really needed, this solution * looks very hacky to me */ - be_set_constr_in(res, be_pos_Reload_frame, arch_no_register_req); + be_set_constr_in(res, n_be_Reload_frame, arch_no_register_req); return res; } @@ -373,35 +326,38 @@ ir_node *be_new_Reload(const arch_register_class_t *cls, ir_node *be_get_Reload_mem(const ir_node *irn) { assert(be_is_Reload(irn)); - return get_irn_n(irn, be_pos_Reload_mem); + return get_irn_n(irn, n_be_Reload_mem); } ir_node *be_get_Reload_frame(const ir_node *irn) { assert(be_is_Reload(irn)); - return get_irn_n(irn, be_pos_Reload_frame); + return get_irn_n(irn, n_be_Reload_frame); } ir_node *be_get_Spill_val(const ir_node *irn) { assert(be_is_Spill(irn)); - return get_irn_n(irn, be_pos_Spill_val); + return get_irn_n(irn, n_be_Spill_val); } ir_node *be_get_Spill_frame(const ir_node *irn) { assert(be_is_Spill(irn)); - return get_irn_n(irn, be_pos_Spill_frame); + return get_irn_n(irn, n_be_Spill_frame); } ir_node *be_new_Perm(const arch_register_class_t *cls, ir_node *block, int n, ir_node *in[]) { - int i; - ir_graph *irg = get_Block_irg(block); + int i; + ir_graph *irg = get_Block_irg(block); + be_node_attr_t *attr; ir_node *irn = new_ir_node(NULL, irg, block, op_be_Perm, mode_T, n, in); init_node_attr(irn, n, n); + attr = (be_node_attr_t*) get_irn_generic_attr(irn); + attr->exc.pin_state = op_pin_state_pinned; for (i = 0; i < n; ++i) { be_node_set_reg_class_in(irn, i, cls); be_node_set_reg_class_out(irn, i, cls); @@ -412,29 +368,28 @@ ir_node *be_new_Perm(const arch_register_class_t *cls, ir_node *block, void be_Perm_reduce(ir_node *perm, int new_size, int *map) { - int arity = get_irn_arity(perm); - be_reg_data_t *old_data = ALLOCAN(be_reg_data_t, arity); - reg_out_info_t *old_infos = ALLOCAN(reg_out_info_t, arity); - be_node_attr_t *attr = get_irn_attr(perm); - backend_info_t *info = be_get_info(perm); + int arity = get_irn_arity(perm); + const arch_register_req_t **old_in_reqs + = ALLOCAN(const arch_register_req_t*, arity); + reg_out_info_t *old_infos = ALLOCAN(reg_out_info_t, arity); + backend_info_t *info = be_get_info(perm); ir_node **new_in; - - int i; + int i; assert(be_is_Perm(perm)); assert(new_size <= arity); - new_in = alloca(new_size * sizeof(*new_in)); + new_in = ALLOCAN(ir_node*, new_size); /* save the old register data */ - memcpy(old_data, attr->reg_data, arity * sizeof(old_data[0])); + memcpy(old_in_reqs, info->in_reqs, arity * sizeof(old_in_reqs[0])); memcpy(old_infos, info->out_infos, arity * sizeof(old_infos[0])); - /* compose the new in array and set the new register data directly in place */ + /* compose the new in array and set the new register data directly */ for (i = 0; i < new_size; ++i) { int idx = map[i]; new_in[i] = get_irn_n(perm, idx); - attr->reg_data[i] = old_data[idx]; + info->in_reqs[i] = old_in_reqs[idx]; info->out_infos[i] = old_infos[idx]; } @@ -460,7 +415,7 @@ ir_node *be_new_MemPerm(ir_node *block, int n, ir_node *in[]) init_node_attr(irn, n + 1, n); be_node_set_reg_class_in(irn, 0, sp->reg_class); - attr = get_irn_attr(irn); + attr = (be_memperm_attr_t*)get_irn_generic_attr(irn); attr->in_entities = OALLOCNZ(irg->obst, ir_entity*, n); attr->out_entities = OALLOCNZ(irg->obst, ir_entity*, n); @@ -472,11 +427,14 @@ ir_node *be_new_Copy(const arch_register_class_t *cls, ir_node *bl, ir_node *op) ir_node *in[1]; ir_node *res; arch_register_req_t *req; + be_node_attr_t *attr; ir_graph *irg = get_Block_irg(bl); in[0] = op; res = new_ir_node(NULL, irg, bl, op_be_Copy, get_irn_mode(op), 1, in); init_node_attr(res, 1, 1); + attr = (be_node_attr_t*) get_irn_generic_attr(res); + attr->exc.pin_state = op_pin_state_floats; be_node_set_reg_class_in(res, 0, cls); be_node_set_reg_class_out(res, 0, cls); @@ -484,6 +442,7 @@ ir_node *be_new_Copy(const arch_register_class_t *cls, ir_node *bl, ir_node *op) req->cls = cls; req->type = arch_register_req_type_should_be_same; req->other_same = 1U << 0; + req->width = 1; be_set_constr_out(res, 0, req); return res; @@ -491,12 +450,12 @@ ir_node *be_new_Copy(const arch_register_class_t *cls, ir_node *bl, ir_node *op) ir_node *be_get_Copy_op(const ir_node *cpy) { - return get_irn_n(cpy, be_pos_Copy_op); + return get_irn_n(cpy, n_be_Copy_op); } void be_set_Copy_op(ir_node *cpy, ir_node *op) { - set_irn_n(cpy, be_pos_Copy_op, op); + set_irn_n(cpy, n_be_Copy_op, op); } ir_node *be_new_Keep(ir_node *block, int n, ir_node *in[]) @@ -504,9 +463,12 @@ ir_node *be_new_Keep(ir_node *block, int n, ir_node *in[]) int i; ir_node *res; ir_graph *irg = get_Block_irg(block); + be_node_attr_t *attr; res = new_ir_node(NULL, irg, block, op_be_Keep, mode_ANY, -1, NULL); init_node_attr(res, -1, 1); + attr = (be_node_attr_t*) get_irn_generic_attr(res); + attr->exc.pin_state = op_pin_state_pinned; for (i = 0; i < n; ++i) { add_irn_n(res, in[i]); @@ -527,75 +489,71 @@ void be_Keep_add_node(ir_node *keep, const arch_register_class_t *cls, ir_node * be_node_set_reg_class_in(keep, n, cls); } -/* creates a be_Call */ ir_node *be_new_Call(dbg_info *dbg, ir_graph *irg, ir_node *bl, ir_node *mem, ir_node *sp, ir_node *ptr, int n_outs, int n, ir_node *in[], ir_type *call_tp) { be_call_attr_t *a; - int real_n = be_pos_Call_first_arg + n; + int real_n = n_be_Call_first_arg + n; ir_node *irn; ir_node **real_in; NEW_ARR_A(ir_node *, real_in, real_n); - real_in[be_pos_Call_mem] = mem; - real_in[be_pos_Call_sp] = sp; - real_in[be_pos_Call_ptr] = ptr; - memcpy(&real_in[be_pos_Call_first_arg], in, n * sizeof(in[0])); + real_in[n_be_Call_mem] = mem; + real_in[n_be_Call_sp] = sp; + real_in[n_be_Call_ptr] = ptr; + memcpy(&real_in[n_be_Call_first_arg], in, n * sizeof(in[0])); irn = new_ir_node(dbg, irg, bl, op_be_Call, mode_T, real_n, real_in); - a = init_node_attr(irn, real_n, n_outs); - a->ent = NULL; - a->call_tp = call_tp; - a->pop = 0; + init_node_attr(irn, real_n, n_outs); + a = (be_call_attr_t*)get_irn_generic_attr(irn); + a->ent = NULL; + a->call_tp = call_tp; + a->pop = 0; + a->base.exc.pin_state = op_pin_state_pinned; return irn; } -/* Gets the call entity or NULL if this is no static call. */ ir_entity *be_Call_get_entity(const ir_node *call) { - const be_call_attr_t *a = get_irn_attr_const(call); + const be_call_attr_t *a = (const be_call_attr_t*)get_irn_generic_attr_const(call); assert(be_is_Call(call)); return a->ent; } -/* Sets the call entity. */ void be_Call_set_entity(ir_node *call, ir_entity *ent) { - be_call_attr_t *a = get_irn_attr(call); + be_call_attr_t *a = (be_call_attr_t*)get_irn_generic_attr(call); assert(be_is_Call(call)); a->ent = ent; } -/* Gets the call type. */ ir_type *be_Call_get_type(ir_node *call) { - const be_call_attr_t *a = get_irn_attr_const(call); + const be_call_attr_t *a = (const be_call_attr_t*)get_irn_generic_attr_const(call); assert(be_is_Call(call)); return a->call_tp; } -/* Sets the call type. */ void be_Call_set_type(ir_node *call, ir_type *call_tp) { - be_call_attr_t *a = get_irn_attr(call); + be_call_attr_t *a = (be_call_attr_t*)get_irn_generic_attr(call); assert(be_is_Call(call)); a->call_tp = call_tp; } void be_Call_set_pop(ir_node *call, unsigned pop) { - be_call_attr_t *a = get_irn_attr(call); + be_call_attr_t *a = (be_call_attr_t*)get_irn_generic_attr(call); a->pop = pop; } unsigned be_Call_get_pop(const ir_node *call) { - const be_call_attr_t *a = get_irn_attr_const(call); + const be_call_attr_t *a = (const be_call_attr_t*)get_irn_generic_attr_const(call); return a->pop; } -/* Construct a new be_Return. */ ir_node *be_new_Return(dbg_info *dbg, ir_graph *irg, ir_node *block, int n_res, unsigned pop, int n, ir_node *in[]) { @@ -611,39 +569,36 @@ ir_node *be_new_Return(dbg_info *dbg, ir_graph *irg, ir_node *block, int n_res, } be_set_constr_out(res, 0, arch_no_register_req); - a = get_irn_attr(res); - a->num_ret_vals = n_res; - a->pop = pop; - a->emit_pop = 0; + a = (be_return_attr_t*)get_irn_generic_attr(res); + a->num_ret_vals = n_res; + a->pop = pop; + a->emit_pop = 0; + a->base.exc.pin_state = op_pin_state_pinned; return res; } -/* Returns the number of real returns values */ int be_Return_get_n_rets(const ir_node *ret) { - const be_return_attr_t *a = get_irn_generic_attr_const(ret); + const be_return_attr_t *a = (const be_return_attr_t*)get_irn_generic_attr_const(ret); return a->num_ret_vals; } -/* return the number of bytes that should be popped from stack when executing the Return. */ unsigned be_Return_get_pop(const ir_node *ret) { - const be_return_attr_t *a = get_irn_generic_attr_const(ret); + const be_return_attr_t *a = (const be_return_attr_t*)get_irn_generic_attr_const(ret); return a->pop; } -/* return non-zero, if number of popped bytes must be always emitted */ int be_Return_get_emit_pop(const ir_node *ret) { - const be_return_attr_t *a = get_irn_generic_attr_const(ret); + const be_return_attr_t *a = (const be_return_attr_t*)get_irn_generic_attr_const(ret); return a->emit_pop; } -/* return non-zero, if number of popped bytes must be always emitted */ void be_Return_set_emit_pop(ir_node *ret, int emit_pop) { - be_return_attr_t *a = get_irn_generic_attr(ret); + be_return_attr_t *a = (be_return_attr_t*)get_irn_generic_attr(ret); a->emit_pop = emit_pop; } @@ -668,9 +623,11 @@ ir_node *be_new_IncSP(const arch_register_t *sp, ir_node *bl, in[0] = old_sp; irn = new_ir_node(NULL, irg, bl, op_be_IncSP, sp->reg_class->mode, sizeof(in) / sizeof(in[0]), in); - a = init_node_attr(irn, 1, 1); - a->offset = offset; - a->align = align; + init_node_attr(irn, 1, 1); + a = (be_incsp_attr_t*)get_irn_generic_attr(irn); + a->offset = offset; + a->align = align; + a->base.exc.pin_state = op_pin_state_pinned; /* Set output constraint to stack register. */ be_node_set_reg_class_in(irn, 0, sp->reg_class); @@ -682,46 +639,50 @@ ir_node *be_new_IncSP(const arch_register_t *sp, ir_node *bl, ir_node *be_new_AddSP(const arch_register_t *sp, ir_node *bl, ir_node *old_sp, ir_node *sz) { - be_node_attr_t *a; ir_node *irn; - ir_node *in[be_pos_AddSP_last]; - const arch_register_class_t *cls; + ir_node *in[n_be_AddSP_last]; ir_graph *irg; + be_node_attr_t *attr; - in[be_pos_AddSP_old_sp] = old_sp; - in[be_pos_AddSP_size] = sz; + in[n_be_AddSP_old_sp] = old_sp; + in[n_be_AddSP_size] = sz; irg = get_Block_irg(bl); - irn = new_ir_node(NULL, irg, bl, op_be_AddSP, mode_T, be_pos_AddSP_last, in); - a = init_node_attr(irn, be_pos_AddSP_last, pn_be_AddSP_last); + irn = new_ir_node(NULL, irg, bl, op_be_AddSP, mode_T, n_be_AddSP_last, in); + init_node_attr(irn, n_be_AddSP_last, pn_be_AddSP_last); + attr = (be_node_attr_t*) get_irn_generic_attr(irn); + attr->exc.pin_state = op_pin_state_pinned; /* Set output constraint to stack register. */ - be_set_constr_single_reg_in(irn, be_pos_AddSP_old_sp, sp, 0); - be_node_set_reg_class_in(irn, be_pos_AddSP_size, arch_register_get_class(sp)); - be_set_constr_single_reg_out(irn, pn_be_AddSP_sp, sp, arch_register_req_type_produces_sp); - - cls = arch_register_get_class(sp); + be_set_constr_single_reg_in(irn, n_be_AddSP_old_sp, sp, + arch_register_req_type_none); + be_node_set_reg_class_in(irn, n_be_AddSP_size, arch_register_get_class(sp)); + be_set_constr_single_reg_out(irn, pn_be_AddSP_sp, sp, + arch_register_req_type_produces_sp); return irn; } ir_node *be_new_SubSP(const arch_register_t *sp, ir_node *bl, ir_node *old_sp, ir_node *sz) { - be_node_attr_t *a; ir_node *irn; - ir_node *in[be_pos_SubSP_last]; + ir_node *in[n_be_SubSP_last]; ir_graph *irg; + be_node_attr_t *attr; - in[be_pos_SubSP_old_sp] = old_sp; - in[be_pos_SubSP_size] = sz; + in[n_be_SubSP_old_sp] = old_sp; + in[n_be_SubSP_size] = sz; irg = get_Block_irg(bl); - irn = new_ir_node(NULL, irg, bl, op_be_SubSP, mode_T, be_pos_SubSP_last, in); - a = init_node_attr(irn, be_pos_SubSP_last, pn_be_SubSP_last); + irn = new_ir_node(NULL, irg, bl, op_be_SubSP, mode_T, n_be_SubSP_last, in); + init_node_attr(irn, n_be_SubSP_last, pn_be_SubSP_last); + attr = (be_node_attr_t*) get_irn_generic_attr(irn); + attr->exc.pin_state = op_pin_state_pinned; /* Set output constraint to stack register. */ - be_set_constr_single_reg_in(irn, be_pos_SubSP_old_sp, sp, 0); - be_node_set_reg_class_in(irn, be_pos_SubSP_size, arch_register_get_class(sp)); + be_set_constr_single_reg_in(irn, n_be_SubSP_old_sp, sp, + arch_register_req_type_none); + be_node_set_reg_class_in(irn, n_be_SubSP_size, arch_register_get_class(sp)); be_set_constr_single_reg_out(irn, pn_be_SubSP_sp, sp, arch_register_req_type_produces_sp); return irn; @@ -732,9 +693,12 @@ ir_node *be_new_Start(dbg_info *dbgi, ir_node *bl, int n_outs) ir_node *res; int i; ir_graph *irg = get_Block_irg(bl); + be_node_attr_t *attr; res = new_ir_node(dbgi, irg, bl, op_be_Start, mode_T, 0, NULL); init_node_attr(res, 0, -1); + attr = (be_node_attr_t*) get_irn_generic_attr(res); + attr->exc.pin_state = op_pin_state_pinned; for (i = 0; i < n_outs; ++i) { add_register_req_out(res); } @@ -751,9 +715,11 @@ ir_node *be_new_FrameAddr(const arch_register_class_t *cls_frame, ir_node *bl, i in[0] = frame; irn = new_ir_node(NULL, irg, bl, op_be_FrameAddr, get_irn_mode(frame), 1, in); - a = init_node_attr(irn, 1, 1); - a->ent = ent; - a->offset = 0; + init_node_attr(irn, 1, 1); + a = (be_frame_attr_t*)get_irn_generic_attr(irn); + a->ent = ent; + a->offset = 0; + a->base.exc.pin_state = op_pin_state_floats; be_node_set_reg_class_in(irn, 0, cls_frame); be_node_set_reg_class_out(irn, 0, cls_frame); @@ -763,12 +729,12 @@ ir_node *be_new_FrameAddr(const arch_register_class_t *cls_frame, ir_node *bl, i ir_node *be_get_FrameAddr_frame(const ir_node *node) { assert(be_is_FrameAddr(node)); - return get_irn_n(node, be_pos_FrameAddr_ptr); + return get_irn_n(node, n_be_FrameAddr_ptr); } ir_entity *be_get_FrameAddr_entity(const ir_node *node) { - const be_frame_attr_t *attr = get_irn_generic_attr_const(node); + const be_frame_attr_t *attr = (const be_frame_attr_t*)get_irn_generic_attr_const(node); return attr->ent; } @@ -777,11 +743,14 @@ ir_node *be_new_CopyKeep(const arch_register_class_t *cls, ir_node *bl, ir_node ir_node *irn; ir_node **in = ALLOCAN(ir_node*, n + 1); ir_graph *irg = get_Block_irg(bl); + be_node_attr_t *attr; in[0] = src; memcpy(&in[1], in_keep, n * sizeof(in[0])); irn = new_ir_node(NULL, irg, bl, op_be_CopyKeep, mode, n + 1, in); init_node_attr(irn, n + 1, 1); + attr = (be_node_attr_t*) get_irn_generic_attr(irn); + attr->exc.pin_state = op_pin_state_floats; be_node_set_reg_class_in(irn, 0, cls); be_node_set_reg_class_out(irn, 0, cls); @@ -795,41 +764,12 @@ ir_node *be_new_CopyKeep_single(const arch_register_class_t *cls, ir_node *bl, i ir_node *be_get_CopyKeep_op(const ir_node *cpy) { - return get_irn_n(cpy, be_pos_CopyKeep_op); + return get_irn_n(cpy, n_be_CopyKeep_op); } void be_set_CopyKeep_op(ir_node *cpy, ir_node *op) { - set_irn_n(cpy, be_pos_CopyKeep_op, op); -} - -ir_node *be_new_Barrier(ir_node *bl, int n, ir_node *in[]) -{ - ir_node *res; - int i; - ir_graph *irg = get_Block_irg(bl); - - res = new_ir_node(NULL, irg, bl, op_be_Barrier, mode_T, -1, NULL); - init_node_attr(res, -1, -1); - for (i = 0; i < n; ++i) { - add_irn_n(res, in[i]); - add_register_req_in(res); - add_register_req_out(res); - } - - return res; -} - -ir_node *be_Barrier_append_node(ir_node *barrier, ir_node *node) -{ - ir_mode *mode = get_irn_mode(node); - int n = add_irn_n(barrier, node); - - ir_node *proj = new_r_Proj(barrier, mode, n); - add_register_req_in(barrier); - add_register_req_out(barrier); - - return proj; + set_irn_n(cpy, n_be_CopyKeep_op, op); } static bool be_has_frame_entity(const ir_node *irn) @@ -847,7 +787,7 @@ static bool be_has_frame_entity(const ir_node *irn) ir_entity *be_get_frame_entity(const ir_node *irn) { if (be_has_frame_entity(irn)) { - const be_frame_attr_t *a = get_irn_attr_const(irn); + const be_frame_attr_t *a = (const be_frame_attr_t*)get_irn_generic_attr_const(irn); return a->ent; } return NULL; @@ -857,7 +797,7 @@ int be_get_frame_offset(const ir_node *irn) { assert(is_be_node(irn)); if (be_has_frame_entity(irn)) { - const be_frame_attr_t *a = get_irn_attr_const(irn); + const be_frame_attr_t *a = (const be_frame_attr_t*)get_irn_generic_attr_const(irn); return a->offset; } return 0; @@ -865,7 +805,7 @@ int be_get_frame_offset(const ir_node *irn) void be_set_MemPerm_in_entity(const ir_node *irn, int n, ir_entity *ent) { - const be_memperm_attr_t *attr = get_irn_attr_const(irn); + const be_memperm_attr_t *attr = (const be_memperm_attr_t*)get_irn_generic_attr_const(irn); assert(be_is_MemPerm(irn)); assert(n < be_get_MemPerm_entity_arity(irn)); @@ -875,7 +815,7 @@ void be_set_MemPerm_in_entity(const ir_node *irn, int n, ir_entity *ent) ir_entity* be_get_MemPerm_in_entity(const ir_node* irn, int n) { - const be_memperm_attr_t *attr = get_irn_attr_const(irn); + const be_memperm_attr_t *attr = (const be_memperm_attr_t*)get_irn_generic_attr_const(irn); assert(be_is_MemPerm(irn)); assert(n < be_get_MemPerm_entity_arity(irn)); @@ -885,7 +825,7 @@ ir_entity* be_get_MemPerm_in_entity(const ir_node* irn, int n) void be_set_MemPerm_out_entity(const ir_node *irn, int n, ir_entity *ent) { - const be_memperm_attr_t *attr = get_irn_attr_const(irn); + const be_memperm_attr_t *attr = (const be_memperm_attr_t*)get_irn_generic_attr_const(irn); assert(be_is_MemPerm(irn)); assert(n < be_get_MemPerm_entity_arity(irn)); @@ -895,7 +835,7 @@ void be_set_MemPerm_out_entity(const ir_node *irn, int n, ir_entity *ent) ir_entity* be_get_MemPerm_out_entity(const ir_node* irn, int n) { - const be_memperm_attr_t *attr = get_irn_attr_const(irn); + const be_memperm_attr_t *attr = (const be_memperm_attr_t*)get_irn_generic_attr_const(irn); assert(be_is_MemPerm(irn)); assert(n < be_get_MemPerm_entity_arity(irn)); @@ -911,7 +851,7 @@ int be_get_MemPerm_entity_arity(const ir_node *irn) const arch_register_req_t *be_create_reg_req(struct obstack *obst, const arch_register_t *reg, arch_register_req_type_t additional_types) { - arch_register_req_t *req = obstack_alloc(obst, sizeof(*req)); + arch_register_req_t *req = OALLOC(obst, arch_register_req_t); const arch_register_class_t *cls = arch_register_get_class(reg); unsigned *limited_bitset; @@ -921,6 +861,7 @@ const arch_register_req_t *be_create_reg_req(struct obstack *obst, req->type = arch_register_req_type_limited | additional_types; req->cls = cls; req->limited = limited_bitset; + req->width = 1; return req; } @@ -942,10 +883,12 @@ void be_set_constr_single_reg_in(ir_node *node, int pos, void be_set_constr_single_reg_out(ir_node *node, int pos, const arch_register_t *reg, arch_register_req_type_t additional_types) { + ir_graph *irg = get_irn_irg(node); + be_irg_t *birg = be_birg_from_irg(irg); const arch_register_req_t *req; /* if we have an ignore register, add ignore flag and just assign it */ - if (reg->type & arch_register_type_ignore) { + if (!rbitset_is_set(birg->allocatable_regs, reg->global_index)) { additional_types |= arch_register_req_type_ignore; } @@ -987,21 +930,21 @@ void be_set_IncSP_pred(ir_node *incsp, ir_node *pred) void be_set_IncSP_offset(ir_node *irn, int offset) { - be_incsp_attr_t *a = get_irn_attr(irn); + be_incsp_attr_t *a = (be_incsp_attr_t*)get_irn_generic_attr(irn); assert(be_is_IncSP(irn)); a->offset = offset; } int be_get_IncSP_offset(const ir_node *irn) { - const be_incsp_attr_t *a = get_irn_attr_const(irn); + const be_incsp_attr_t *a = (const be_incsp_attr_t*)get_irn_generic_attr_const(irn); assert(be_is_IncSP(irn)); return a->offset; } int be_get_IncSP_align(const ir_node *irn) { - const be_incsp_attr_t *a = get_irn_attr_const(irn); + const be_incsp_attr_t *a = (const be_incsp_attr_t*)get_irn_generic_attr_const(irn); assert(be_is_IncSP(irn)); return a->align; } @@ -1031,7 +974,9 @@ ir_node *be_reload(const arch_register_class_t *cls, ir_node *insert, ir_mode *m reload = be_new_Reload(cls, cls_frame, bl, frame, spill, mode); if (is_Block(insert)) { - insert = sched_skip(insert, 0, sched_skip_cf_predicator, NULL); + do { + insert = sched_prev(insert); + } while (is_cfop(insert)); sched_add_after(insert, reload); } else { sched_add_before(insert, reload); @@ -1040,28 +985,6 @@ ir_node *be_reload(const arch_register_class_t *cls, ir_node *insert, ir_mode *m return reload; } -/* - ____ ____ - | _ \ ___ __ _ | _ \ ___ __ _ ___ - | |_) / _ \/ _` | | |_) / _ \/ _` / __| - | _ < __/ (_| | | _ < __/ (_| \__ \ - |_| \_\___|\__, | |_| \_\___|\__, |___/ - |___/ |_| - -*/ - - -static const arch_register_req_t *be_node_get_in_reg_req( - const ir_node *irn, int pos) -{ - const be_node_attr_t *a = get_irn_attr_const(irn); - - assert(pos >= 0); - if (pos >= get_irn_arity(irn) || pos >= ARR_LEN(a->reg_data)) - return arch_no_register_req; - - return a->reg_data[pos].in_req; -} static arch_irn_class_t be_node_classify(const ir_node *irn) { @@ -1070,7 +993,7 @@ static arch_irn_class_t be_node_classify(const ir_node *irn) case beo_Reload: return arch_irn_class_reload; case beo_Perm: return arch_irn_class_perm; case beo_Copy: return arch_irn_class_copy; - default: return 0; + default: return arch_irn_class_none; } } @@ -1085,7 +1008,7 @@ void be_node_set_frame_entity(ir_node *irn, ir_entity *ent) assert(be_has_frame_entity(irn)); - a = get_irn_attr(irn); + a = (be_frame_attr_t*)get_irn_generic_attr(irn); a->ent = ent; } @@ -1096,7 +1019,7 @@ static void be_node_set_frame_offset(ir_node *irn, int offset) if (!be_has_frame_entity(irn)) return; - a = get_irn_attr(irn); + a = (be_frame_attr_t*)get_irn_generic_attr(irn); a->offset = offset; } @@ -1110,18 +1033,10 @@ static int be_node_get_sp_bias(const ir_node *irn) return 0; } -/* - ___ ____ _ _ _ _ _ _ - |_ _| _ \| \ | | | | | | __ _ _ __ __| | | ___ _ __ - | || |_) | \| | | |_| |/ _` | '_ \ / _` | |/ _ \ '__| - | || _ <| |\ | | _ | (_| | | | | (_| | | __/ | - |___|_| \_\_| \_| |_| |_|\__,_|_| |_|\__,_|_|\___|_| -*/ /* for be nodes */ static const arch_irn_ops_t be_node_irn_ops = { - be_node_get_in_reg_req, be_node_classify, be_node_get_frame_entity, be_node_set_frame_offset, @@ -1132,18 +1047,67 @@ static const arch_irn_ops_t be_node_irn_ops = { NULL, /* perform_memory_operand */ }; -static const arch_register_req_t *dummy_reg_req( - const ir_node *node, int pos) +static int get_start_reg_index(ir_graph *irg, const arch_register_t *reg) { - (void) node; - (void) pos; - return arch_no_register_req; + ir_node *start = get_irg_start(irg); + unsigned n_outs = arch_irn_get_n_outs(start); + int i; + + /* do a naive linear search... */ + for (i = 0; i < (int)n_outs; ++i) { + const arch_register_req_t *out_req + = arch_get_out_register_req(start, i); + if (! (out_req->type & arch_register_req_type_limited)) + continue; + if (out_req->cls != arch_register_get_class(reg)) + continue; + if (!rbitset_is_set(out_req->limited, reg->index)) + continue; + return i; + } + panic("Tried querying undefined register '%s' at Start", reg->name); +} + +ir_node *be_get_initial_reg_value(ir_graph *irg, const arch_register_t *reg) +{ + int i = get_start_reg_index(irg, reg); + ir_node *start = get_irg_start(irg); + ir_mode *mode = arch_register_class_mode(arch_register_get_class(reg)); + const ir_edge_t *edge; + + foreach_out_edge(start, edge) { + ir_node *proj = get_edge_src_irn(edge); + if (!is_Proj(proj)) // maybe End/Anchor + continue; + if (get_Proj_proj(proj) == i) { + return proj; + } + } + return new_r_Proj(start, mode, i); +} + +int be_find_return_reg_input(ir_node *ret, const arch_register_t *reg) +{ + int arity = get_irn_arity(ret); + int i; + /* do a naive linear search... */ + for (i = 0; i < arity; ++i) { + const arch_register_req_t *req = arch_get_in_register_req(ret, i); + if (! (req->type & arch_register_req_type_limited)) + continue; + if (req->cls != arch_register_get_class(reg)) + continue; + if (!rbitset_is_set(req->limited, reg->index)) + continue; + return i; + } + panic("Tried querying undefined register '%s' at Return", reg->name); } static arch_irn_class_t dummy_classify(const ir_node *node) { (void) node; - return 0; + return arch_irn_class_none; } static ir_entity* dummy_get_frame_entity(const ir_node *node) @@ -1167,7 +1131,6 @@ static int dummy_get_sp_bias(const ir_node *node) /* for "middleend" nodes */ static const arch_irn_ops_t dummy_be_irn_ops = { - dummy_reg_req, dummy_classify, dummy_get_frame_entity, dummy_set_frame_offset, @@ -1178,105 +1141,46 @@ static const arch_irn_ops_t dummy_be_irn_ops = { NULL, /* perform_memory_operand */ }; -/* - ____ _ _ ___ ____ _ _ _ _ _ _ - | _ \| |__ (_) |_ _| _ \| \ | | | | | | __ _ _ __ __| | | ___ _ __ - | |_) | '_ \| | | || |_) | \| | | |_| |/ _` | '_ \ / _` | |/ _ \ '__| - | __/| | | | | | || _ <| |\ | | _ | (_| | | | | (_| | | __/ | - |_| |_| |_|_| |___|_| \_\_| \_| |_| |_|\__,_|_| |_|\__,_|_|\___|_| -*/ ir_node *be_new_Phi(ir_node *block, int n_ins, ir_node **ins, ir_mode *mode, const arch_register_class_t *cls) { - struct obstack *obst = be_get_be_obst(get_irn_irg(block)); + const arch_register_req_t *req; + ir_graph *irg = get_irn_irg(block); + struct obstack *obst = be_get_be_obst(irg); backend_info_t *info; + int i; ir_node *phi = new_r_Phi(block, n_ins, ins, mode); info = be_get_info(phi); info->out_infos = NEW_ARR_D(reg_out_info_t, obst, 1); memset(info->out_infos, 0, 1 * sizeof(info->out_infos[0])); + info->in_reqs = OALLOCN(obst, const arch_register_req_t*, n_ins); + if (cls == NULL) { - info->out_infos[0].req = arch_no_register_req; + req = arch_no_register_req; } else { - info->out_infos[0].req = cls->class_req; + req = cls->class_req; + } + info->out_infos[0].req = req; + for (i = 0; i < n_ins; ++i) { + info->in_reqs[i] = req; } return phi; } -/** - * Guess correct register class of a phi node by looking at its arguments - */ -static const arch_register_req_t *get_Phi_reg_req_recursive(const ir_node *phi, - pset **visited) +void be_set_phi_reg_req(ir_node *node, const arch_register_req_t *req) { - int n = get_irn_arity(phi); - ir_node *op; + int arity = get_irn_arity(node); int i; - if (*visited && pset_find_ptr(*visited, phi)) - return NULL; - - for (i = 0; i < n; ++i) { - op = get_irn_n(phi, i); - /* Matze: don't we unnecessary constraint our phis with this? - * we only need to take the regclass IMO*/ - if (!is_Phi(op)) - return arch_get_register_req_out(op); - } - - /* - * The operands of that Phi were all Phis themselves. - * We have to start a DFS for a non-Phi argument now. - */ - if (!*visited) - *visited = pset_new_ptr(16); - - pset_insert_ptr(*visited, phi); - - for (i = 0; i < n; ++i) { - const arch_register_req_t *req; - op = get_irn_n(phi, i); - req = get_Phi_reg_req_recursive(op, visited); - if (req != NULL) - return req; - } - - return NULL; -} - -static const arch_register_req_t *phi_get_irn_reg_req(const ir_node *node, - int pos) -{ - backend_info_t *info = be_get_info(node); - const arch_register_req_t *req = info->out_infos[0].req; - (void) pos; - - if (req == NULL) { - if (!mode_is_datab(get_irn_mode(node))) { - req = arch_no_register_req; - } else { - pset *visited = NULL; - - req = get_Phi_reg_req_recursive(node, &visited); - assert(req->cls != NULL); - req = req->cls->class_req; - - if (visited != NULL) - del_pset(visited); - } - info->out_infos[0].req = req; - } - - return req; -} - -void be_set_phi_reg_req(ir_node *node, const arch_register_req_t *req) -{ backend_info_t *info = be_get_info(node); info->out_infos[0].req = req; + for (i = 0; i < arity; ++i) { + info->in_reqs[i] = req; + } assert(mode_is_datab(get_irn_mode(node))); } @@ -1307,7 +1211,6 @@ void be_dump_phi_reg_reqs(FILE *F, ir_node *node, dump_reason_t reason) } static const arch_irn_ops_t phi_irn_ops = { - phi_get_irn_reg_req, dummy_classify, dummy_get_frame_entity, dummy_set_frame_offset, @@ -1318,14 +1221,7 @@ static const arch_irn_ops_t phi_irn_ops = { NULL, /* perform_memory_operand */ }; -/* - _ _ _ ____ _ - | \ | | ___ __| | ___ | _ \ _ _ _ __ ___ _ __ (_)_ __ __ _ - | \| |/ _ \ / _` |/ _ \ | | | | | | | '_ ` _ \| '_ \| | '_ \ / _` | - | |\ | (_) | (_| | __/ | |_| | |_| | | | | | | |_) | | | | | (_| | - |_| \_|\___/ \__,_|\___| |____/ \__,_|_| |_| |_| .__/|_|_| |_|\__, | - |_| |___/ -*/ + /** * ir_op-Operation: dump a be node to file @@ -1345,26 +1241,20 @@ static void dump_node(FILE *f, ir_node *irn, dump_reason_t reason) break; case dump_node_nodeattr_txt: if (be_is_Call(irn)) { - const be_call_attr_t *a = get_irn_generic_attr_const(irn); + const be_call_attr_t *a = (const be_call_attr_t*)get_irn_generic_attr_const(irn); if (a->ent) fprintf(f, " [%s] ", get_entity_name(a->ent)); } if (be_is_IncSP(irn)) { - const be_incsp_attr_t *attr = get_irn_generic_attr_const(irn); - if (attr->offset == BE_STACK_FRAME_SIZE_EXPAND) { - fprintf(f, " [Setup Stackframe] "); - } else if (attr->offset == BE_STACK_FRAME_SIZE_SHRINK) { - fprintf(f, " [Destroy Stackframe] "); - } else { - fprintf(f, " [%d] ", attr->offset); - } + const be_incsp_attr_t *attr = (const be_incsp_attr_t*)get_irn_generic_attr_const(irn); + fprintf(f, " [%d] ", attr->offset); } break; case dump_node_info_txt: arch_dump_reqs_and_registers(f, irn); if (be_has_frame_entity(irn)) { - const be_frame_attr_t *a = get_irn_generic_attr_const(irn); + const be_frame_attr_t *a = (const be_frame_attr_t*)get_irn_generic_attr_const(irn); if (a->ent) { unsigned size = get_type_size_bytes(get_entity_type(a->ent)); ir_fprintf(f, "frame entity: %+F, offset 0x%x (%d), size 0x%x (%d) bytes\n", @@ -1375,18 +1265,13 @@ static void dump_node(FILE *f, ir_node *irn, dump_reason_t reason) switch (get_irn_opcode(irn)) { case beo_IncSP: { - const be_incsp_attr_t *a = get_irn_generic_attr_const(irn); + const be_incsp_attr_t *a = (const be_incsp_attr_t*)get_irn_generic_attr_const(irn); fprintf(f, "align: %d\n", a->align); - if (a->offset == BE_STACK_FRAME_SIZE_EXPAND) - fprintf(f, "offset: FRAME_SIZE\n"); - else if (a->offset == BE_STACK_FRAME_SIZE_SHRINK) - fprintf(f, "offset: -FRAME SIZE\n"); - else - fprintf(f, "offset: %d\n", a->offset); + fprintf(f, "offset: %d\n", a->offset); break; } case beo_Call: { - const be_call_attr_t *a = get_irn_generic_attr_const(irn); + const be_call_attr_t *a = (const be_call_attr_t*)get_irn_generic_attr_const(irn); if (a->ent) fprintf(f, "\ncalling: %s\n", get_entity_name(a->ent)); @@ -1420,8 +1305,8 @@ static void dump_node(FILE *f, ir_node *irn, dump_reason_t reason) */ static void copy_attr(ir_graph *irg, const ir_node *old_node, ir_node *new_node) { - const be_node_attr_t *old_attr = get_irn_attr_const(old_node); - be_node_attr_t *new_attr = get_irn_attr(new_node); + const void *old_attr = get_irn_generic_attr_const(old_node); + void *new_attr = get_irn_generic_attr(new_node); struct obstack *obst = be_get_be_obst(irg); backend_info_t *old_info = be_get_info(old_node); backend_info_t *new_info = be_get_info(new_node); @@ -1431,10 +1316,11 @@ static void copy_attr(ir_graph *irg, const ir_node *old_node, ir_node *new_node) memcpy(new_attr, old_attr, get_op_attr_size(get_irn_op(old_node))); + new_info->flags = old_info->flags; if (old_info->out_infos != NULL) { - unsigned n_outs = ARR_LEN(old_info->out_infos); + size_t n_outs = ARR_LEN(old_info->out_infos); /* need dyanmic out infos? */ - if (be_is_Barrier(new_node) || be_is_Perm(new_node)) { + if (be_is_Perm(new_node)) { new_info->out_infos = NEW_ARR_F(reg_out_info_t, n_outs); } else { new_info->out_infos = NEW_ARR_D(reg_out_info_t, obst, n_outs); @@ -1446,18 +1332,18 @@ static void copy_attr(ir_graph *irg, const ir_node *old_node, ir_node *new_node) } /* input infos */ - if (old_attr->reg_data != NULL) { - unsigned n_ins = ARR_LEN(old_attr->reg_data); + if (old_info->in_reqs != NULL) { + unsigned n_ins = get_irn_arity(old_node); /* need dynamic in infos? */ if (get_irn_op(old_node)->opar == oparity_dynamic) { - new_attr->reg_data = NEW_ARR_F(be_reg_data_t, n_ins); + new_info->in_reqs = NEW_ARR_F(const arch_register_req_t*, n_ins); } else { - new_attr->reg_data = NEW_ARR_D(be_reg_data_t, obst, n_ins); + new_info->in_reqs = OALLOCN(obst,const arch_register_req_t*, n_ins); } - memcpy(new_attr->reg_data, old_attr->reg_data, - n_ins * sizeof(be_reg_data_t)); + memcpy(new_info->in_reqs, old_info->in_reqs, + n_ins * sizeof(new_info->in_reqs[0])); } else { - new_attr->reg_data = NULL; + new_info->in_reqs = NULL; } } @@ -1476,7 +1362,6 @@ static const ir_op_ops be_node_op_ops = { NULL, NULL, NULL, - NULL, dump_node, NULL, &be_node_irn_ops @@ -1489,40 +1374,39 @@ int is_be_node(const ir_node *irn) void be_init_op(void) { - ir_opcode opc; + unsigned opc; /* Acquire all needed opcodes. */ - op_be_Spill = new_ir_op(beo_Spill, "be_Spill", op_pin_state_pinned, N, oparity_unary, 0, sizeof(be_frame_attr_t), &be_node_op_ops); - op_be_Reload = new_ir_op(beo_Reload, "be_Reload", op_pin_state_pinned, N, oparity_zero, 0, sizeof(be_frame_attr_t), &be_node_op_ops); - op_be_Perm = new_ir_op(beo_Perm, "be_Perm", op_pin_state_pinned, N, oparity_variable, 0, sizeof(be_node_attr_t), &be_node_op_ops); - op_be_MemPerm = new_ir_op(beo_MemPerm, "be_MemPerm", op_pin_state_pinned, N, oparity_variable, 0, sizeof(be_memperm_attr_t), &be_node_op_ops); - op_be_Copy = new_ir_op(beo_Copy, "be_Copy", op_pin_state_floats, N, oparity_unary, 0, sizeof(be_node_attr_t), &be_node_op_ops); - op_be_Keep = new_ir_op(beo_Keep, "be_Keep", op_pin_state_floats, K, oparity_dynamic, 0, sizeof(be_node_attr_t), &be_node_op_ops); - op_be_CopyKeep = new_ir_op(beo_CopyKeep, "be_CopyKeep", op_pin_state_floats, K, oparity_variable, 0, sizeof(be_node_attr_t), &be_node_op_ops); - op_be_Call = new_ir_op(beo_Call, "be_Call", op_pin_state_pinned, F|M, oparity_variable, 0, sizeof(be_call_attr_t), &be_node_op_ops); - op_be_Return = new_ir_op(beo_Return, "be_Return", op_pin_state_pinned, X, oparity_dynamic, 0, sizeof(be_return_attr_t), &be_node_op_ops); - op_be_AddSP = new_ir_op(beo_AddSP, "be_AddSP", op_pin_state_pinned, N, oparity_unary, 0, sizeof(be_node_attr_t), &be_node_op_ops); - op_be_SubSP = new_ir_op(beo_SubSP, "be_SubSP", op_pin_state_pinned, N, oparity_unary, 0, sizeof(be_node_attr_t), &be_node_op_ops); - op_be_IncSP = new_ir_op(beo_IncSP, "be_IncSP", op_pin_state_pinned, N, oparity_unary, 0, sizeof(be_incsp_attr_t), &be_node_op_ops); - op_be_Start = new_ir_op(beo_Start, "be_Start", op_pin_state_pinned, N, oparity_zero, 0, sizeof(be_node_attr_t), &be_node_op_ops); - op_be_FrameAddr = new_ir_op(beo_FrameAddr, "be_FrameAddr", op_pin_state_floats, N, oparity_unary, 0, sizeof(be_frame_attr_t), &be_node_op_ops); - op_be_Barrier = new_ir_op(beo_Barrier, "be_Barrier", op_pin_state_pinned, N, oparity_dynamic, 0, sizeof(be_node_attr_t), &be_node_op_ops); + op_be_Spill = new_ir_op(beo_Spill, "be_Spill", op_pin_state_exc_pinned, irop_flag_none, oparity_unary, 0, sizeof(be_frame_attr_t), &be_node_op_ops); + op_be_Reload = new_ir_op(beo_Reload, "be_Reload", op_pin_state_exc_pinned, irop_flag_none, oparity_zero, 0, sizeof(be_frame_attr_t), &be_node_op_ops); + op_be_Perm = new_ir_op(beo_Perm, "be_Perm", op_pin_state_exc_pinned, irop_flag_none, oparity_variable, 0, sizeof(be_node_attr_t), &be_node_op_ops); + op_be_MemPerm = new_ir_op(beo_MemPerm, "be_MemPerm", op_pin_state_exc_pinned, irop_flag_none, oparity_variable, 0, sizeof(be_memperm_attr_t), &be_node_op_ops); + op_be_Copy = new_ir_op(beo_Copy, "be_Copy", op_pin_state_exc_pinned, irop_flag_none, oparity_unary, 0, sizeof(be_node_attr_t), &be_node_op_ops); + op_be_Keep = new_ir_op(beo_Keep, "be_Keep", op_pin_state_exc_pinned, irop_flag_keep, oparity_dynamic, 0, sizeof(be_node_attr_t), &be_node_op_ops); + op_be_CopyKeep = new_ir_op(beo_CopyKeep, "be_CopyKeep", op_pin_state_exc_pinned, irop_flag_keep, oparity_variable, 0, sizeof(be_node_attr_t), &be_node_op_ops); + op_be_Call = new_ir_op(beo_Call, "be_Call", op_pin_state_exc_pinned, irop_flag_fragile|irop_flag_uses_memory, oparity_variable, 0, sizeof(be_call_attr_t), &be_node_op_ops); + ir_op_set_fragile_indices(op_be_Call, n_be_Call_mem, pn_be_Call_X_regular, pn_be_Call_X_except); + op_be_Return = new_ir_op(beo_Return, "be_Return", op_pin_state_exc_pinned, irop_flag_cfopcode, oparity_dynamic, 0, sizeof(be_return_attr_t), &be_node_op_ops); + op_be_AddSP = new_ir_op(beo_AddSP, "be_AddSP", op_pin_state_exc_pinned, irop_flag_none, oparity_unary, 0, sizeof(be_node_attr_t), &be_node_op_ops); + op_be_SubSP = new_ir_op(beo_SubSP, "be_SubSP", op_pin_state_exc_pinned, irop_flag_none, oparity_unary, 0, sizeof(be_node_attr_t), &be_node_op_ops); + op_be_IncSP = new_ir_op(beo_IncSP, "be_IncSP", op_pin_state_exc_pinned, irop_flag_none, oparity_unary, 0, sizeof(be_incsp_attr_t), &be_node_op_ops); + op_be_Start = new_ir_op(beo_Start, "be_Start", op_pin_state_exc_pinned, irop_flag_none, oparity_zero, 0, sizeof(be_node_attr_t), &be_node_op_ops); + op_be_FrameAddr = new_ir_op(beo_FrameAddr, "be_FrameAddr", op_pin_state_exc_pinned, irop_flag_none, oparity_unary, 0, sizeof(be_frame_attr_t), &be_node_op_ops); op_be_Spill->ops.node_cmp_attr = FrameAddr_cmp_attr; op_be_Reload->ops.node_cmp_attr = FrameAddr_cmp_attr; - op_be_Perm->ops.node_cmp_attr = node_cmp_attr; - op_be_MemPerm->ops.node_cmp_attr = node_cmp_attr; - op_be_Copy->ops.node_cmp_attr = node_cmp_attr; - op_be_Keep->ops.node_cmp_attr = node_cmp_attr; - op_be_CopyKeep->ops.node_cmp_attr = node_cmp_attr; + op_be_Perm->ops.node_cmp_attr = be_nodes_equal; + op_be_MemPerm->ops.node_cmp_attr = be_nodes_equal; + op_be_Copy->ops.node_cmp_attr = be_nodes_equal; + op_be_Keep->ops.node_cmp_attr = be_nodes_equal; + op_be_CopyKeep->ops.node_cmp_attr = be_nodes_equal; op_be_Call->ops.node_cmp_attr = Call_cmp_attr; op_be_Return->ops.node_cmp_attr = Return_cmp_attr; - op_be_AddSP->ops.node_cmp_attr = node_cmp_attr; - op_be_SubSP->ops.node_cmp_attr = node_cmp_attr; + op_be_AddSP->ops.node_cmp_attr = be_nodes_equal; + op_be_SubSP->ops.node_cmp_attr = be_nodes_equal; op_be_IncSP->ops.node_cmp_attr = IncSP_cmp_attr; - op_be_Start->ops.node_cmp_attr = node_cmp_attr; + op_be_Start->ops.node_cmp_attr = be_nodes_equal; op_be_FrameAddr->ops.node_cmp_attr = FrameAddr_cmp_attr; - op_be_Barrier->ops.node_cmp_attr = node_cmp_attr; /* attach out dummy_ops to middle end nodes */ for (opc = iro_First; opc <= iro_Last; ++opc) {