X-Git-Url: http://nsz.repo.hu/git/?a=blobdiff_plain;f=ir%2Fbe%2Fbenode.c;h=215aea191706e2d866ba46e28efd5a919dd441ba;hb=ca21c59ea00ff05918de26952e91ac39f1589e01;hp=2ad91a7000eec22f334f10494a8a226f1610886f;hpb=55766c7570f253b4e38b169bf178e9bf722316ee;p=libfirm diff --git a/ir/be/benode.c b/ir/be/benode.c index 2ad91a700..215aea191 100644 --- a/ir/be/benode.c +++ b/ir/be/benode.c @@ -53,9 +53,9 @@ #include "be_t.h" #include "belive_t.h" -#include "besched_t.h" -#include "benode_t.h" -#include "bearch_t.h" +#include "besched.h" +#include "benode.h" +#include "bearch.h" #include "beirgmod.h" @@ -63,8 +63,7 @@ #define get_irn_attr_const(irn) get_irn_generic_attr_const(irn) typedef struct { - arch_register_req_t req; - arch_register_req_t in_req; + const arch_register_req_t *in_req; } be_reg_data_t; /** The generic be nodes attribute type. */ @@ -121,10 +120,9 @@ ir_op *op_be_Return; ir_op *op_be_IncSP; ir_op *op_be_AddSP; ir_op *op_be_SubSP; -ir_op *op_be_RegParams; +ir_op *op_be_Start; ir_op *op_be_FrameAddr; ir_op *op_be_Barrier; -ir_op *op_be_Unwind; static const ir_op_ops be_node_op_ops; @@ -145,49 +143,41 @@ static const ir_op_ops be_node_op_ops; * * @return zero if both attributes are identically */ -static int _node_cmp_attr(const be_node_attr_t *a, const be_node_attr_t *b) { - int i, len = ARR_LEN(a->reg_data); +static int node_cmp_attr(ir_node *a, ir_node *b) +{ + const be_node_attr_t *a_attr = get_irn_attr_const(a); + const be_node_attr_t *b_attr = get_irn_attr_const(b); + int i, len = ARR_LEN(a_attr->reg_data); + + if (len != ARR_LEN(b_attr->reg_data)) + return 1; - if (len != ARR_LEN(b->reg_data)) + if (!be_nodes_equal(a, b)) return 1; for (i = len - 1; i >= 0; --i) { - if (!reg_reqs_equal(&a->reg_data[i].in_req, &b->reg_data[i].in_req) || - !reg_reqs_equal(&a->reg_data[i].req, &b->reg_data[i].req)) + if (!reg_reqs_equal(a_attr->reg_data[i].in_req, + b_attr->reg_data[i].in_req)) return 1; } return 0; } -/** - * Compare the node attributes of two be_node's. - * - * @return zero if both nodes have identically attributes - */ -static int node_cmp_attr(ir_node *a, ir_node *b) { - const be_node_attr_t *a_attr = get_irn_attr_const(a); - const be_node_attr_t *b_attr = get_irn_attr_const(b); - - if (_node_cmp_attr(a_attr, b_attr) != 0) - return 1; - - return !be_info_equal(a, b); -} - /** * Compare the attributes of two be_FrameAddr nodes. * * @return zero if both nodes have identically attributes */ -static int FrameAddr_cmp_attr(ir_node *a, ir_node *b) { +static int FrameAddr_cmp_attr(ir_node *a, ir_node *b) +{ const be_frame_attr_t *a_attr = get_irn_attr_const(a); const be_frame_attr_t *b_attr = get_irn_attr_const(b); if (a_attr->ent != b_attr->ent || a_attr->offset != b_attr->offset) return 1; - return _node_cmp_attr(&a_attr->node_attr, &b_attr->node_attr); + return node_cmp_attr(a, b); } /** @@ -195,7 +185,8 @@ static int FrameAddr_cmp_attr(ir_node *a, ir_node *b) { * * @return zero if both nodes have identically attributes */ -static int Return_cmp_attr(ir_node *a, ir_node *b) { +static int Return_cmp_attr(ir_node *a, ir_node *b) +{ const be_return_attr_t *a_attr = get_irn_attr_const(a); const be_return_attr_t *b_attr = get_irn_attr_const(b); @@ -206,7 +197,7 @@ static int Return_cmp_attr(ir_node *a, ir_node *b) { if (a_attr->emit_pop != b_attr->emit_pop) return 1; - return _node_cmp_attr(&a_attr->node_attr, &b_attr->node_attr); + return node_cmp_attr(a, b); } /** @@ -214,14 +205,15 @@ static int Return_cmp_attr(ir_node *a, ir_node *b) { * * @return zero if both nodes have identically attributes */ -static int IncSP_cmp_attr(ir_node *a, ir_node *b) { +static int IncSP_cmp_attr(ir_node *a, ir_node *b) +{ const be_incsp_attr_t *a_attr = get_irn_attr_const(a); const be_incsp_attr_t *b_attr = get_irn_attr_const(b); if (a_attr->offset != b_attr->offset) return 1; - return _node_cmp_attr(&a_attr->node_attr, &b_attr->node_attr); + return node_cmp_attr(a, b); } /** @@ -229,7 +221,8 @@ static int IncSP_cmp_attr(ir_node *a, ir_node *b) { * * @return zero if both nodes have identically attributes */ -static int Call_cmp_attr(ir_node *a, ir_node *b) { +static int Call_cmp_attr(ir_node *a, ir_node *b) +{ const be_call_attr_t *a_attr = get_irn_attr_const(a); const be_call_attr_t *b_attr = get_irn_attr_const(b); @@ -237,131 +230,144 @@ static int Call_cmp_attr(ir_node *a, ir_node *b) { a_attr->call_tp != b_attr->call_tp) return 1; - return _node_cmp_attr(&a_attr->node_attr, &b_attr->node_attr); + return node_cmp_attr(a, b); } -static inline arch_register_req_t *get_be_req(const ir_node *node, int pos) +static arch_register_req_t *allocate_reg_req(const ir_node *node) { - int idx; - const be_node_attr_t *attr; - be_reg_data_t *rd; + ir_graph *irg = get_irn_irg(node); + struct obstack *obst = be_get_birg_obst(irg); - assert(is_be_node(node)); - attr = get_irn_attr_const(node); + arch_register_req_t *req = obstack_alloc(obst, sizeof(*req)); + memset(req, 0, sizeof(*req)); + return req; +} - if (pos < 0) { - idx = -(pos + 1); - } else { - idx = pos; - assert(idx < get_irn_arity(node)); - } - assert(idx < ARR_LEN(attr->reg_data)); - rd = &attr->reg_data[idx]; +void be_set_constr_in(ir_node *node, int pos, const arch_register_req_t *req) +{ + const be_node_attr_t *attr = get_irn_attr_const(node); + be_reg_data_t *rd = &attr->reg_data[pos]; + assert(pos < ARR_LEN(attr->reg_data)); + rd->in_req = req; +} - return pos < 0 ? &rd->req : &rd->in_req; +void be_set_constr_out(ir_node *node, int pos, const arch_register_req_t *req) +{ + backend_info_t *info = be_get_info(node); + info->out_infos[pos].req = req; } /** * Initializes the generic attribute of all be nodes and return it. */ -static void *init_node_attr(ir_node *node, int max_reg_data) +static void *init_node_attr(ir_node *node, int n_inputs, int n_outputs) { - ir_graph *irg = get_irn_irg(node); - struct obstack *obst = get_irg_obstack(irg); - be_node_attr_t *a = get_irn_attr(node); + ir_graph *irg = get_irn_irg(node); + struct obstack *obst = be_get_birg_obst(irg); + be_node_attr_t *a = get_irn_attr(node); + backend_info_t *info = be_get_info(node); memset(a, 0, sizeof(get_op_attr_size(get_irn_op(node)))); - if(max_reg_data >= 0) { - backend_info_t *info = be_get_info(node); - info->out_infos = NEW_ARR_D(reg_out_info_t, obst, max_reg_data); - memset(info->out_infos, 0, max_reg_data * sizeof(info->out_infos[0])); + if (n_inputs >= 0) { + int i; + a->reg_data = NEW_ARR_D(be_reg_data_t, obst, n_inputs); + for (i = 0; i < n_inputs; ++i) { + a->reg_data[i].in_req = arch_no_register_req; + } + } else { + a->reg_data = NEW_ARR_F(be_reg_data_t, 0); + } - a->reg_data = NEW_ARR_D(be_reg_data_t, obst, max_reg_data); - memset(a->reg_data, 0, max_reg_data * sizeof(a->reg_data[0])); + if (n_outputs >= 0) { + int i; + info->out_infos = NEW_ARR_D(reg_out_info_t, obst, n_outputs); + memset(info->out_infos, 0, n_outputs * sizeof(info->out_infos[0])); + for (i = 0; i < n_outputs; ++i) { + info->out_infos[i].req = arch_no_register_req; + } } else { - backend_info_t *info = be_get_info(node); info->out_infos = NEW_ARR_F(reg_out_info_t, 0); - - a->reg_data = NEW_ARR_F(be_reg_data_t, 0); } return a; } -static void add_register_req(ir_node *node) +static void add_register_req_out(ir_node *node) { backend_info_t *info = be_get_info(node); - be_node_attr_t *a = get_irn_attr(node); - be_reg_data_t regreq; reg_out_info_t out_info; - memset(®req, 0, sizeof(regreq)); memset(&out_info, 0, sizeof(out_info)); - ARR_APP1(be_reg_data_t, a->reg_data, regreq); + out_info.req = arch_no_register_req; ARR_APP1(reg_out_info_t, info->out_infos, out_info); } -/** - * Skip Proj nodes and return their Proj numbers. - * - * If *node is a Proj or Proj(Proj) node, skip it. - * - * @param node points to the node to be skipped - * - * @return 0 if *node was no Proj node, its Proj number else. - */ -static int redir_proj(const ir_node **node) +static void add_register_req_in(ir_node *node) { - const ir_node *n = *node; - - if(is_Proj(n)) { - ir_node *irn; - - *node = irn = get_Proj_pred(n); - if(is_Proj(irn)) { - assert(get_irn_mode(irn) == mode_T); - *node = get_Proj_pred(irn); - } - return get_Proj_proj(n); - } - - return 0; + be_node_attr_t *a = get_irn_attr(node); + be_reg_data_t regreq; + memset(®req, 0, sizeof(regreq)); + regreq.in_req = arch_no_register_req; + ARR_APP1(be_reg_data_t, a->reg_data, regreq); } -ir_node *be_new_Spill(const arch_register_class_t *cls, const arch_register_class_t *cls_frame, - ir_graph *irg, ir_node *bl, ir_node *frame, ir_node *to_spill) +ir_node *be_new_Spill(const arch_register_class_t *cls, + const arch_register_class_t *cls_frame, ir_node *bl, + ir_node *frame, ir_node *to_spill) { be_frame_attr_t *a; ir_node *in[2]; ir_node *res; + ir_graph *irg = get_Block_irg(bl); in[0] = frame; in[1] = to_spill; res = new_ir_node(NULL, irg, bl, op_be_Spill, mode_M, 2, in); - a = init_node_attr(res, 2); + a = init_node_attr(res, 2, 1); a->ent = NULL; a->offset = 0; be_node_set_reg_class_in(res, be_pos_Spill_frame, cls_frame); be_node_set_reg_class_in(res, be_pos_Spill_val, cls); + /* + * For spills and reloads, we return "none" as requirement for frame + * pointer, so every input is ok. Some backends need this (STA). + * Matze: we should investigate if this is really needed, this solution + * looks very hacky to me + */ + be_set_constr_in(res, be_pos_Spill_frame, arch_no_register_req); + + arch_set_out_register_req(res, 0, arch_no_register_req); + return res; } ir_node *be_new_Reload(const arch_register_class_t *cls, - const arch_register_class_t *cls_frame, ir_graph *irg, ir_node *block, + const arch_register_class_t *cls_frame, ir_node *block, ir_node *frame, ir_node *mem, ir_mode *mode) { - ir_node *in[2]; - ir_node *res; + ir_node *in[2]; + ir_node *res; + ir_graph *irg = get_Block_irg(block); in[0] = frame; in[1] = mem; res = new_ir_node(NULL, irg, block, op_be_Reload, mode, 2, in); - init_node_attr(res, 2); + init_node_attr(res, 2, 1); be_node_set_reg_class_out(res, 0, cls); + be_node_set_reg_class_in(res, be_pos_Reload_frame, cls_frame); arch_irn_set_flags(res, arch_irn_flags_rematerializable); + + /* + * For spills and reloads, we return "none" as requirement for frame + * pointer, so every input is ok. Some backends need this (e.g. STA). + * Matze: we should investigate if this is really needed, this solution + * looks very hacky to me + */ + be_set_constr_in(res, be_pos_Reload_frame, arch_no_register_req); + return res; } @@ -389,13 +395,15 @@ ir_node *be_get_Spill_frame(const ir_node *irn) return get_irn_n(irn, be_pos_Spill_frame); } -ir_node *be_new_Perm(const arch_register_class_t *cls, ir_graph *irg, - ir_node *block, int n, ir_node *in[]) +ir_node *be_new_Perm(const arch_register_class_t *cls, ir_node *block, + int n, ir_node *in[]) { - int i; + int i; + ir_graph *irg = get_Block_irg(block); + ir_node *irn = new_ir_node(NULL, irg, block, op_be_Perm, mode_T, n, in); - init_node_attr(irn, n); - for(i = 0; i < n; ++i) { + init_node_attr(irn, n, n); + for (i = 0; i < n; ++i) { be_node_set_reg_class_in(irn, i, cls); be_node_set_reg_class_out(irn, i, cls); } @@ -434,80 +442,75 @@ void be_Perm_reduce(ir_node *perm, int new_size, int *map) set_irn_in(perm, new_size, new_in); } -ir_node *be_new_MemPerm(const arch_env_t *arch_env, ir_graph *irg, ir_node *bl, int n, ir_node *in[]) +ir_node *be_new_MemPerm(const arch_env_t *arch_env, ir_node *bl, int n, ir_node *in[]) { + ir_graph *irg = get_Block_irg(bl); ir_node *frame = get_irg_frame(irg); - const arch_register_class_t *cls_frame = arch_get_irn_reg_class_out(frame); const arch_register_t *sp = arch_env->sp; ir_node *irn; be_memperm_attr_t *attr; ir_node **real_in; - int i; real_in = ALLOCAN(ir_node*, n + 1); real_in[0] = frame; memcpy(&real_in[1], in, n * sizeof(real_in[0])); - irn = new_ir_node(NULL, irg, bl, op_be_MemPerm, mode_T, n+1, real_in); + irn = new_ir_node(NULL, irg, bl, op_be_MemPerm, mode_T, n+1, real_in); - init_node_attr(irn, n + 1); + init_node_attr(irn, n + 1, n); be_node_set_reg_class_in(irn, 0, sp->reg_class); - for (i = 0; i < n; ++i) { - be_node_set_reg_class_in(irn, i + 1, cls_frame); - be_node_set_reg_class_out(irn, i, cls_frame); - } - - attr = get_irn_attr(irn); - attr->in_entities = obstack_alloc(irg->obst, n * sizeof(attr->in_entities[0])); - memset(attr->in_entities, 0, n * sizeof(attr->in_entities[0])); - attr->out_entities = obstack_alloc(irg->obst, n*sizeof(attr->out_entities[0])); - memset(attr->out_entities, 0, n*sizeof(attr->out_entities[0])); + attr = get_irn_attr(irn); + attr->in_entities = OALLOCNZ(irg->obst, ir_entity*, n); + attr->out_entities = OALLOCNZ(irg->obst, ir_entity*, n); return irn; } - -ir_node *be_new_Copy(const arch_register_class_t *cls, ir_graph *irg, ir_node *bl, ir_node *op) +ir_node *be_new_Copy(const arch_register_class_t *cls, ir_node *bl, ir_node *op) { ir_node *in[1]; ir_node *res; arch_register_req_t *req; + ir_graph *irg = get_Block_irg(bl); in[0] = op; res = new_ir_node(NULL, irg, bl, op_be_Copy, get_irn_mode(op), 1, in); - init_node_attr(res, 1); + init_node_attr(res, 1, 1); be_node_set_reg_class_in(res, 0, cls); be_node_set_reg_class_out(res, 0, cls); - req = get_be_req(res, BE_OUT_POS(0)); - req->cls = cls; - req->type = arch_register_req_type_should_be_same; + req = allocate_reg_req(res); + req->cls = cls; + req->type = arch_register_req_type_should_be_same; req->other_same = 1U << 0; + be_set_constr_out(res, 0, req); return res; } -ir_node *be_get_Copy_op(const ir_node *cpy) { +ir_node *be_get_Copy_op(const ir_node *cpy) +{ return get_irn_n(cpy, be_pos_Copy_op); } -void be_set_Copy_op(ir_node *cpy, ir_node *op) { +void be_set_Copy_op(ir_node *cpy, ir_node *op) +{ set_irn_n(cpy, be_pos_Copy_op, op); } -ir_node *be_new_Keep(const arch_register_class_t *cls, ir_graph *irg, ir_node *bl, int n, ir_node *in[]) +ir_node *be_new_Keep(ir_node *block, int n, ir_node *in[]) { int i; ir_node *res; + ir_graph *irg = get_Block_irg(block); - res = new_ir_node(NULL, irg, bl, op_be_Keep, mode_ANY, -1, NULL); - init_node_attr(res, -1); + res = new_ir_node(NULL, irg, block, op_be_Keep, mode_ANY, -1, NULL); + init_node_attr(res, -1, 1); - for(i = 0; i < n; ++i) { + for (i = 0; i < n; ++i) { add_irn_n(res, in[i]); - add_register_req(res); - be_node_set_reg_class_in(res, i, cls); + add_register_req_in(res); } keep_alive(res); @@ -520,13 +523,14 @@ void be_Keep_add_node(ir_node *keep, const arch_register_class_t *cls, ir_node * assert(be_is_Keep(keep)); n = add_irn_n(keep, node); - add_register_req(keep); + add_register_req_in(keep); be_node_set_reg_class_in(keep, n, cls); } /* creates a be_Call */ -ir_node *be_new_Call(dbg_info *dbg, ir_graph *irg, ir_node *bl, ir_node *mem, ir_node *sp, ir_node *ptr, - int n_outs, int n, ir_node *in[], ir_type *call_tp) +ir_node *be_new_Call(dbg_info *dbg, ir_graph *irg, ir_node *bl, ir_node *mem, + ir_node *sp, ir_node *ptr, int n_outs, int n, ir_node *in[], + ir_type *call_tp) { be_call_attr_t *a; int real_n = be_pos_Call_first_arg + n; @@ -540,7 +544,7 @@ ir_node *be_new_Call(dbg_info *dbg, ir_graph *irg, ir_node *bl, ir_node *mem, ir memcpy(&real_in[be_pos_Call_first_arg], in, n * sizeof(in[0])); irn = new_ir_node(dbg, irg, bl, op_be_Call, mode_T, real_n, real_in); - a = init_node_attr(irn, (n_outs > real_n ? n_outs : real_n)); + a = init_node_attr(irn, real_n, n_outs); a->ent = NULL; a->call_tp = call_tp; a->pop = 0; @@ -548,39 +552,45 @@ ir_node *be_new_Call(dbg_info *dbg, ir_graph *irg, ir_node *bl, ir_node *mem, ir } /* Gets the call entity or NULL if this is no static call. */ -ir_entity *be_Call_get_entity(const ir_node *call) { +ir_entity *be_Call_get_entity(const ir_node *call) +{ const be_call_attr_t *a = get_irn_attr_const(call); assert(be_is_Call(call)); return a->ent; } /* Sets the call entity. */ -void be_Call_set_entity(ir_node *call, ir_entity *ent) { +void be_Call_set_entity(ir_node *call, ir_entity *ent) +{ be_call_attr_t *a = get_irn_attr(call); assert(be_is_Call(call)); a->ent = ent; } /* Gets the call type. */ -ir_type *be_Call_get_type(ir_node *call) { +ir_type *be_Call_get_type(ir_node *call) +{ const be_call_attr_t *a = get_irn_attr_const(call); assert(be_is_Call(call)); return a->call_tp; } /* Sets the call type. */ -void be_Call_set_type(ir_node *call, ir_type *call_tp) { +void be_Call_set_type(ir_node *call, ir_type *call_tp) +{ be_call_attr_t *a = get_irn_attr(call); assert(be_is_Call(call)); a->call_tp = call_tp; } -void be_Call_set_pop(ir_node *call, unsigned pop) { +void be_Call_set_pop(ir_node *call, unsigned pop) +{ be_call_attr_t *a = get_irn_attr(call); a->pop = pop; } -unsigned be_Call_get_pop(const ir_node *call) { +unsigned be_Call_get_pop(const ir_node *call) +{ const be_call_attr_t *a = get_irn_attr_const(call); return a->pop; } @@ -594,11 +604,12 @@ ir_node *be_new_Return(dbg_info *dbg, ir_graph *irg, ir_node *block, int n_res, int i; res = new_ir_node(dbg, irg, block, op_be_Return, mode_X, -1, NULL); - init_node_attr(res, -1); - for(i = 0; i < n; ++i) { + init_node_attr(res, -1, 1); + for (i = 0; i < n; ++i) { add_irn_n(res, in[i]); - add_register_req(res); + add_register_req_in(res); } + be_set_constr_out(res, 0, arch_no_register_req); a = get_irn_attr(res); a->num_ret_vals = n_res; @@ -609,49 +620,55 @@ ir_node *be_new_Return(dbg_info *dbg, ir_graph *irg, ir_node *block, int n_res, } /* Returns the number of real returns values */ -int be_Return_get_n_rets(const ir_node *ret) { +int be_Return_get_n_rets(const ir_node *ret) +{ const be_return_attr_t *a = get_irn_generic_attr_const(ret); return a->num_ret_vals; } /* return the number of bytes that should be popped from stack when executing the Return. */ -unsigned be_Return_get_pop(const ir_node *ret) { +unsigned be_Return_get_pop(const ir_node *ret) +{ const be_return_attr_t *a = get_irn_generic_attr_const(ret); return a->pop; } /* return non-zero, if number of popped bytes must be always emitted */ -int be_Return_get_emit_pop(const ir_node *ret) { +int be_Return_get_emit_pop(const ir_node *ret) +{ const be_return_attr_t *a = get_irn_generic_attr_const(ret); return a->emit_pop; } /* return non-zero, if number of popped bytes must be always emitted */ -void be_Return_set_emit_pop(ir_node *ret, int emit_pop) { +void be_Return_set_emit_pop(ir_node *ret, int emit_pop) +{ be_return_attr_t *a = get_irn_generic_attr(ret); a->emit_pop = emit_pop; } -int be_Return_append_node(ir_node *ret, ir_node *node) { +int be_Return_append_node(ir_node *ret, ir_node *node) +{ int pos; pos = add_irn_n(ret, node); - add_register_req(ret); + add_register_req_in(ret); return pos; } -ir_node *be_new_IncSP(const arch_register_t *sp, ir_graph *irg, ir_node *bl, +ir_node *be_new_IncSP(const arch_register_t *sp, ir_node *bl, ir_node *old_sp, int offset, int align) { be_incsp_attr_t *a; ir_node *irn; ir_node *in[1]; + ir_graph *irg = get_Block_irg(bl); in[0] = old_sp; irn = new_ir_node(NULL, irg, bl, op_be_IncSP, sp->reg_class->mode, sizeof(in) / sizeof(in[0]), in); - a = init_node_attr(irn, 1); + a = init_node_attr(irn, 1, 1); a->offset = offset; a->align = align; @@ -662,18 +679,21 @@ ir_node *be_new_IncSP(const arch_register_t *sp, ir_graph *irg, ir_node *bl, return irn; } -ir_node *be_new_AddSP(const arch_register_t *sp, ir_graph *irg, ir_node *bl, ir_node *old_sp, ir_node *sz) +ir_node *be_new_AddSP(const arch_register_t *sp, ir_node *bl, ir_node *old_sp, + ir_node *sz) { be_node_attr_t *a; ir_node *irn; ir_node *in[be_pos_AddSP_last]; const arch_register_class_t *cls; + ir_graph *irg; in[be_pos_AddSP_old_sp] = old_sp; in[be_pos_AddSP_size] = sz; + irg = get_Block_irg(bl); irn = new_ir_node(NULL, irg, bl, op_be_AddSP, mode_T, be_pos_AddSP_last, in); - a = init_node_attr(irn, be_pos_AddSP_last); + a = init_node_attr(irn, be_pos_AddSP_last, pn_be_AddSP_last); /* Set output constraint to stack register. */ be_set_constr_single_reg_in(irn, be_pos_AddSP_old_sp, sp, 0); @@ -685,17 +705,19 @@ ir_node *be_new_AddSP(const arch_register_t *sp, ir_graph *irg, ir_node *bl, ir_ return irn; } -ir_node *be_new_SubSP(const arch_register_t *sp, ir_graph *irg, ir_node *bl, ir_node *old_sp, ir_node *sz) +ir_node *be_new_SubSP(const arch_register_t *sp, ir_node *bl, ir_node *old_sp, ir_node *sz) { be_node_attr_t *a; ir_node *irn; ir_node *in[be_pos_SubSP_last]; + ir_graph *irg; in[be_pos_SubSP_old_sp] = old_sp; in[be_pos_SubSP_size] = sz; + irg = get_Block_irg(bl); irn = new_ir_node(NULL, irg, bl, op_be_SubSP, mode_T, be_pos_SubSP_last, in); - a = init_node_attr(irn, be_pos_SubSP_last); + a = init_node_attr(irn, be_pos_SubSP_last, pn_be_SubSP_last); /* Set output constraint to stack register. */ be_set_constr_single_reg_in(irn, be_pos_SubSP_old_sp, sp, 0); @@ -705,28 +727,31 @@ ir_node *be_new_SubSP(const arch_register_t *sp, ir_graph *irg, ir_node *bl, ir_ return irn; } -ir_node *be_new_RegParams(ir_graph *irg, ir_node *bl, int n_outs) +ir_node *be_new_Start(dbg_info *dbgi, ir_node *bl, int n_outs) { ir_node *res; int i; + ir_graph *irg = get_Block_irg(bl); - res = new_ir_node(NULL, irg, bl, op_be_RegParams, mode_T, 0, NULL); - init_node_attr(res, -1); - for(i = 0; i < n_outs; ++i) - add_register_req(res); + res = new_ir_node(dbgi, irg, bl, op_be_Start, mode_T, 0, NULL); + init_node_attr(res, 0, -1); + for (i = 0; i < n_outs; ++i) { + add_register_req_out(res); + } return res; } -ir_node *be_new_FrameAddr(const arch_register_class_t *cls_frame, ir_graph *irg, ir_node *bl, ir_node *frame, ir_entity *ent) +ir_node *be_new_FrameAddr(const arch_register_class_t *cls_frame, ir_node *bl, ir_node *frame, ir_entity *ent) { be_frame_attr_t *a; ir_node *irn; ir_node *in[1]; + ir_graph *irg = get_Block_irg(bl); in[0] = frame; irn = new_ir_node(NULL, irg, bl, op_be_FrameAddr, get_irn_mode(frame), 1, in); - a = init_node_attr(irn, 1); + a = init_node_attr(irn, 1, 1); a->ent = ent; a->offset = 0; be_node_set_reg_class_in(irn, 0, cls_frame); @@ -735,7 +760,8 @@ ir_node *be_new_FrameAddr(const arch_register_class_t *cls_frame, ir_graph *irg, return optimize_node(irn); } -ir_node *be_get_FrameAddr_frame(const ir_node *node) { +ir_node *be_get_FrameAddr_frame(const ir_node *node) +{ assert(be_is_FrameAddr(node)); return get_irn_n(node, be_pos_FrameAddr_ptr); } @@ -746,44 +772,49 @@ ir_entity *be_get_FrameAddr_entity(const ir_node *node) return attr->ent; } -ir_node *be_new_CopyKeep(const arch_register_class_t *cls, ir_graph *irg, ir_node *bl, ir_node *src, int n, ir_node *in_keep[], ir_mode *mode) +ir_node *be_new_CopyKeep(const arch_register_class_t *cls, ir_node *bl, ir_node *src, int n, ir_node *in_keep[], ir_mode *mode) { ir_node *irn; ir_node **in = ALLOCAN(ir_node*, n + 1); + ir_graph *irg = get_Block_irg(bl); in[0] = src; memcpy(&in[1], in_keep, n * sizeof(in[0])); irn = new_ir_node(NULL, irg, bl, op_be_CopyKeep, mode, n + 1, in); - init_node_attr(irn, n + 1); + init_node_attr(irn, n + 1, 1); be_node_set_reg_class_in(irn, 0, cls); be_node_set_reg_class_out(irn, 0, cls); return irn; } -ir_node *be_new_CopyKeep_single(const arch_register_class_t *cls, ir_graph *irg, ir_node *bl, ir_node *src, ir_node *keep, ir_mode *mode) +ir_node *be_new_CopyKeep_single(const arch_register_class_t *cls, ir_node *bl, ir_node *src, ir_node *keep, ir_mode *mode) { - return be_new_CopyKeep(cls, irg, bl, src, 1, &keep, mode); + return be_new_CopyKeep(cls, bl, src, 1, &keep, mode); } -ir_node *be_get_CopyKeep_op(const ir_node *cpy) { +ir_node *be_get_CopyKeep_op(const ir_node *cpy) +{ return get_irn_n(cpy, be_pos_CopyKeep_op); } -void be_set_CopyKeep_op(ir_node *cpy, ir_node *op) { +void be_set_CopyKeep_op(ir_node *cpy, ir_node *op) +{ set_irn_n(cpy, be_pos_CopyKeep_op, op); } -ir_node *be_new_Barrier(ir_graph *irg, ir_node *bl, int n, ir_node *in[]) +ir_node *be_new_Barrier(ir_node *bl, int n, ir_node *in[]) { ir_node *res; int i; + ir_graph *irg = get_Block_irg(bl); res = new_ir_node(NULL, irg, bl, op_be_Barrier, mode_T, -1, NULL); - init_node_attr(res, -1); - for(i = 0; i < n; ++i) { + init_node_attr(res, -1, -1); + for (i = 0; i < n; ++i) { add_irn_n(res, in[i]); - add_register_req(res); + add_register_req_in(res); + add_register_req_out(res); } return res; @@ -791,32 +822,16 @@ ir_node *be_new_Barrier(ir_graph *irg, ir_node *bl, int n, ir_node *in[]) ir_node *be_Barrier_append_node(ir_node *barrier, ir_node *node) { - ir_graph *irg = get_irn_irg(barrier); - ir_node *block = get_nodes_block(barrier); ir_mode *mode = get_irn_mode(node); int n = add_irn_n(barrier, node); - ir_node *proj = new_r_Proj(irg, block, barrier, mode, n); - add_register_req(barrier); + ir_node *proj = new_r_Proj(barrier, mode, n); + add_register_req_in(barrier); + add_register_req_out(barrier); return proj; } -/* Construct a new be_Unwind. */ -ir_node *be_new_Unwind(dbg_info *dbg, ir_graph *irg, ir_node *block, - ir_node *mem, ir_node *sp) -{ - ir_node *res; - ir_node *in[2]; - - in[be_pos_Unwind_mem] = mem; - in[be_pos_Unwind_sp] = sp; - res = new_ir_node(dbg, irg, block, op_be_Unwind, mode_X, 2, in); - init_node_attr(res, -1); - - return res; -} - int be_has_frame_entity(const ir_node *irn) { switch (get_irn_opcode(irn)) { @@ -893,97 +908,79 @@ int be_get_MemPerm_entity_arity(const ir_node *irn) return get_irn_arity(irn) - 1; } -static void set_req_single(struct obstack *obst, arch_register_req_t *req, +static const arch_register_req_t *get_single_req(struct obstack *obst, const arch_register_t *reg, arch_register_req_type_t additional_types) { + arch_register_req_t *req = obstack_alloc(obst, sizeof(*req)); const arch_register_class_t *cls = arch_register_get_class(reg); unsigned *limited_bitset; limited_bitset = rbitset_obstack_alloc(obst, arch_register_class_n_regs(cls)); rbitset_set(limited_bitset, arch_register_get_index(reg)); - req->cls = cls; - req->type |= arch_register_req_type_limited | additional_types; + req->type = arch_register_req_type_limited | additional_types; + req->cls = cls; req->limited = limited_bitset; - + return req; } void be_set_constr_single_reg_in(ir_node *node, int pos, const arch_register_t *reg, arch_register_req_type_t additional_types) { - arch_register_req_t *req = get_be_req(node, pos); - ir_graph *irg = get_irn_irg(node); - struct obstack *obst = get_irg_obstack(irg); + const arch_register_req_t *req; - set_req_single(obst, req, reg, additional_types); + if (additional_types == 0) { + req = reg->single_req; + } else { + ir_graph *irg = get_irn_irg(node); + struct obstack *obst = be_get_birg_obst(irg); + req = get_single_req(obst, reg, additional_types); + } + be_set_constr_in(node, pos, req); } void be_set_constr_single_reg_out(ir_node *node, int pos, const arch_register_t *reg, arch_register_req_type_t additional_types) { - arch_register_req_t *req = get_be_req(node, BE_OUT_POS(pos)); - ir_graph *irg = get_irn_irg(node); - struct obstack *obst = get_irg_obstack(irg); + const arch_register_req_t *req; /* if we have an ignore register, add ignore flag and just assign it */ if (reg->type & arch_register_type_ignore) { additional_types |= arch_register_req_type_ignore; } + if (additional_types == 0) { + req = reg->single_req; + } else { + ir_graph *irg = get_irn_irg(node); + struct obstack *obst = be_get_birg_obst(irg); + req = get_single_req(obst, reg, additional_types); + } + arch_irn_set_register(node, pos, reg); - set_req_single(obst, req, reg, additional_types); + be_set_constr_out(node, pos, req); } -void be_set_constr_limited(ir_node *node, int pos, const arch_register_req_t *req) +void be_node_set_reg_class_in(ir_node *irn, int pos, + const arch_register_class_t *cls) { - ir_graph *irg = get_irn_irg(node); - struct obstack *obst = get_irg_obstack(irg); - arch_register_req_t *r = get_be_req(node, pos); - - assert(arch_register_req_is(req, limited)); - assert(!(req->type & (arch_register_req_type_should_be_same | arch_register_req_type_must_be_different))); - memcpy(r, req, sizeof(r[0])); - r->limited = rbitset_duplicate_obstack_alloc(obst, req->limited, req->cls->n_regs); + be_set_constr_in(irn, pos, cls->class_req); } -void be_node_set_reg_class_in(ir_node *irn, int pos, const arch_register_class_t *cls) +void be_node_set_reg_class_out(ir_node *irn, int pos, + const arch_register_class_t *cls) { - arch_register_req_t *req = get_be_req(irn, pos); - - req->cls = cls; - - if (cls == NULL) { - req->type = arch_register_req_type_none; - } else if (req->type == arch_register_req_type_none) { - req->type = arch_register_req_type_normal; - } -} - -void be_node_set_reg_class_out(ir_node *irn, int pos, const arch_register_class_t *cls) -{ - arch_register_req_t *req = get_be_req(irn, BE_OUT_POS(pos)); - - req->cls = cls; - - if (cls == NULL) { - req->type = arch_register_req_type_none; - } else if (req->type == arch_register_req_type_none) { - req->type = arch_register_req_type_normal; - } + be_set_constr_out(irn, pos, cls->class_req); } -void be_node_set_req_type(ir_node *irn, int pos, arch_register_req_type_t type) +ir_node *be_get_IncSP_pred(ir_node *irn) { - arch_register_req_t *req = get_be_req(irn, pos); - req->type = type; -} - -ir_node *be_get_IncSP_pred(ir_node *irn) { assert(be_is_IncSP(irn)); return get_irn_n(irn, 0); } -void be_set_IncSP_pred(ir_node *incsp, ir_node *pred) { +void be_set_IncSP_pred(ir_node *incsp, ir_node *pred) +{ assert(be_is_IncSP(incsp)); set_irn_n(incsp, 0, pred); } @@ -1011,13 +1008,13 @@ int be_get_IncSP_align(const ir_node *irn) ir_node *be_spill(ir_node *block, ir_node *irn) { - ir_graph *irg = get_irn_irg(block); + ir_graph *irg = get_Block_irg(block); ir_node *frame = get_irg_frame(irg); const arch_register_class_t *cls = arch_get_irn_reg_class_out(irn); const arch_register_class_t *cls_frame = arch_get_irn_reg_class_out(frame); ir_node *spill; - spill = be_new_Spill(cls, cls_frame, irg, block, frame, irn); + spill = be_new_Spill(cls, cls_frame, block, frame, irn); return spill; } @@ -1025,13 +1022,13 @@ ir_node *be_reload(const arch_register_class_t *cls, ir_node *insert, ir_mode *m { ir_node *reload; ir_node *bl = is_Block(insert) ? insert : get_nodes_block(insert); - ir_graph *irg = get_irn_irg(bl); + ir_graph *irg = get_Block_irg(bl); ir_node *frame = get_irg_frame(irg); const arch_register_class_t *cls_frame = arch_get_irn_reg_class_out(frame); assert(be_is_Spill(spill) || (is_Phi(spill) && get_irn_mode(spill) == mode_M)); - reload = be_new_Reload(cls, cls_frame, irg, bl, frame, spill, mode); + reload = be_new_Reload(cls, cls_frame, bl, frame, spill, mode); if (is_Block(insert)) { insert = sched_skip(insert, 0, sched_skip_cf_predicator, NULL); @@ -1054,78 +1051,26 @@ ir_node *be_reload(const arch_register_class_t *cls, ir_node *insert, ir_mode *m */ -static const -arch_register_req_t *get_out_reg_req(const ir_node *irn, int out_pos) -{ - const be_node_attr_t *a = get_irn_attr_const(irn); - - if (out_pos >= ARR_LEN(a->reg_data)) { - return arch_no_register_req; - } - - return &a->reg_data[out_pos].req; -} - -static const -arch_register_req_t *get_in_reg_req(const ir_node *irn, int pos) +static const arch_register_req_t *be_node_get_in_reg_req( + const ir_node *irn, int pos) { const be_node_attr_t *a = get_irn_attr_const(irn); + assert(pos >= 0); if (pos >= get_irn_arity(irn) || pos >= ARR_LEN(a->reg_data)) return arch_no_register_req; - return &a->reg_data[pos].in_req; -} - -static const arch_register_req_t * -be_node_get_irn_reg_req(const ir_node *irn, int pos) -{ - int out_pos = pos; - - if (pos < 0) { - if (get_irn_mode(irn) == mode_T) - return arch_no_register_req; - - assert(pos == -1); - out_pos = redir_proj((const ir_node **)&irn); - assert(is_be_node(irn)); - return get_out_reg_req(irn, out_pos); - } else if (is_be_node(irn)) { - /* - * For spills and reloads, we return "none" as requirement for frame - * pointer, so every input is ok. Some backends need this (e.g. STA). - */ - if ((pos == be_pos_Spill_frame && be_is_Spill(irn)) || - (pos == be_pos_Reload_frame && be_is_Reload(irn))) - return arch_no_register_req; - - return get_in_reg_req(irn, pos); - } - - return arch_no_register_req; + return a->reg_data[pos].in_req; } static arch_irn_class_t be_node_classify(const ir_node *irn) { -restart: switch (get_irn_opcode(irn)) { -#define XXX(a,b) case a: return b - XXX(beo_Spill, arch_irn_class_spill); - XXX(beo_Reload, arch_irn_class_reload); - XXX(beo_Perm, arch_irn_class_perm); - XXX(beo_Copy, arch_irn_class_copy); - XXX(beo_Return, arch_irn_class_branch); -#undef XXX - case iro_Proj: - irn = get_Proj_pred(irn); - if (is_Proj(irn)) { - assert(get_irn_mode(irn) == mode_T); - irn = get_Proj_pred(irn); - } - goto restart; - - default: - return 0; + case beo_Spill: return arch_irn_class_spill; + case beo_Reload: return arch_irn_class_reload; + case beo_Perm: return arch_irn_class_perm; + case beo_Copy: return arch_irn_class_copy; + default: return 0; } } @@ -1146,17 +1091,20 @@ static void be_node_set_frame_entity(ir_node *irn, ir_entity *ent) static void be_node_set_frame_offset(ir_node *irn, int offset) { - if(be_has_frame_entity(irn)) { - be_frame_attr_t *a = get_irn_attr(irn); - a->offset = offset; - } + be_frame_attr_t *a; + + if (!be_has_frame_entity(irn)) + return; + + a = get_irn_attr(irn); + a->offset = offset; } static int be_node_get_sp_bias(const ir_node *irn) { - if(be_is_IncSP(irn)) + if (be_is_IncSP(irn)) return be_get_IncSP_offset(irn); - if(be_is_Call(irn)) + if (be_is_Call(irn)) return -(int)be_Call_get_pop(irn); return 0; @@ -1171,8 +1119,9 @@ static int be_node_get_sp_bias(const ir_node *irn) */ +/* for be nodes */ static const arch_irn_ops_t be_node_irn_ops = { - be_node_get_irn_reg_req, + be_node_get_in_reg_req, be_node_classify, be_node_get_frame_entity, be_node_set_frame_entity, @@ -1184,6 +1133,60 @@ static const arch_irn_ops_t be_node_irn_ops = { NULL, /* perform_memory_operand */ }; +static const arch_register_req_t *dummy_reg_req( + const ir_node *node, int pos) +{ + (void) node; + (void) pos; + return arch_no_register_req; +} + +static arch_irn_class_t dummy_classify(const ir_node *node) +{ + (void) node; + return 0; +} + +static ir_entity* dummy_get_frame_entity(const ir_node *node) +{ + (void) node; + return NULL; +} + +static void dummy_set_frame_entity(ir_node *node, ir_entity *entity) +{ + (void) node; + (void) entity; + panic("dummy_set_frame_entity() should not be called"); +} + +static void dummy_set_frame_offset(ir_node *node, int bias) +{ + (void) node; + (void) bias; + panic("dummy_set_frame_offset() should not be called"); +} + +static int dummy_get_sp_bias(const ir_node *node) +{ + (void) node; + return 0; +} + +/* for "middleend" nodes */ +static const arch_irn_ops_t dummy_be_irn_ops = { + dummy_reg_req, + dummy_classify, + dummy_get_frame_entity, + dummy_set_frame_entity, + dummy_set_frame_offset, + dummy_get_sp_bias, + NULL, /* get_inverse */ + NULL, /* get_op_estimated_cost */ + NULL, /* possible_memory_operand */ + NULL, /* perform_memory_operand */ +}; + /* ____ _ _ ___ ____ _ _ _ _ _ _ | _ \| |__ (_) |_ _| _ \| \ | | | | | | __ _ _ __ __| | | ___ _ __ @@ -1193,52 +1196,43 @@ static const arch_irn_ops_t be_node_irn_ops = { */ -typedef struct { - const arch_register_t *reg; - arch_register_req_t req; - arch_irn_flags_t flags; -} phi_attr_t; - -static struct { - pmap *phi_attrs; -} phi_handler; - -#define get_phi_handler_from_ops(h) container_of(h, phi_handler_t, irn_ops) - -static inline -phi_attr_t *get_Phi_attr(const ir_node *phi) -{ - phi_attr_t *attr = pmap_get(phi_handler.phi_attrs, (void*) phi); - if(attr == NULL) { - ir_graph *irg = get_irn_irg(phi); - struct obstack *obst = get_irg_obstack(irg); - attr = obstack_alloc(obst, sizeof(attr[0])); - memset(attr, 0, sizeof(attr[0])); - pmap_insert(phi_handler.phi_attrs, phi, attr); +ir_node *be_new_Phi(ir_node *block, int n_ins, ir_node **ins, ir_mode *mode, + const arch_register_class_t *cls) +{ + struct obstack *obst = be_get_birg_obst(get_irn_irg(block)); + backend_info_t *info; + + ir_node *phi = new_r_Phi(block, n_ins, ins, mode); + info = be_get_info(phi); + info->out_infos = NEW_ARR_D(reg_out_info_t, obst, 1); + memset(info->out_infos, 0, 1 * sizeof(info->out_infos[0])); + if (cls == NULL) { + info->out_infos[0].req = arch_no_register_req; + } else { + info->out_infos[0].req = cls->class_req; } - return attr; + return phi; } /** - * Get register class of a Phi. + * Guess correct register class of a phi node by looking at its arguments */ -static -const arch_register_req_t *get_Phi_reg_req_recursive(const ir_node *phi, - pset **visited) +static const arch_register_req_t *get_Phi_reg_req_recursive(const ir_node *phi, + pset **visited) { int n = get_irn_arity(phi); ir_node *op; int i; - if(*visited && pset_find_ptr(*visited, phi)) + if (*visited && pset_find_ptr(*visited, phi)) return NULL; - for(i = 0; i < n; ++i) { + for (i = 0; i < n; ++i) { op = get_irn_n(phi, i); /* Matze: don't we unnecessary constraint our phis with this? * we only need to take the regclass IMO*/ - if(!is_Phi(op)) + if (!is_Phi(op)) return arch_get_register_req_out(op); } @@ -1246,136 +1240,96 @@ const arch_register_req_t *get_Phi_reg_req_recursive(const ir_node *phi, * The operands of that Phi were all Phis themselves. * We have to start a DFS for a non-Phi argument now. */ - if(!*visited) + if (!*visited) *visited = pset_new_ptr(16); pset_insert_ptr(*visited, phi); - for(i = 0; i < n; ++i) { + for (i = 0; i < n; ++i) { const arch_register_req_t *req; op = get_irn_n(phi, i); req = get_Phi_reg_req_recursive(op, visited); - if(req != NULL) + if (req != NULL) return req; } return NULL; } -static -const arch_register_req_t *phi_get_irn_reg_req(const ir_node *irn, int pos) +static const arch_register_req_t *phi_get_irn_reg_req(const ir_node *node, + int pos) { - phi_attr_t *attr; + backend_info_t *info = be_get_info(node); + const arch_register_req_t *req = info->out_infos[0].req; (void) pos; - if(!mode_is_datab(get_irn_mode(irn))) - return arch_no_register_req; + if (req == NULL) { + if (!mode_is_datab(get_irn_mode(node))) { + req = arch_no_register_req; + } else { + pset *visited = NULL; - attr = get_Phi_attr(irn); + req = get_Phi_reg_req_recursive(node, &visited); + assert(req->cls != NULL); + req = req->cls->class_req; - if(attr->req.type == arch_register_req_type_none) { - pset *visited = NULL; - const arch_register_req_t *req; - req = get_Phi_reg_req_recursive(irn, &visited); - - attr->req = *req; - assert(attr->req.cls != NULL); - attr->req.type = arch_register_req_type_normal; - - if(visited != NULL) - del_pset(visited); + if (visited != NULL) + del_pset(visited); + } + info->out_infos[0].req = req; } - return &attr->req; -} - -void be_set_phi_reg_req(ir_node *node, const arch_register_req_t *req, - arch_register_req_type_t additional_types) -{ - phi_attr_t *attr; - - assert(mode_is_datab(get_irn_mode(node))); - - attr = get_Phi_attr(node); - attr->req = *req; - attr->req.type |= additional_types; + return req; } -void be_set_phi_flags(ir_node *node, arch_irn_flags_t flags) +void be_set_phi_reg_req(ir_node *node, const arch_register_req_t *req) { - phi_attr_t *attr; + backend_info_t *info = be_get_info(node); + info->out_infos[0].req = req; assert(mode_is_datab(get_irn_mode(node))); - - attr = get_Phi_attr(node); - attr->flags = flags; -} - -static arch_irn_class_t phi_classify(const ir_node *irn) -{ - (void) irn; - return 0; -} - -static ir_entity *phi_get_frame_entity(const ir_node *irn) -{ - (void) irn; - return NULL; } -static void phi_set_frame_entity(ir_node *irn, ir_entity *ent) +int be_dump_phi_reg_reqs(ir_node *node, FILE *F, dump_reason_t reason) { - (void) irn; - (void) ent; - panic("phi_set_frame_entity() should not be called"); -} + switch (reason) { + case dump_node_opcode_txt: + fputs(get_op_name(get_irn_op(node)), F); + break; + case dump_node_mode_txt: + fprintf(F, "%s", get_mode_name(get_irn_mode(node))); + break; + case dump_node_nodeattr_txt: + break; + case dump_node_info_txt: + { + backend_info_t *info = be_get_info(node); + if (info != NULL && info->out_infos[0].req != NULL) { + arch_dump_reqs_and_registers(F, node); + } + break; + } -static void phi_set_frame_offset(ir_node *irn, int bias) -{ - (void) irn; - (void) bias; - panic("phi_set_frame_offset() should not be called"); -} + default: + break; + } -static int phi_get_sp_bias(const ir_node *irn) -{ - (void) irn; return 0; } static const arch_irn_ops_t phi_irn_ops = { phi_get_irn_reg_req, - phi_classify, - phi_get_frame_entity, - phi_set_frame_entity, - phi_set_frame_offset, - phi_get_sp_bias, + dummy_classify, + dummy_get_frame_entity, + dummy_set_frame_entity, + dummy_set_frame_offset, + dummy_get_sp_bias, NULL, /* get_inverse */ NULL, /* get_op_estimated_cost */ NULL, /* possible_memory_operand */ NULL, /* perform_memory_operand */ }; -void be_phi_handler_new(void) -{ - phi_handler.phi_attrs = pmap_create(); - op_Phi->ops.be_ops = &phi_irn_ops; -} - -void be_phi_handler_free(void) -{ - pmap_destroy(phi_handler.phi_attrs); - phi_handler.phi_attrs = NULL; - op_Phi->ops.be_ops = NULL; -} - -void be_phi_handler_reset(void) -{ - if(phi_handler.phi_attrs) - pmap_destroy(phi_handler.phi_attrs); - phi_handler.phi_attrs = pmap_create(); -} - /* _ _ _ ____ _ | \ | | ___ __| | ___ | _ \ _ _ _ __ ___ _ __ (_)_ __ __ _ @@ -1385,46 +1339,6 @@ void be_phi_handler_reset(void) |_| |___/ */ -/** - * Dumps a register requirement to a file. - */ -static void dump_node_req(FILE *f, int idx, const arch_register_req_t *req, - const ir_node *node) -{ - char tmp[256]; - - if (req->cls == NULL) return; - - arch_register_req_format(tmp, sizeof(tmp), req, node); - fprintf(f, "#%d %s\n", idx, tmp); -} - -/** - * Dumps node register requirements to a file. - */ -static void dump_node_reqs(FILE *f, ir_node *node) -{ - int i; - be_node_attr_t *a = get_irn_attr(node); - int len = ARR_LEN(a->reg_data); - - fprintf(f, "registers: \n"); - for (i = 0; i < len; ++i) { - const arch_register_t *reg = arch_irn_get_register(node, i); - fprintf(f, "#%d: %s\n", i, reg != NULL ? reg->name : "n/a"); - } - - fprintf(f, "in requirements:\n"); - for (i = 0; i < len; ++i) { - dump_node_req(f, i, &a->reg_data[i].in_req, node); - } - - fprintf(f, "\nout requirements:\n"); - for (i = 0; i < len; ++i) { - dump_node_req(f, i, &a->reg_data[i].req, node); - } -} - /** * ir_op-Operation: dump a be node to file */ @@ -1434,26 +1348,26 @@ static int dump_node(ir_node *irn, FILE *f, dump_reason_t reason) assert(is_be_node(irn)); - switch(reason) { + switch (reason) { case dump_node_opcode_txt: fputs(get_op_name(get_irn_op(irn)), f); break; case dump_node_mode_txt: - if(be_is_Perm(irn) || be_is_Copy(irn) || be_is_CopyKeep(irn)) { + if (be_is_Perm(irn) || be_is_Copy(irn) || be_is_CopyKeep(irn)) { fprintf(f, " %s", get_mode_name(get_irn_mode(irn))); } break; case dump_node_nodeattr_txt: - if(be_is_Call(irn)) { + if (be_is_Call(irn)) { be_call_attr_t *a = (be_call_attr_t *) at; if (a->ent) fprintf(f, " [%s] ", get_entity_name(a->ent)); } - if(be_is_IncSP(irn)) { + if (be_is_IncSP(irn)) { const be_incsp_attr_t *attr = get_irn_generic_attr_const(irn); - if(attr->offset == BE_STACK_FRAME_SIZE_EXPAND) { + if (attr->offset == BE_STACK_FRAME_SIZE_EXPAND) { fprintf(f, " [Setup Stackframe] "); - } else if(attr->offset == BE_STACK_FRAME_SIZE_SHRINK) { + } else if (attr->offset == BE_STACK_FRAME_SIZE_SHRINK) { fprintf(f, " [Destroy Stackframe] "); } else { fprintf(f, " [%d] ", attr->offset); @@ -1461,9 +1375,9 @@ static int dump_node(ir_node *irn, FILE *f, dump_reason_t reason) } break; case dump_node_info_txt: - dump_node_reqs(f, irn); + arch_dump_reqs_and_registers(f, irn); - if(be_has_frame_entity(irn)) { + if (be_has_frame_entity(irn)) { be_frame_attr_t *a = (be_frame_attr_t *) at; if (a->ent) { unsigned size = get_type_size_bytes(get_entity_type(a->ent)); @@ -1479,7 +1393,7 @@ static int dump_node(ir_node *irn, FILE *f, dump_reason_t reason) be_incsp_attr_t *a = (be_incsp_attr_t *) at; if (a->offset == BE_STACK_FRAME_SIZE_EXPAND) fprintf(f, "offset: FRAME_SIZE\n"); - else if(a->offset == BE_STACK_FRAME_SIZE_SHRINK) + else if (a->offset == BE_STACK_FRAME_SIZE_SHRINK) fprintf(f, "offset: -FRAME SIZE\n"); else fprintf(f, "offset: %u\n", a->offset); @@ -1496,14 +1410,14 @@ static int dump_node(ir_node *irn, FILE *f, dump_reason_t reason) case beo_MemPerm: { int i; - for(i = 0; i < be_get_MemPerm_entity_arity(irn); ++i) { + for (i = 0; i < be_get_MemPerm_entity_arity(irn); ++i) { ir_entity *in, *out; in = be_get_MemPerm_in_entity(irn, i); out = be_get_MemPerm_out_entity(irn, i); - if(in) { + if (in) { fprintf(f, "\nin[%d]: %s\n", i, get_entity_name(in)); } - if(out) { + if (out) { fprintf(f, "\nout[%d]: %s\n", i, get_entity_name(out)); } } @@ -1526,50 +1440,43 @@ static void copy_attr(const ir_node *old_node, ir_node *new_node) { const be_node_attr_t *old_attr = get_irn_attr_const(old_node); be_node_attr_t *new_attr = get_irn_attr(new_node); - struct obstack *obst = get_irg_obstack(get_irn_irg(new_node)); + ir_graph *irg = get_irn_irg(new_node); + struct obstack *obst = be_get_birg_obst(irg); backend_info_t *old_info = be_get_info(old_node); backend_info_t *new_info = be_get_info(new_node); - unsigned i, len; assert(is_be_node(old_node)); assert(is_be_node(new_node)); memcpy(new_attr, old_attr, get_op_attr_size(get_irn_op(old_node))); - new_attr->reg_data = NULL; - - if(old_attr->reg_data != NULL) - len = ARR_LEN(old_attr->reg_data); - else - len = 0; - if(get_irn_op(old_node)->opar == oparity_dynamic - || be_is_RegParams(old_node)) { - new_attr->reg_data = NEW_ARR_F(be_reg_data_t, len); - new_info->out_infos = NEW_ARR_F(reg_out_info_t, len); + if (old_info->out_infos != NULL) { + unsigned n_outs = ARR_LEN(old_info->out_infos); + /* need dyanmic out infos? */ + if (be_is_Barrier(new_node) || be_is_Perm(new_node)) { + new_info->out_infos = NEW_ARR_F(reg_out_info_t, n_outs); + } else { + new_info->out_infos = NEW_ARR_D(reg_out_info_t, obst, n_outs); + } + memcpy(new_info->out_infos, old_info->out_infos, + n_outs * sizeof(new_info->out_infos[0])); } else { - new_attr->reg_data = NEW_ARR_D(be_reg_data_t, obst, len); - new_info->out_infos = NEW_ARR_D(reg_out_info_t, obst, len); + new_info->out_infos = NULL; } - if(len > 0) { - memcpy(new_attr->reg_data, old_attr->reg_data, len * sizeof(be_reg_data_t)); - memcpy(new_info->out_infos, old_info->out_infos, len * sizeof(new_info->out_infos[0])); - for(i = 0; i < len; ++i) { - const be_reg_data_t *rd = &old_attr->reg_data[i]; - be_reg_data_t *newrd = &new_attr->reg_data[i]; - if (arch_register_req_is(&rd->req, limited)) { - const arch_register_req_t *req = &rd->req; - arch_register_req_t *new_req = &newrd->req; - new_req->limited - = rbitset_duplicate_obstack_alloc(obst, req->limited, req->cls->n_regs); - } - if(arch_register_req_is(&rd->in_req, limited)) { - const arch_register_req_t *req = &rd->in_req; - arch_register_req_t *new_req = &newrd->in_req; - new_req->limited - = rbitset_duplicate_obstack_alloc(obst, req->limited, req->cls->n_regs); - } + /* input infos */ + if (old_attr->reg_data != NULL) { + unsigned n_ins = ARR_LEN(old_attr->reg_data); + /* need dynamic in infos? */ + if (get_irn_op(old_node)->opar == oparity_dynamic) { + new_attr->reg_data = NEW_ARR_F(be_reg_data_t, n_ins); + } else { + new_attr->reg_data = NEW_ARR_D(be_reg_data_t, obst, n_ins); } + memcpy(new_attr->reg_data, old_attr->reg_data, + n_ins * sizeof(be_reg_data_t)); + } else { + new_attr->reg_data = NULL; } } @@ -1599,13 +1506,9 @@ int is_be_node(const ir_node *irn) return get_op_ops(get_irn_op(irn))->be_ops == &be_node_irn_ops; } -void be_node_init(void) { - static int inited = 0; - - if(inited) - return; - - inited = 1; +void be_init_op(void) +{ + ir_opcode opc; /* Acquire all needed opcodes. */ op_be_Spill = new_ir_op(beo_Spill, "be_Spill", op_pin_state_pinned, N, oparity_unary, 0, sizeof(be_frame_attr_t), &be_node_op_ops); @@ -1620,10 +1523,9 @@ void be_node_init(void) { op_be_AddSP = new_ir_op(beo_AddSP, "be_AddSP", op_pin_state_pinned, N, oparity_unary, 0, sizeof(be_node_attr_t), &be_node_op_ops); op_be_SubSP = new_ir_op(beo_SubSP, "be_SubSP", op_pin_state_pinned, N, oparity_unary, 0, sizeof(be_node_attr_t), &be_node_op_ops); op_be_IncSP = new_ir_op(beo_IncSP, "be_IncSP", op_pin_state_pinned, N, oparity_unary, 0, sizeof(be_incsp_attr_t), &be_node_op_ops); - op_be_RegParams = new_ir_op(beo_RegParams, "be_RegParams", op_pin_state_pinned, N, oparity_zero, 0, sizeof(be_node_attr_t), &be_node_op_ops); + op_be_Start = new_ir_op(beo_Start, "be_Start", op_pin_state_pinned, N, oparity_zero, 0, sizeof(be_node_attr_t), &be_node_op_ops); op_be_FrameAddr = new_ir_op(beo_FrameAddr, "be_FrameAddr", op_pin_state_floats, N, oparity_unary, 0, sizeof(be_frame_attr_t), &be_node_op_ops); op_be_Barrier = new_ir_op(beo_Barrier, "be_Barrier", op_pin_state_pinned, N, oparity_dynamic, 0, sizeof(be_node_attr_t), &be_node_op_ops); - op_be_Unwind = new_ir_op(beo_Unwind, "be_Unwind", op_pin_state_pinned, X, oparity_zero, 0, sizeof(be_node_attr_t), &be_node_op_ops); op_be_Spill->ops.node_cmp_attr = FrameAddr_cmp_attr; op_be_Reload->ops.node_cmp_attr = FrameAddr_cmp_attr; @@ -1637,8 +1539,16 @@ void be_node_init(void) { op_be_AddSP->ops.node_cmp_attr = node_cmp_attr; op_be_SubSP->ops.node_cmp_attr = node_cmp_attr; op_be_IncSP->ops.node_cmp_attr = IncSP_cmp_attr; - op_be_RegParams->ops.node_cmp_attr = node_cmp_attr; + op_be_Start->ops.node_cmp_attr = node_cmp_attr; op_be_FrameAddr->ops.node_cmp_attr = FrameAddr_cmp_attr; op_be_Barrier->ops.node_cmp_attr = node_cmp_attr; - op_be_Unwind->ops.node_cmp_attr = node_cmp_attr; + + /* attach out dummy_ops to middle end nodes */ + for (opc = iro_First; opc <= iro_Last; ++opc) { + ir_op *op = get_irp_opcode(opc); + assert(op->ops.be_ops == NULL); + op->ops.be_ops = &dummy_be_irn_ops; + } + + op_Phi->ops.be_ops = &phi_irn_ops; }