X-Git-Url: http://nsz.repo.hu/git/?a=blobdiff_plain;f=ir%2Fbe%2Fbenode.c;h=00d5c44bbd1fba1fbd98c3aead2b2bf87f24b358;hb=39e27d1fe336c1238bfe4a7b614d103762e39ba2;hp=9846b1c5691cc8892df002549d55b4ca6c155077;hpb=9cf5785dacb19acf921835a9dab9decf75dc44aa;p=libfirm diff --git a/ir/be/benode.c b/ir/be/benode.c index 9846b1c56..00d5c44bb 100644 --- a/ir/be/benode.c +++ b/ir/be/benode.c @@ -32,6 +32,7 @@ #include "ircons_t.h" #include "irprintf.h" #include "irgwalk.h" +#include "iropt_t.h" #include "be_t.h" #include "belive_t.h" @@ -47,17 +48,6 @@ static unsigned be_node_tag = FOURCC('B', 'E', 'N', 'O'); -#if 0 -typedef enum _node_kind_t { - node_kind_spill, - node_kind_reload, - node_kind_perm, - node_kind_copy, - node_kind_kill, - node_kind_last -} node_kind_t; -#endif - typedef enum { be_req_kind_old_limited, be_req_kind_negate_old_limited, @@ -90,11 +80,16 @@ typedef struct { be_reg_data_t *reg_data; } be_node_attr_t; +/** The be_Return nodes attribute type. */ +typedef struct { + be_node_attr_t node_attr; + int num_ret_vals; /**< number of return values */ +} be_return_attr_t; + /** The be_Stack attribute type. */ typedef struct { be_node_attr_t node_attr; - int offset; /**< The offset by which the stack shall be increased/decreased. */ - be_stack_dir_t dir; /**< The direction in which the stack shall be modified (expand or shrink). */ + int offset; /**< The offset by which the stack shall be expanded/shrinked. */ } be_stack_attr_t; /** The be_Frame attribute type. */ @@ -111,15 +106,16 @@ typedef struct { ir_type *call_tp; /**< The call type, copied from the original Call node. */ } be_call_attr_t; -/** The be_Spill attribute type. */ typedef struct { - be_frame_attr_t frame_attr; - ir_node *spill_ctx; /**< The node in whose context this spill was introduced. */ -} be_spill_attr_t; + be_node_attr_t node_attr; + entity **in_entities; + entity **out_entities; +} be_memperm_attr_t; ir_op *op_be_Spill; ir_op *op_be_Reload; ir_op *op_be_Perm; +ir_op *op_be_MemPerm; ir_op *op_be_Copy; ir_op *op_be_Keep; ir_op *op_be_CopyKeep; @@ -152,6 +148,40 @@ static const ir_op_ops be_node_op_ops; #define M irop_flag_machine +/** + * Compare two node attributes. + * + * @return zero if both attributes are identically + */ +static int cmp_node_attr(be_node_attr_t *a, be_node_attr_t *b) { + if (a->max_reg_data == b->max_reg_data) { + int i; + + for (i = 0; i < a->max_reg_data; ++i) { + if (a->reg_data[i].reg != b->reg_data[i].reg || + memcmp(&a->reg_data[i].in_req, &b->reg_data[i].in_req, sizeof(b->reg_data[i].in_req)) || + memcmp(&a->reg_data[i].req, &b->reg_data[i].req, sizeof(a->reg_data[i].req))) + return 1; + } + return 0; + } + return 1; +} + +/** + * Compare the attributes of two FrameAddr nodes. + * + * @return zero if both attributes are identically + */ +static int FrameAddr_cmp_attr(ir_node *a, ir_node *b) { + be_frame_attr_t *a_attr = get_irn_attr(a); + be_frame_attr_t *b_attr = get_irn_attr(b); + + if (a_attr->ent == b_attr->ent && a_attr->offset == b_attr->offset) + return cmp_node_attr(&a_attr->node_attr, &b_attr->node_attr); + return 1; +} + void be_node_init(void) { static int inited = 0; @@ -163,27 +193,29 @@ void be_node_init(void) { /* Acquire all needed opcodes. */ beo_base = get_next_ir_opcodes(beo_Last - 1); - op_be_Spill = new_ir_op(beo_base + beo_Spill, "be_Spill", op_pin_state_mem_pinned, N, oparity_unary, 0, sizeof(be_spill_attr_t), &be_node_op_ops); - op_be_Reload = new_ir_op(beo_base + beo_Reload, "be_Reload", op_pin_state_mem_pinned, N, oparity_zero, 0, sizeof(be_frame_attr_t), &be_node_op_ops); - op_be_Perm = new_ir_op(beo_base + beo_Perm, "be_Perm", op_pin_state_pinned, N, oparity_variable, 0, sizeof(be_node_attr_t), &be_node_op_ops); - op_be_Copy = new_ir_op(beo_base + beo_Copy, "be_Copy", op_pin_state_floats, N, oparity_unary, 0, sizeof(be_node_attr_t), &be_node_op_ops); - op_be_Keep = new_ir_op(beo_base + beo_Keep, "be_Keep", op_pin_state_pinned, K, oparity_variable, 0, sizeof(be_node_attr_t), &be_node_op_ops); - op_be_CopyKeep = new_ir_op(beo_base + beo_CopyKeep, "be_CopyKeep", op_pin_state_pinned, K, oparity_variable, 0, sizeof(be_node_attr_t), &be_node_op_ops); - op_be_Call = new_ir_op(beo_base + beo_Call, "be_Call", op_pin_state_pinned, N, oparity_variable, 0, sizeof(be_call_attr_t), &be_node_op_ops); - op_be_Return = new_ir_op(beo_base + beo_Return, "be_Return", op_pin_state_pinned, X, oparity_variable, 0, sizeof(be_node_attr_t), &be_node_op_ops); - op_be_AddSP = new_ir_op(beo_base + beo_AddSP, "be_AddSP", op_pin_state_pinned, N, oparity_unary, 0, sizeof(be_node_attr_t), &be_node_op_ops); - op_be_SetSP = new_ir_op(beo_base + beo_SetSP, "be_SetSP", op_pin_state_pinned, N, oparity_binary, 0, sizeof(be_stack_attr_t), &be_node_op_ops); - op_be_IncSP = new_ir_op(beo_base + beo_IncSP, "be_IncSP", op_pin_state_pinned, N, oparity_binary, 0, sizeof(be_stack_attr_t), &be_node_op_ops); - op_be_RegParams = new_ir_op(beo_base + beo_RegParams, "be_RegParams", op_pin_state_pinned, N, oparity_zero, 0, sizeof(be_node_attr_t), &be_node_op_ops); - op_be_StackParam = new_ir_op(beo_base + beo_StackParam, "be_StackParam", op_pin_state_pinned, N, oparity_unary, 0, sizeof(be_frame_attr_t), &be_node_op_ops); - op_be_FrameAddr = new_ir_op(beo_base + beo_FrameAddr, "be_FrameAddr", op_pin_state_pinned, N, oparity_binary, 0, sizeof(be_frame_attr_t), &be_node_op_ops); - op_be_FrameLoad = new_ir_op(beo_base + beo_FrameLoad, "be_FrameLoad", op_pin_state_pinned, N, oparity_any, 0, sizeof(be_frame_attr_t), &be_node_op_ops); - op_be_FrameStore = new_ir_op(beo_base + beo_FrameStore, "be_FrameStore", op_pin_state_pinned, N, oparity_any, 0, sizeof(be_frame_attr_t), &be_node_op_ops); - op_be_Barrier = new_ir_op(beo_base + beo_Barrier, "be_Barrier", op_pin_state_pinned, N, oparity_any, 0, sizeof(be_node_attr_t), &be_node_op_ops); + op_be_Spill = new_ir_op(beo_base + beo_Spill, "be_Spill", op_pin_state_mem_pinned, N, oparity_unary, 0, sizeof(be_frame_attr_t), &be_node_op_ops); + op_be_Reload = new_ir_op(beo_base + beo_Reload, "be_Reload", op_pin_state_mem_pinned, N, oparity_zero, 0, sizeof(be_frame_attr_t), &be_node_op_ops); + op_be_Perm = new_ir_op(beo_base + beo_Perm, "be_Perm", op_pin_state_pinned, N, oparity_variable, 0, sizeof(be_node_attr_t), &be_node_op_ops); + op_be_MemPerm = new_ir_op(beo_base + beo_MemPerm, "be_MemPerm", op_pin_state_mem_pinned, N, oparity_variable, 0, sizeof(be_memperm_attr_t), &be_node_op_ops); + op_be_Copy = new_ir_op(beo_base + beo_Copy, "be_Copy", op_pin_state_floats, N, oparity_unary, 0, sizeof(be_node_attr_t), &be_node_op_ops); + op_be_Keep = new_ir_op(beo_base + beo_Keep, "be_Keep", op_pin_state_pinned, K, oparity_variable, 0, sizeof(be_node_attr_t), &be_node_op_ops); + op_be_CopyKeep = new_ir_op(beo_base + beo_CopyKeep, "be_CopyKeep", op_pin_state_pinned, K, oparity_variable, 0, sizeof(be_node_attr_t), &be_node_op_ops); + op_be_Call = new_ir_op(beo_base + beo_Call, "be_Call", op_pin_state_pinned, N, oparity_variable, 0, sizeof(be_call_attr_t), &be_node_op_ops); + op_be_Return = new_ir_op(beo_base + beo_Return, "be_Return", op_pin_state_pinned, X, oparity_variable, 0, sizeof(be_return_attr_t), &be_node_op_ops); + op_be_AddSP = new_ir_op(beo_base + beo_AddSP, "be_AddSP", op_pin_state_pinned, N, oparity_unary, 0, sizeof(be_node_attr_t), &be_node_op_ops); + op_be_SetSP = new_ir_op(beo_base + beo_SetSP, "be_SetSP", op_pin_state_pinned, N, oparity_binary, 0, sizeof(be_stack_attr_t), &be_node_op_ops); + op_be_IncSP = new_ir_op(beo_base + beo_IncSP, "be_IncSP", op_pin_state_pinned, N, oparity_binary, 0, sizeof(be_stack_attr_t), &be_node_op_ops); + op_be_RegParams = new_ir_op(beo_base + beo_RegParams, "be_RegParams", op_pin_state_pinned, N, oparity_zero, 0, sizeof(be_node_attr_t), &be_node_op_ops); + op_be_StackParam = new_ir_op(beo_base + beo_StackParam, "be_StackParam", op_pin_state_pinned, N, oparity_unary, 0, sizeof(be_frame_attr_t), &be_node_op_ops); + op_be_FrameAddr = new_ir_op(beo_base + beo_FrameAddr, "be_FrameAddr", op_pin_state_pinned, N, oparity_unary, 0, sizeof(be_frame_attr_t), &be_node_op_ops); + op_be_FrameLoad = new_ir_op(beo_base + beo_FrameLoad, "be_FrameLoad", op_pin_state_pinned, N, oparity_any, 0, sizeof(be_frame_attr_t), &be_node_op_ops); + op_be_FrameStore = new_ir_op(beo_base + beo_FrameStore, "be_FrameStore", op_pin_state_pinned, N, oparity_any, 0, sizeof(be_frame_attr_t), &be_node_op_ops); + op_be_Barrier = new_ir_op(beo_base + beo_Barrier, "be_Barrier", op_pin_state_pinned, N, oparity_any, 0, sizeof(be_node_attr_t), &be_node_op_ops); set_op_tag(op_be_Spill, &be_node_tag); set_op_tag(op_be_Reload, &be_node_tag); set_op_tag(op_be_Perm, &be_node_tag); + set_op_tag(op_be_MemPerm, &be_node_tag); set_op_tag(op_be_Copy, &be_node_tag); set_op_tag(op_be_Keep, &be_node_tag); set_op_tag(op_be_CopyKeep, &be_node_tag); @@ -198,6 +230,8 @@ void be_node_init(void) { set_op_tag(op_be_FrameStore, &be_node_tag); set_op_tag(op_be_FrameAddr, &be_node_tag); set_op_tag(op_be_Barrier, &be_node_tag); + + op_be_FrameAddr->ops.node_cmp_attr = FrameAddr_cmp_attr; } /** @@ -226,7 +260,7 @@ static void *init_node_attr(ir_node* irn, int max_reg_data) return a; } -static INLINE int is_be_node(const ir_node *irn) +int is_be_node(const ir_node *irn) { return get_op_tag(get_irn_op(irn)) == &be_node_tag; } @@ -243,7 +277,6 @@ static int redir_proj(const ir_node **node, int pos) if(is_Proj(n)) { ir_node *irn; - assert(pos == -1 && "Illegal pos for a Proj"); *node = irn = get_Proj_pred(n); if(is_Proj(irn)) { assert(get_irn_mode(irn) == mode_T); @@ -255,24 +288,56 @@ static int redir_proj(const ir_node **node, int pos) return 0; } +static be_node_attr_t *retrieve_irn_attr(const ir_node *irn, int *the_pos) +{ + int dummy; + be_node_attr_t *res = NULL; + int *pos = the_pos ? the_pos : &dummy; + + *pos = -1; + if(is_Proj(irn)) { + ir_node *pred = get_Proj_pred(irn); + int p = get_Proj_proj(irn); + + if(is_be_node(pred)) { + assert(get_irn_mode(pred) == mode_T); + *pos = p; + res = get_irn_attr(pred); + assert(p >= 0 && p < res->max_reg_data && "illegal proj number"); + } + } + + else if(is_be_node(irn) && get_irn_mode(irn) != mode_T) { + be_node_attr_t *a = get_irn_attr(irn); + if(a->max_reg_data > 0) { + res = a; + *pos = 0; + } + } + + return res; +} + +static be_reg_data_t *retrieve_reg_data(const ir_node *irn) +{ + int pos; + be_node_attr_t *a = retrieve_irn_attr(irn, &pos); + return a ? &a->reg_data[pos] : NULL; +} + static void be_node_set_irn_reg(const void *_self, ir_node *irn, const arch_register_t *reg) { - int out_pos; - be_node_attr_t *a; - - out_pos = redir_proj((const ir_node **) &irn, -1); - a = get_irn_attr(irn); + be_reg_data_t *r = retrieve_reg_data(irn); - assert(is_be_node(irn)); - assert(out_pos < a->max_reg_data && "position too high"); - a->reg_data[out_pos].reg = reg; + if(r) + r->reg = reg; } -ir_node *be_new_Spill(const arch_register_class_t *cls, const arch_register_class_t *cls_frame, ir_graph *irg, ir_node *bl, ir_node *frame, ir_node *to_spill, ir_node *ctx) +ir_node *be_new_Spill(const arch_register_class_t *cls, const arch_register_class_t *cls_frame, ir_graph *irg, ir_node *bl, ir_node *frame, ir_node *to_spill) { - be_spill_attr_t *a; + be_frame_attr_t *a; ir_node *in[2]; ir_node *res; @@ -280,9 +345,8 @@ ir_node *be_new_Spill(const arch_register_class_t *cls, const arch_register_clas in[1] = to_spill; res = new_ir_node(NULL, irg, bl, op_be_Spill, mode_M, 2, in); a = init_node_attr(res, 2); - a->frame_attr.ent = NULL; - a->frame_attr.offset = 0; - a->spill_ctx = ctx; + a->ent = NULL; + a->offset = 0; be_node_set_reg_class(res, 0, cls_frame); be_node_set_reg_class(res, 1, cls); @@ -303,12 +367,30 @@ ir_node *be_new_Reload(const arch_register_class_t *cls, const arch_register_cla return res; } -ir_node *(be_get_Reload_mem)(const ir_node *irn) +ir_node *be_get_Reload_mem(const ir_node *irn) { assert(be_is_Reload(irn)); return get_irn_n(irn, be_pos_Reload_mem); } +ir_node *be_get_Reload_frame(const ir_node *irn) +{ + assert(be_is_Reload(irn)); + return get_irn_n(irn, be_pos_Reload_frame); +} + +ir_node *be_get_Spill_val(const ir_node *irn) +{ + assert(be_is_Spill(irn)); + return get_irn_n(irn, be_pos_Spill_val); +} + +ir_node *be_get_Spill_frame(const ir_node *irn) +{ + assert(be_is_Spill(irn)); + return get_irn_n(irn, be_pos_Spill_frame); +} + ir_node *be_new_Perm(const arch_register_class_t *cls, ir_graph *irg, ir_node *bl, int n, ir_node *in[]) { int i; @@ -322,6 +404,40 @@ ir_node *be_new_Perm(const arch_register_class_t *cls, ir_graph *irg, ir_node *b return irn; } +ir_node *be_new_MemPerm(const arch_env_t *arch_env, ir_graph *irg, ir_node *bl, int n, ir_node *in[]) +{ + int i; + ir_node *frame = get_irg_frame(irg); + const arch_register_class_t *cls_frame = arch_get_irn_reg_class(arch_env, frame, -1); + ir_node *irn; + const arch_register_t *sp = arch_env->isa->sp; + be_memperm_attr_t *attr; + ir_node **real_in; + + real_in = alloca((n+1) * sizeof(real_in[0])); + real_in[0] = frame; + memcpy(&real_in[1], in, n * sizeof(real_in[0])); + + irn = new_ir_node(NULL, irg, bl, op_be_MemPerm, mode_T, n+1, real_in); + + init_node_attr(irn, n + 1); + be_node_set_reg_class(irn, 0, sp->reg_class); + for(i = 0; i < n; ++i) { + be_node_set_reg_class(irn, i + 1, cls_frame); + be_node_set_reg_class(irn, OUT_POS(i), cls_frame); + } + + attr = get_irn_attr(irn); + + attr->in_entities = obstack_alloc(irg->obst, n * sizeof(attr->in_entities[0])); + memset(attr->in_entities, 0, n * sizeof(attr->in_entities[0])); + attr->out_entities = obstack_alloc(irg->obst, n*sizeof(attr->out_entities[0])); + memset(attr->out_entities, 0, n*sizeof(attr->out_entities[0])); + + return irn; +} + + ir_node *be_new_Copy(const arch_register_class_t *cls, ir_graph *irg, ir_node *bl, ir_node *op) { ir_node *in[1]; @@ -339,6 +455,10 @@ ir_node *be_get_Copy_op(const ir_node *cpy) { return get_irn_n(cpy, be_pos_Copy_op); } +void be_set_Copy_op(ir_node *cpy, ir_node *op) { + set_irn_n(cpy, be_pos_Copy_op, op); +} + ir_node *be_new_Keep(const arch_register_class_t *cls, ir_graph *irg, ir_node *bl, int n, ir_node *in[]) { int i; @@ -367,7 +487,7 @@ ir_node *be_new_Call(dbg_info *dbg, ir_graph *irg, ir_node *bl, ir_node *mem, ir real_in[be_pos_Call_ptr] = ptr; memcpy(&real_in[be_pos_Call_first_arg], in, n * sizeof(in[0])); - irn = new_ir_node(NULL, irg, bl, op_be_Call, mode_T, real_n, real_in); + irn = new_ir_node(dbg, irg, bl, op_be_Call, mode_T, real_n, real_in); a = init_node_attr(irn, (n_outs > real_n ? n_outs : real_n)); a->ent = NULL; a->call_tp = call_tp; @@ -402,33 +522,42 @@ void be_Call_set_type(ir_node *call, ir_type *call_tp) { a->call_tp = call_tp; } - -ir_node *be_new_Return(dbg_info *dbg, ir_graph *irg, ir_node *bl, int n, ir_node *in[]) +/* Construct a new be_Return. */ +ir_node *be_new_Return(dbg_info *dbg, ir_graph *irg, ir_node *bl, int n_res, int n, ir_node *in[]) { + be_return_attr_t *a; ir_node *irn = new_ir_node(dbg, irg, bl, op_be_Return, mode_X, n, in); init_node_attr(irn, n); + a = get_irn_attr(irn); + a->num_ret_vals = n_res; return irn; } -ir_node *be_new_IncSP(const arch_register_t *sp, ir_graph *irg, ir_node *bl, ir_node *old_sp, ir_node *mem, unsigned offset, be_stack_dir_t dir) +/* Returns the number of real returns values */ +int be_Return_get_n_rets(ir_node *ret) +{ + be_return_attr_t *a = get_irn_attr(ret); + return a->num_ret_vals; +} + +ir_node *be_new_IncSP(const arch_register_t *sp, ir_graph *irg, ir_node *bl, ir_node *old_sp, ir_node *mem, int offset) { be_stack_attr_t *a; ir_node *irn; - ir_node *in[1]; + ir_node *in[2]; in[0] = old_sp; in[1] = mem; irn = new_ir_node(NULL, irg, bl, op_be_IncSP, sp->reg_class->mode, 2, in); a = init_node_attr(irn, 1); - a->dir = dir; a->offset = offset; - be_node_set_flags(irn, -1, arch_irn_flags_ignore); + be_node_set_flags(irn, -1, arch_irn_flags_ignore | arch_irn_flags_modify_sp); /* Set output constraint to stack register. */ be_node_set_reg_class(irn, 0, sp->reg_class); - be_set_constr_single_reg(irn, -1, sp); + be_set_constr_single_reg(irn, BE_OUT_POS(0), sp); be_node_set_irn_reg(NULL, irn, sp); return irn; @@ -443,13 +572,16 @@ ir_node *be_new_AddSP(const arch_register_t *sp, ir_graph *irg, ir_node *bl, ir_ in[be_pos_AddSP_old_sp] = old_sp; in[be_pos_AddSP_size] = sz; - irn = new_ir_node(NULL, irg, bl, op_be_AddSP, mode_T, be_pos_AddSP_last, in); - a = init_node_attr(irn, be_pos_AddSP_last); + irn = new_ir_node(NULL, irg, bl, op_be_AddSP, mode_T, be_pos_AddSP_last, in); + a = init_node_attr(irn, be_pos_AddSP_last); - be_node_set_flags(irn, OUT_POS(0), arch_irn_flags_ignore); + be_node_set_flags(irn, OUT_POS(pn_be_AddSP_res), arch_irn_flags_ignore | arch_irn_flags_modify_sp); /* Set output constraint to stack register. */ - be_set_constr_single_reg(irn, OUT_POS(0), sp); + be_set_constr_single_reg(irn, be_pos_AddSP_old_sp, sp); + be_node_set_reg_class(irn, be_pos_AddSP_size, arch_register_get_class(sp)); + be_set_constr_single_reg(irn, OUT_POS(pn_be_AddSP_res), sp); + a->reg_data[pn_be_AddSP_res].reg = sp; return irn; } @@ -466,7 +598,7 @@ ir_node *be_new_SetSP(const arch_register_t *sp, ir_graph *irg, ir_node *bl, ir_ irn = new_ir_node(NULL, irg, bl, op_be_SetSP, get_irn_mode(old_sp), 3, in); a = init_node_attr(irn, 3); - be_node_set_flags(irn, OUT_POS(0), arch_irn_flags_ignore); + be_node_set_flags(irn, OUT_POS(0), arch_irn_flags_ignore | arch_irn_flags_modify_sp); /* Set output constraint to stack register. */ be_set_constr_single_reg(irn, OUT_POS(0), sp); @@ -553,7 +685,8 @@ ir_node *be_new_FrameAddr(const arch_register_class_t *cls_frame, ir_graph *irg, a->offset = 0; be_node_set_reg_class(irn, 0, cls_frame); be_node_set_reg_class(irn, OUT_POS(0), cls_frame); - return irn; + + return optimize_node(irn); } ir_node *be_new_CopyKeep(const arch_register_class_t *cls, ir_graph *irg, ir_node *bl, ir_node *src, int n, ir_node *in_keep[], ir_mode *mode) @@ -579,6 +712,14 @@ ir_node *be_new_CopyKeep_single(const arch_register_class_t *cls, ir_graph *irg, return be_new_CopyKeep(cls, irg, bl, src, 1, in, mode); } +ir_node *be_get_CopyKeep_op(const ir_node *cpy) { + return get_irn_n(cpy, be_pos_CopyKeep_op); +} + +void be_set_CopyKeep_op(ir_node *cpy, ir_node *op) { + set_irn_n(cpy, be_pos_CopyKeep_op, op); +} + ir_node *be_new_Barrier(ir_graph *irg, ir_node *bl, int n, ir_node *in[]) { ir_node *irn; @@ -591,7 +732,9 @@ ir_node *be_new_Barrier(ir_graph *irg, ir_node *bl, int n, ir_node *in[]) int be_is_Spill (const ir_node *irn) { return be_get_irn_opcode(irn) == beo_Spill ; } int be_is_Reload (const ir_node *irn) { return be_get_irn_opcode(irn) == beo_Reload ; } int be_is_Copy (const ir_node *irn) { return be_get_irn_opcode(irn) == beo_Copy ; } +int be_is_CopyKeep (const ir_node *irn) { return be_get_irn_opcode(irn) == beo_CopyKeep ; } int be_is_Perm (const ir_node *irn) { return be_get_irn_opcode(irn) == beo_Perm ; } +int be_is_MemPerm (const ir_node *irn) { return be_get_irn_opcode(irn) == beo_MemPerm ; } int be_is_Keep (const ir_node *irn) { return be_get_irn_opcode(irn) == beo_Keep ; } int be_is_Call (const ir_node *irn) { return be_get_irn_opcode(irn) == beo_Call ; } int be_is_Return (const ir_node *irn) { return be_get_irn_opcode(irn) == beo_Return ; } @@ -620,7 +763,7 @@ int be_has_frame_entity(const ir_node *irn) } } -entity *be_get_frame_entity(const ir_node *irn) +entity* be_get_frame_entity(const ir_node *irn) { if(be_has_frame_entity(irn)) { be_frame_attr_t *a = get_irn_attr(irn); @@ -629,6 +772,51 @@ entity *be_get_frame_entity(const ir_node *irn) return NULL; } +void be_set_MemPerm_in_entity(const ir_node *irn, int n, entity *ent) +{ + be_memperm_attr_t *attr = get_irn_attr(irn); + + assert(be_is_MemPerm(irn)); + assert(n < be_get_MemPerm_entity_arity(irn)); + + attr->in_entities[n] = ent; +} + +entity* be_get_MemPerm_in_entity(const ir_node* irn, int n) +{ + be_memperm_attr_t *attr = get_irn_attr(irn); + + assert(be_is_MemPerm(irn)); + assert(n < be_get_MemPerm_entity_arity(irn)); + + return attr->in_entities[n]; +} + +void be_set_MemPerm_out_entity(const ir_node *irn, int n, entity *ent) +{ + be_memperm_attr_t *attr = get_irn_attr(irn); + + assert(be_is_MemPerm(irn)); + assert(n < be_get_MemPerm_entity_arity(irn)); + + attr->out_entities[n] = ent; +} + +entity* be_get_MemPerm_out_entity(const ir_node* irn, int n) +{ + be_memperm_attr_t *attr = get_irn_attr(irn); + + assert(be_is_MemPerm(irn)); + assert(n < be_get_MemPerm_entity_arity(irn)); + + return attr->out_entities[n]; +} + +int be_get_MemPerm_entity_arity(const ir_node *irn) +{ + return get_irn_arity(irn) - 1; +} + static void be_limited(void *data, bitset_t *bs) { be_req_t *req = data; @@ -714,7 +902,7 @@ ir_node *be_get_IncSP_pred(ir_node *irn) { return get_irn_n(irn, 0); } -void *be_set_IncSP_pred(ir_node *incsp, ir_node *pred) { +void be_set_IncSP_pred(ir_node *incsp, ir_node *pred) { assert(be_is_IncSP(incsp)); set_irn_n(incsp, 0, pred); } @@ -724,162 +912,31 @@ ir_node *be_get_IncSP_mem(ir_node *irn) { return get_irn_n(irn, 1); } -void be_set_IncSP_offset(ir_node *irn, unsigned offset) +void be_set_IncSP_offset(ir_node *irn, int offset) { be_stack_attr_t *a = get_irn_attr(irn); assert(be_is_IncSP(irn)); a->offset = offset; } -unsigned be_get_IncSP_offset(const ir_node *irn) +int be_get_IncSP_offset(const ir_node *irn) { be_stack_attr_t *a = get_irn_attr(irn); assert(be_is_IncSP(irn)); return a->offset; } -void be_set_IncSP_direction(ir_node *irn, be_stack_dir_t dir) -{ - be_stack_attr_t *a = get_irn_attr(irn); - assert(be_is_IncSP(irn)); - a->dir = dir; -} - -be_stack_dir_t be_get_IncSP_direction(const ir_node *irn) -{ - be_stack_attr_t *a = get_irn_attr(irn); - assert(be_is_IncSP(irn)); - return a->dir; -} - -void be_set_Spill_entity(ir_node *irn, entity *ent) -{ - be_spill_attr_t *a = get_irn_attr(irn); - assert(be_is_Spill(irn)); - a->frame_attr.ent = ent; -} - -static ir_node *find_a_spill_walker(ir_node *irn, unsigned visited_nr) -{ - unsigned nr = get_irn_visited(irn); - - set_irn_visited(irn, visited_nr); - - if(is_Phi(irn)) { - int i, n; - if(nr < visited_nr) { - for(i = 0, n = get_irn_arity(irn); i < n; ++i) { - ir_node *n = find_a_spill_walker(get_irn_n(irn, i), visited_nr); - if(n != NULL) - return n; - } - } - } - - else if(be_get_irn_opcode(irn) == beo_Spill) - return irn; - - return NULL; -} - -ir_node *be_get_Spill_context(const ir_node *irn) { - const be_spill_attr_t *a = get_irn_attr(irn); - assert(be_is_Spill(irn)); - return a->spill_ctx; -} - -/** - * Finds a spill for a reload. - * If the reload is directly using the spill, this is simple, - * else we perform DFS from the reload (over all PhiMs) and return - * the first spill node we find. - */ -static INLINE ir_node *find_a_spill(const ir_node *irn) -{ - ir_graph *irg = get_irn_irg(irn); - unsigned visited_nr = get_irg_visited(irg) + 1; - - assert(be_is_Reload(irn)); - set_irg_visited(irg, visited_nr); - return find_a_spill_walker(be_get_Reload_mem(irn), visited_nr); -} - -entity *be_get_spill_entity(const ir_node *irn) -{ - switch(be_get_irn_opcode(irn)) { - case beo_Reload: - { - ir_node *spill = find_a_spill(irn); - return be_get_spill_entity(spill); - } - case beo_Spill: - { - be_spill_attr_t *a = get_irn_attr(irn); - return a->frame_attr.ent; - } - default: - assert(0 && "Must give spill/reload node"); - break; - } - - return NULL; -} - -static void link_reload_walker(ir_node *irn, void *data) -{ - ir_node **root = (ir_node **) data; - if(be_is_Reload(irn)) { - set_irn_link(irn, *root); - *root = irn; - } -} - -void be_copy_entities_to_reloads(ir_graph *irg) -{ - ir_node *irn = NULL; - irg_walk_graph(irg, link_reload_walker, NULL, (void *) &irn); - - while(irn) { - be_frame_attr_t *a = get_irn_attr(irn); - entity *ent = be_get_spill_entity(irn); - a->ent = ent; - irn = get_irn_link(irn); - } -} - -ir_node *be_spill(const arch_env_t *arch_env, ir_node *irn, ir_node *ctx) +ir_node *be_spill(const arch_env_t *arch_env, ir_node *irn) { ir_node *bl = get_nodes_block(irn); ir_graph *irg = get_irn_irg(bl); ir_node *frame = get_irg_frame(irg); - ir_node *insert = bl; ir_node *spill; const arch_register_class_t *cls = arch_get_irn_reg_class(arch_env, irn, -1); const arch_register_class_t *cls_frame = arch_get_irn_reg_class(arch_env, frame, -1); - spill = be_new_Spill(cls, cls_frame, irg, bl, frame, irn, ctx); - - /* - * search the right insertion point. a spill of a phi cannot be put - * directly after the phi, if there are some phis behind the one which - * is spilled. Also, a spill of a Proj must be after all Projs of the - * same tuple node. - * - * Here's one special case: - * If the spill is in the start block, the spill must be after the frame - * pointer is set up. This is done by setting insert to the end of the block - * which is its default initialization (see above). - */ - - insert = sched_next(irn); - if(insert != bl && bl == get_irg_start_block(irg) && sched_get_time_step(frame) >= sched_get_time_step(insert)) - insert = sched_next(frame); - - while((is_Phi(insert) || is_Proj(insert)) && !sched_is_end(insert)) - insert = sched_next(insert); - - sched_add_before(insert, spill); + spill = be_new_Spill(cls, cls_frame, irg, bl, frame, irn); return spill; } @@ -976,16 +1033,8 @@ be_node_get_irn_reg_req(const void *self, arch_register_req_t *req, const ir_nod const arch_register_t * be_node_get_irn_reg(const void *_self, const ir_node *irn) { - int out_pos; - be_node_attr_t *a; - - out_pos = redir_proj((const ir_node **) &irn, -1); - a = get_irn_attr(irn); - - assert(is_be_node(irn)); - assert(out_pos < a->max_reg_data && "position too high"); - - return a->reg_data[out_pos].reg; + be_reg_data_t *r = retrieve_reg_data(irn); + return r ? r->reg : NULL; } static arch_irn_class_t be_node_classify(const void *_self, const ir_node *irn) @@ -999,9 +1048,10 @@ static arch_irn_class_t be_node_classify(const void *_self, const ir_node *irn) XXX(Perm, perm); XXX(Copy, copy); XXX(Return, branch); + XXX(StackParam, stackparam); #undef XXX default: - return 0; + return arch_irn_class_normal; } return 0; @@ -1009,16 +1059,8 @@ static arch_irn_class_t be_node_classify(const void *_self, const ir_node *irn) static arch_irn_flags_t be_node_get_flags(const void *_self, const ir_node *irn) { - int out_pos; - be_node_attr_t *a; - - out_pos = redir_proj((const ir_node **) &irn, -1); - a = get_irn_attr(irn); - - assert(is_be_node(irn)); - assert(out_pos < a->max_reg_data && "position too high"); - - return a->reg_data[out_pos].req.flags; + be_reg_data_t *r = retrieve_reg_data(irn); + return r ? r->req.flags : 0; } static entity *be_node_get_frame_entity(const void *self, const ir_node *irn) @@ -1026,6 +1068,16 @@ static entity *be_node_get_frame_entity(const void *self, const ir_node *irn) return be_get_frame_entity(irn); } +static void be_node_set_frame_entity(const void *self, ir_node *irn, entity *ent) +{ + be_frame_attr_t *a; + + assert(be_has_frame_entity(irn)); + + a = get_irn_attr(irn); + a->ent = ent; +} + static void be_node_set_frame_offset(const void *self, ir_node *irn, int offset) { if(be_has_frame_entity(irn)) { @@ -1034,6 +1086,17 @@ static void be_node_set_frame_offset(const void *self, ir_node *irn, int offset) } } +static int be_node_get_sp_bias(const void *self, const ir_node *irn) +{ + int result = 0; + + if(be_is_IncSP(irn)) { + result = be_get_IncSP_offset(irn); + } + + return result; +} + /* ___ ____ _ _ _ _ _ _ |_ _| _ \| \ | | | | | | __ _ _ __ __| | | ___ _ __ @@ -1050,7 +1113,13 @@ static const arch_irn_ops_if_t be_node_irn_ops_if = { be_node_classify, be_node_get_flags, be_node_get_frame_entity, - be_node_set_frame_offset + be_node_set_frame_entity, + be_node_set_frame_offset, + be_node_get_sp_bias, + NULL, /* get_inverse */ + NULL, /* get_op_estimated_cost */ + NULL, /* possible_memory_operand */ + NULL, /* perform_memory_operand */ }; static const arch_irn_ops_t be_node_irn_ops = { @@ -1170,10 +1239,19 @@ static entity *phi_get_frame_entity(const void *_self, const ir_node *irn) return NULL; } +static void phi_set_frame_entity(const void *_self, ir_node *irn, entity *ent) +{ +} + static void phi_set_frame_offset(const void *_self, ir_node *irn, int bias) { } +static int phi_get_sp_bias(const void* self, const ir_node *irn) +{ + return 0; +} + static const arch_irn_ops_if_t phi_irn_ops = { phi_get_irn_reg_req, phi_set_irn_reg, @@ -1181,7 +1259,13 @@ static const arch_irn_ops_if_t phi_irn_ops = { phi_classify, phi_get_flags, phi_get_frame_entity, - phi_set_frame_offset + phi_set_frame_entity, + phi_set_frame_offset, + phi_get_sp_bias, + NULL, /* get_inverse */ + NULL, /* get_op_estimated_cost */ + NULL, /* possible_memory_operand */ + NULL, /* perform_memory_operand */ }; static const arch_irn_handler_t phi_irn_handler = { @@ -1219,7 +1303,6 @@ void be_phi_handler_reset(arch_irn_handler_t *handler) h->regs = pmap_create(); } - /* _ _ _ ____ _ | \ | | ___ __| | ___ | _ \ _ _ _ __ ___ _ __ (_)_ __ __ _ @@ -1240,6 +1323,7 @@ static void dump_node_req(FILE *f, int idx, be_req_t *req) const char *prefix = buf; snprintf(buf, sizeof(buf), "#%d ", idx); + buf[sizeof(buf) - 1] = '\0'; if(req->flags != arch_irn_flags_none) { fprintf(f, "%sflags: ", prefix); @@ -1317,28 +1401,22 @@ static int dump_node(ir_node *irn, FILE *f, dump_reason_t reason) be_frame_attr_t *a = (be_frame_attr_t *) at; if (a->ent) { int bits = get_type_size_bits(get_entity_type(a->ent)); - ir_fprintf(f, "frame entity: %+F offset 0x%x (%d) size 0x%x %d\n", + ir_fprintf(f, "frame entity: %+F, offset 0x%x (%d), size 0x%x (%d) bits\n", a->ent, a->offset, a->offset, bits, bits); } } switch(be_get_irn_opcode(irn)) { - case beo_Spill: - { - be_spill_attr_t *a = (be_spill_attr_t *) at; - ir_fprintf(f, "spill context: %+F\n", a->spill_ctx); - } - break; - case beo_IncSP: { be_stack_attr_t *a = (be_stack_attr_t *) at; - if (a->offset == BE_STACK_FRAME_SIZE) + if (a->offset == BE_STACK_FRAME_SIZE_EXPAND) fprintf(f, "offset: FRAME_SIZE\n"); + else if(a->offset == BE_STACK_FRAME_SIZE_SHRINK) + fprintf(f, "offset: -FRAME SIZE\n"); else fprintf(f, "offset: %u\n", a->offset); - fprintf(f, "direction: %s\n", a->dir == be_stack_dir_expand ? "expand" : "shrink"); } break; case beo_Call: @@ -1348,6 +1426,24 @@ static int dump_node(ir_node *irn, FILE *f, dump_reason_t reason) if (a->ent) fprintf(f, "\ncalling: %s\n", get_entity_name(a->ent)); } + break; + case beo_MemPerm: + { + int i; + for(i = 0; i < be_get_MemPerm_entity_arity(irn); ++i) { + entity *in, *out; + in = be_get_MemPerm_in_entity(irn, i); + out = be_get_MemPerm_out_entity(irn, i); + if(in) { + fprintf(f, "\nin[%d]: %s\n", i, get_entity_name(in)); + } + if(out) { + fprintf(f, "\nout[%d]: %s\n", i, get_entity_name(out)); + } + } + } + break; + default: break; }