X-Git-Url: http://nsz.repo.hu/git/?a=blobdiff_plain;f=ir%2Fir%2Firnode.c;h=2c3a5a6f0bd9a9d2027791d7c36711b8674e11b4;hb=2080968df998886ef494ff80bac44cac5773eb14;hp=481cab2752dafcb9409d99aef042aa6a0c855785;hpb=1529b8f31d7ad60e576db43318e1c917ee0a95ac;p=libfirm diff --git a/ir/ir/irnode.c b/ir/ir/irnode.c index 481cab275..2c3a5a6f0 100644 --- a/ir/ir/irnode.c +++ b/ir/ir/irnode.c @@ -27,8 +27,10 @@ #include "irdump.h" #include "irop_t.h" #include "irprog_t.h" +#include "iredges_t.h" #include "irhooks.h" +#include "irtools.h" /* some constants fixing the positions of nodes predecessors in the in array */ @@ -39,10 +41,10 @@ #define END_KEEPALIVE_OFFSET 0 static const char *pnc_name_arr [] = { - "False", "Eq", "Lt", "Le", - "Gt", "Ge", "Lg", "Leg", "Uo", - "Ue", "Ul", "Ule", "Ug", "Uge", - "Ne", "True" + "pn_Cmp_False", "pn_Cmp_Eq", "pn_Cmp_Lt", "pn_Cmp_Le", + "pn_Cmp_Gt", "pn_Cmp_Ge", "pn_Cmp_Lg", "pn_Cmp_Leg", + "pn_Cmp_Uo", "pn_Cmp_Ue", "pn_Cmp_Ul", "pn_Cmp_Ule", + "pn_Cmp_Ug", "pn_Cmp_Uge", "pn_Cmp_Ne", "pn_Cmp_True" }; /** @@ -52,30 +54,29 @@ const char *get_pnc_string(int pnc) { return pnc_name_arr[pnc]; } -/** - * Calculates the negated pnc condition. +/* + * Calculates the negated (Complement(R)) pnc condition. */ +int get_negated_pnc(int pnc, ir_mode *mode) { + pnc ^= pn_Cmp_True; + + /* do NOT add the Uo bit for non-floating point values */ + if (! mode_is_float(mode)) + pnc &= ~pn_Cmp_Uo; + + return pnc; +} + +/* Calculates the inversed (R^-1) pnc condition, i.e., "<" --> ">" */ int -get_negated_pnc(int pnc) { - switch (pnc) { - case False: return True; break; - case Eq: return Ne; break; - case Lt: return Uge; break; - case Le: return Ug; break; - case Gt: return Ule; break; - case Ge: return Ul; break; - case Lg: return Ue; break; - case Leg: return Uo; break; - case Uo: return Leg; break; - case Ue: return Lg; break; - case Ul: return Ge; break; - case Ule: return Gt; break; - case Ug: return Le; break; - case Uge: return Lt; break; - case Ne: return Eq; break; - case True: return False; break; - } - return 99; /* to shut up gcc */ +get_inversed_pnc(int pnc) { + int code = pnc & ~(pn_Cmp_Lt|pn_Cmp_Gt); + int lesser = pnc & pn_Cmp_Lt; + int greater = pnc & pn_Cmp_Gt; + + code |= (lesser ? pn_Cmp_Gt : 0) | (greater ? pn_Cmp_Lt : 0); + + return code; } const char *pns_name_arr [] = { @@ -136,17 +137,19 @@ new_ir_node (dbg_info *db, ir_graph *irg, ir_node *block, ir_op *op, ir_mode *mo memset(p, 0, node_size); res = (ir_node *) (p + firm_add_node_size); - res->kind = k_ir_node; - res->op = op; - res->mode = mode; - res->visited = 0; - res->link = NULL; + res->kind = k_ir_node; + res->op = op; + res->mode = mode; + res->visited = 0; + res->node_idx = get_irg_next_node_idx(irg); + res->link = NULL; if (arity < 0) { res->in = NEW_ARR_F (ir_node *, 1); /* 1: space for block */ } else { res->in = NEW_ARR_D (ir_node *, irg->obst, (arity+1)); memcpy (&res->in[1], in, sizeof (ir_node *) * arity); } + res->in[0] = block; set_irn_dbg_info(res, db); res->out = NULL; @@ -155,7 +158,22 @@ new_ir_node (dbg_info *db, ir_graph *irg, ir_node *block, ir_op *op, ir_mode *mo res->node_nr = get_irp_new_node_nr(); #endif - hook_new_node(res); +#if FIRM_EDGES_INPLACE + { + int i; + int is_bl = is_Block(res); + + INIT_LIST_HEAD(&res->edge_info.outs_head); + if(is_bl) + INIT_LIST_HEAD(&res->attr.block.succ_head); + + + for (i = is_bl; i <= arity; ++i) + edges_notify_edge(res, i - 1, res->in[i], NULL, irg); + } +#endif + + hook_new_node(irg, res); return res; } @@ -207,7 +225,9 @@ get_irn_in (const ir_node *node) { void set_irn_in (ir_node *node, int arity, ir_node **in) { + int i; ir_node *** arr; + ir_graph *irg = current_ir_graph; assert(node); if (get_interprocedural_view()) { /* handle Filter and Block specially */ if (get_irn_opcode(node) == iro_Filter) { @@ -221,12 +241,24 @@ set_irn_in (ir_node *node, int arity, ir_node **in) { } else { arr = &node->in; } - if (arity != ARR_LEN(*arr) - 1) { + + for (i = 0; i < arity; i++) { + if (i < ARR_LEN(*arr)-1) + edges_notify_edge(node, i, in[i], (*arr)[i+1], irg); + else + edges_notify_edge(node, i, in[i], NULL, irg); + } + for(;i < ARR_LEN(*arr)-1; i++) { + edges_notify_edge(node, i, NULL, (*arr)[i+1], irg); + } + + if (arity != ARR_LEN(*arr) - 1) { ir_node * block = (*arr)[0]; - *arr = NEW_ARR_D(ir_node *, current_ir_graph->obst, arity + 1); + *arr = NEW_ARR_D(ir_node *, irg->obst, arity + 1); (*arr)[0] = block; } - fix_backedges(current_ir_graph->obst, node); + fix_backedges(irg->obst, node); + memcpy((*arr) + 1, in, sizeof(ir_node *) * arity); } @@ -249,7 +281,9 @@ ir_node * void set_irn_n (ir_node *node, int n, ir_node *in) { - assert(node && node->kind == k_ir_node && -1 <= n && n < get_irn_arity(node)); + assert(node && node->kind == k_ir_node); + assert(-1 <= n); + assert(n < get_irn_arity(node)); assert(in && in->kind == k_ir_node); if ((n == -1) && (get_irn_opcode(node) == iro_Filter)) { @@ -270,6 +304,13 @@ set_irn_n (ir_node *node, int n, ir_node *in) { } /* else fall through */ } + + /* Call the hook */ + hook_set_irn_n(node, n, in, node->in[n + 1]); + + /* Here, we rely on src and tgt being in the current ir graph */ + edges_notify_edge(node, n, in, node->in[n + 1], current_ir_graph); + node->in[n + 1] = in; } @@ -307,17 +348,14 @@ get_irn_modeident (const ir_node *node) } ir_op * -(get_irn_op)(const ir_node *node) -{ +(get_irn_op)(const ir_node *node) { return _get_irn_op(node); } /* should be private to the library: */ void -set_irn_op (ir_node *node, ir_op *op) -{ - assert (node); - node->op = op; +(set_irn_op)(ir_node *node, ir_op *op) { + _set_irn_op(node, op); } opcode @@ -385,6 +423,11 @@ op_pin_state return _get_irn_pinned(node); } +op_pin_state +(is_irn_pinned_in_irg) (const ir_node *node) { + return _is_irn_pinned_in_irg(node); +} + void set_irn_pinned(ir_node *node, op_pin_state state) { /* due to optimization an opt may be turned into a Tuple */ if (get_irn_op(node) == op_Tuple) @@ -422,13 +465,12 @@ void firm_set_irn_section(ir_node *n, struct section *s) {} /* Outputs a unique number for this node */ -long -get_irn_node_nr(const ir_node *node) { +long get_irn_node_nr(const ir_node *node) { assert(node); #ifdef DEBUG_libfirm return node->node_nr; #else - return (long)node; + return (long)PTR_TO_INT(node); #endif } @@ -453,11 +495,11 @@ get_irn_alloc_attr (ir_node *node) return node->attr.a; } -type * +free_attr get_irn_free_attr (ir_node *node) { assert (node->op == op_Free); - return node->attr.f = skip_tid(node->attr.f); + return node->attr.f; } symconst_attr @@ -467,7 +509,7 @@ get_irn_symconst_attr (ir_node *node) return node->attr.i; } -type * +ir_type * get_irn_call_attr (ir_node *node) { assert (node->op == op_Call); @@ -517,12 +559,23 @@ get_irn_except_attr (ir_node *node) return node->attr.except; } +void * +get_irn_generic_attr (ir_node *node) { + return &node->attr; +} + +unsigned (get_irn_idx)(const ir_node *node) { + assert(is_ir_node(node)); + return _get_irn_idx(node); +} + /** manipulate fields of individual nodes **/ /* this works for all except Block */ ir_node * get_nodes_block (const ir_node *node) { assert (!(node->op == op_Block)); + assert (is_irn_pinned_in_irg(node) && "block info may be incorrect"); return get_irn_n(node, -1); } @@ -534,7 +587,7 @@ set_nodes_block (ir_node *node, ir_node *block) { /* Test whether arbitrary node is frame pointer, i.e. Proj(pn_Start_P_frame_base) * from Start. If so returns frame type, else Null. */ -type *is_frame_pointer(ir_node *n) { +ir_type *is_frame_pointer(ir_node *n) { if ((get_irn_op(n) == op_Proj) && (get_Proj_proj(n) == pn_Start_P_frame_base)) { ir_node *start = get_Proj_pred(n); @@ -547,7 +600,7 @@ type *is_frame_pointer(ir_node *n) { /* Test whether arbitrary node is globals pointer, i.e. Proj(pn_Start_P_globals) * from Start. If so returns global type, else Null. */ -type *is_globals_pointer(ir_node *n) { +ir_type *is_globals_pointer(ir_node *n) { if ((get_irn_op(n) == op_Proj) && (get_Proj_proj(n) == pn_Start_P_globals)) { ir_node *start = get_Proj_pred(n); @@ -579,18 +632,14 @@ get_Block_cfgpred_arr (ir_node *node) return (ir_node **)&(get_irn_in(node)[1]); } - int -get_Block_n_cfgpreds (ir_node *node) { - assert ((node->op == op_Block)); - return get_irn_arity(node); +(get_Block_n_cfgpreds)(ir_node *node) { + return _get_Block_n_cfgpreds(node); } ir_node * -get_Block_cfgpred (ir_node *node, int pos) { - assert(-1 <= pos && pos < get_irn_arity(node)); - assert(node->op == op_Block); - return get_irn_n(node, pos); +(get_Block_cfgpred)(ir_node *node, int pos) { + return _get_Block_cfgpred(node, pos); } void @@ -599,40 +648,42 @@ set_Block_cfgpred (ir_node *node, int pos, ir_node *pred) { set_irn_n(node, pos, pred); } -bool +ir_node * +(get_Block_cfgpred_block)(ir_node *node, int pos) { + return _get_Block_cfgpred_block(node, pos); +} + +int get_Block_matured (ir_node *node) { assert (node->op == op_Block); - return node->attr.block.matured; + return (int)node->attr.block.matured; } void -set_Block_matured (ir_node *node, bool matured) { +set_Block_matured (ir_node *node, int matured) { assert (node->op == op_Block); node->attr.block.matured = matured; } + unsigned long -get_Block_block_visited (ir_node *node) { - assert (node->op == op_Block); - return node->attr.block.block_visited; +(get_Block_block_visited)(ir_node *node) { + return _get_Block_block_visited(node); } void -set_Block_block_visited (ir_node *node, unsigned long visit) { - assert (node->op == op_Block); - node->attr.block.block_visited = visit; +(set_Block_block_visited)(ir_node *node, unsigned long visit) { + _set_Block_block_visited(node, visit); } /* For this current_ir_graph must be set. */ void -mark_Block_block_visited (ir_node *node) { - assert (node->op == op_Block); - node->attr.block.block_visited = get_irg_block_visited(current_ir_graph); +(mark_Block_block_visited)(ir_node *node) { + _mark_Block_block_visited(node); } int -Block_not_block_visited(ir_node *node) { - assert (node->op == op_Block); - return (node->attr.block.block_visited < get_irg_block_visited(current_ir_graph)); +(Block_not_block_visited)(ir_node *node) { + return _Block_not_block_visited(node); } ir_node * @@ -654,10 +705,10 @@ void set_Block_cg_cfgpred_arr(ir_node * node, int arity, ir_node ** in) { node->attr.block.in_cg[0] = NULL; node->attr.block.cg_backedge = new_backedge_arr(current_ir_graph->obst, arity); { - /* Fix backedge array. fix_backedges operates depending on + /* Fix backedge array. fix_backedges() operates depending on interprocedural_view. */ int ipv = get_interprocedural_view(); - set_interprocedural_view(true); + set_interprocedural_view(1); fix_backedges(current_ir_graph->obst, node); set_interprocedural_view(ipv); } @@ -700,11 +751,18 @@ int (is_Block_dead)(const ir_node *block) { return _is_Block_dead(block); } -void -set_Start_irg(ir_node *node, ir_graph *irg) { - assert(node->op == op_Start); - assert(is_ir_graph(irg)); - assert(0 && " Why set irg? -- use set_irn_irg"); +ir_extblk *get_Block_extbb(const ir_node *block) { + ir_extblk *res; + assert(is_Block(block)); + res = block->attr.block.extblk; + assert(res == NULL || is_ir_extbb(res)); + return res; +} + +void set_Block_extbb(ir_node *block, ir_extblk *extblk) { + assert(is_Block(block)); + assert(extblk == NULL || is_ir_extbb(extblk)); + block->attr.block.extblk = extblk; } int @@ -740,6 +798,17 @@ free_End (ir_node *end) { in array afterwards ... */ } +/* Return the target address of an IJmp */ +ir_node *get_IJmp_target(ir_node *ijmp) { + assert(ijmp->op == op_IJmp); + return get_irn_n(ijmp, 0); +} + +/** Sets the target address of an IJmp */ +void set_IJmp_target(ir_node *ijmp, ir_node *tgt) { + assert(ijmp->op == op_IJmp); + set_irn_n(ijmp, 0, tgt); +} /* > Implementing the case construct (which is where the constant Proj node is @@ -850,33 +919,8 @@ set_Return_res (ir_node *node, int pos, ir_node *res){ set_irn_n(node, pos + RETURN_RESULT_OFFSET, res); } -ir_node * -get_Raise_mem (ir_node *node) { - assert (node->op == op_Raise); - return get_irn_n(node, 0); -} - -void -set_Raise_mem (ir_node *node, ir_node *mem) { - assert (node->op == op_Raise); - set_irn_n(node, 0, mem); -} - -ir_node * -get_Raise_exo_ptr (ir_node *node) { - assert (node->op == op_Raise); - return get_irn_n(node, 1); -} - -void -set_Raise_exo_ptr (ir_node *node, ir_node *exo_ptr) { - assert (node->op == op_Raise); - set_irn_n(node, 1, exo_ptr); -} - -tarval *get_Const_tarval (ir_node *node) { - assert (node->op == op_Const); - return node->attr.con.tv; +tarval *(get_Const_tarval)(ir_node *node) { + return _get_Const_tarval(node); } void @@ -885,18 +929,23 @@ set_Const_tarval (ir_node *node, tarval *con) { node->attr.con.tv = con; } +cnst_classify_t (classify_Const)(ir_node *node) +{ + return _classify_Const(node); +} + /* The source language type. Must be an atomic type. Mode of type must be mode of node. For tarvals from entities type must be pointer to entity type. */ -type * +ir_type * get_Const_type (ir_node *node) { assert (node->op == op_Const); return node->attr.con.tp; } void -set_Const_type (ir_node *node, type *tp) { +set_Const_type (ir_node *node, ir_type *tp) { assert (node->op == op_Const); if (tp != firm_unknown_type) { assert (is_atomic_type(tp)); @@ -918,7 +967,7 @@ set_SymConst_kind (ir_node *node, symconst_kind num) { node->attr.i.num = num; } -type * +ir_type * get_SymConst_type (ir_node *node) { assert ( (node->op == op_SymConst) && ( get_SymConst_kind(node) == symconst_type_tag @@ -927,7 +976,7 @@ get_SymConst_type (ir_node *node) { } void -set_SymConst_type (ir_node *node, type *tp) { +set_SymConst_type (ir_node *node, ir_type *tp) { assert ( (node->op == op_SymConst) && ( get_SymConst_kind(node) == symconst_type_tag || get_SymConst_kind(node) == symconst_size)); @@ -975,7 +1024,7 @@ set_SymConst_symbol (ir_node *node, union symconst_symbol sym) { node->attr.i.sym = sym; } -type * +ir_type * get_SymConst_value_type (ir_node *node) { assert (node->op == op_SymConst); if (node->attr.i.tp) node->attr.i.tp = skip_tid(node->attr.i.tp); @@ -983,7 +1032,7 @@ get_SymConst_value_type (ir_node *node) { } void -set_SymConst_value_type (ir_node *node, type *tp) { +set_SymConst_value_type (ir_node *node, ir_type *tp) { assert (node->op == op_SymConst); node->attr.i.tp = tp; } @@ -1052,42 +1101,6 @@ set_Sel_entity (ir_node *node, entity *ent) { node->attr.s.ent = ent; } -type * -get_InstOf_ent (ir_node *node) { - assert (node->op = op_InstOf); - return (node->attr.io.ent); -} - -void -set_InstOf_ent (ir_node *node, type *ent) { - assert (node->op = op_InstOf); - node->attr.io.ent = ent; -} - -ir_node * -get_InstOf_store (ir_node *node) { - assert (node->op = op_InstOf); - return (get_irn_n (node, 0)); -} - -void -set_InstOf_store (ir_node *node, ir_node *obj) { - assert (node->op = op_InstOf); - set_irn_n (node, 0, obj); -} - -ir_node * -get_InstOf_obj (ir_node *node) { - assert (node->op = op_InstOf); - return (get_irn_n (node, 1)); -} - -void -set_InstOf_obj (ir_node *node, ir_node *obj) { - assert (node->op = op_InstOf); - set_irn_n (node, 1, obj); -} - /* For unary and binary arithmetic operations the access to the operands can be factored out. Left is the first, right the @@ -1157,14 +1170,14 @@ set_Call_param (ir_node *node, int pos, ir_node *param) { set_irn_n(node, pos + CALL_PARAM_OFFSET, param); } -type * +ir_type * get_Call_type (ir_node *node) { assert (node->op == op_Call); return node->attr.call.cld_tp = skip_tid(node->attr.call.cld_tp); } void -set_Call_type (ir_node *node, type *tp) { +set_Call_type (ir_node *node, ir_type *tp) { assert (node->op == op_Call); assert ((get_unknown_type() == tp) || is_Method_type(tp)); node->attr.call.cld_tp = tp; @@ -1318,18 +1331,64 @@ BINOP(Cmp) UNOP(Conv) UNOP(Cast) -type * +ir_type * get_Cast_type (ir_node *node) { assert (node->op == op_Cast); return node->attr.cast.totype; } void -set_Cast_type (ir_node *node, type *to_tp) { +set_Cast_type (ir_node *node, ir_type *to_tp) { assert (node->op == op_Cast); node->attr.cast.totype = to_tp; } + +/* Checks for upcast. + * + * Returns true if the Cast node casts a class type to a super type. + */ +int is_Cast_upcast(ir_node *node) { + ir_type *totype = get_Cast_type(node); + ir_type *fromtype = get_irn_typeinfo_type(get_Cast_op(node)); + ir_graph *myirg = get_irn_irg(node); + + assert(get_irg_typeinfo_state(myirg) == ir_typeinfo_consistent); + assert(fromtype); + + while (is_Pointer_type(totype) && is_Pointer_type(fromtype)) { + totype = get_pointer_points_to_type(totype); + fromtype = get_pointer_points_to_type(fromtype); + } + + assert(fromtype); + + if (!is_Class_type(totype)) return 0; + return is_SubClass_of(fromtype, totype); +} + +/* Checks for downcast. + * + * Returns true if the Cast node casts a class type to a sub type. + */ +int is_Cast_downcast(ir_node *node) { + ir_type *totype = get_Cast_type(node); + ir_type *fromtype = get_irn_typeinfo_type(get_Cast_op(node)); + + assert(get_irg_typeinfo_state(get_irn_irg(node)) == ir_typeinfo_consistent); + assert(fromtype); + + while (is_Pointer_type(totype) && is_Pointer_type(fromtype)) { + totype = get_pointer_points_to_type(totype); + fromtype = get_pointer_points_to_type(fromtype); + } + + assert(fromtype); + + if (!is_Class_type(totype)) return 0; + return is_SubClass_of(totype, fromtype); +} + int (is_unop)(const ir_node *node) { return _is_unop(node); @@ -1591,14 +1650,14 @@ set_Alloc_size (ir_node *node, ir_node *size) { set_irn_n(node, 1, size); } -type * +ir_type * get_Alloc_type (ir_node *node) { assert (node->op == op_Alloc); return node->attr.a.type = skip_tid(node->attr.a.type); } void -set_Alloc_type (ir_node *node, type *tp) { +set_Alloc_type (ir_node *node, ir_type *tp) { assert (node->op == op_Alloc); node->attr.a.type = tp; } @@ -1652,16 +1711,28 @@ set_Free_size (ir_node *node, ir_node *size) { set_irn_n(node, 2, size); } -type * +ir_type * get_Free_type (ir_node *node) { assert (node->op == op_Free); - return node->attr.f = skip_tid(node->attr.f); + return node->attr.f.type = skip_tid(node->attr.f.type); } void -set_Free_type (ir_node *node, type *tp) { +set_Free_type (ir_node *node, ir_type *tp) { assert (node->op == op_Free); - node->attr.f = tp; + node->attr.f.type = tp; +} + +where_alloc +get_Free_where (ir_node *node) { + assert (node->op == op_Free); + return node->attr.f.where; +} + +void +set_Free_where (ir_node *node, where_alloc where) { + assert (node->op == op_Free); + node->attr.f.where = where; } ir_node ** @@ -1695,8 +1766,40 @@ set_Sync_pred (ir_node *node, int pos, ir_node *pred) { set_irn_n(node, pos, pred); } +ir_type *get_Proj_type(ir_node *n) +{ + ir_type *tp = NULL; + ir_node *pred = get_Proj_pred(n); + + switch (get_irn_opcode(pred)) { + case iro_Proj: { + ir_node *pred_pred; + /* Deal with Start / Call here: we need to know the Proj Nr. */ + assert(get_irn_mode(pred) == mode_T); + pred_pred = get_Proj_pred(pred); + if (get_irn_op(pred_pred) == op_Start) { + ir_type *mtp = get_entity_type(get_irg_entity(get_irn_irg(pred_pred))); + tp = get_method_param_type(mtp, get_Proj_proj(n)); + } else if (get_irn_op(pred_pred) == op_Call) { + ir_type *mtp = get_Call_type(pred_pred); + tp = get_method_res_type(mtp, get_Proj_proj(n)); + } + } break; + case iro_Start: break; + case iro_Call: break; + case iro_Load: { + ir_node *a = get_Load_ptr(pred); + if (is_Sel(a)) + tp = get_entity_type(get_Sel_entity(a)); + } break; + default: + break; + } + return tp; +} + ir_node * -get_Proj_pred (ir_node *node) { +get_Proj_pred (const ir_node *node) { assert (is_Proj(node)); return get_irn_n(node, 0); } @@ -1708,7 +1811,7 @@ set_Proj_pred (ir_node *node, ir_node *pred) { } long -get_Proj_proj (ir_node *node) { +get_Proj_proj (const ir_node *node) { assert (is_Proj(node)); if (get_irn_opcode(node) == iro_Proj) { return node->attr.proj; @@ -1847,39 +1950,269 @@ ir_node *get_Filter_cg_pred(ir_node *node, int pos) { /* Mux support */ ir_node *get_Mux_sel (ir_node *node) { + if (node->op == op_Psi) { + assert(get_irn_arity(node) == 3); + return get_Psi_cond(node, 0); + } assert(node->op == op_Mux); return node->in[1]; } void set_Mux_sel (ir_node *node, ir_node *sel) { - assert(node->op == op_Mux); - node->in[1] = sel; + if (node->op == op_Psi) { + assert(get_irn_arity(node) == 3); + set_Psi_cond(node, 0, sel); + } + else { + assert(node->op == op_Mux); + node->in[1] = sel; + } } ir_node *get_Mux_false (ir_node *node) { + if (node->op == op_Psi) { + assert(get_irn_arity(node) == 3); + return get_Psi_default(node); + } assert(node->op == op_Mux); return node->in[2]; } void set_Mux_false (ir_node *node, ir_node *ir_false) { - assert(node->op == op_Mux); - node->in[2] = ir_false; + if (node->op == op_Psi) { + assert(get_irn_arity(node) == 3); + set_Psi_default(node, ir_false); + } + else { + assert(node->op == op_Mux); + node->in[2] = ir_false; + } } ir_node *get_Mux_true (ir_node *node) { + if (node->op == op_Psi) { + assert(get_irn_arity(node) == 3); + return get_Psi_val(node, 0); + } assert(node->op == op_Mux); return node->in[3]; } void set_Mux_true (ir_node *node, ir_node *ir_true) { - assert(node->op == op_Mux); - node->in[3] = ir_true; + if (node->op == op_Psi) { + assert(get_irn_arity(node) == 3); + set_Psi_val(node, 0, ir_true); + } + else { + assert(node->op == op_Mux); + node->in[3] = ir_true; + } +} + +/* Psi support */ +ir_node *get_Psi_cond (ir_node *node, int pos) { + int num_conds = get_Psi_n_conds(node); + assert(node->op == op_Psi); + assert(pos < num_conds); + return node->in[1 + 2 * pos]; +} + +void set_Psi_cond (ir_node *node, int pos, ir_node *cond) { + int num_conds = get_Psi_n_conds(node); + assert(node->op == op_Psi); + assert(pos < num_conds); + node->in[1 + 2 * pos] = cond; +} + +ir_node *get_Psi_val (ir_node *node, int pos) { + int num_vals = get_Psi_n_conds(node); + assert(node->op == op_Psi); + assert(pos < num_vals); + return node->in[1 + 2 * pos + 1]; +} + +void set_Psi_val (ir_node *node, int pos, ir_node *val) { + int num_vals = get_Psi_n_conds(node); + assert(node->op == op_Psi); + assert(pos < num_vals); + node->in[1 + 2 * pos + 1] = val; +} + +ir_node *get_Psi_default(ir_node *node) { + int def_pos = get_irn_arity(node); + assert(node->op == op_Psi); + return node->in[def_pos]; +} + +void set_Psi_default(ir_node *node, ir_node *val) { + int def_pos = get_irn_arity(node); + assert(node->op == op_Psi); + node->in[def_pos] = node; +} + +int (get_Psi_n_conds)(ir_node *node) { + return _get_Psi_n_conds(node); +} + +/* CopyB support */ +ir_node *get_CopyB_mem (ir_node *node) { + assert (node->op == op_CopyB); + return get_irn_n(node, 0); +} + +void set_CopyB_mem (ir_node *node, ir_node *mem) { + assert (node->op == op_CopyB); + set_irn_n(node, 0, mem); +} + +ir_node *get_CopyB_dst (ir_node *node) { + assert (node->op == op_CopyB); + return get_irn_n(node, 1); +} + +void set_CopyB_dst (ir_node *node, ir_node *dst) { + assert (node->op == op_CopyB); + set_irn_n(node, 1, dst); +} + +ir_node *get_CopyB_src (ir_node *node) { + assert (node->op == op_CopyB); + return get_irn_n(node, 2); +} + +void set_CopyB_src (ir_node *node, ir_node *src) { + assert (node->op == op_CopyB); + set_irn_n(node, 2, src); +} + +ir_type *get_CopyB_type(ir_node *node) { + assert (node->op == op_CopyB); + return node->attr.copyb.data_type; +} + +void set_CopyB_type(ir_node *node, ir_type *data_type) { + assert (node->op == op_CopyB && data_type); + node->attr.copyb.data_type = data_type; +} + + +ir_type * +get_InstOf_type (ir_node *node) { + assert (node->op = op_InstOf); + return node->attr.io.type; +} + +void +set_InstOf_type (ir_node *node, ir_type *type) { + assert (node->op = op_InstOf); + node->attr.io.type = type; +} + +ir_node * +get_InstOf_store (ir_node *node) { + assert (node->op = op_InstOf); + return get_irn_n(node, 0); +} + +void +set_InstOf_store (ir_node *node, ir_node *obj) { + assert (node->op = op_InstOf); + set_irn_n(node, 0, obj); +} + +ir_node * +get_InstOf_obj (ir_node *node) { + assert (node->op = op_InstOf); + return get_irn_n(node, 1); +} + +void +set_InstOf_obj (ir_node *node, ir_node *obj) { + assert (node->op = op_InstOf); + set_irn_n(node, 1, obj); +} + +/* Returns the memory input of a Raise operation. */ +ir_node * +get_Raise_mem (ir_node *node) { + assert (node->op == op_Raise); + return get_irn_n(node, 0); +} + +void +set_Raise_mem (ir_node *node, ir_node *mem) { + assert (node->op == op_Raise); + set_irn_n(node, 0, mem); +} + +ir_node * +get_Raise_exo_ptr (ir_node *node) { + assert (node->op == op_Raise); + return get_irn_n(node, 1); +} + +void +set_Raise_exo_ptr (ir_node *node, ir_node *exo_ptr) { + assert (node->op == op_Raise); + set_irn_n(node, 1, exo_ptr); +} + +/* Bound support */ + +/* Returns the memory input of a Bound operation. */ +ir_node *get_Bound_mem(ir_node *bound) { + assert (bound->op == op_Bound); + return get_irn_n(bound, 0); +} + +void set_Bound_mem (ir_node *bound, ir_node *mem) { + assert (bound->op == op_Bound); + set_irn_n(bound, 0, mem); +} + +/* Returns the index input of a Bound operation. */ +ir_node *get_Bound_index(ir_node *bound) { + assert (bound->op == op_Bound); + return get_irn_n(bound, 1); +} + +void set_Bound_index(ir_node *bound, ir_node *idx) { + assert (bound->op == op_Bound); + set_irn_n(bound, 1, idx); +} + +/* Returns the lower bound input of a Bound operation. */ +ir_node *get_Bound_lower(ir_node *bound) { + assert (bound->op == op_Bound); + return get_irn_n(bound, 2); +} + +void set_Bound_lower(ir_node *bound, ir_node *lower) { + assert (bound->op == op_Bound); + set_irn_n(bound, 2, lower); } +/* Returns the upper bound input of a Bound operation. */ +ir_node *get_Bound_upper(ir_node *bound) { + assert (bound->op == op_Bound); + return get_irn_n(bound, 3); +} + +void set_Bound_upper(ir_node *bound, ir_node *upper) { + assert (bound->op == op_Bound); + set_irn_n(bound, 3, upper); +} +/* returns the graph of a node */ ir_graph * get_irn_irg(const ir_node *node) { + /* + * Do not use get_nodes_Block() here, because this + * will check the pinned state. + * However even a 'wrong' block is always in the proper + * irg. + */ if (! is_Block(node)) - node = get_nodes_block(node); + node = get_irn_n(node, -1); if (is_Bad(node)) /* sometimes bad is predecessor of nodes instead of block: in case of optimization */ - node = get_nodes_block(node); + node = get_irn_n(node, -1); assert(get_irn_op(node) == op_Block); return node->attr.block.irg; } @@ -1902,39 +2235,69 @@ skip_Proj (ir_node *node) { ir_node * skip_Tuple (ir_node *node) { ir_node *pred; + ir_op *op; if (!get_opt_normalize()) return node; +restart: node = skip_Id(node); if (get_irn_op(node) == op_Proj) { pred = skip_Id(get_Proj_pred(node)); - if (get_irn_op(pred) == op_Proj) /* nested Tuple ? */ + op = get_irn_op(pred); + + /* + * Looks strange but calls get_irn_op() only once + * in most often cases. + */ + if (op == op_Proj) { /* nested Tuple ? */ pred = skip_Id(skip_Tuple(pred)); - if (get_irn_op(pred) == op_Tuple) - return get_Tuple_pred(pred, get_Proj_proj(node)); + op = get_irn_op(pred); + + if (op == op_Tuple) { + node = get_Tuple_pred(pred, get_Proj_proj(node)); + goto restart; + } + } + else if (op == op_Tuple) { + node = get_Tuple_pred(pred, get_Proj_proj(node)); + goto restart; + } } return node; } -/** returns operand of node if node is a Cast */ +/* returns operand of node if node is a Cast */ ir_node *skip_Cast (ir_node *node) { - if (node && get_irn_op(node) == op_Cast) { - return skip_Id(get_irn_n(node, 0)); - } else { - return node; - } + if (node && get_irn_op(node) == op_Cast) + return get_Cast_op(node); + return node; +} + +/* returns operand of node if node is a Confirm */ +ir_node *skip_Confirm (ir_node *node) { + if (node && get_irn_op(node) == op_Confirm) + return get_Confirm_value(node); + return node; +} + +/* skip all high-level ops */ +ir_node *skip_HighLevel(ir_node *node) { + if (node && is_op_highlevel(get_irn_op(node))) + return get_irn_n(node, 0); + return node; } #if 0 /* This should compact Id-cycles to self-cycles. It has the same (or less?) complexity - than any other approach, as Id chains are resolved and all point to the real node, or - all id's are self loops. */ + * than any other approach, as Id chains are resolved and all point to the real node, or + * all id's are self loops. + * + * Moreover, it CANNOT be switched off using get_opt_normalize() ... + */ ir_node * skip_Id (ir_node *node) { /* don't assert node !!! */ - if (!get_opt_normalize()) return node; - /* Don't use get_Id_pred: We get into an endless loop for self-referencing Ids. */ if (node && (node->op == op_Id) && (node != node->in[0+1])) { @@ -1955,8 +2318,14 @@ skip_Id (ir_node *node) { } #else /* This should compact Id-cycles to self-cycles. It has the same (or less?) complexity - than any other approach, as Id chains are resolved and all point to the real node, or - all id's are self loops. */ + * than any other approach, as Id chains are resolved and all point to the real node, or + * all id's are self loops. + * + * Note: This function takes 10% of mostly ANY the compiler run, so it's + * a little bit "hand optimized". + * + * Moreover, it CANNOT be switched off using get_opt_normalize() ... + */ ir_node * skip_Id (ir_node *node) { ir_node *pred; @@ -1964,9 +2333,7 @@ skip_Id (ir_node *node) { if (!node || (node->op != op_Id)) return node; - if (!get_opt_normalize()) return node; - - /* Don't use get_Id_pred: We get into an endless loop for + /* Don't use get_Id_pred(): We get into an endless loop for self-referencing Ids. */ pred = node->in[0+1]; @@ -1992,11 +2359,26 @@ skip_Id (ir_node *node) { } #endif +void skip_Id_and_store(ir_node **node) { + ir_node *n = *node; + + if (!n || (n->op != op_Id)) return; + + /* Don't use get_Id_pred(): We get into an endless loop for + self-referencing Ids. */ + *node = skip_Id(n); +} + int (is_Bad)(const ir_node *node) { return _is_Bad(node); } +int +(is_Const)(const ir_node *node) { + return _is_Const(node); +} + int (is_no_Block)(const ir_node *node) { return _is_no_Block(node); @@ -2007,11 +2389,34 @@ int return _is_Block(node); } -/* returns true if node is a Unknown node. */ +/* returns true if node is an Unknown node. */ int -is_Unknown (const ir_node *node) { - assert(node); - return (get_irn_op(node) == op_Unknown); +(is_Unknown)(const ir_node *node) { + return _is_Unknown(node); +} + +/* returns true if node is a Return node. */ +int +(is_Return)(const ir_node *node) { + return _is_Return(node); +} + +/* returns true if node is a Call node. */ +int +(is_Call)(const ir_node *node) { + return _is_Call(node); +} + +/* returns true if node is a Sel node. */ +int +(is_Sel)(const ir_node *node) { + return _is_Sel(node); +} + +/* returns true if node is a Mux node or a Psi with only one condition. */ +int +(is_Mux)(const ir_node *node) { + return _is_Mux(node); } int @@ -2064,11 +2469,158 @@ ir_node *get_fragile_op_mem(ir_node *node) { } /* Returns true if the operation is a forking control flow operation. */ -int -is_forking_op(const ir_node *node) { - return is_op_forking(get_irn_op(node)); +int (is_irn_forking)(const ir_node *node) { + return _is_irn_forking(node); +} + +/* Return the type associated with the value produced by n + * if the node remarks this type as it is the case for + * Cast, Const, SymConst and some Proj nodes. */ +ir_type *(get_irn_type)(ir_node *node) { + return _get_irn_type(node); +} + +/* Return the type attribute of a node n (SymConst, Call, Alloc, Free, + Cast) or NULL.*/ +ir_type *(get_irn_type_attr)(ir_node *node) { + return _get_irn_type_attr(node); +} + +/* Return the entity attribute of a node n (SymConst, Sel) or NULL. */ +entity *(get_irn_entity_attr)(ir_node *node) { + return _get_irn_entity_attr(node); +} + +/* Returns non-zero for constant-like nodes. */ +int (is_irn_constlike)(const ir_node *node) { + return _is_irn_constlike(node); } +/* + * Returns non-zero for nodes that are allowed to have keep-alives and + * are neither Block nor PhiM. + */ +int (is_irn_keep)(const ir_node *node) { + return _is_irn_keep(node); +} + +/* Returns non-zero for nodes that are machine operations. */ +int (is_irn_machine_op)(const ir_node *node) { + return _is_irn_machine_op(node); +} + +/* Returns non-zero for nodes that are machine operands. */ +int (is_irn_machine_operand)(const ir_node *node) { + return _is_irn_machine_operand(node); +} + +/* Returns non-zero for nodes that have the n'th user machine flag set. */ +int (is_irn_machine_user)(const ir_node *node, unsigned n) { + return _is_irn_machine_user(node, n); +} + + +/* Gets the string representation of the jump prediction .*/ +const char *get_cond_jmp_predicate_name(cond_jmp_predicate pred) +{ + switch (pred) { + default: + case COND_JMP_PRED_NONE: return "no prediction"; + case COND_JMP_PRED_TRUE: return "true taken"; + case COND_JMP_PRED_FALSE: return "false taken"; + } +} + +/* Returns the conditional jump prediction of a Cond node. */ +cond_jmp_predicate (get_Cond_jmp_pred)(ir_node *cond) { + return _get_Cond_jmp_pred(cond); +} + +/* Sets a new conditional jump prediction. */ +void (set_Cond_jmp_pred)(ir_node *cond, cond_jmp_predicate pred) { + _set_Cond_jmp_pred(cond, pred); +} + +/** the get_type operation must be always implemented and return a firm type */ +static ir_type *get_Default_type(ir_node *n) { + return get_unknown_type(); +} + +/* Sets the get_type operation for an ir_op_ops. */ +ir_op_ops *firm_set_default_get_type(opcode code, ir_op_ops *ops) +{ + switch (code) { + case iro_Const: ops->get_type = get_Const_type; break; + case iro_SymConst: ops->get_type = get_SymConst_value_type; break; + case iro_Cast: ops->get_type = get_Cast_type; break; + case iro_Proj: ops->get_type = get_Proj_type; break; + default: + /* not allowed to be NULL */ + if (! ops->get_type) + ops->get_type = get_Default_type; + break; + } + return ops; +} + +/** Return the attribute type of a SymConst node if exists */ +static ir_type *get_SymConst_attr_type(ir_node *self) { + symconst_kind kind = get_SymConst_kind(self); + if (kind == symconst_type_tag || kind == symconst_size) + return get_SymConst_type(self); + return NULL; +} + +/** Return the attribute entity of a SymConst node if exists */ +static entity *get_SymConst_attr_entity(ir_node *self) { + symconst_kind kind = get_SymConst_kind(self); + if (kind == symconst_addr_ent) + return get_SymConst_entity(self); + return NULL; +} + +/** the get_type_attr operation must be always implemented */ +static ir_type *get_Null_type(ir_node *n) { + return firm_unknown_type; +} + +/* Sets the get_type operation for an ir_op_ops. */ +ir_op_ops *firm_set_default_get_type_attr(opcode code, ir_op_ops *ops) +{ + switch (code) { + case iro_SymConst: ops->get_type_attr = get_SymConst_attr_type; break; + case iro_Call: ops->get_type_attr = get_Call_type; break; + case iro_Alloc: ops->get_type_attr = get_Alloc_type; break; + case iro_Free: ops->get_type_attr = get_Free_type; break; + case iro_Cast: ops->get_type_attr = get_Cast_type; break; + default: + /* not allowed to be NULL */ + if (! ops->get_type_attr) + ops->get_type_attr = get_Null_type; + break; + } + return ops; +} + +/** the get_entity_attr operation must be always implemented */ +static entity *get_Null_ent(ir_node *n) { + return NULL; +} + +/* Sets the get_type operation for an ir_op_ops. */ +ir_op_ops *firm_set_default_get_entity_attr(opcode code, ir_op_ops *ops) +{ + switch (code) { + case iro_SymConst: ops->get_entity_attr = get_SymConst_attr_entity; break; + case iro_Sel: ops->get_entity_attr = get_Sel_entity; break; + default: + /* not allowed to be NULL */ + if (! ops->get_entity_attr) + ops->get_entity_attr = get_Null_ent; + break; + } + return ops; +} #ifdef DEBUG_libfirm void dump_irn (ir_node *n) {