X-Git-Url: http://nsz.repo.hu/git/?a=blobdiff_plain;f=ir%2Fir%2Firopt.c;h=0c7cf979ae8f470c81f26de68611afad451fc02c;hb=997941bd3c7913ecf8ffe4a4a08836fda815982b;hp=5f4c08f3c0639eb7f3dca624e695a95c15f5eb89;hpb=cc9831d8dbcdd84f2a270c0c23134dd1745e9392;p=libfirm diff --git a/ir/ir/iropt.c b/ir/ir/iropt.c index 5f4c08f3c..0c7cf979a 100644 --- a/ir/ir/iropt.c +++ b/ir/ir/iropt.c @@ -47,6 +47,8 @@ #include "irtools.h" #include "irhooks.h" #include "array_t.h" +#include "vrp.h" +#include "firm_types.h" /* Make types visible to allow most efficient access */ #include "entity_t.h" @@ -54,7 +56,8 @@ /** * Returns the tarval of a Const node or tarval_bad for all other nodes. */ -static tarval *default_value_of(const ir_node *n) { +static tarval *default_value_of(const ir_node *n) +{ if (is_Const(n)) return get_Const_tarval(n); /* might return tarval_bad */ else @@ -64,7 +67,8 @@ static tarval *default_value_of(const ir_node *n) { value_of_func value_of_ptr = default_value_of; /* * Set a new value_of function. */ -void set_value_of_func(value_of_func func) { +void set_value_of_func(value_of_func func) +{ if (func != NULL) value_of_ptr = func; else @@ -74,14 +78,16 @@ void set_value_of_func(value_of_func func) { /** * Return the value of a Constant. */ -static tarval *computed_value_Const(const ir_node *n) { +static tarval *computed_value_Const(const ir_node *n) +{ return get_Const_tarval(n); } /* computed_value_Const */ /** * Return the value of a 'sizeof', 'alignof' or 'offsetof' SymConst. */ -static tarval *computed_value_SymConst(const ir_node *n) { +static tarval *computed_value_SymConst(const ir_node *n) +{ ir_type *type; ir_entity *ent; @@ -111,7 +117,8 @@ static tarval *computed_value_SymConst(const ir_node *n) { /** * Return the value of an Add. */ -static tarval *computed_value_Add(const ir_node *n) { +static tarval *computed_value_Add(const ir_node *n) +{ ir_node *a = get_Add_left(n); ir_node *b = get_Add_right(n); @@ -128,7 +135,8 @@ static tarval *computed_value_Add(const ir_node *n) { * Return the value of a Sub. * Special case: a - a */ -static tarval *computed_value_Sub(const ir_node *n) { +static tarval *computed_value_Sub(const ir_node *n) +{ ir_mode *mode = get_irn_mode(n); ir_node *a = get_Sub_left(n); ir_node *b = get_Sub_right(n); @@ -155,7 +163,8 @@ static tarval *computed_value_Sub(const ir_node *n) { * Return the value of a Carry. * Special : a op 0, 0 op b */ -static tarval *computed_value_Carry(const ir_node *n) { +static tarval *computed_value_Carry(const ir_node *n) +{ ir_node *a = get_binop_left(n); ir_node *b = get_binop_right(n); ir_mode *m = get_irn_mode(n); @@ -177,7 +186,8 @@ static tarval *computed_value_Carry(const ir_node *n) { * Return the value of a Borrow. * Special : a op 0 */ -static tarval *computed_value_Borrow(const ir_node *n) { +static tarval *computed_value_Borrow(const ir_node *n) +{ ir_node *a = get_binop_left(n); ir_node *b = get_binop_right(n); ir_mode *m = get_irn_mode(n); @@ -196,7 +206,8 @@ static tarval *computed_value_Borrow(const ir_node *n) { /** * Return the value of an unary Minus. */ -static tarval *computed_value_Minus(const ir_node *n) { +static tarval *computed_value_Minus(const ir_node *n) +{ ir_node *a = get_Minus_op(n); tarval *ta = value_of(a); @@ -209,7 +220,8 @@ static tarval *computed_value_Minus(const ir_node *n) { /** * Return the value of a Mul. */ -static tarval *computed_value_Mul(const ir_node *n) { +static tarval *computed_value_Mul(const ir_node *n) +{ ir_node *a = get_Mul_left(n); ir_node *b = get_Mul_right(n); ir_mode *mode; @@ -242,7 +254,8 @@ static tarval *computed_value_Mul(const ir_node *n) { /** * Return the value of an Abs. */ -static tarval *computed_value_Abs(const ir_node *n) { +static tarval *computed_value_Abs(const ir_node *n) +{ ir_node *a = get_Abs_op(n); tarval *ta = value_of(a); @@ -256,7 +269,8 @@ static tarval *computed_value_Abs(const ir_node *n) { * Return the value of an And. * Special case: a & 0, 0 & b */ -static tarval *computed_value_And(const ir_node *n) { +static tarval *computed_value_And(const ir_node *n) +{ ir_node *a = get_And_left(n); ir_node *b = get_And_right(n); @@ -276,7 +290,8 @@ static tarval *computed_value_And(const ir_node *n) { * Return the value of an Or. * Special case: a | 1...1, 1...1 | b */ -static tarval *computed_value_Or(const ir_node *n) { +static tarval *computed_value_Or(const ir_node *n) +{ ir_node *a = get_Or_left(n); ir_node *b = get_Or_right(n); @@ -295,7 +310,8 @@ static tarval *computed_value_Or(const ir_node *n) { /** * Return the value of an Eor. */ -static tarval *computed_value_Eor(const ir_node *n) { +static tarval *computed_value_Eor(const ir_node *n) +{ ir_node *a = get_Eor_left(n); ir_node *b = get_Eor_right(n); @@ -316,7 +332,8 @@ static tarval *computed_value_Eor(const ir_node *n) { /** * Return the value of a Not. */ -static tarval *computed_value_Not(const ir_node *n) { +static tarval *computed_value_Not(const ir_node *n) +{ ir_node *a = get_Not_op(n); tarval *ta = value_of(a); @@ -329,7 +346,8 @@ static tarval *computed_value_Not(const ir_node *n) { /** * Return the value of a Shl. */ -static tarval *computed_value_Shl(const ir_node *n) { +static tarval *computed_value_Shl(const ir_node *n) +{ ir_node *a = get_Shl_left(n); ir_node *b = get_Shl_right(n); @@ -345,7 +363,8 @@ static tarval *computed_value_Shl(const ir_node *n) { /** * Return the value of a Shr. */ -static tarval *computed_value_Shr(const ir_node *n) { +static tarval *computed_value_Shr(const ir_node *n) +{ ir_node *a = get_Shr_left(n); ir_node *b = get_Shr_right(n); @@ -361,7 +380,8 @@ static tarval *computed_value_Shr(const ir_node *n) { /** * Return the value of a Shrs. */ -static tarval *computed_value_Shrs(const ir_node *n) { +static tarval *computed_value_Shrs(const ir_node *n) +{ ir_node *a = get_Shrs_left(n); ir_node *b = get_Shrs_right(n); @@ -377,7 +397,8 @@ static tarval *computed_value_Shrs(const ir_node *n) { /** * Return the value of a Rotl. */ -static tarval *computed_value_Rotl(const ir_node *n) { +static tarval *computed_value_Rotl(const ir_node *n) +{ ir_node *a = get_Rotl_left(n); ir_node *b = get_Rotl_right(n); @@ -393,7 +414,8 @@ static tarval *computed_value_Rotl(const ir_node *n) { /** * Return the value of a Conv. */ -static tarval *computed_value_Conv(const ir_node *n) { +static tarval *computed_value_Conv(const ir_node *n) +{ ir_node *a = get_Conv_op(n); tarval *ta = value_of(a); @@ -407,7 +429,8 @@ static tarval *computed_value_Conv(const ir_node *n) { * Calculate the value of a Mux: can be evaluated, if the * sel and the right input are known. */ -static tarval *computed_value_Mux(const ir_node *n) { +static tarval *computed_value_Mux(const ir_node *n) +{ ir_node *sel = get_Mux_sel(n); tarval *ts = value_of(sel); @@ -426,7 +449,8 @@ static tarval *computed_value_Mux(const ir_node *n) { * Calculate the value of a Confirm: can be evaluated, * if it has the form Confirm(x, '=', Const). */ -static tarval *computed_value_Confirm(const ir_node *n) { +static tarval *computed_value_Confirm(const ir_node *n) +{ /* * Beware: we might produce Phi(Confirm(x == true), Confirm(x == false)). * Do NOT optimize them away (jump threading wants them), so wait until @@ -451,113 +475,102 @@ static tarval *computed_value_Confirm(const ir_node *n) { * only 1 is used. * There are several case where we can evaluate a Cmp node, see later. */ -static tarval *computed_value_Proj_Cmp(const ir_node *n) { - ir_node *a = get_Proj_pred(n); - ir_node *aa = get_Cmp_left(a); - ir_node *ab = get_Cmp_right(a); - long proj_nr = get_Proj_proj(n); +static tarval *computed_value_Proj_Cmp(const ir_node *n) +{ + ir_node *cmp = get_Proj_pred(n); + ir_node *left = get_Cmp_left(cmp); + ir_node *right = get_Cmp_right(cmp); + long pn_cmp = get_Proj_proj(n); + ir_mode *mode = get_irn_mode(left); + tarval *tv_l, *tv_r; /* * BEWARE: a == a is NOT always True for floating Point values, as * NaN != NaN is defined, so we must check this here. */ - if (aa == ab && ( - !mode_is_float(get_irn_mode(aa)) || proj_nr == pn_Cmp_Lt || proj_nr == pn_Cmp_Gt) - ) { /* 1.: */ - + if (left == right && (!mode_is_float(mode) || pn_cmp == pn_Cmp_Lt || pn_cmp == pn_Cmp_Gt)) { /* This is a trick with the bits used for encoding the Cmp Proj numbers, the following statement is not the same: - return new_tarval_from_long (proj_nr == pn_Cmp_Eq, mode_b) */ - return new_tarval_from_long (proj_nr & pn_Cmp_Eq, mode_b); + return new_tarval_from_long(pn_cmp == pn_Cmp_Eq, mode_b) */ + return new_tarval_from_long(pn_cmp & pn_Cmp_Eq, mode_b); } - else { - tarval *taa = value_of(aa); - tarval *tab = value_of(ab); - ir_mode *mode = get_irn_mode(aa); + tv_l = value_of(left); + tv_r = value_of(right); + if ((tv_l != tarval_bad) && (tv_r != tarval_bad)) { /* * The predecessors of Cmp are target values. We can evaluate * the Cmp. */ - if ((taa != tarval_bad) && (tab != tarval_bad)) { - /* strange checks... */ - pn_Cmp flags = tarval_cmp(taa, tab); - if (flags != pn_Cmp_False) { - return new_tarval_from_long (proj_nr & flags, mode_b); - } + pn_Cmp flags = tarval_cmp(tv_l, tv_r); + if (flags != pn_Cmp_False) { + return new_tarval_from_long (pn_cmp & flags, mode_b); } + } else if (mode_is_int(mode)) { /* for integer values, we can check against MIN/MAX */ - else if (mode_is_int(mode)) { + pn_Cmp cmp_result; + + if (tv_l == get_mode_min(mode)) { /* MIN <=/> x. This results in true/false. */ - if (taa == get_mode_min(mode)) { - /* a compare with the MIN value */ - if (proj_nr == pn_Cmp_Le) - return get_tarval_b_true(); - else if (proj_nr == pn_Cmp_Gt) - return get_tarval_b_false(); - } + if (pn_cmp == pn_Cmp_Le) + return tarval_b_true; + else if (pn_cmp == pn_Cmp_Gt) + return tarval_b_false; + } else if (tv_r == get_mode_min(mode)) { /* x >=/< MIN. This results in true/false. */ - else - if (tab == get_mode_min(mode)) { - /* a compare with the MIN value */ - if (proj_nr == pn_Cmp_Ge) - return get_tarval_b_true(); - else if (proj_nr == pn_Cmp_Lt) - return get_tarval_b_false(); - } - /* MAX >=/< x. This results in true/false. */ - else if (taa == get_mode_max(mode)) { - if (proj_nr == pn_Cmp_Ge) - return get_tarval_b_true(); - else if (proj_nr == pn_Cmp_Lt) - return get_tarval_b_false(); - } - /* x <=/> MAX. This results in true/false. */ - else if (tab == get_mode_max(mode)) { - if (proj_nr == pn_Cmp_Le) - return get_tarval_b_true(); - else if (proj_nr == pn_Cmp_Gt) - return get_tarval_b_false(); + if (pn_cmp == pn_Cmp_Ge) + return tarval_b_true; + else if (pn_cmp == pn_Cmp_Lt) + return tarval_b_false; + } else if (tv_l == get_mode_max(mode)) { + /* MAX >=/< x. This results in true/false. */ + if (pn_cmp == pn_Cmp_Ge) + return tarval_b_true; + else if (pn_cmp == pn_Cmp_Lt) + return tarval_b_false; + } else if (tv_r == get_mode_max(mode)) { + /* x <=/> MAX. This results in true/false. */ + if (pn_cmp == pn_Cmp_Le) + return tarval_b_true; + else if (pn_cmp == pn_Cmp_Gt) + return tarval_b_false; + } + + cmp_result = vrp_cmp(left, right); + if (cmp_result != pn_Cmp_False) { + if (cmp_result == pn_Cmp_Lg) { + if (pn_cmp == pn_Cmp_Eq) { + return tarval_b_false; + } else if (pn_cmp == pn_Cmp_Lg) { + return tarval_b_true; } + } else { + return new_tarval_from_long(cmp_result & pn_cmp, mode_b); + } } - /* - * The predecessors are Allocs or (void*)(0) constants. Allocs never - * return NULL, they raise an exception. Therefore we can predict - * the Cmp result. - */ - else { - ir_node *aaa = skip_Proj(aa); - ir_node *aba = skip_Proj(ab); - - if ( ( (/* aa is ProjP and aaa is Alloc */ - is_Proj(aa) - && mode_is_reference(get_irn_mode(aa)) - && is_Alloc(aaa)) - && ( (/* ab is NULL */ - mode_is_reference(get_irn_mode(ab)) - && tarval_is_null(tab)) - || (/* ab is other Alloc */ - is_Proj(ab) - && mode_is_reference(get_irn_mode(ab)) - && is_Alloc(aba) - && (aaa != aba)))) - || (/* aa is NULL and aba is Alloc */ - mode_is_reference(get_irn_mode(aa)) - && tarval_is_null(taa) - && is_Proj(ab) - && mode_is_reference(get_irn_mode(ab)) - && is_Alloc(aba))) - /* 3.: */ - return new_tarval_from_long(proj_nr & pn_Cmp_Lg, mode_b); - } - } - return computed_value_Cmp_Confirm(a, aa, ab, proj_nr); + } else if (mode_is_reference(mode)) { + /* pointer compare */ + ir_node *s_l = skip_Proj(left); + ir_node *s_r = skip_Proj(right); + + if ((is_Alloc(s_l) && tarval_is_null(tv_r)) || + (tarval_is_null(tv_l) && is_Alloc(s_r))) { + /* + * The predecessors are Allocs and (void*)(0) constants. In Firm Allocs never + * return NULL, they raise an exception. Therefore we can predict + * the Cmp result. + */ + return new_tarval_from_long(pn_cmp & pn_Cmp_Lg, mode_b); + } + } + return computed_value_Cmp_Confirm(cmp, left, right, pn_cmp); } /* computed_value_Proj_Cmp */ /** * Return the value of a floating point Quot. */ -static tarval *do_computed_value_Quot(const ir_node *a, const ir_node *b) { +static tarval *do_computed_value_Quot(const ir_node *a, const ir_node *b) +{ tarval *ta = value_of(a); tarval *tb = value_of(b); @@ -571,7 +584,8 @@ static tarval *do_computed_value_Quot(const ir_node *a, const ir_node *b) { * Calculate the value of an integer Div of two nodes. * Special case: 0 / b */ -static tarval *do_computed_value_Div(const ir_node *a, const ir_node *b) { +static tarval *do_computed_value_Div(const ir_node *a, const ir_node *b) +{ tarval *ta = value_of(a); tarval *tb; const ir_node *dummy; @@ -589,7 +603,8 @@ static tarval *do_computed_value_Div(const ir_node *a, const ir_node *b) { * Calculate the value of an integer Mod of two nodes. * Special case: a % 1 */ -static tarval *do_computed_value_Mod(const ir_node *a, const ir_node *b) { +static tarval *do_computed_value_Mod(const ir_node *a, const ir_node *b) +{ tarval *ta = value_of(a); tarval *tb = value_of(b); @@ -604,7 +619,8 @@ static tarval *do_computed_value_Mod(const ir_node *a, const ir_node *b) { /** * Return the value of a Proj(DivMod). */ -static tarval *computed_value_Proj_DivMod(const ir_node *n) { +static tarval *computed_value_Proj_DivMod(const ir_node *n) +{ long proj_nr = get_Proj_proj(n); /* compute either the Div or the Mod part */ @@ -621,7 +637,8 @@ static tarval *computed_value_Proj_DivMod(const ir_node *n) { /** * Return the value of a Proj(Div). */ -static tarval *computed_value_Proj_Div(const ir_node *n) { +static tarval *computed_value_Proj_Div(const ir_node *n) +{ long proj_nr = get_Proj_proj(n); if (proj_nr == pn_Div_res) { @@ -634,7 +651,8 @@ static tarval *computed_value_Proj_Div(const ir_node *n) { /** * Return the value of a Proj(Mod). */ -static tarval *computed_value_Proj_Mod(const ir_node *n) { +static tarval *computed_value_Proj_Mod(const ir_node *n) +{ long proj_nr = get_Proj_proj(n); if (proj_nr == pn_Mod_res) { @@ -647,7 +665,8 @@ static tarval *computed_value_Proj_Mod(const ir_node *n) { /** * Return the value of a Proj(Quot). */ -static tarval *computed_value_Proj_Quot(const ir_node *n) { +static tarval *computed_value_Proj_Quot(const ir_node *n) +{ long proj_nr = get_Proj_proj(n); if (proj_nr == pn_Quot_res) { @@ -660,7 +679,8 @@ static tarval *computed_value_Proj_Quot(const ir_node *n) { /** * Return the value of a Proj. */ -static tarval *computed_value_Proj(const ir_node *proj) { +static tarval *computed_value_Proj(const ir_node *proj) +{ ir_node *n = get_Proj_pred(proj); if (n->op->ops.computed_value_Proj != NULL) @@ -674,7 +694,12 @@ static tarval *computed_value_Proj(const ir_node *proj) { * * @param n The node this should be evaluated */ -tarval *computed_value(const ir_node *n) { +tarval *computed_value(const ir_node *n) +{ + vrp_attr *vrp = vrp_get_info(n); + if (vrp && vrp->valid && tarval_cmp(vrp->bits_set, vrp->bits_not_set) == pn_Cmp_Eq) { + return vrp->bits_set; + } if (n->op->ops.computed_value) return n->op->ops.computed_value(n); return tarval_bad; @@ -728,7 +753,8 @@ static ir_op_ops *firm_set_default_computed_value(ir_opcode code, ir_op_ops *ops CASE_PROJ(Quot); CASE(Proj); default: - /* leave NULL */; + /* leave NULL */ + break; } return ops; @@ -849,7 +875,8 @@ static ir_node *equivalent_node_Block(ir_node *n) * Returns a equivalent node for a Jmp, a Bad :-) * Of course this only happens if the Block of the Jmp is dead. */ -static ir_node *equivalent_node_Jmp(ir_node *n) { +static ir_node *equivalent_node_Jmp(ir_node *n) +{ ir_node *oldn = n; /* unreachable code elimination */ @@ -871,7 +898,8 @@ static ir_node *equivalent_node_Jmp(ir_node *n) { * Optimize operations that are commutative and have neutral 0, * so a op 0 = 0 op a = a. */ -static ir_node *equivalent_node_neutral_zero(ir_node *n) { +static ir_node *equivalent_node_neutral_zero(ir_node *n) +{ ir_node *oldn = n; ir_node *a = get_binop_left(n); @@ -908,7 +936,8 @@ static ir_node *equivalent_node_neutral_zero(ir_node *n) { /** * Eor is commutative and has neutral 0. */ -static ir_node *equivalent_node_Eor(ir_node *n) { +static ir_node *equivalent_node_Eor(ir_node *n) +{ ir_node *oldn = n; ir_node *a; ir_node *b; @@ -963,7 +992,8 @@ static ir_node *equivalent_node_Eor(ir_node *n) { * Beware: The Mode of an Add may be different than the mode of its * predecessors, so we could not return a predecessors in all cases. */ -static ir_node *equivalent_node_Add(ir_node *n) { +static ir_node *equivalent_node_Add(ir_node *n) +{ ir_node *oldn = n; ir_node *left, *right; ir_mode *mode = get_irn_mode(n); @@ -1008,7 +1038,8 @@ static ir_node *equivalent_node_Add(ir_node *n) { * optimize operations that are not commutative but have neutral 0 on left, * so a op 0 = a. */ -static ir_node *equivalent_node_left_zero(ir_node *n) { +static ir_node *equivalent_node_left_zero(ir_node *n) +{ ir_node *oldn = n; ir_node *a = get_binop_left(n); @@ -1037,7 +1068,8 @@ static ir_node *equivalent_node_left_zero(ir_node *n) { * Beware: The Mode of a Sub may be different than the mode of its * predecessors, so we could not return a predecessors in all cases. */ -static ir_node *equivalent_node_Sub(ir_node *n) { +static ir_node *equivalent_node_Sub(ir_node *n) +{ ir_node *oldn = n; ir_node *b; ir_mode *mode = get_irn_mode(n); @@ -1071,7 +1103,8 @@ static ir_node *equivalent_node_Sub(ir_node *n) { * We handle it anyway here but the better way would be a * flag. This would be needed for Pascal for instance. */ -static ir_node *equivalent_node_idempotent_unop(ir_node *n) { +static ir_node *equivalent_node_idempotent_unop(ir_node *n) +{ ir_node *oldn = n; ir_node *pred = get_unop_op(n); @@ -1093,7 +1126,8 @@ static ir_node *equivalent_node_idempotent_unop(ir_node *n) { /** * Optimize a * 1 = 1 * a = a. */ -static ir_node *equivalent_node_Mul(ir_node *n) { +static ir_node *equivalent_node_Mul(ir_node *n) +{ ir_node *oldn = n; ir_node *a = get_Mul_left(n); @@ -1124,7 +1158,8 @@ static ir_node *equivalent_node_Mul(ir_node *n) { /** * Use algebraic simplification a | a = a | 0 = 0 | a = a. */ -static ir_node *equivalent_node_Or(ir_node *n) { +static ir_node *equivalent_node_Or(ir_node *n) +{ ir_node *oldn = n; ir_node *a = get_Or_left(n); @@ -1156,7 +1191,8 @@ static ir_node *equivalent_node_Or(ir_node *n) { /** * Optimize a & 0b1...1 = 0b1...1 & a = a & a = (a|X) & a = a. */ -static ir_node *equivalent_node_And(ir_node *n) { +static ir_node *equivalent_node_And(ir_node *n) +{ ir_node *oldn = n; ir_node *a = get_And_left(n); @@ -1218,7 +1254,8 @@ static ir_node *equivalent_node_And(ir_node *n) { /** * Try to remove useless Conv's: */ -static ir_node *equivalent_node_Conv(ir_node *n) { +static ir_node *equivalent_node_Conv(ir_node *n) +{ ir_node *oldn = n; ir_node *a = get_Conv_op(n); @@ -1351,7 +1388,8 @@ restart: * A Cast may be removed if the type of the previous node * is already the type of the Cast. */ -static ir_node *equivalent_node_Cast(ir_node *n) { +static ir_node *equivalent_node_Cast(ir_node *n) +{ ir_node *oldn = n; ir_node *pred = get_Cast_op(n); @@ -1366,7 +1404,8 @@ static ir_node *equivalent_node_Cast(ir_node *n) { * - fold Phi-nodes, iff they have only one predecessor except * themselves. */ -static ir_node *equivalent_node_Phi(ir_node *n) { +static ir_node *equivalent_node_Phi(ir_node *n) +{ int i, n_preds; ir_node *oldn = n; @@ -1436,7 +1475,8 @@ static ir_node *equivalent_node_Phi(ir_node *n) { * - fold Sync-nodes, iff they have only one predecessor except * themselves. */ -static ir_node *equivalent_node_Sync(ir_node *n) { +static ir_node *equivalent_node_Sync(ir_node *n) +{ int arity = get_Sync_n_preds(n); int i; @@ -1473,7 +1513,8 @@ static ir_node *equivalent_node_Sync(ir_node *n) { /** * Optimize Proj(Tuple). */ -static ir_node *equivalent_node_Proj_Tuple(ir_node *proj) { +static ir_node *equivalent_node_Proj_Tuple(ir_node *proj) +{ ir_node *oldn = proj; ir_node *tuple = get_Proj_pred(proj); @@ -1487,7 +1528,8 @@ static ir_node *equivalent_node_Proj_Tuple(ir_node *proj) { /** * Optimize a / 1 = a. */ -static ir_node *equivalent_node_Proj_Div(ir_node *proj) { +static ir_node *equivalent_node_Proj_Div(ir_node *proj) +{ ir_node *oldn = proj; ir_node *div = get_Proj_pred(proj); ir_node *b = get_Div_right(div); @@ -1518,7 +1560,8 @@ static ir_node *equivalent_node_Proj_Div(ir_node *proj) { /** * Optimize a / 1.0 = a. */ -static ir_node *equivalent_node_Proj_Quot(ir_node *proj) { +static ir_node *equivalent_node_Proj_Quot(ir_node *proj) +{ ir_node *oldn = proj; ir_node *quot = get_Proj_pred(proj); ir_node *b = get_Quot_right(quot); @@ -1549,7 +1592,8 @@ static ir_node *equivalent_node_Proj_Quot(ir_node *proj) { /** * Optimize a / 1 = a. */ -static ir_node *equivalent_node_Proj_DivMod(ir_node *proj) { +static ir_node *equivalent_node_Proj_DivMod(ir_node *proj) +{ ir_node *oldn = proj; ir_node *divmod = get_Proj_pred(proj); ir_node *b = get_DivMod_right(divmod); @@ -1582,7 +1626,8 @@ static ir_node *equivalent_node_Proj_DivMod(ir_node *proj) { /** * Optimize CopyB(mem, x, x) into a Nop. */ -static ir_node *equivalent_node_Proj_CopyB(ir_node *proj) { +static ir_node *equivalent_node_Proj_CopyB(ir_node *proj) +{ ir_node *oldn = proj; ir_node *copyb = get_Proj_pred(proj); ir_node *a = get_CopyB_dst(copyb); @@ -1608,7 +1653,8 @@ static ir_node *equivalent_node_Proj_CopyB(ir_node *proj) { /** * Optimize Bounds(idx, idx, upper) into idx. */ -static ir_node *equivalent_node_Proj_Bound(ir_node *proj) { +static ir_node *equivalent_node_Proj_Bound(ir_node *proj) +{ ir_node *oldn = proj; ir_node *bound = get_Proj_pred(proj); ir_node *idx = get_Bound_index(bound); @@ -1653,7 +1699,7 @@ static ir_node *equivalent_node_Proj_Bound(ir_node *proj) { break; default: /* cannot optimize pn_Bound_X_regular, handled in transform ... */ - ; + break; } } return proj; @@ -1662,7 +1708,8 @@ static ir_node *equivalent_node_Proj_Bound(ir_node *proj) { /** * Optimize an Exception Proj(Load) with a non-null address. */ -static ir_node *equivalent_node_Proj_Load(ir_node *proj) { +static ir_node *equivalent_node_Proj_Load(ir_node *proj) +{ if (get_opt_ldst_only_null_ptr_exceptions()) { if (get_irn_mode(proj) == mode_X) { ir_node *load = get_Proj_pred(proj); @@ -1685,7 +1732,8 @@ static ir_node *equivalent_node_Proj_Load(ir_node *proj) { /** * Optimize an Exception Proj(Store) with a non-null address. */ -static ir_node *equivalent_node_Proj_Store(ir_node *proj) { +static ir_node *equivalent_node_Proj_Store(ir_node *proj) +{ if (get_opt_ldst_only_null_ptr_exceptions()) { if (get_irn_mode(proj) == mode_X) { ir_node *store = get_Proj_pred(proj); @@ -1709,7 +1757,8 @@ static ir_node *equivalent_node_Proj_Store(ir_node *proj) { * Does all optimizations on nodes that must be done on it's Proj's * because of creating new nodes. */ -static ir_node *equivalent_node_Proj(ir_node *proj) { +static ir_node *equivalent_node_Proj(ir_node *proj) +{ ir_node *n = get_Proj_pred(proj); if (get_irn_mode(proj) == mode_X) { @@ -1726,7 +1775,8 @@ static ir_node *equivalent_node_Proj(ir_node *proj) { /** * Remove Id's. */ -static ir_node *equivalent_node_Id(ir_node *n) { +static ir_node *equivalent_node_Id(ir_node *n) +{ ir_node *oldn = n; do { @@ -1844,7 +1894,8 @@ static ir_node *equivalent_node_Mux(ir_node *n) * Remove Confirm nodes if setting is on. * Replace Confirms(x, '=', Constlike) by Constlike. */ -static ir_node *equivalent_node_Confirm(ir_node *n) { +static ir_node *equivalent_node_Confirm(ir_node *n) +{ ir_node *pred = get_Confirm_value(n); pn_Cmp pnc = get_Confirm_cmp(n); @@ -1869,7 +1920,8 @@ static ir_node *equivalent_node_Confirm(ir_node *n) { * If a node returns a Tuple we can not just skip it. If the size of the * in array fits, we transform n into a tuple (e.g., Div). */ -ir_node *equivalent_node(ir_node *n) { +ir_node *equivalent_node(ir_node *n) +{ if (n->op->ops.equivalent_node) return n->op->ops.equivalent_node(n); return n; @@ -1928,7 +1980,8 @@ static ir_op_ops *firm_set_default_equivalent_node(ir_opcode code, ir_op_ops *op CASE(Mux); CASE(Confirm); default: - /* leave NULL */; + /* leave NULL */ + break; } return ops; @@ -1940,7 +1993,8 @@ static ir_op_ops *firm_set_default_equivalent_node(ir_opcode code, ir_op_ops *op * Returns non-zero if a node is a Phi node * with all predecessors constant. */ -static int is_const_Phi(ir_node *n) { +static int is_const_Phi(ir_node *n) +{ int i; if (! is_Phi(n) || get_irn_arity(n) == 0) @@ -1988,7 +2042,8 @@ static tarval *do_eval(eval_func eval, tarval *a, tarval *b, ir_mode *mode) * * @return a new Phi node if the conversion was successful, NULL else */ -static ir_node *apply_binop_on_phi(ir_node *phi, tarval *other, eval_func eval, ir_mode *mode, int left) { +static ir_node *apply_binop_on_phi(ir_node *phi, tarval *other, eval_func eval, ir_mode *mode, int left) +{ tarval *tv; void **res; ir_node *pred; @@ -2039,7 +2094,8 @@ static ir_node *apply_binop_on_phi(ir_node *phi, tarval *other, eval_func eval, * * @return a new Phi node if the conversion was successful, NULL else */ -static ir_node *apply_binop_on_2_phis(ir_node *a, ir_node *b, eval_func eval, ir_mode *mode) { +static ir_node *apply_binop_on_2_phis(ir_node *a, ir_node *b, eval_func eval, ir_mode *mode) +{ tarval *tv_l, *tv_r, *tv; void **res; ir_node *pred; @@ -2081,7 +2137,8 @@ static ir_node *apply_binop_on_2_phis(ir_node *a, ir_node *b, eval_func eval, ir * * @return a new Phi node if the conversion was successful, NULL else */ -static ir_node *apply_unop_on_phi(ir_node *phi, tarval *(*eval)(tarval *)) { +static ir_node *apply_unop_on_phi(ir_node *phi, tarval *(*eval)(tarval *)) +{ tarval *tv; void **res; ir_node *pred; @@ -2117,7 +2174,8 @@ static ir_node *apply_unop_on_phi(ir_node *phi, tarval *(*eval)(tarval *)) { * * @return a new Phi node if the conversion was successful, NULL else */ -static ir_node *apply_conv_on_phi(ir_node *phi, ir_mode *mode) { +static ir_node *apply_conv_on_phi(ir_node *phi, ir_mode *mode) +{ tarval *tv; void **res; ir_node *pred; @@ -2149,7 +2207,8 @@ static ir_node *apply_conv_on_phi(ir_node *phi, ir_mode *mode) { * SubP(P, ConvIs(Iu)), SubP(P, ConvIu(Is)). * If possible, remove the Conv's. */ -static ir_node *transform_node_AddSub(ir_node *n) { +static ir_node *transform_node_AddSub(ir_node *n) +{ ir_mode *mode = get_irn_mode(n); if (mode_is_reference(mode)) { @@ -2221,6 +2280,7 @@ static ir_node *transform_node_AddSub(ir_node *n) { } /* transform_node_AddSub */ #define HANDLE_BINOP_PHI(eval, a, b, c, mode) \ + do { \ c = NULL; \ if (is_Const(b) && is_const_Phi(a)) { \ /* check for Op(Phi, Const) */ \ @@ -2237,9 +2297,11 @@ static ir_node *transform_node_AddSub(ir_node *n) { if (c) { \ DBG_OPT_ALGSIM0(oldn, c, FS_OPT_CONST_PHI); \ return c; \ - } + } \ + } while(0) #define HANDLE_UNOP_PHI(eval, a, c) \ + do { \ c = NULL; \ if (is_const_Phi(a)) { \ /* check for Op(Phi) */ \ @@ -2248,7 +2310,8 @@ static ir_node *transform_node_AddSub(ir_node *n) { DBG_OPT_ALGSIM0(oldn, c, FS_OPT_CONST_PHI); \ return c; \ } \ - } + } \ + } while(0) /** * Do the AddSub optimization, then Transform @@ -2259,9 +2322,11 @@ static ir_node *transform_node_AddSub(ir_node *n) { * Transform Add(a,-b) into Sub(a,b). * Reassociation might fold this further. */ -static ir_node *transform_node_Add(ir_node *n) { +static ir_node *transform_node_Add(ir_node *n) +{ ir_mode *mode; ir_node *a, *b, *c, *oldn = n; + vrp_attr *a_vrp, *b_vrp; n = transform_node_AddSub(n); @@ -2273,7 +2338,7 @@ static ir_node *transform_node_Add(ir_node *n) { if (mode_is_reference(mode)) { ir_mode *lmode = get_irn_mode(a); - if (is_Const(b) && is_Const_null(b) && mode_is_int(lmode)) { + if (is_Const(b) && is_Const_null(b) && mode_is_int(lmode)) { /* an Add(a, NULL) is a hidden Conv */ dbg_info *dbg = get_irn_dbg_info(n); return new_rd_Conv(dbg, get_nodes_block(n), a, mode); @@ -2352,13 +2417,30 @@ static ir_node *transform_node_Add(ir_node *n) { } } } + + a_vrp = vrp_get_info(a); + b_vrp = vrp_get_info(b); + + if (a_vrp && b_vrp) { + tarval *c = tarval_and( + a_vrp->bits_not_set, + b_vrp->bits_not_set + ); + + if (tarval_is_null(c)) { + dbg_info *dbgi = get_irn_dbg_info(n); + return new_rd_Or(dbgi, get_nodes_block(n), + a, b, mode); + } + } return n; } /* transform_node_Add */ /** * returns -cnst or NULL if impossible */ -static ir_node *const_negate(ir_node *cnst) { +static ir_node *const_negate(ir_node *cnst) +{ tarval *tv = tarval_neg(get_Const_tarval(cnst)); dbg_info *dbgi = get_irn_dbg_info(cnst); ir_graph *irg = get_irn_irg(cnst); @@ -2376,7 +2458,8 @@ static ir_node *const_negate(ir_node *cnst) { * Sub(x, Add(x, a)) -> -a * Sub(x, Const) -> Add(x, -Const) */ -static ir_node *transform_node_Sub(ir_node *n) { +static ir_node *transform_node_Sub(ir_node *n) +{ ir_mode *mode; ir_node *oldn = n; ir_node *a, *b, *c; @@ -2594,7 +2677,7 @@ restart: get_irn_dbg_info(n), blk, mb, - new_Const_long(mode, 1), + new_Const(get_mode_one(mode)), mode), mode); DBG_OPT_ALGSIM0(oldn, n, FS_OPT_SUB_MUL_A_X_A); @@ -2609,7 +2692,7 @@ restart: get_irn_dbg_info(n), blk, ma, - new_Const_long(mode, 1), + new_Const(get_mode_one(mode)), mode), mode); DBG_OPT_ALGSIM0(oldn, n, FS_OPT_SUB_MUL_A_X_A); @@ -2671,7 +2754,8 @@ restart: * Several transformation done on n*n=2n bits mul. * These transformations must be done here because new nodes may be produced. */ -static ir_node *transform_node_Mul2n(ir_node *n, ir_mode *mode) { +static ir_node *transform_node_Mul2n(ir_node *n, ir_mode *mode) +{ ir_node *oldn = n; ir_node *a = get_Mul_left(n); ir_node *b = get_Mul_right(n); @@ -2717,7 +2801,8 @@ static ir_node *transform_node_Mul2n(ir_node *n, ir_mode *mode) { * Do constant evaluation of Phi nodes. * Do architecture dependent optimizations on Mul nodes */ -static ir_node *transform_node_Mul(ir_node *n) { +static ir_node *transform_node_Mul(ir_node *n) +{ ir_node *c, *oldn = n; ir_mode *mode = get_irn_mode(n); ir_node *a = get_Mul_left(n); @@ -2832,7 +2917,8 @@ static ir_node *transform_node_Mul(ir_node *n) { /** * Transform a Div Node. */ -static ir_node *transform_node_Div(ir_node *n) { +static ir_node *transform_node_Div(ir_node *n) +{ ir_mode *mode = get_Div_resmode(n); ir_node *a = get_Div_left(n); ir_node *b = get_Div_right(n); @@ -2908,7 +2994,8 @@ make_tuple: /** * Transform a Mod node. */ -static ir_node *transform_node_Mod(ir_node *n) { +static ir_node *transform_node_Mod(ir_node *n) +{ ir_mode *mode = get_Mod_resmode(n); ir_node *a = get_Mod_left(n); ir_node *b = get_Mod_right(n); @@ -2995,7 +3082,8 @@ make_tuple: /** * Transform a DivMod node. */ -static ir_node *transform_node_DivMod(ir_node *n) { +static ir_node *transform_node_DivMod(ir_node *n) +{ const ir_node *dummy; ir_node *a = get_DivMod_left(n); ir_node *b = get_DivMod_right(n); @@ -3108,7 +3196,8 @@ make_tuple: /** * Optimize x / c to x * (1/c) */ -static ir_node *transform_node_Quot(ir_node *n) { +static ir_node *transform_node_Quot(ir_node *n) +{ ir_mode *mode = get_Quot_resmode(n); ir_node *oldn = n; @@ -3158,7 +3247,8 @@ static ir_node *transform_node_Quot(ir_node *n) { * Optimize Abs(x) into -x if x is Confirmed <= 0 * Optimize Abs(-x) int Abs(x) */ -static ir_node *transform_node_Abs(ir_node *n) { +static ir_node *transform_node_Abs(ir_node *n) +{ ir_node *c, *oldn = n; ir_node *a = get_Abs_op(n); ir_mode *mode; @@ -3209,7 +3299,8 @@ static ir_node *transform_node_Abs(ir_node *n) { * * For == and != can be handled in Proj(Cmp) */ -static ir_node *transform_node_Cmp(ir_node *n) { +static ir_node *transform_node_Cmp(ir_node *n) +{ ir_node *oldn = n; ir_node *left = get_Cmp_left(n); ir_node *right = get_Cmp_right(n); @@ -3231,7 +3322,8 @@ static ir_node *transform_node_Cmp(ir_node *n) { * Replace the Cond by a Jmp if it branches on a constant * condition. */ -static ir_node *transform_node_Cond(ir_node *n) { +static ir_node *transform_node_Cond(ir_node *n) +{ ir_node *jmp; ir_node *a = get_Cond_selector(n); @@ -3282,7 +3374,7 @@ static ir_node *transform_bitwise_distributive(ir_node *n, ir_op *op = get_irn_op(a); ir_op *op_root = get_irn_op(n); - if(op != get_irn_op(b)) + if (op != get_irn_op(b)) return n; /* and(conv(a), conv(b)) -> conv(and(a,b)) */ @@ -3291,7 +3383,7 @@ static ir_node *transform_bitwise_distributive(ir_node *n, ir_node *b_op = get_Conv_op(b); ir_mode *a_mode = get_irn_mode(a_op); ir_mode *b_mode = get_irn_mode(b_op); - if(a_mode == b_mode && (mode_is_int(a_mode) || a_mode == mode_b)) { + if (a_mode == b_mode && (mode_is_int(a_mode) || a_mode == mode_b)) { ir_node *blk = get_nodes_block(n); n = exact_copy(n); @@ -3326,17 +3418,17 @@ static ir_node *transform_bitwise_distributive(ir_node *n, c = a_left; op1 = a_right; op2 = b_right; - } else if(a_left == b_right) { + } else if (a_left == b_right) { c = a_left; op1 = a_right; op2 = b_left; - } else if(a_right == b_left) { + } else if (a_right == b_left) { c = a_right; op1 = a_left; op2 = b_right; } } - if(a_right == b_right) { + if (a_right == b_right) { c = a_right; op1 = a_left; op2 = b_left; @@ -3351,7 +3443,7 @@ static ir_node *transform_bitwise_distributive(ir_node *n, set_binop_right(new_n, op2); new_n = trans_func(new_n); - if(op_root == op_Eor && op == op_Or) { + if (op_root == op_Eor && op == op_Or) { dbg_info *dbgi = get_irn_dbg_info(n); ir_mode *mode = get_irn_mode(c); @@ -3376,11 +3468,13 @@ static ir_node *transform_bitwise_distributive(ir_node *n, /** * Transform an And. */ -static ir_node *transform_node_And(ir_node *n) { +static ir_node *transform_node_And(ir_node *n) +{ ir_node *c, *oldn = n; ir_node *a = get_And_left(n); ir_node *b = get_And_right(n); ir_mode *mode; + vrp_attr *a_vrp, *b_vrp; mode = get_irn_mode(n); HANDLE_BINOP_PHI((eval_func) tarval_and, a, b, c, mode); @@ -3391,13 +3485,12 @@ static ir_node *transform_node_And(ir_node *n) { ir_node *pred_b = get_Proj_pred(b); if (pred_a == pred_b) { dbg_info *dbgi = get_irn_dbg_info(n); - ir_node *block = get_nodes_block(pred_a); pn_Cmp pn_a = get_Proj_proj(a); pn_Cmp pn_b = get_Proj_proj(b); /* yes, we can simply calculate with pncs */ pn_Cmp new_pnc = pn_a & pn_b; - return new_rd_Proj(dbgi, block, pred_a, mode_b, new_pnc); + return new_rd_Proj(dbgi, pred_a, mode_b, new_pnc); } } if (is_Or(a)) { @@ -3501,6 +3594,20 @@ static ir_node *transform_node_And(ir_node *n) { return n; } + b_vrp = vrp_get_info(b); + if (is_Const(a) && b_vrp && (tarval_cmp(tarval_or(get_Const_tarval(a), + b_vrp->bits_not_set), get_Const_tarval(a)) == pn_Cmp_Eq)) { + + return b; + + } + + a_vrp = vrp_get_info(a); + if (is_Const(b) && a_vrp && (tarval_cmp(tarval_or(get_Const_tarval(b), + a_vrp->bits_not_set), get_Const_tarval(b)) == pn_Cmp_Eq)) { + return a; + } + n = transform_bitwise_distributive(n, transform_node_And); return n; @@ -3509,7 +3616,8 @@ static ir_node *transform_node_And(ir_node *n) { /** * Transform an Eor. */ -static ir_node *transform_node_Eor(ir_node *n) { +static ir_node *transform_node_Eor(ir_node *n) +{ ir_node *c, *oldn = n; ir_node *a = get_Eor_left(n); ir_node *b = get_Eor_right(n); @@ -3521,15 +3629,14 @@ static ir_node *transform_node_Eor(ir_node *n) { if (mode == mode_b && is_Proj(a) && is_Proj(b)) { ir_node *pred_a = get_Proj_pred(a); ir_node *pred_b = get_Proj_pred(b); - if(pred_a == pred_b) { + if (pred_a == pred_b) { dbg_info *dbgi = get_irn_dbg_info(n); - ir_node *block = get_nodes_block(pred_a); pn_Cmp pn_a = get_Proj_proj(a); pn_Cmp pn_b = get_Proj_proj(b); /* yes, we can simply calculate with pncs */ pn_Cmp new_pnc = pn_a ^ pn_b; - return new_rd_Proj(dbgi, block, pred_a, mode_b, new_pnc); + return new_rd_Proj(dbgi, pred_a, mode_b, new_pnc); } } @@ -3561,7 +3668,8 @@ static ir_node *transform_node_Eor(ir_node *n) { /** * Transform a Not. */ -static ir_node *transform_node_Not(ir_node *n) { +static ir_node *transform_node_Not(ir_node *n) +{ ir_node *c, *oldn = n; ir_node *a = get_Not_op(n); ir_mode *mode = get_irn_mode(n); @@ -3572,15 +3680,14 @@ static ir_node *transform_node_Not(ir_node *n) { if (mode == mode_b && is_Proj(a)) { ir_node *a_pred = get_Proj_pred(a); if (is_Cmp(a_pred)) { - ir_node *cmp_block = get_nodes_block(a_pred); /* We negate a Cmp. The Cmp has the negated result anyways! */ - n = new_r_Proj(cmp_block, get_Proj_pred(a), + n = new_r_Proj(get_Proj_pred(a), mode_b, get_negated_pnc(get_Proj_proj(a), mode_b)); DBG_OPT_ALGSIM0(oldn, n, FS_OPT_NOT_CMP); return n; } } - if (is_Eor(a)) { + if (is_Eor(a)) { ir_node *eor_b = get_Eor_right(a); if (is_Const(eor_b)) { /* ~(x ^ const) -> x ^ ~const */ ir_node *cnst = new_Const(tarval_not(get_Const_tarval(eor_b))); @@ -3623,7 +3730,8 @@ static ir_node *transform_node_Not(ir_node *n) { * -(a >>s (size-1)) = a >>u (size-1) * -(a * const) -> a * -const */ -static ir_node *transform_node_Minus(ir_node *n) { +static ir_node *transform_node_Minus(ir_node *n) +{ ir_node *c, *oldn = n; ir_node *a = get_Minus_op(n); ir_mode *mode; @@ -3710,7 +3818,8 @@ static ir_node *transform_node_Minus(ir_node *n) { /** * Transform a Cast_type(Const) into a new Const_type */ -static ir_node *transform_node_Cast(ir_node *n) { +static ir_node *transform_node_Cast(ir_node *n) +{ ir_node *oldn = n; ir_node *pred = get_Cast_op(n); ir_type *tp = get_irn_type(n); @@ -3730,7 +3839,8 @@ static ir_node *transform_node_Cast(ir_node *n) { /** * Transform a Proj(Load) with a non-null address. */ -static ir_node *transform_node_Proj_Load(ir_node *proj) { +static ir_node *transform_node_Proj_Load(ir_node *proj) +{ if (get_opt_ldst_only_null_ptr_exceptions()) { if (get_irn_mode(proj) == mode_X) { ir_node *load = get_Proj_pred(proj); @@ -3760,7 +3870,8 @@ static ir_node *transform_node_Proj_Load(ir_node *proj) { /** * Transform a Proj(Store) with a non-null address. */ -static ir_node *transform_node_Proj_Store(ir_node *proj) { +static ir_node *transform_node_Proj_Store(ir_node *proj) +{ if (get_opt_ldst_only_null_ptr_exceptions()) { if (get_irn_mode(proj) == mode_X) { ir_node *store = get_Proj_pred(proj); @@ -3791,7 +3902,8 @@ static ir_node *transform_node_Proj_Store(ir_node *proj) { * Transform a Proj(Div) with a non-zero value. * Removes the exceptions and routes the memory to the NoMem node. */ -static ir_node *transform_node_Proj_Div(ir_node *proj) { +static ir_node *transform_node_Proj_Div(ir_node *proj) +{ ir_node *div = get_Proj_pred(proj); ir_node *b = get_Div_right(div); ir_node *res, *new_mem; @@ -3839,7 +3951,8 @@ static ir_node *transform_node_Proj_Div(ir_node *proj) { * Transform a Proj(Mod) with a non-zero value. * Removes the exceptions and routes the memory to the NoMem node. */ -static ir_node *transform_node_Proj_Mod(ir_node *proj) { +static ir_node *transform_node_Proj_Mod(ir_node *proj) +{ ir_node *mod = get_Proj_pred(proj); ir_node *b = get_Mod_right(mod); ir_node *res, *new_mem; @@ -3897,7 +4010,8 @@ static ir_node *transform_node_Proj_Mod(ir_node *proj) { * Transform a Proj(DivMod) with a non-zero value. * Removes the exceptions and routes the memory to the NoMem node. */ -static ir_node *transform_node_Proj_DivMod(ir_node *proj) { +static ir_node *transform_node_Proj_DivMod(ir_node *proj) +{ ir_node *divmod = get_Proj_pred(proj); ir_node *b = get_DivMod_right(divmod); ir_node *res, *new_mem; @@ -3955,7 +4069,8 @@ static ir_node *transform_node_Proj_DivMod(ir_node *proj) { /** * Optimizes jump tables (CondIs or CondIu) by removing all impossible cases. */ -static ir_node *transform_node_Proj_Cond(ir_node *proj) { +static ir_node *transform_node_Proj_Cond(ir_node *proj) +{ if (get_opt_unreachable_code()) { ir_node *n = get_Proj_pred(proj); ir_node *b = get_Cond_selector(n); @@ -3976,6 +4091,51 @@ static ir_node *transform_node_Proj_Cond(ir_node *proj) { /* this case will NEVER be taken, kill it */ return get_irg_bad(current_ir_graph); } + } + } else { + long num = get_Proj_proj(proj); + vrp_attr *b_vrp = vrp_get_info(b); + if (num != get_Cond_default_proj(n) && b_vrp) { + /* Try handling with vrp data. We only remove dead parts. */ + tarval *tp = new_tarval_from_long(num, get_irn_mode(b)); + + if (b_vrp->range_type == VRP_RANGE) { + pn_Cmp cmp_result = tarval_cmp(b_vrp->range_bottom, tp); + pn_Cmp cmp_result2 = tarval_cmp(b_vrp->range_top, tp); + + if ((cmp_result & pn_Cmp_Gt) == cmp_result && (cmp_result2 + & pn_Cmp_Lt) == cmp_result2) { + return get_irg_bad(current_ir_graph); + } + } else if (b_vrp->range_type == VRP_ANTIRANGE) { + pn_Cmp cmp_result = tarval_cmp(b_vrp->range_bottom, tp); + pn_Cmp cmp_result2 = tarval_cmp(b_vrp->range_top, tp); + + if ((cmp_result & pn_Cmp_Le) == cmp_result && (cmp_result2 + & pn_Cmp_Ge) == cmp_result2) { + return get_irg_bad(current_ir_graph); + } + } + + if (!(tarval_cmp( + tarval_and( b_vrp->bits_set, tp), + b_vrp->bits_set + ) == pn_Cmp_Eq)) { + + return get_irg_bad(current_ir_graph); + } + + if (!(tarval_cmp( + tarval_and( + tarval_not(tp), + tarval_not(b_vrp->bits_not_set)), + tarval_not(b_vrp->bits_not_set)) + == pn_Cmp_Eq)) { + + return get_irg_bad(current_ir_graph); + } + + } } } @@ -3986,7 +4146,8 @@ static ir_node *transform_node_Proj_Cond(ir_node *proj) { /** * Create a 0 constant of given mode. */ -static ir_node *create_zero_const(ir_mode *mode) { +static ir_node *create_zero_const(ir_mode *mode) +{ tarval *tv = get_mode_null(mode); ir_node *cnst = new_Const(tv); @@ -4023,7 +4184,8 @@ static int operands_are_normalized(const ir_node *l, const ir_node *r) /** * Normalizes and optimizes Cmp nodes. */ -static ir_node *transform_node_Proj_Cmp(ir_node *proj) { +static ir_node *transform_node_Proj_Cmp(ir_node *proj) +{ ir_node *n = get_Proj_pred(proj); ir_node *left = get_Cmp_left(n); ir_node *right = get_Cmp_right(n); @@ -4282,10 +4444,13 @@ static ir_node *transform_node_Proj_Cmp(ir_node *proj) { /* * UpConv(x) REL 0 ==> x REL 0 + * Don't do this for float values as it's unclear whether it is a + * win. (on the other side it makes detection/creation of fabs hard) */ if (get_mode_size_bits(mode) > get_mode_size_bits(op_mode) && ((proj_nr == pn_Cmp_Eq || proj_nr == pn_Cmp_Lg) || - mode_is_signed(mode) || !mode_is_signed(op_mode))) { + mode_is_signed(mode) || !mode_is_signed(op_mode)) && + !mode_is_float(mode)) { tv = get_mode_null(op_mode); left = op; mode = op_mode; @@ -4630,7 +4795,7 @@ static ir_node *transform_node_Proj_Cmp(ir_node *proj) { /* create a new compare */ n = new_rd_Cmp(get_irn_dbg_info(n), block, left, right); - proj = new_rd_Proj(get_irn_dbg_info(proj), block, n, get_irn_mode(proj), proj_nr); + proj = new_rd_Proj(get_irn_dbg_info(proj), n, get_irn_mode(proj), proj_nr); } return proj; @@ -4639,7 +4804,8 @@ static ir_node *transform_node_Proj_Cmp(ir_node *proj) { /** * Optimize CopyB(mem, x, x) into a Nop. */ -static ir_node *transform_node_Proj_CopyB(ir_node *proj) { +static ir_node *transform_node_Proj_CopyB(ir_node *proj) +{ ir_node *copyb = get_Proj_pred(proj); ir_node *a = get_CopyB_dst(copyb); ir_node *b = get_CopyB_src(copyb); @@ -4665,7 +4831,8 @@ static ir_node *transform_node_Proj_CopyB(ir_node *proj) { /** * Optimize Bounds(idx, idx, upper) into idx. */ -static ir_node *transform_node_Proj_Bound(ir_node *proj) { +static ir_node *transform_node_Proj_Bound(ir_node *proj) +{ ir_node *oldn = proj; ir_node *bound = get_Proj_pred(proj); ir_node *idx = get_Bound_index(bound); @@ -4723,7 +4890,8 @@ static ir_node *transform_node_Proj_Bound(ir_node *proj) { * Does all optimizations on nodes that must be done on it's Proj's * because of creating new nodes. */ -static ir_node *transform_node_Proj(ir_node *proj) { +static ir_node *transform_node_Proj(ir_node *proj) +{ ir_node *n = get_Proj_pred(proj); if (n->op->ops.transform_node_Proj) @@ -4734,7 +4902,8 @@ static ir_node *transform_node_Proj(ir_node *proj) { /** * Move Confirms down through Phi nodes. */ -static ir_node *transform_node_Phi(ir_node *phi) { +static ir_node *transform_node_Phi(ir_node *phi) +{ int i, n; ir_mode *mode = get_irn_mode(phi); @@ -4778,7 +4947,8 @@ static ir_node *transform_node_Phi(ir_node *phi) { * Returns the operands of a commutative bin-op, if one operand is * a const, it is returned as the second one. */ -static void get_comm_Binop_Ops(ir_node *binop, ir_node **a, ir_node **c) { +static void get_comm_Binop_Ops(ir_node *binop, ir_node **a, ir_node **c) +{ ir_node *op_a = get_binop_left(binop); ir_node *op_b = get_binop_right(binop); @@ -4808,7 +4978,8 @@ static void get_comm_Binop_Ops(ir_node *binop, ir_node **a, ir_node **c) { * AND c1 ===> OR if (c1 | c2) == 0x111..11 * OR */ -static ir_node *transform_node_Or_bf_store(ir_node *or) { +static ir_node *transform_node_Or_bf_store(ir_node *or) +{ ir_node *and, *c1; ir_node *or_l, *c2; ir_node *and_l, *c3; @@ -4816,9 +4987,9 @@ static ir_node *transform_node_Or_bf_store(ir_node *or) { ir_node *new_and, *new_const, *block; ir_mode *mode = get_irn_mode(or); - tarval *tv1, *tv2, *tv3, *tv4, *tv, *n_tv4, *n_tv2; + tarval *tv1, *tv2, *tv3, *tv4, *tv; - while (1) { + for (;;) { get_comm_Binop_Ops(or, &and, &c1); if (!is_Const(c1) || !is_And(and)) return or; @@ -4865,14 +5036,12 @@ static ir_node *transform_node_Or_bf_store(ir_node *or) { return or; } - n_tv4 = tarval_not(tv4); - if (tv3 != tarval_and(tv3, n_tv4)) { + if (tv3 != tarval_andnot(tv3, tv4)) { /* bit in the or_mask is outside the and_mask */ return or; } - n_tv2 = tarval_not(tv2); - if (tv1 != tarval_and(tv1, n_tv2)) { + if (tv1 != tarval_andnot(tv1, tv2)) { /* bit in the or_mask is outside the and_mask */ return or; } @@ -4894,7 +5063,8 @@ static ir_node *transform_node_Or_bf_store(ir_node *or) { /** * Optimize an Or(shl(x, c), shr(x, bits - c)) into a Rotl */ -static ir_node *transform_node_Or_Rotl(ir_node *or) { +static ir_node *transform_node_Or_Rotl(ir_node *or) +{ ir_mode *mode = get_irn_mode(or); ir_node *shl, *shr, *block; ir_node *irn, *x, *c1, *c2, *v, *sub, *n, *rotval; @@ -4982,7 +5152,8 @@ static ir_node *transform_node_Or_Rotl(ir_node *or) { /** * Transform an Or. */ -static ir_node *transform_node_Or(ir_node *n) { +static ir_node *transform_node_Or(ir_node *n) +{ ir_node *c, *oldn = n; ir_node *a = get_Or_left(n); ir_node *b = get_Or_right(n); @@ -5007,13 +5178,12 @@ static ir_node *transform_node_Or(ir_node *n) { ir_node *pred_b = get_Proj_pred(b); if (pred_a == pred_b) { dbg_info *dbgi = get_irn_dbg_info(n); - ir_node *block = get_nodes_block(pred_a); pn_Cmp pn_a = get_Proj_proj(a); pn_Cmp pn_b = get_Proj_proj(b); /* yes, we can simply calculate with pncs */ pn_Cmp new_pnc = pn_a | pn_b; - return new_rd_Proj(dbgi, block, pred_a, mode_b, new_pnc); + return new_rd_Proj(dbgi, pred_a, mode_b, new_pnc); } } @@ -5039,7 +5209,8 @@ static ir_node *transform_node(ir_node *n); * * Should be moved to reassociation? */ -static ir_node *transform_node_shift(ir_node *n) { +static ir_node *transform_node_shift(ir_node *n) +{ ir_node *left, *right; ir_mode *mode; tarval *tv1, *tv2, *res; @@ -5109,7 +5280,8 @@ static ir_node *transform_node_shift(ir_node *n) { * - Shl, Shr, Shrs, rotl instead of >> * (with a special case for Or/Xor + Shrs) */ -static ir_node *transform_node_bitop_shift(ir_node *n) { +static ir_node *transform_node_bitop_shift(ir_node *n) +{ ir_node *left; ir_node *right = get_binop_right(n); ir_mode *mode = get_irn_mode(n); @@ -5157,10 +5329,10 @@ static ir_node *transform_node_bitop_shift(ir_node *n) { if (is_Shl(n)) { new_shift = new_rd_Shl(dbgi, block, bitop_left, right, mode); tv_shift = tarval_shl(tv1, tv2); - } else if(is_Shr(n)) { + } else if (is_Shr(n)) { new_shift = new_rd_Shr(dbgi, block, bitop_left, right, mode); tv_shift = tarval_shr(tv1, tv2); - } else if(is_Shrs(n)) { + } else if (is_Shrs(n)) { new_shift = new_rd_Shrs(dbgi, block, bitop_left, right, mode); tv_shift = tarval_shrs(tv1, tv2); } else { @@ -5174,7 +5346,7 @@ static ir_node *transform_node_bitop_shift(ir_node *n) { if (op_left == op_And) { new_bitop = new_rd_And(dbgi, block, new_shift, new_const, mode); - } else if(op_left == op_Or) { + } else if (op_left == op_Or) { new_bitop = new_rd_Or(dbgi, block, new_shift, new_const, mode); } else { assert(op_left == op_Eor); @@ -5191,11 +5363,11 @@ static ir_node *transform_node_bitop_shift(ir_node *n) { * (x >> c1) << c2 <=> x OP (c2-c1) & ((-1 >> c1) << c2) * (also with x >>s c1 when c1>=c2) */ -static ir_node *transform_node_shl_shr(ir_node *n) { +static ir_node *transform_node_shl_shr(ir_node *n) +{ ir_node *left; ir_node *right = get_binop_right(n); ir_node *x; - ir_graph *irg; ir_node *block; ir_mode *mode; dbg_info *dbgi; @@ -5237,7 +5409,7 @@ static ir_node *transform_node_shl_shr(ir_node *n) { tv_mask = tarval_shr(get_mode_all_one(mode), tv_shr); } tv_mask = tarval_shl(tv_mask, tv_shl); - } else if(is_Shr(n) && is_Shl(left)) { + } else if (is_Shr(n) && is_Shl(left)) { ir_node *shl_right = get_Shl_right(left); if (!is_Const(shl_right)) @@ -5260,7 +5432,6 @@ static ir_node *transform_node_shl_shr(ir_node *n) { assert(tv_mask != tarval_bad); assert(get_tarval_mode(tv_mask) == mode); - irg = get_irn_irg(n); block = get_nodes_block(n); dbgi = get_irn_dbg_info(n); @@ -5289,7 +5460,8 @@ static ir_node *transform_node_shl_shr(ir_node *n) { /** * Transform a Shr. */ -static ir_node *transform_node_Shr(ir_node *n) { +static ir_node *transform_node_Shr(ir_node *n) +{ ir_node *c, *oldn = n; ir_node *left = get_Shr_left(n); ir_node *right = get_Shr_right(n); @@ -5309,7 +5481,8 @@ static ir_node *transform_node_Shr(ir_node *n) { /** * Transform a Shrs. */ -static ir_node *transform_node_Shrs(ir_node *n) { +static ir_node *transform_node_Shrs(ir_node *n) +{ ir_node *c, *oldn = n; ir_node *a = get_Shrs_left(n); ir_node *b = get_Shrs_right(n); @@ -5327,7 +5500,8 @@ static ir_node *transform_node_Shrs(ir_node *n) { /** * Transform a Shl. */ -static ir_node *transform_node_Shl(ir_node *n) { +static ir_node *transform_node_Shl(ir_node *n) +{ ir_node *c, *oldn = n; ir_node *a = get_Shl_left(n); ir_node *b = get_Shl_right(n); @@ -5347,7 +5521,8 @@ static ir_node *transform_node_Shl(ir_node *n) { /** * Transform a Rotl. */ -static ir_node *transform_node_Rotl(ir_node *n) { +static ir_node *transform_node_Rotl(ir_node *n) +{ ir_node *c, *oldn = n; ir_node *a = get_Rotl_left(n); ir_node *b = get_Rotl_right(n); @@ -5365,7 +5540,8 @@ static ir_node *transform_node_Rotl(ir_node *n) { /** * Transform a Conv. */ -static ir_node *transform_node_Conv(ir_node *n) { +static ir_node *transform_node_Conv(ir_node *n) +{ ir_node *c, *oldn = n; ir_mode *mode = get_irn_mode(n); ir_node *a = get_Conv_op(n); @@ -5393,17 +5569,17 @@ static ir_node *transform_node_Conv(ir_node *n) { ir_node *r = get_Add_right(a); dbg_info *dbgi = get_irn_dbg_info(a); ir_node *block = get_nodes_block(n); - if(is_Conv(l)) { + if (is_Conv(l)) { ir_node *lop = get_Conv_op(l); - if(get_irn_mode(lop) == mode) { + if (get_irn_mode(lop) == mode) { /* ConvP(AddI(ConvI(P), x)) -> AddP(P, x) */ n = new_rd_Add(dbgi, block, lop, r, mode); return n; } } - if(is_Conv(r)) { + if (is_Conv(r)) { ir_node *rop = get_Conv_op(r); - if(get_irn_mode(rop) == mode) { + if (get_irn_mode(rop) == mode) { /* ConvP(AddI(x, ConvI(P))) -> AddP(x, P) */ n = new_rd_Add(dbgi, block, l, rop, mode); return n; @@ -5418,7 +5594,8 @@ static ir_node *transform_node_Conv(ir_node *n) { * Remove dead blocks and nodes in dead blocks * in keep alive list. We do not generate a new End node. */ -static ir_node *transform_node_End(ir_node *n) { +static ir_node *transform_node_End(ir_node *n) +{ int i, j, n_keepalives = get_End_n_keepalives(n); ir_node **in; @@ -5444,12 +5621,12 @@ static ir_node *transform_node_End(ir_node *n) { return n; } /* transform_node_End */ -/** returns 1 if a == -b */ -static int is_negated_value(ir_node *a, ir_node *b) { +bool is_negated_value(ir_node *a, ir_node *b) +{ if (is_Minus(a) && get_Minus_op(a) == b) - return 1; + return true; if (is_Minus(b) && get_Minus_op(b) == a) - return 1; + return true; if (is_Sub(a) && is_Sub(b)) { ir_node *a_left = get_Sub_left(a); ir_node *a_right = get_Sub_right(a); @@ -5457,16 +5634,17 @@ static int is_negated_value(ir_node *a, ir_node *b) { ir_node *b_right = get_Sub_right(b); if (a_left == b_right && a_right == b_left) - return 1; + return true; } - return 0; + return false; } /** * Optimize a Mux into some simpler cases. */ -static ir_node *transform_node_Mux(ir_node *n) { +static ir_node *transform_node_Mux(ir_node *n) +{ ir_node *oldn = n, *sel = get_Mux_sel(n); ir_mode *mode = get_irn_mode(n); ir_node *t = get_Mux_true(n); @@ -5540,7 +5718,7 @@ static ir_node *transform_node_Mux(ir_node *n) { /* Mux(x, 0, y) => Mux(x, y, 0) */ pn_Cmp pnc = get_Proj_proj(sel); - sel = new_r_Proj(get_nodes_block(cmp), cmp, mode_b, + sel = new_r_Proj(cmp, mode_b, get_negated_pnc(pnc, get_irn_mode(get_Cmp_left(cmp)))); n = new_rd_Mux(get_irn_dbg_info(n), get_nodes_block(n), sel, t, f, mode); tmp = t; @@ -5651,15 +5829,18 @@ static ir_node *transform_node_Mux(ir_node *n) { if (!mode_honor_signed_zeros(mode) && is_negated_value(f, t)) { /* f = -t */ - if ( (cmp_l == t && (pn == pn_Cmp_Ge || pn == pn_Cmp_Gt)) - || (cmp_l == f && (pn == pn_Cmp_Le || pn == pn_Cmp_Lt))) + /* NaN's work fine with abs, so it is ok to remove Uo */ + long pnc = pn & ~pn_Cmp_Uo; + + if ( (cmp_l == t && (pnc == pn_Cmp_Ge || pnc == pn_Cmp_Gt)) + || (cmp_l == f && (pnc == pn_Cmp_Le || pnc == pn_Cmp_Lt))) { /* Mux(a >/>= 0, a, -a) = Mux(a Abs(a) */ n = new_rd_Abs(get_irn_dbg_info(n), block, cmp_l, mode); DBG_OPT_ALGSIM1(oldn, cmp, sel, n, FS_OPT_MUX_TO_ABS); return n; - } else if ((cmp_l == t && (pn == pn_Cmp_Le || pn == pn_Cmp_Lt)) - || (cmp_l == f && (pn == pn_Cmp_Ge || pn == pn_Cmp_Gt))) + } else if ((cmp_l == t && (pnc == pn_Cmp_Le || pnc == pn_Cmp_Lt)) + || (cmp_l == f && (pnc == pn_Cmp_Ge || pnc == pn_Cmp_Gt))) { /* Mux(a />= 0, -a, a) ==> -Abs(a) */ n = new_rd_Abs(get_irn_dbg_info(n), block, cmp_l, mode); @@ -5741,7 +5922,8 @@ static ir_node *transform_node_Mux(ir_node *n) { * optimize Sync nodes that have other syncs as input we simply add the inputs * of the other sync to our own inputs */ -static ir_node *transform_node_Sync(ir_node *n) { +static ir_node *transform_node_Sync(ir_node *n) +{ int arity = get_Sync_n_preds(n); int i; @@ -5783,7 +5965,8 @@ static ir_node *transform_node_Sync(ir_node *n) { /** * optimize a trampoline Call into a direct Call */ -static ir_node *transform_node_Call(ir_node *call) { +static ir_node *transform_node_Call(ir_node *call) +{ ir_node *callee = get_Call_ptr(call); ir_node *adr, *mem, *res, *bl, **in; ir_type *ctp, *mtp, *tp; @@ -5856,7 +6039,8 @@ static ir_node *transform_node_Call(ir_node *call) { * transformations _do_ generate new nodes, and thus the old node must * not be freed even if the equivalent node isn't the old one. */ -static ir_node *transform_node(ir_node *n) { +static ir_node *transform_node(ir_node *n) +{ ir_node *oldn; /* @@ -5949,92 +6133,94 @@ static ir_op_ops *firm_set_default_transform_node(ir_opcode code, ir_op_ops *ops #define N_IR_NODES 512 /** Compares the attributes of two Const nodes. */ -static int node_cmp_attr_Const(ir_node *a, ir_node *b) { +static int node_cmp_attr_Const(ir_node *a, ir_node *b) +{ return (get_Const_tarval(a) != get_Const_tarval(b)) || (get_Const_type(a) != get_Const_type(b)); } /* node_cmp_attr_Const */ /** Compares the attributes of two Proj nodes. */ -static int node_cmp_attr_Proj(ir_node *a, ir_node *b) { - return get_irn_proj_attr(a) != get_irn_proj_attr(b); +static int node_cmp_attr_Proj(ir_node *a, ir_node *b) +{ + return a->attr.proj != b->attr.proj; } /* node_cmp_attr_Proj */ /** Compares the attributes of two Filter nodes. */ -static int node_cmp_attr_Filter(ir_node *a, ir_node *b) { +static int node_cmp_attr_Filter(ir_node *a, ir_node *b) +{ return get_Filter_proj(a) != get_Filter_proj(b); } /* node_cmp_attr_Filter */ /** Compares the attributes of two Alloc nodes. */ -static int node_cmp_attr_Alloc(ir_node *a, ir_node *b) { - const alloc_attr *pa = get_irn_alloc_attr(a); - const alloc_attr *pb = get_irn_alloc_attr(b); +static int node_cmp_attr_Alloc(ir_node *a, ir_node *b) +{ + const alloc_attr *pa = &a->attr.alloc; + const alloc_attr *pb = &b->attr.alloc; return (pa->where != pb->where) || (pa->type != pb->type); } /* node_cmp_attr_Alloc */ /** Compares the attributes of two Free nodes. */ -static int node_cmp_attr_Free(ir_node *a, ir_node *b) { - const free_attr *pa = get_irn_free_attr(a); - const free_attr *pb = get_irn_free_attr(b); +static int node_cmp_attr_Free(ir_node *a, ir_node *b) +{ + const free_attr *pa = &a->attr.free; + const free_attr *pb = &b->attr.free; return (pa->where != pb->where) || (pa->type != pb->type); } /* node_cmp_attr_Free */ /** Compares the attributes of two SymConst nodes. */ -static int node_cmp_attr_SymConst(ir_node *a, ir_node *b) { - const symconst_attr *pa = get_irn_symconst_attr(a); - const symconst_attr *pb = get_irn_symconst_attr(b); +static int node_cmp_attr_SymConst(ir_node *a, ir_node *b) +{ + const symconst_attr *pa = &a->attr.symc; + const symconst_attr *pb = &b->attr.symc; return (pa->kind != pb->kind) || (pa->sym.type_p != pb->sym.type_p) || (pa->tp != pb->tp); } /* node_cmp_attr_SymConst */ /** Compares the attributes of two Call nodes. */ -static int node_cmp_attr_Call(ir_node *a, ir_node *b) { - const call_attr *pa = get_irn_call_attr(a); - const call_attr *pb = get_irn_call_attr(b); +static int node_cmp_attr_Call(ir_node *a, ir_node *b) +{ + const call_attr *pa = &a->attr.call; + const call_attr *pb = &b->attr.call; return (pa->type != pb->type) || (pa->tail_call != pb->tail_call); } /* node_cmp_attr_Call */ /** Compares the attributes of two Sel nodes. */ -static int node_cmp_attr_Sel(ir_node *a, ir_node *b) { +static int node_cmp_attr_Sel(ir_node *a, ir_node *b) +{ const ir_entity *a_ent = get_Sel_entity(a); const ir_entity *b_ent = get_Sel_entity(b); -#if 0 - return - (a_ent->kind != b_ent->kind) || - (a_ent->name != b_ent->name) || - (a_ent->owner != b_ent->owner) || - (a_ent->ld_name != b_ent->ld_name) || - (a_ent->type != b_ent->type); -#endif - /* Matze: inlining of functions can produce 2 entities with same type, - * name, etc. */ return a_ent != b_ent; } /* node_cmp_attr_Sel */ /** Compares the attributes of two Phi nodes. */ -static int node_cmp_attr_Phi(ir_node *a, ir_node *b) { +static int node_cmp_attr_Phi(ir_node *a, ir_node *b) +{ /* we can only enter this function if both nodes have the same number of inputs, hence it is enough to check if one of them is a Phi0 */ if (is_Phi0(a)) { /* check the Phi0 pos attribute */ - return get_irn_phi_attr(a)->u.pos != get_irn_phi_attr(b)->u.pos; + return a->attr.phi.u.pos != b->attr.phi.u.pos; } return 0; } /* node_cmp_attr_Phi */ /** Compares the attributes of two Conv nodes. */ -static int node_cmp_attr_Conv(ir_node *a, ir_node *b) { +static int node_cmp_attr_Conv(ir_node *a, ir_node *b) +{ return get_Conv_strict(a) != get_Conv_strict(b); } /* node_cmp_attr_Conv */ /** Compares the attributes of two Cast nodes. */ -static int node_cmp_attr_Cast(ir_node *a, ir_node *b) { +static int node_cmp_attr_Cast(ir_node *a, ir_node *b) +{ return get_Cast_type(a) != get_Cast_type(b); } /* node_cmp_attr_Cast */ /** Compares the attributes of two Load nodes. */ -static int node_cmp_attr_Load(ir_node *a, ir_node *b) { +static int node_cmp_attr_Load(ir_node *a, ir_node *b) +{ if (get_Load_volatility(a) == volatility_is_volatile || get_Load_volatility(b) == volatility_is_volatile) /* NEVER do CSE on volatile Loads */ @@ -6047,7 +6233,8 @@ static int node_cmp_attr_Load(ir_node *a, ir_node *b) { } /* node_cmp_attr_Load */ /** Compares the attributes of two Store nodes. */ -static int node_cmp_attr_Store(ir_node *a, ir_node *b) { +static int node_cmp_attr_Store(ir_node *a, ir_node *b) +{ /* do not CSE Stores with different alignment. Be conservative. */ if (get_Store_align(a) != get_Store_align(b)) return 1; @@ -6058,9 +6245,10 @@ static int node_cmp_attr_Store(ir_node *a, ir_node *b) { } /* node_cmp_attr_Store */ /** Compares two exception attributes */ -static int node_cmp_exception(ir_node *a, ir_node *b) { - const except_attr *ea = get_irn_except_attr(a); - const except_attr *eb = get_irn_except_attr(b); +static int node_cmp_exception(ir_node *a, ir_node *b) +{ + const except_attr *ea = &a->attr.except; + const except_attr *eb = &b->attr.except; return ea->pin_state != eb->pin_state; } @@ -6068,55 +6256,53 @@ static int node_cmp_exception(ir_node *a, ir_node *b) { #define node_cmp_attr_Bound node_cmp_exception /** Compares the attributes of two Div nodes. */ -static int node_cmp_attr_Div(ir_node *a, ir_node *b) { - const divmod_attr *ma = get_irn_divmod_attr(a); - const divmod_attr *mb = get_irn_divmod_attr(b); +static int node_cmp_attr_Div(ir_node *a, ir_node *b) +{ + const divmod_attr *ma = &a->attr.divmod; + const divmod_attr *mb = &b->attr.divmod; return ma->exc.pin_state != mb->exc.pin_state || ma->resmode != mb->resmode || ma->no_remainder != mb->no_remainder; } /* node_cmp_attr_Div */ /** Compares the attributes of two DivMod nodes. */ -static int node_cmp_attr_DivMod(ir_node *a, ir_node *b) { - const divmod_attr *ma = get_irn_divmod_attr(a); - const divmod_attr *mb = get_irn_divmod_attr(b); +static int node_cmp_attr_DivMod(ir_node *a, ir_node *b) +{ + const divmod_attr *ma = &a->attr.divmod; + const divmod_attr *mb = &b->attr.divmod; return ma->exc.pin_state != mb->exc.pin_state || ma->resmode != mb->resmode; } /* node_cmp_attr_DivMod */ /** Compares the attributes of two Mod nodes. */ -static int node_cmp_attr_Mod(ir_node *a, ir_node *b) { - const divmod_attr *ma = get_irn_divmod_attr(a); - const divmod_attr *mb = get_irn_divmod_attr(b); - return ma->exc.pin_state != mb->exc.pin_state || - ma->resmode != mb->resmode; +static int node_cmp_attr_Mod(ir_node *a, ir_node *b) +{ + return node_cmp_attr_DivMod(a, b); } /* node_cmp_attr_Mod */ /** Compares the attributes of two Quot nodes. */ -static int node_cmp_attr_Quot(ir_node *a, ir_node *b) { - const divmod_attr *ma = get_irn_divmod_attr(a); - const divmod_attr *mb = get_irn_divmod_attr(b); - return ma->exc.pin_state != mb->exc.pin_state || - ma->resmode != mb->resmode; +static int node_cmp_attr_Quot(ir_node *a, ir_node *b) +{ + return node_cmp_attr_DivMod(a, b); } /* node_cmp_attr_Quot */ /** Compares the attributes of two Confirm nodes. */ -static int node_cmp_attr_Confirm(ir_node *a, ir_node *b) { +static int node_cmp_attr_Confirm(ir_node *a, ir_node *b) +{ /* no need to compare the bound, as this is a input */ return (get_Confirm_cmp(a) != get_Confirm_cmp(b)); } /* node_cmp_attr_Confirm */ /** Compares the attributes of two Builtin nodes. */ -static int node_cmp_attr_Builtin(ir_node *a, ir_node *b) { - const builtin_attr *ma = get_irn_builtin_attr(a); - const builtin_attr *mb = get_irn_builtin_attr(b); - +static int node_cmp_attr_Builtin(ir_node *a, ir_node *b) +{ /* no need to compare the type, equal kind means equal type */ - return ma->kind != mb->kind; + return get_Builtin_kind(a) != get_Builtin_kind(b); } /* node_cmp_attr_Builtin */ /** Compares the attributes of two ASM nodes. */ -static int node_cmp_attr_ASM(ir_node *a, ir_node *b) { +static int node_cmp_attr_ASM(ir_node *a, ir_node *b) +{ int i, n; const ir_asm_constraint *ca; const ir_asm_constraint *cb; @@ -6210,7 +6396,8 @@ static ir_op_ops *firm_set_default_node_cmp_attr(ir_opcode code, ir_op_ops *ops) CASE(Dummy); /* FIXME CopyB */ default: - /* leave NULL */; + /* leave NULL */ + break; } return ops; @@ -6221,7 +6408,8 @@ static ir_op_ops *firm_set_default_node_cmp_attr(ir_opcode code, ir_op_ops *ops) * Compare function for two nodes in the value table. Gets two * nodes as parameters. Returns 0 if the nodes are a Common Sub Expression. */ -int identities_cmp(const void *elt, const void *key) { +int identities_cmp(const void *elt, const void *key) +{ ir_node *a = (ir_node *)elt; ir_node *b = (ir_node *)key; int i, irn_arity_a; @@ -6247,9 +6435,15 @@ int identities_cmp(const void *elt, const void *key) { } /* compare a->in[0..ins] with b->in[0..ins] */ - for (i = 0; i < irn_arity_a; i++) - if (get_irn_intra_n(a, i) != get_irn_intra_n(b, i)) - return 1; + for (i = 0; i < irn_arity_a; ++i) { + ir_node *pred_a = get_irn_intra_n(a, i); + ir_node *pred_b = get_irn_intra_n(b, i); + if (pred_a != pred_b) { + /* if both predecessors are CSE neutral they might be different */ + if (!is_irn_cse_neutral(pred_a) || !is_irn_cse_neutral(pred_b)) + return 1; + } + } /* * here, we already now that the nodes are identical except their @@ -6266,22 +6460,26 @@ int identities_cmp(const void *elt, const void *key) { * * @param node The IR-node */ -unsigned ir_node_hash(const ir_node *node) { +unsigned ir_node_hash(const ir_node *node) +{ return node->op->ops.hash(node); } /* ir_node_hash */ -pset *new_identities(void) { +pset *new_identities(void) +{ return new_pset(identities_cmp, N_IR_NODES); } /* new_identities */ -void del_identities(pset *value_table) { +void del_identities(pset *value_table) +{ del_pset(value_table); } /* del_identities */ /* Normalize a node by putting constants (and operands with larger * node index) on the right (operator side). */ -void ir_normalize_node(ir_node *n) { +void ir_normalize_node(ir_node *n) +{ if (is_op_commutative(get_irn_op(n))) { ir_node *l = get_binop_left(n); ir_node *r = get_binop_right(n); @@ -6307,7 +6505,8 @@ void ir_normalize_node(ir_node *n) { * dominance info here: We known, that one block must dominate the other and * following the only block input will allow to find it. */ -static void update_known_irn(ir_node *known_irn, const ir_node *new_ir_node) { +static void update_known_irn(ir_node *known_irn, const ir_node *new_ir_node) +{ ir_node *known_blk, *new_block, *block, *mbh; if (get_opt_global_cse()) { @@ -6357,7 +6556,8 @@ static void update_known_irn(ir_node *known_irn, const ir_node *new_ir_node) { * @return a node that computes the same value as n or n if no such * node could be found */ -ir_node *identify_remember(pset *value_table, ir_node *n) { +ir_node *identify_remember(pset *value_table, ir_node *n) +{ ir_node *nn = NULL; if (!value_table) return n; @@ -6384,7 +6584,8 @@ ir_node *identify_remember(pset *value_table, ir_node *n) { * @param value_table The value table * @param n The node to lookup */ -static inline ir_node *identify_cons(pset *value_table, ir_node *n) { +static inline ir_node *identify_cons(pset *value_table, ir_node *n) +{ ir_node *old = n; n = identify_remember(value_table, n); @@ -6394,13 +6595,15 @@ static inline ir_node *identify_cons(pset *value_table, ir_node *n) { } /* identify_cons */ /* Add a node to the identities value table. */ -void add_identities(pset *value_table, ir_node *node) { +void add_identities(pset *value_table, ir_node *node) +{ if (get_opt_cse() && is_no_Block(node)) identify_remember(value_table, node); } /* add_identities */ /* Visit each node in the value table of a graph. */ -void visit_all_identities(ir_graph *irg, irg_walk_func visit, void *env) { +void visit_all_identities(ir_graph *irg, irg_walk_func visit, void *env) +{ ir_node *node; ir_graph *rem = current_ir_graph; @@ -6414,7 +6617,8 @@ void visit_all_identities(ir_graph *irg, irg_walk_func visit, void *env) { * Garbage in, garbage out. If a node has a dead input, i.e., the * Bad node is input to the node, return the Bad node. */ -static ir_node *gigo(ir_node *node) { +static ir_node *gigo(ir_node *node) +{ int i, irn_arity; ir_op *op = get_irn_op(node); @@ -6510,7 +6714,8 @@ static ir_node *gigo(ir_node *node) { * * current_ir_graph must be set to the graph of the node! */ -ir_node *optimize_node(ir_node *n) { +ir_node *optimize_node(ir_node *n) +{ tarval *tv; ir_node *oldn = n; ir_opcode iro = get_irn_opcode(n); @@ -6626,7 +6831,8 @@ ir_node *optimize_node(ir_node *n) { * nodes lying on the obstack. Remove these by a dead node elimination, * i.e., a copying garbage collection. */ -ir_node *optimize_in_place_2(ir_node *n) { +ir_node *optimize_in_place_2(ir_node *n) +{ tarval *tv; ir_node *oldn = n; ir_opcode iro = get_irn_opcode(n); @@ -6717,7 +6923,8 @@ ir_node *optimize_in_place_2(ir_node *n) { /** * Wrapper for external use, set proper status bits after optimization. */ -ir_node *optimize_in_place(ir_node *n) { +ir_node *optimize_in_place(ir_node *n) +{ /* Handle graph state */ assert(get_irg_phase_state(current_ir_graph) != phase_building); @@ -6735,11 +6942,12 @@ ir_node *optimize_in_place(ir_node *n) { /** * Calculate a hash value of a Const node. */ -static unsigned hash_Const(const ir_node *node) { +static unsigned hash_Const(const ir_node *node) +{ unsigned h; /* special value for const, as they only differ in their tarval. */ - h = HASH_PTR(node->attr.con.tv); + h = HASH_PTR(node->attr.con.tarval); return h; } /* hash_Const */ @@ -6747,7 +6955,8 @@ static unsigned hash_Const(const ir_node *node) { /** * Calculate a hash value of a SymConst node. */ -static unsigned hash_SymConst(const ir_node *node) { +static unsigned hash_SymConst(const ir_node *node) +{ unsigned h; /* all others are pointers */ @@ -6791,7 +7000,8 @@ static ir_op_ops *firm_set_default_hash(ir_opcode code, ir_op_ops *ops) /* * Sets the default operation for an ir_ops. */ -ir_op_ops *firm_set_default_operations(ir_opcode code, ir_op_ops *ops) { +ir_op_ops *firm_set_default_operations(ir_opcode code, ir_op_ops *ops) +{ ops = firm_set_default_hash(code, ops); ops = firm_set_default_computed_value(code, ops); ops = firm_set_default_equivalent_node(code, ops);