X-Git-Url: http://nsz.repo.hu/git/?a=blobdiff_plain;f=ir%2Fir%2Firopt.c;h=0c7cf979ae8f470c81f26de68611afad451fc02c;hb=7719bf6a7bd442f4763731d56d856cb082156877;hp=110afbe01eb9dc7fde4ca436a4e790fa15731d99;hpb=46998ae95efac7c88ff94697e104e07de15aa13e;p=libfirm diff --git a/ir/ir/iropt.c b/ir/ir/iropt.c index 110afbe01..0c7cf979a 100644 --- a/ir/ir/iropt.c +++ b/ir/ir/iropt.c @@ -56,7 +56,8 @@ /** * Returns the tarval of a Const node or tarval_bad for all other nodes. */ -static tarval *default_value_of(const ir_node *n) { +static tarval *default_value_of(const ir_node *n) +{ if (is_Const(n)) return get_Const_tarval(n); /* might return tarval_bad */ else @@ -66,7 +67,8 @@ static tarval *default_value_of(const ir_node *n) { value_of_func value_of_ptr = default_value_of; /* * Set a new value_of function. */ -void set_value_of_func(value_of_func func) { +void set_value_of_func(value_of_func func) +{ if (func != NULL) value_of_ptr = func; else @@ -76,14 +78,16 @@ void set_value_of_func(value_of_func func) { /** * Return the value of a Constant. */ -static tarval *computed_value_Const(const ir_node *n) { +static tarval *computed_value_Const(const ir_node *n) +{ return get_Const_tarval(n); } /* computed_value_Const */ /** * Return the value of a 'sizeof', 'alignof' or 'offsetof' SymConst. */ -static tarval *computed_value_SymConst(const ir_node *n) { +static tarval *computed_value_SymConst(const ir_node *n) +{ ir_type *type; ir_entity *ent; @@ -113,7 +117,8 @@ static tarval *computed_value_SymConst(const ir_node *n) { /** * Return the value of an Add. */ -static tarval *computed_value_Add(const ir_node *n) { +static tarval *computed_value_Add(const ir_node *n) +{ ir_node *a = get_Add_left(n); ir_node *b = get_Add_right(n); @@ -130,7 +135,8 @@ static tarval *computed_value_Add(const ir_node *n) { * Return the value of a Sub. * Special case: a - a */ -static tarval *computed_value_Sub(const ir_node *n) { +static tarval *computed_value_Sub(const ir_node *n) +{ ir_mode *mode = get_irn_mode(n); ir_node *a = get_Sub_left(n); ir_node *b = get_Sub_right(n); @@ -157,7 +163,8 @@ static tarval *computed_value_Sub(const ir_node *n) { * Return the value of a Carry. * Special : a op 0, 0 op b */ -static tarval *computed_value_Carry(const ir_node *n) { +static tarval *computed_value_Carry(const ir_node *n) +{ ir_node *a = get_binop_left(n); ir_node *b = get_binop_right(n); ir_mode *m = get_irn_mode(n); @@ -179,7 +186,8 @@ static tarval *computed_value_Carry(const ir_node *n) { * Return the value of a Borrow. * Special : a op 0 */ -static tarval *computed_value_Borrow(const ir_node *n) { +static tarval *computed_value_Borrow(const ir_node *n) +{ ir_node *a = get_binop_left(n); ir_node *b = get_binop_right(n); ir_mode *m = get_irn_mode(n); @@ -198,7 +206,8 @@ static tarval *computed_value_Borrow(const ir_node *n) { /** * Return the value of an unary Minus. */ -static tarval *computed_value_Minus(const ir_node *n) { +static tarval *computed_value_Minus(const ir_node *n) +{ ir_node *a = get_Minus_op(n); tarval *ta = value_of(a); @@ -211,7 +220,8 @@ static tarval *computed_value_Minus(const ir_node *n) { /** * Return the value of a Mul. */ -static tarval *computed_value_Mul(const ir_node *n) { +static tarval *computed_value_Mul(const ir_node *n) +{ ir_node *a = get_Mul_left(n); ir_node *b = get_Mul_right(n); ir_mode *mode; @@ -244,7 +254,8 @@ static tarval *computed_value_Mul(const ir_node *n) { /** * Return the value of an Abs. */ -static tarval *computed_value_Abs(const ir_node *n) { +static tarval *computed_value_Abs(const ir_node *n) +{ ir_node *a = get_Abs_op(n); tarval *ta = value_of(a); @@ -258,7 +269,8 @@ static tarval *computed_value_Abs(const ir_node *n) { * Return the value of an And. * Special case: a & 0, 0 & b */ -static tarval *computed_value_And(const ir_node *n) { +static tarval *computed_value_And(const ir_node *n) +{ ir_node *a = get_And_left(n); ir_node *b = get_And_right(n); @@ -278,7 +290,8 @@ static tarval *computed_value_And(const ir_node *n) { * Return the value of an Or. * Special case: a | 1...1, 1...1 | b */ -static tarval *computed_value_Or(const ir_node *n) { +static tarval *computed_value_Or(const ir_node *n) +{ ir_node *a = get_Or_left(n); ir_node *b = get_Or_right(n); @@ -297,7 +310,8 @@ static tarval *computed_value_Or(const ir_node *n) { /** * Return the value of an Eor. */ -static tarval *computed_value_Eor(const ir_node *n) { +static tarval *computed_value_Eor(const ir_node *n) +{ ir_node *a = get_Eor_left(n); ir_node *b = get_Eor_right(n); @@ -318,7 +332,8 @@ static tarval *computed_value_Eor(const ir_node *n) { /** * Return the value of a Not. */ -static tarval *computed_value_Not(const ir_node *n) { +static tarval *computed_value_Not(const ir_node *n) +{ ir_node *a = get_Not_op(n); tarval *ta = value_of(a); @@ -331,7 +346,8 @@ static tarval *computed_value_Not(const ir_node *n) { /** * Return the value of a Shl. */ -static tarval *computed_value_Shl(const ir_node *n) { +static tarval *computed_value_Shl(const ir_node *n) +{ ir_node *a = get_Shl_left(n); ir_node *b = get_Shl_right(n); @@ -347,7 +363,8 @@ static tarval *computed_value_Shl(const ir_node *n) { /** * Return the value of a Shr. */ -static tarval *computed_value_Shr(const ir_node *n) { +static tarval *computed_value_Shr(const ir_node *n) +{ ir_node *a = get_Shr_left(n); ir_node *b = get_Shr_right(n); @@ -363,7 +380,8 @@ static tarval *computed_value_Shr(const ir_node *n) { /** * Return the value of a Shrs. */ -static tarval *computed_value_Shrs(const ir_node *n) { +static tarval *computed_value_Shrs(const ir_node *n) +{ ir_node *a = get_Shrs_left(n); ir_node *b = get_Shrs_right(n); @@ -379,7 +397,8 @@ static tarval *computed_value_Shrs(const ir_node *n) { /** * Return the value of a Rotl. */ -static tarval *computed_value_Rotl(const ir_node *n) { +static tarval *computed_value_Rotl(const ir_node *n) +{ ir_node *a = get_Rotl_left(n); ir_node *b = get_Rotl_right(n); @@ -395,7 +414,8 @@ static tarval *computed_value_Rotl(const ir_node *n) { /** * Return the value of a Conv. */ -static tarval *computed_value_Conv(const ir_node *n) { +static tarval *computed_value_Conv(const ir_node *n) +{ ir_node *a = get_Conv_op(n); tarval *ta = value_of(a); @@ -409,7 +429,8 @@ static tarval *computed_value_Conv(const ir_node *n) { * Calculate the value of a Mux: can be evaluated, if the * sel and the right input are known. */ -static tarval *computed_value_Mux(const ir_node *n) { +static tarval *computed_value_Mux(const ir_node *n) +{ ir_node *sel = get_Mux_sel(n); tarval *ts = value_of(sel); @@ -428,7 +449,8 @@ static tarval *computed_value_Mux(const ir_node *n) { * Calculate the value of a Confirm: can be evaluated, * if it has the form Confirm(x, '=', Const). */ -static tarval *computed_value_Confirm(const ir_node *n) { +static tarval *computed_value_Confirm(const ir_node *n) +{ /* * Beware: we might produce Phi(Confirm(x == true), Confirm(x == false)). * Do NOT optimize them away (jump threading wants them), so wait until @@ -453,116 +475,102 @@ static tarval *computed_value_Confirm(const ir_node *n) { * only 1 is used. * There are several case where we can evaluate a Cmp node, see later. */ -static tarval *computed_value_Proj_Cmp(const ir_node *n) { - ir_node *a = get_Proj_pred(n); - ir_node *aa = get_Cmp_left(a); - ir_node *ab = get_Cmp_right(a); - long proj_nr = get_Proj_proj(n); +static tarval *computed_value_Proj_Cmp(const ir_node *n) +{ + ir_node *cmp = get_Proj_pred(n); + ir_node *left = get_Cmp_left(cmp); + ir_node *right = get_Cmp_right(cmp); + long pn_cmp = get_Proj_proj(n); + ir_mode *mode = get_irn_mode(left); + tarval *tv_l, *tv_r; /* * BEWARE: a == a is NOT always True for floating Point values, as * NaN != NaN is defined, so we must check this here. */ - if (aa == ab && ( - !mode_is_float(get_irn_mode(aa)) || proj_nr == pn_Cmp_Lt || proj_nr == pn_Cmp_Gt) - ) { /* 1.: */ - + if (left == right && (!mode_is_float(mode) || pn_cmp == pn_Cmp_Lt || pn_cmp == pn_Cmp_Gt)) { /* This is a trick with the bits used for encoding the Cmp Proj numbers, the following statement is not the same: - return new_tarval_from_long (proj_nr == pn_Cmp_Eq, mode_b) */ - return new_tarval_from_long (proj_nr & pn_Cmp_Eq, mode_b); + return new_tarval_from_long(pn_cmp == pn_Cmp_Eq, mode_b) */ + return new_tarval_from_long(pn_cmp & pn_Cmp_Eq, mode_b); } - tarval *taa = value_of(aa); - tarval *tab = value_of(ab); - ir_mode *mode = get_irn_mode(aa); + tv_l = value_of(left); + tv_r = value_of(right); - /* - * The predecessors of Cmp are target values. We can evaluate - * the Cmp. - */ - if ((taa != tarval_bad) && (tab != tarval_bad)) { - /* strange checks... */ - pn_Cmp flags = tarval_cmp(taa, tab); + if ((tv_l != tarval_bad) && (tv_r != tarval_bad)) { + /* + * The predecessors of Cmp are target values. We can evaluate + * the Cmp. + */ + pn_Cmp flags = tarval_cmp(tv_l, tv_r); if (flags != pn_Cmp_False) { - return new_tarval_from_long (proj_nr & flags, mode_b); - } - } - /* for integer values, we can check against MIN/MAX */ - else if (mode_is_int(mode)) { - /* MIN <=/> x. This results in true/false. */ - if (taa == get_mode_min(mode)) { - /* a compare with the MIN value */ - if (proj_nr == pn_Cmp_Le) - return get_tarval_b_true(); - else if (proj_nr == pn_Cmp_Gt) - return get_tarval_b_false(); - } - /* x >=/< MIN. This results in true/false. */ - else - if (tab == get_mode_min(mode)) { - /* a compare with the MIN value */ - if (proj_nr == pn_Cmp_Ge) - return get_tarval_b_true(); - else if (proj_nr == pn_Cmp_Lt) - return get_tarval_b_false(); - } - /* MAX >=/< x. This results in true/false. */ - else if (taa == get_mode_max(mode)) { - if (proj_nr == pn_Cmp_Ge) - return get_tarval_b_true(); - else if (proj_nr == pn_Cmp_Lt) - return get_tarval_b_false(); - } - /* x <=/> MAX. This results in true/false. */ - else if (tab == get_mode_max(mode)) { - if (proj_nr == pn_Cmp_Le) - return get_tarval_b_true(); - else if (proj_nr == pn_Cmp_Gt) - return get_tarval_b_false(); + return new_tarval_from_long (pn_cmp & flags, mode_b); + } + } else if (mode_is_int(mode)) { + /* for integer values, we can check against MIN/MAX */ + pn_Cmp cmp_result; + + if (tv_l == get_mode_min(mode)) { + /* MIN <=/> x. This results in true/false. */ + if (pn_cmp == pn_Cmp_Le) + return tarval_b_true; + else if (pn_cmp == pn_Cmp_Gt) + return tarval_b_false; + } else if (tv_r == get_mode_min(mode)) { + /* x >=/< MIN. This results in true/false. */ + if (pn_cmp == pn_Cmp_Ge) + return tarval_b_true; + else if (pn_cmp == pn_Cmp_Lt) + return tarval_b_false; + } else if (tv_l == get_mode_max(mode)) { + /* MAX >=/< x. This results in true/false. */ + if (pn_cmp == pn_Cmp_Ge) + return tarval_b_true; + else if (pn_cmp == pn_Cmp_Lt) + return tarval_b_false; + } else if (tv_r == get_mode_max(mode)) { + /* x <=/> MAX. This results in true/false. */ + if (pn_cmp == pn_Cmp_Le) + return tarval_b_true; + else if (pn_cmp == pn_Cmp_Gt) + return tarval_b_false; + } + + cmp_result = vrp_cmp(left, right); + if (cmp_result != pn_Cmp_False) { + if (cmp_result == pn_Cmp_Lg) { + if (pn_cmp == pn_Cmp_Eq) { + return tarval_b_false; + } else if (pn_cmp == pn_Cmp_Lg) { + return tarval_b_true; + } + } else { + return new_tarval_from_long(cmp_result & pn_cmp, mode_b); } + } + } else if (mode_is_reference(mode)) { + /* pointer compare */ + ir_node *s_l = skip_Proj(left); + ir_node *s_r = skip_Proj(right); - pn_Cmp cmp_result = vrp_cmp(aa, ab); - if (cmp_result != pn_Cmp_False) { - return new_tarval_from_long(cmp_result & proj_nr, mode_b); + if ((is_Alloc(s_l) && tarval_is_null(tv_r)) || + (tarval_is_null(tv_l) && is_Alloc(s_r))) { + /* + * The predecessors are Allocs and (void*)(0) constants. In Firm Allocs never + * return NULL, they raise an exception. Therefore we can predict + * the Cmp result. + */ + return new_tarval_from_long(pn_cmp & pn_Cmp_Lg, mode_b); } } - /* - * The predecessors are Allocs or (void*)(0) constants. Allocs never - * return NULL, they raise an exception. Therefore we can predict - * the Cmp result. - */ - else { - ir_node *aaa = skip_Proj(aa); - ir_node *aba = skip_Proj(ab); - - if ( ( (/* aa is ProjP and aaa is Alloc */ - is_Proj(aa) - && mode_is_reference(get_irn_mode(aa)) - && is_Alloc(aaa)) - && ( (/* ab is NULL */ - mode_is_reference(get_irn_mode(ab)) - && tarval_is_null(tab)) - || (/* ab is other Alloc */ - is_Proj(ab) - && mode_is_reference(get_irn_mode(ab)) - && is_Alloc(aba) - && (aaa != aba)))) - || (/* aa is NULL and aba is Alloc */ - mode_is_reference(get_irn_mode(aa)) - && tarval_is_null(taa) - && is_Proj(ab) - && mode_is_reference(get_irn_mode(ab)) - && is_Alloc(aba))) - /* 3.: */ - return new_tarval_from_long(proj_nr & pn_Cmp_Lg, mode_b); - } - return computed_value_Cmp_Confirm(a, aa, ab, proj_nr); + return computed_value_Cmp_Confirm(cmp, left, right, pn_cmp); } /* computed_value_Proj_Cmp */ /** * Return the value of a floating point Quot. */ -static tarval *do_computed_value_Quot(const ir_node *a, const ir_node *b) { +static tarval *do_computed_value_Quot(const ir_node *a, const ir_node *b) +{ tarval *ta = value_of(a); tarval *tb = value_of(b); @@ -576,7 +584,8 @@ static tarval *do_computed_value_Quot(const ir_node *a, const ir_node *b) { * Calculate the value of an integer Div of two nodes. * Special case: 0 / b */ -static tarval *do_computed_value_Div(const ir_node *a, const ir_node *b) { +static tarval *do_computed_value_Div(const ir_node *a, const ir_node *b) +{ tarval *ta = value_of(a); tarval *tb; const ir_node *dummy; @@ -594,7 +603,8 @@ static tarval *do_computed_value_Div(const ir_node *a, const ir_node *b) { * Calculate the value of an integer Mod of two nodes. * Special case: a % 1 */ -static tarval *do_computed_value_Mod(const ir_node *a, const ir_node *b) { +static tarval *do_computed_value_Mod(const ir_node *a, const ir_node *b) +{ tarval *ta = value_of(a); tarval *tb = value_of(b); @@ -609,7 +619,8 @@ static tarval *do_computed_value_Mod(const ir_node *a, const ir_node *b) { /** * Return the value of a Proj(DivMod). */ -static tarval *computed_value_Proj_DivMod(const ir_node *n) { +static tarval *computed_value_Proj_DivMod(const ir_node *n) +{ long proj_nr = get_Proj_proj(n); /* compute either the Div or the Mod part */ @@ -626,7 +637,8 @@ static tarval *computed_value_Proj_DivMod(const ir_node *n) { /** * Return the value of a Proj(Div). */ -static tarval *computed_value_Proj_Div(const ir_node *n) { +static tarval *computed_value_Proj_Div(const ir_node *n) +{ long proj_nr = get_Proj_proj(n); if (proj_nr == pn_Div_res) { @@ -639,7 +651,8 @@ static tarval *computed_value_Proj_Div(const ir_node *n) { /** * Return the value of a Proj(Mod). */ -static tarval *computed_value_Proj_Mod(const ir_node *n) { +static tarval *computed_value_Proj_Mod(const ir_node *n) +{ long proj_nr = get_Proj_proj(n); if (proj_nr == pn_Mod_res) { @@ -652,7 +665,8 @@ static tarval *computed_value_Proj_Mod(const ir_node *n) { /** * Return the value of a Proj(Quot). */ -static tarval *computed_value_Proj_Quot(const ir_node *n) { +static tarval *computed_value_Proj_Quot(const ir_node *n) +{ long proj_nr = get_Proj_proj(n); if (proj_nr == pn_Quot_res) { @@ -665,7 +679,8 @@ static tarval *computed_value_Proj_Quot(const ir_node *n) { /** * Return the value of a Proj. */ -static tarval *computed_value_Proj(const ir_node *proj) { +static tarval *computed_value_Proj(const ir_node *proj) +{ ir_node *n = get_Proj_pred(proj); if (n->op->ops.computed_value_Proj != NULL) @@ -679,10 +694,11 @@ static tarval *computed_value_Proj(const ir_node *proj) { * * @param n The node this should be evaluated */ -tarval *computed_value(const ir_node *n) { - if(mode_is_int(get_irn_mode(n)) && tarval_is_all_one( - tarval_or(n->vrp.bits_set, n->vrp.bits_not_set))) { - return n->vrp.bits_set; +tarval *computed_value(const ir_node *n) +{ + vrp_attr *vrp = vrp_get_info(n); + if (vrp && vrp->valid && tarval_cmp(vrp->bits_set, vrp->bits_not_set) == pn_Cmp_Eq) { + return vrp->bits_set; } if (n->op->ops.computed_value) return n->op->ops.computed_value(n); @@ -737,7 +753,8 @@ static ir_op_ops *firm_set_default_computed_value(ir_opcode code, ir_op_ops *ops CASE_PROJ(Quot); CASE(Proj); default: - /* leave NULL */; + /* leave NULL */ + break; } return ops; @@ -858,7 +875,8 @@ static ir_node *equivalent_node_Block(ir_node *n) * Returns a equivalent node for a Jmp, a Bad :-) * Of course this only happens if the Block of the Jmp is dead. */ -static ir_node *equivalent_node_Jmp(ir_node *n) { +static ir_node *equivalent_node_Jmp(ir_node *n) +{ ir_node *oldn = n; /* unreachable code elimination */ @@ -880,7 +898,8 @@ static ir_node *equivalent_node_Jmp(ir_node *n) { * Optimize operations that are commutative and have neutral 0, * so a op 0 = 0 op a = a. */ -static ir_node *equivalent_node_neutral_zero(ir_node *n) { +static ir_node *equivalent_node_neutral_zero(ir_node *n) +{ ir_node *oldn = n; ir_node *a = get_binop_left(n); @@ -917,7 +936,8 @@ static ir_node *equivalent_node_neutral_zero(ir_node *n) { /** * Eor is commutative and has neutral 0. */ -static ir_node *equivalent_node_Eor(ir_node *n) { +static ir_node *equivalent_node_Eor(ir_node *n) +{ ir_node *oldn = n; ir_node *a; ir_node *b; @@ -972,7 +992,8 @@ static ir_node *equivalent_node_Eor(ir_node *n) { * Beware: The Mode of an Add may be different than the mode of its * predecessors, so we could not return a predecessors in all cases. */ -static ir_node *equivalent_node_Add(ir_node *n) { +static ir_node *equivalent_node_Add(ir_node *n) +{ ir_node *oldn = n; ir_node *left, *right; ir_mode *mode = get_irn_mode(n); @@ -1017,7 +1038,8 @@ static ir_node *equivalent_node_Add(ir_node *n) { * optimize operations that are not commutative but have neutral 0 on left, * so a op 0 = a. */ -static ir_node *equivalent_node_left_zero(ir_node *n) { +static ir_node *equivalent_node_left_zero(ir_node *n) +{ ir_node *oldn = n; ir_node *a = get_binop_left(n); @@ -1046,7 +1068,8 @@ static ir_node *equivalent_node_left_zero(ir_node *n) { * Beware: The Mode of a Sub may be different than the mode of its * predecessors, so we could not return a predecessors in all cases. */ -static ir_node *equivalent_node_Sub(ir_node *n) { +static ir_node *equivalent_node_Sub(ir_node *n) +{ ir_node *oldn = n; ir_node *b; ir_mode *mode = get_irn_mode(n); @@ -1080,7 +1103,8 @@ static ir_node *equivalent_node_Sub(ir_node *n) { * We handle it anyway here but the better way would be a * flag. This would be needed for Pascal for instance. */ -static ir_node *equivalent_node_idempotent_unop(ir_node *n) { +static ir_node *equivalent_node_idempotent_unop(ir_node *n) +{ ir_node *oldn = n; ir_node *pred = get_unop_op(n); @@ -1102,7 +1126,8 @@ static ir_node *equivalent_node_idempotent_unop(ir_node *n) { /** * Optimize a * 1 = 1 * a = a. */ -static ir_node *equivalent_node_Mul(ir_node *n) { +static ir_node *equivalent_node_Mul(ir_node *n) +{ ir_node *oldn = n; ir_node *a = get_Mul_left(n); @@ -1133,7 +1158,8 @@ static ir_node *equivalent_node_Mul(ir_node *n) { /** * Use algebraic simplification a | a = a | 0 = 0 | a = a. */ -static ir_node *equivalent_node_Or(ir_node *n) { +static ir_node *equivalent_node_Or(ir_node *n) +{ ir_node *oldn = n; ir_node *a = get_Or_left(n); @@ -1165,7 +1191,8 @@ static ir_node *equivalent_node_Or(ir_node *n) { /** * Optimize a & 0b1...1 = 0b1...1 & a = a & a = (a|X) & a = a. */ -static ir_node *equivalent_node_And(ir_node *n) { +static ir_node *equivalent_node_And(ir_node *n) +{ ir_node *oldn = n; ir_node *a = get_And_left(n); @@ -1227,7 +1254,8 @@ static ir_node *equivalent_node_And(ir_node *n) { /** * Try to remove useless Conv's: */ -static ir_node *equivalent_node_Conv(ir_node *n) { +static ir_node *equivalent_node_Conv(ir_node *n) +{ ir_node *oldn = n; ir_node *a = get_Conv_op(n); @@ -1360,7 +1388,8 @@ restart: * A Cast may be removed if the type of the previous node * is already the type of the Cast. */ -static ir_node *equivalent_node_Cast(ir_node *n) { +static ir_node *equivalent_node_Cast(ir_node *n) +{ ir_node *oldn = n; ir_node *pred = get_Cast_op(n); @@ -1375,7 +1404,8 @@ static ir_node *equivalent_node_Cast(ir_node *n) { * - fold Phi-nodes, iff they have only one predecessor except * themselves. */ -static ir_node *equivalent_node_Phi(ir_node *n) { +static ir_node *equivalent_node_Phi(ir_node *n) +{ int i, n_preds; ir_node *oldn = n; @@ -1445,7 +1475,8 @@ static ir_node *equivalent_node_Phi(ir_node *n) { * - fold Sync-nodes, iff they have only one predecessor except * themselves. */ -static ir_node *equivalent_node_Sync(ir_node *n) { +static ir_node *equivalent_node_Sync(ir_node *n) +{ int arity = get_Sync_n_preds(n); int i; @@ -1482,7 +1513,8 @@ static ir_node *equivalent_node_Sync(ir_node *n) { /** * Optimize Proj(Tuple). */ -static ir_node *equivalent_node_Proj_Tuple(ir_node *proj) { +static ir_node *equivalent_node_Proj_Tuple(ir_node *proj) +{ ir_node *oldn = proj; ir_node *tuple = get_Proj_pred(proj); @@ -1496,7 +1528,8 @@ static ir_node *equivalent_node_Proj_Tuple(ir_node *proj) { /** * Optimize a / 1 = a. */ -static ir_node *equivalent_node_Proj_Div(ir_node *proj) { +static ir_node *equivalent_node_Proj_Div(ir_node *proj) +{ ir_node *oldn = proj; ir_node *div = get_Proj_pred(proj); ir_node *b = get_Div_right(div); @@ -1527,7 +1560,8 @@ static ir_node *equivalent_node_Proj_Div(ir_node *proj) { /** * Optimize a / 1.0 = a. */ -static ir_node *equivalent_node_Proj_Quot(ir_node *proj) { +static ir_node *equivalent_node_Proj_Quot(ir_node *proj) +{ ir_node *oldn = proj; ir_node *quot = get_Proj_pred(proj); ir_node *b = get_Quot_right(quot); @@ -1558,7 +1592,8 @@ static ir_node *equivalent_node_Proj_Quot(ir_node *proj) { /** * Optimize a / 1 = a. */ -static ir_node *equivalent_node_Proj_DivMod(ir_node *proj) { +static ir_node *equivalent_node_Proj_DivMod(ir_node *proj) +{ ir_node *oldn = proj; ir_node *divmod = get_Proj_pred(proj); ir_node *b = get_DivMod_right(divmod); @@ -1591,7 +1626,8 @@ static ir_node *equivalent_node_Proj_DivMod(ir_node *proj) { /** * Optimize CopyB(mem, x, x) into a Nop. */ -static ir_node *equivalent_node_Proj_CopyB(ir_node *proj) { +static ir_node *equivalent_node_Proj_CopyB(ir_node *proj) +{ ir_node *oldn = proj; ir_node *copyb = get_Proj_pred(proj); ir_node *a = get_CopyB_dst(copyb); @@ -1617,7 +1653,8 @@ static ir_node *equivalent_node_Proj_CopyB(ir_node *proj) { /** * Optimize Bounds(idx, idx, upper) into idx. */ -static ir_node *equivalent_node_Proj_Bound(ir_node *proj) { +static ir_node *equivalent_node_Proj_Bound(ir_node *proj) +{ ir_node *oldn = proj; ir_node *bound = get_Proj_pred(proj); ir_node *idx = get_Bound_index(bound); @@ -1662,7 +1699,7 @@ static ir_node *equivalent_node_Proj_Bound(ir_node *proj) { break; default: /* cannot optimize pn_Bound_X_regular, handled in transform ... */ - ; + break; } } return proj; @@ -1671,7 +1708,8 @@ static ir_node *equivalent_node_Proj_Bound(ir_node *proj) { /** * Optimize an Exception Proj(Load) with a non-null address. */ -static ir_node *equivalent_node_Proj_Load(ir_node *proj) { +static ir_node *equivalent_node_Proj_Load(ir_node *proj) +{ if (get_opt_ldst_only_null_ptr_exceptions()) { if (get_irn_mode(proj) == mode_X) { ir_node *load = get_Proj_pred(proj); @@ -1694,7 +1732,8 @@ static ir_node *equivalent_node_Proj_Load(ir_node *proj) { /** * Optimize an Exception Proj(Store) with a non-null address. */ -static ir_node *equivalent_node_Proj_Store(ir_node *proj) { +static ir_node *equivalent_node_Proj_Store(ir_node *proj) +{ if (get_opt_ldst_only_null_ptr_exceptions()) { if (get_irn_mode(proj) == mode_X) { ir_node *store = get_Proj_pred(proj); @@ -1718,7 +1757,8 @@ static ir_node *equivalent_node_Proj_Store(ir_node *proj) { * Does all optimizations on nodes that must be done on it's Proj's * because of creating new nodes. */ -static ir_node *equivalent_node_Proj(ir_node *proj) { +static ir_node *equivalent_node_Proj(ir_node *proj) +{ ir_node *n = get_Proj_pred(proj); if (get_irn_mode(proj) == mode_X) { @@ -1735,7 +1775,8 @@ static ir_node *equivalent_node_Proj(ir_node *proj) { /** * Remove Id's. */ -static ir_node *equivalent_node_Id(ir_node *n) { +static ir_node *equivalent_node_Id(ir_node *n) +{ ir_node *oldn = n; do { @@ -1853,7 +1894,8 @@ static ir_node *equivalent_node_Mux(ir_node *n) * Remove Confirm nodes if setting is on. * Replace Confirms(x, '=', Constlike) by Constlike. */ -static ir_node *equivalent_node_Confirm(ir_node *n) { +static ir_node *equivalent_node_Confirm(ir_node *n) +{ ir_node *pred = get_Confirm_value(n); pn_Cmp pnc = get_Confirm_cmp(n); @@ -1878,7 +1920,8 @@ static ir_node *equivalent_node_Confirm(ir_node *n) { * If a node returns a Tuple we can not just skip it. If the size of the * in array fits, we transform n into a tuple (e.g., Div). */ -ir_node *equivalent_node(ir_node *n) { +ir_node *equivalent_node(ir_node *n) +{ if (n->op->ops.equivalent_node) return n->op->ops.equivalent_node(n); return n; @@ -1937,7 +1980,8 @@ static ir_op_ops *firm_set_default_equivalent_node(ir_opcode code, ir_op_ops *op CASE(Mux); CASE(Confirm); default: - /* leave NULL */; + /* leave NULL */ + break; } return ops; @@ -1949,7 +1993,8 @@ static ir_op_ops *firm_set_default_equivalent_node(ir_opcode code, ir_op_ops *op * Returns non-zero if a node is a Phi node * with all predecessors constant. */ -static int is_const_Phi(ir_node *n) { +static int is_const_Phi(ir_node *n) +{ int i; if (! is_Phi(n) || get_irn_arity(n) == 0) @@ -1997,7 +2042,8 @@ static tarval *do_eval(eval_func eval, tarval *a, tarval *b, ir_mode *mode) * * @return a new Phi node if the conversion was successful, NULL else */ -static ir_node *apply_binop_on_phi(ir_node *phi, tarval *other, eval_func eval, ir_mode *mode, int left) { +static ir_node *apply_binop_on_phi(ir_node *phi, tarval *other, eval_func eval, ir_mode *mode, int left) +{ tarval *tv; void **res; ir_node *pred; @@ -2048,7 +2094,8 @@ static ir_node *apply_binop_on_phi(ir_node *phi, tarval *other, eval_func eval, * * @return a new Phi node if the conversion was successful, NULL else */ -static ir_node *apply_binop_on_2_phis(ir_node *a, ir_node *b, eval_func eval, ir_mode *mode) { +static ir_node *apply_binop_on_2_phis(ir_node *a, ir_node *b, eval_func eval, ir_mode *mode) +{ tarval *tv_l, *tv_r, *tv; void **res; ir_node *pred; @@ -2090,7 +2137,8 @@ static ir_node *apply_binop_on_2_phis(ir_node *a, ir_node *b, eval_func eval, ir * * @return a new Phi node if the conversion was successful, NULL else */ -static ir_node *apply_unop_on_phi(ir_node *phi, tarval *(*eval)(tarval *)) { +static ir_node *apply_unop_on_phi(ir_node *phi, tarval *(*eval)(tarval *)) +{ tarval *tv; void **res; ir_node *pred; @@ -2126,7 +2174,8 @@ static ir_node *apply_unop_on_phi(ir_node *phi, tarval *(*eval)(tarval *)) { * * @return a new Phi node if the conversion was successful, NULL else */ -static ir_node *apply_conv_on_phi(ir_node *phi, ir_mode *mode) { +static ir_node *apply_conv_on_phi(ir_node *phi, ir_mode *mode) +{ tarval *tv; void **res; ir_node *pred; @@ -2158,7 +2207,8 @@ static ir_node *apply_conv_on_phi(ir_node *phi, ir_mode *mode) { * SubP(P, ConvIs(Iu)), SubP(P, ConvIu(Is)). * If possible, remove the Conv's. */ -static ir_node *transform_node_AddSub(ir_node *n) { +static ir_node *transform_node_AddSub(ir_node *n) +{ ir_mode *mode = get_irn_mode(n); if (mode_is_reference(mode)) { @@ -2230,6 +2280,7 @@ static ir_node *transform_node_AddSub(ir_node *n) { } /* transform_node_AddSub */ #define HANDLE_BINOP_PHI(eval, a, b, c, mode) \ + do { \ c = NULL; \ if (is_Const(b) && is_const_Phi(a)) { \ /* check for Op(Phi, Const) */ \ @@ -2246,9 +2297,11 @@ static ir_node *transform_node_AddSub(ir_node *n) { if (c) { \ DBG_OPT_ALGSIM0(oldn, c, FS_OPT_CONST_PHI); \ return c; \ - } + } \ + } while(0) #define HANDLE_UNOP_PHI(eval, a, c) \ + do { \ c = NULL; \ if (is_const_Phi(a)) { \ /* check for Op(Phi) */ \ @@ -2257,7 +2310,8 @@ static ir_node *transform_node_AddSub(ir_node *n) { DBG_OPT_ALGSIM0(oldn, c, FS_OPT_CONST_PHI); \ return c; \ } \ - } + } \ + } while(0) /** * Do the AddSub optimization, then Transform @@ -2268,9 +2322,11 @@ static ir_node *transform_node_AddSub(ir_node *n) { * Transform Add(a,-b) into Sub(a,b). * Reassociation might fold this further. */ -static ir_node *transform_node_Add(ir_node *n) { +static ir_node *transform_node_Add(ir_node *n) +{ ir_mode *mode; ir_node *a, *b, *c, *oldn = n; + vrp_attr *a_vrp, *b_vrp; n = transform_node_AddSub(n); @@ -2282,7 +2338,7 @@ static ir_node *transform_node_Add(ir_node *n) { if (mode_is_reference(mode)) { ir_mode *lmode = get_irn_mode(a); - if (is_Const(b) && is_Const_null(b) && mode_is_int(lmode)) { + if (is_Const(b) && is_Const_null(b) && mode_is_int(lmode)) { /* an Add(a, NULL) is a hidden Conv */ dbg_info *dbg = get_irn_dbg_info(n); return new_rd_Conv(dbg, get_nodes_block(n), a, mode); @@ -2361,13 +2417,17 @@ static ir_node *transform_node_Add(ir_node *n) { } } } - if (mode_is_int(mode)) { + + a_vrp = vrp_get_info(a); + b_vrp = vrp_get_info(b); + + if (a_vrp && b_vrp) { tarval *c = tarval_and( - tarval_not(a->vrp.bits_not_set), - tarval_not(b->vrp.bits_not_set) + a_vrp->bits_not_set, + b_vrp->bits_not_set ); - if(tarval_is_null(c)) { + if (tarval_is_null(c)) { dbg_info *dbgi = get_irn_dbg_info(n); return new_rd_Or(dbgi, get_nodes_block(n), a, b, mode); @@ -2379,7 +2439,8 @@ static ir_node *transform_node_Add(ir_node *n) { /** * returns -cnst or NULL if impossible */ -static ir_node *const_negate(ir_node *cnst) { +static ir_node *const_negate(ir_node *cnst) +{ tarval *tv = tarval_neg(get_Const_tarval(cnst)); dbg_info *dbgi = get_irn_dbg_info(cnst); ir_graph *irg = get_irn_irg(cnst); @@ -2397,7 +2458,8 @@ static ir_node *const_negate(ir_node *cnst) { * Sub(x, Add(x, a)) -> -a * Sub(x, Const) -> Add(x, -Const) */ -static ir_node *transform_node_Sub(ir_node *n) { +static ir_node *transform_node_Sub(ir_node *n) +{ ir_mode *mode; ir_node *oldn = n; ir_node *a, *b, *c; @@ -2615,7 +2677,7 @@ restart: get_irn_dbg_info(n), blk, mb, - new_Const_long(mode, 1), + new_Const(get_mode_one(mode)), mode), mode); DBG_OPT_ALGSIM0(oldn, n, FS_OPT_SUB_MUL_A_X_A); @@ -2630,7 +2692,7 @@ restart: get_irn_dbg_info(n), blk, ma, - new_Const_long(mode, 1), + new_Const(get_mode_one(mode)), mode), mode); DBG_OPT_ALGSIM0(oldn, n, FS_OPT_SUB_MUL_A_X_A); @@ -2692,7 +2754,8 @@ restart: * Several transformation done on n*n=2n bits mul. * These transformations must be done here because new nodes may be produced. */ -static ir_node *transform_node_Mul2n(ir_node *n, ir_mode *mode) { +static ir_node *transform_node_Mul2n(ir_node *n, ir_mode *mode) +{ ir_node *oldn = n; ir_node *a = get_Mul_left(n); ir_node *b = get_Mul_right(n); @@ -2738,7 +2801,8 @@ static ir_node *transform_node_Mul2n(ir_node *n, ir_mode *mode) { * Do constant evaluation of Phi nodes. * Do architecture dependent optimizations on Mul nodes */ -static ir_node *transform_node_Mul(ir_node *n) { +static ir_node *transform_node_Mul(ir_node *n) +{ ir_node *c, *oldn = n; ir_mode *mode = get_irn_mode(n); ir_node *a = get_Mul_left(n); @@ -2853,7 +2917,8 @@ static ir_node *transform_node_Mul(ir_node *n) { /** * Transform a Div Node. */ -static ir_node *transform_node_Div(ir_node *n) { +static ir_node *transform_node_Div(ir_node *n) +{ ir_mode *mode = get_Div_resmode(n); ir_node *a = get_Div_left(n); ir_node *b = get_Div_right(n); @@ -2929,7 +2994,8 @@ make_tuple: /** * Transform a Mod node. */ -static ir_node *transform_node_Mod(ir_node *n) { +static ir_node *transform_node_Mod(ir_node *n) +{ ir_mode *mode = get_Mod_resmode(n); ir_node *a = get_Mod_left(n); ir_node *b = get_Mod_right(n); @@ -3016,7 +3082,8 @@ make_tuple: /** * Transform a DivMod node. */ -static ir_node *transform_node_DivMod(ir_node *n) { +static ir_node *transform_node_DivMod(ir_node *n) +{ const ir_node *dummy; ir_node *a = get_DivMod_left(n); ir_node *b = get_DivMod_right(n); @@ -3129,7 +3196,8 @@ make_tuple: /** * Optimize x / c to x * (1/c) */ -static ir_node *transform_node_Quot(ir_node *n) { +static ir_node *transform_node_Quot(ir_node *n) +{ ir_mode *mode = get_Quot_resmode(n); ir_node *oldn = n; @@ -3179,7 +3247,8 @@ static ir_node *transform_node_Quot(ir_node *n) { * Optimize Abs(x) into -x if x is Confirmed <= 0 * Optimize Abs(-x) int Abs(x) */ -static ir_node *transform_node_Abs(ir_node *n) { +static ir_node *transform_node_Abs(ir_node *n) +{ ir_node *c, *oldn = n; ir_node *a = get_Abs_op(n); ir_mode *mode; @@ -3230,7 +3299,8 @@ static ir_node *transform_node_Abs(ir_node *n) { * * For == and != can be handled in Proj(Cmp) */ -static ir_node *transform_node_Cmp(ir_node *n) { +static ir_node *transform_node_Cmp(ir_node *n) +{ ir_node *oldn = n; ir_node *left = get_Cmp_left(n); ir_node *right = get_Cmp_right(n); @@ -3252,7 +3322,8 @@ static ir_node *transform_node_Cmp(ir_node *n) { * Replace the Cond by a Jmp if it branches on a constant * condition. */ -static ir_node *transform_node_Cond(ir_node *n) { +static ir_node *transform_node_Cond(ir_node *n) +{ ir_node *jmp; ir_node *a = get_Cond_selector(n); @@ -3303,7 +3374,7 @@ static ir_node *transform_bitwise_distributive(ir_node *n, ir_op *op = get_irn_op(a); ir_op *op_root = get_irn_op(n); - if(op != get_irn_op(b)) + if (op != get_irn_op(b)) return n; /* and(conv(a), conv(b)) -> conv(and(a,b)) */ @@ -3312,7 +3383,7 @@ static ir_node *transform_bitwise_distributive(ir_node *n, ir_node *b_op = get_Conv_op(b); ir_mode *a_mode = get_irn_mode(a_op); ir_mode *b_mode = get_irn_mode(b_op); - if(a_mode == b_mode && (mode_is_int(a_mode) || a_mode == mode_b)) { + if (a_mode == b_mode && (mode_is_int(a_mode) || a_mode == mode_b)) { ir_node *blk = get_nodes_block(n); n = exact_copy(n); @@ -3347,17 +3418,17 @@ static ir_node *transform_bitwise_distributive(ir_node *n, c = a_left; op1 = a_right; op2 = b_right; - } else if(a_left == b_right) { + } else if (a_left == b_right) { c = a_left; op1 = a_right; op2 = b_left; - } else if(a_right == b_left) { + } else if (a_right == b_left) { c = a_right; op1 = a_left; op2 = b_right; } } - if(a_right == b_right) { + if (a_right == b_right) { c = a_right; op1 = a_left; op2 = b_left; @@ -3372,7 +3443,7 @@ static ir_node *transform_bitwise_distributive(ir_node *n, set_binop_right(new_n, op2); new_n = trans_func(new_n); - if(op_root == op_Eor && op == op_Or) { + if (op_root == op_Eor && op == op_Or) { dbg_info *dbgi = get_irn_dbg_info(n); ir_mode *mode = get_irn_mode(c); @@ -3397,11 +3468,13 @@ static ir_node *transform_bitwise_distributive(ir_node *n, /** * Transform an And. */ -static ir_node *transform_node_And(ir_node *n) { +static ir_node *transform_node_And(ir_node *n) +{ ir_node *c, *oldn = n; ir_node *a = get_And_left(n); ir_node *b = get_And_right(n); ir_mode *mode; + vrp_attr *a_vrp, *b_vrp; mode = get_irn_mode(n); HANDLE_BINOP_PHI((eval_func) tarval_and, a, b, c, mode); @@ -3412,13 +3485,12 @@ static ir_node *transform_node_And(ir_node *n) { ir_node *pred_b = get_Proj_pred(b); if (pred_a == pred_b) { dbg_info *dbgi = get_irn_dbg_info(n); - ir_node *block = get_nodes_block(pred_a); pn_Cmp pn_a = get_Proj_proj(a); pn_Cmp pn_b = get_Proj_proj(b); /* yes, we can simply calculate with pncs */ pn_Cmp new_pnc = pn_a & pn_b; - return new_rd_Proj(dbgi, block, pred_a, mode_b, new_pnc); + return new_rd_Proj(dbgi, pred_a, mode_b, new_pnc); } } if (is_Or(a)) { @@ -3522,18 +3594,18 @@ static ir_node *transform_node_And(ir_node *n) { return n; } - if (is_Const(a) && (tarval_is_all_one(tarval_or(get_Const_tarval(a), - b->vrp.bits_not_set)))) { - return new_rd_Id(get_irn_dbg_info(n), get_nodes_block(n), - b, get_irn_mode(n)); + b_vrp = vrp_get_info(b); + if (is_Const(a) && b_vrp && (tarval_cmp(tarval_or(get_Const_tarval(a), + b_vrp->bits_not_set), get_Const_tarval(a)) == pn_Cmp_Eq)) { - } + return b; - if (is_Const(b) && (tarval_is_all_one(tarval_or(get_Const_tarval(b), - a->vrp.bits_not_set)))) { - return new_rd_Id(get_irn_dbg_info(n), get_nodes_block(n), - a, get_irn_mode(n)); + } + a_vrp = vrp_get_info(a); + if (is_Const(b) && a_vrp && (tarval_cmp(tarval_or(get_Const_tarval(b), + a_vrp->bits_not_set), get_Const_tarval(b)) == pn_Cmp_Eq)) { + return a; } n = transform_bitwise_distributive(n, transform_node_And); @@ -3544,7 +3616,8 @@ static ir_node *transform_node_And(ir_node *n) { /** * Transform an Eor. */ -static ir_node *transform_node_Eor(ir_node *n) { +static ir_node *transform_node_Eor(ir_node *n) +{ ir_node *c, *oldn = n; ir_node *a = get_Eor_left(n); ir_node *b = get_Eor_right(n); @@ -3556,15 +3629,14 @@ static ir_node *transform_node_Eor(ir_node *n) { if (mode == mode_b && is_Proj(a) && is_Proj(b)) { ir_node *pred_a = get_Proj_pred(a); ir_node *pred_b = get_Proj_pred(b); - if(pred_a == pred_b) { + if (pred_a == pred_b) { dbg_info *dbgi = get_irn_dbg_info(n); - ir_node *block = get_nodes_block(pred_a); pn_Cmp pn_a = get_Proj_proj(a); pn_Cmp pn_b = get_Proj_proj(b); /* yes, we can simply calculate with pncs */ pn_Cmp new_pnc = pn_a ^ pn_b; - return new_rd_Proj(dbgi, block, pred_a, mode_b, new_pnc); + return new_rd_Proj(dbgi, pred_a, mode_b, new_pnc); } } @@ -3596,7 +3668,8 @@ static ir_node *transform_node_Eor(ir_node *n) { /** * Transform a Not. */ -static ir_node *transform_node_Not(ir_node *n) { +static ir_node *transform_node_Not(ir_node *n) +{ ir_node *c, *oldn = n; ir_node *a = get_Not_op(n); ir_mode *mode = get_irn_mode(n); @@ -3607,15 +3680,14 @@ static ir_node *transform_node_Not(ir_node *n) { if (mode == mode_b && is_Proj(a)) { ir_node *a_pred = get_Proj_pred(a); if (is_Cmp(a_pred)) { - ir_node *cmp_block = get_nodes_block(a_pred); /* We negate a Cmp. The Cmp has the negated result anyways! */ - n = new_r_Proj(cmp_block, get_Proj_pred(a), + n = new_r_Proj(get_Proj_pred(a), mode_b, get_negated_pnc(get_Proj_proj(a), mode_b)); DBG_OPT_ALGSIM0(oldn, n, FS_OPT_NOT_CMP); return n; } } - if (is_Eor(a)) { + if (is_Eor(a)) { ir_node *eor_b = get_Eor_right(a); if (is_Const(eor_b)) { /* ~(x ^ const) -> x ^ ~const */ ir_node *cnst = new_Const(tarval_not(get_Const_tarval(eor_b))); @@ -3658,7 +3730,8 @@ static ir_node *transform_node_Not(ir_node *n) { * -(a >>s (size-1)) = a >>u (size-1) * -(a * const) -> a * -const */ -static ir_node *transform_node_Minus(ir_node *n) { +static ir_node *transform_node_Minus(ir_node *n) +{ ir_node *c, *oldn = n; ir_node *a = get_Minus_op(n); ir_mode *mode; @@ -3745,7 +3818,8 @@ static ir_node *transform_node_Minus(ir_node *n) { /** * Transform a Cast_type(Const) into a new Const_type */ -static ir_node *transform_node_Cast(ir_node *n) { +static ir_node *transform_node_Cast(ir_node *n) +{ ir_node *oldn = n; ir_node *pred = get_Cast_op(n); ir_type *tp = get_irn_type(n); @@ -3765,7 +3839,8 @@ static ir_node *transform_node_Cast(ir_node *n) { /** * Transform a Proj(Load) with a non-null address. */ -static ir_node *transform_node_Proj_Load(ir_node *proj) { +static ir_node *transform_node_Proj_Load(ir_node *proj) +{ if (get_opt_ldst_only_null_ptr_exceptions()) { if (get_irn_mode(proj) == mode_X) { ir_node *load = get_Proj_pred(proj); @@ -3795,7 +3870,8 @@ static ir_node *transform_node_Proj_Load(ir_node *proj) { /** * Transform a Proj(Store) with a non-null address. */ -static ir_node *transform_node_Proj_Store(ir_node *proj) { +static ir_node *transform_node_Proj_Store(ir_node *proj) +{ if (get_opt_ldst_only_null_ptr_exceptions()) { if (get_irn_mode(proj) == mode_X) { ir_node *store = get_Proj_pred(proj); @@ -3826,7 +3902,8 @@ static ir_node *transform_node_Proj_Store(ir_node *proj) { * Transform a Proj(Div) with a non-zero value. * Removes the exceptions and routes the memory to the NoMem node. */ -static ir_node *transform_node_Proj_Div(ir_node *proj) { +static ir_node *transform_node_Proj_Div(ir_node *proj) +{ ir_node *div = get_Proj_pred(proj); ir_node *b = get_Div_right(div); ir_node *res, *new_mem; @@ -3874,7 +3951,8 @@ static ir_node *transform_node_Proj_Div(ir_node *proj) { * Transform a Proj(Mod) with a non-zero value. * Removes the exceptions and routes the memory to the NoMem node. */ -static ir_node *transform_node_Proj_Mod(ir_node *proj) { +static ir_node *transform_node_Proj_Mod(ir_node *proj) +{ ir_node *mod = get_Proj_pred(proj); ir_node *b = get_Mod_right(mod); ir_node *res, *new_mem; @@ -3932,7 +4010,8 @@ static ir_node *transform_node_Proj_Mod(ir_node *proj) { * Transform a Proj(DivMod) with a non-zero value. * Removes the exceptions and routes the memory to the NoMem node. */ -static ir_node *transform_node_Proj_DivMod(ir_node *proj) { +static ir_node *transform_node_Proj_DivMod(ir_node *proj) +{ ir_node *divmod = get_Proj_pred(proj); ir_node *b = get_DivMod_right(divmod); ir_node *res, *new_mem; @@ -3990,7 +4069,8 @@ static ir_node *transform_node_Proj_DivMod(ir_node *proj) { /** * Optimizes jump tables (CondIs or CondIu) by removing all impossible cases. */ -static ir_node *transform_node_Proj_Cond(ir_node *proj) { +static ir_node *transform_node_Proj_Cond(ir_node *proj) +{ if (get_opt_unreachable_code()) { ir_node *n = get_Proj_pred(proj); ir_node *b = get_Cond_selector(n); @@ -4014,31 +4094,32 @@ static ir_node *transform_node_Proj_Cond(ir_node *proj) { } } else { long num = get_Proj_proj(proj); - if (num != get_Cond_default_proj(n)) { + vrp_attr *b_vrp = vrp_get_info(b); + if (num != get_Cond_default_proj(n) && b_vrp) { /* Try handling with vrp data. We only remove dead parts. */ tarval *tp = new_tarval_from_long(num, get_irn_mode(b)); - if (b->vrp.range_type == VRP_RANGE) { - pn_Cmp cmp_result = tarval_cmp(b->vrp.range_bottom, tp); - pn_Cmp cmp_result2 = tarval_cmp(b->vrp.range_top, tp); + if (b_vrp->range_type == VRP_RANGE) { + pn_Cmp cmp_result = tarval_cmp(b_vrp->range_bottom, tp); + pn_Cmp cmp_result2 = tarval_cmp(b_vrp->range_top, tp); - if ((cmp_result & pn_Cmp_Lt) == cmp_result && (cmp_result2 - & pn_Cmp_Gt) == cmp_result2) { + if ((cmp_result & pn_Cmp_Gt) == cmp_result && (cmp_result2 + & pn_Cmp_Lt) == cmp_result2) { return get_irg_bad(current_ir_graph); } - } else if (b->vrp.range_type == VRP_ANTIRANGE) { - pn_Cmp cmp_result = tarval_cmp(b->vrp.range_bottom, tp); - pn_Cmp cmp_result2 = tarval_cmp(b->vrp.range_top, tp); + } else if (b_vrp->range_type == VRP_ANTIRANGE) { + pn_Cmp cmp_result = tarval_cmp(b_vrp->range_bottom, tp); + pn_Cmp cmp_result2 = tarval_cmp(b_vrp->range_top, tp); - if ((cmp_result & pn_Cmp_Ge) == cmp_result && (cmp_result2 - & pn_Cmp_Le) == cmp_result2) { + if ((cmp_result & pn_Cmp_Le) == cmp_result && (cmp_result2 + & pn_Cmp_Ge) == cmp_result2) { return get_irg_bad(current_ir_graph); } } if (!(tarval_cmp( - tarval_and( b->vrp.bits_set, tp), - b->vrp.bits_set + tarval_and( b_vrp->bits_set, tp), + b_vrp->bits_set ) == pn_Cmp_Eq)) { return get_irg_bad(current_ir_graph); @@ -4047,8 +4128,8 @@ static ir_node *transform_node_Proj_Cond(ir_node *proj) { if (!(tarval_cmp( tarval_and( tarval_not(tp), - b->vrp.bits_not_set), - b->vrp.bits_not_set) + tarval_not(b_vrp->bits_not_set)), + tarval_not(b_vrp->bits_not_set)) == pn_Cmp_Eq)) { return get_irg_bad(current_ir_graph); @@ -4065,7 +4146,8 @@ static ir_node *transform_node_Proj_Cond(ir_node *proj) { /** * Create a 0 constant of given mode. */ -static ir_node *create_zero_const(ir_mode *mode) { +static ir_node *create_zero_const(ir_mode *mode) +{ tarval *tv = get_mode_null(mode); ir_node *cnst = new_Const(tv); @@ -4102,7 +4184,8 @@ static int operands_are_normalized(const ir_node *l, const ir_node *r) /** * Normalizes and optimizes Cmp nodes. */ -static ir_node *transform_node_Proj_Cmp(ir_node *proj) { +static ir_node *transform_node_Proj_Cmp(ir_node *proj) +{ ir_node *n = get_Proj_pred(proj); ir_node *left = get_Cmp_left(n); ir_node *right = get_Cmp_right(n); @@ -4712,7 +4795,7 @@ static ir_node *transform_node_Proj_Cmp(ir_node *proj) { /* create a new compare */ n = new_rd_Cmp(get_irn_dbg_info(n), block, left, right); - proj = new_rd_Proj(get_irn_dbg_info(proj), block, n, get_irn_mode(proj), proj_nr); + proj = new_rd_Proj(get_irn_dbg_info(proj), n, get_irn_mode(proj), proj_nr); } return proj; @@ -4721,7 +4804,8 @@ static ir_node *transform_node_Proj_Cmp(ir_node *proj) { /** * Optimize CopyB(mem, x, x) into a Nop. */ -static ir_node *transform_node_Proj_CopyB(ir_node *proj) { +static ir_node *transform_node_Proj_CopyB(ir_node *proj) +{ ir_node *copyb = get_Proj_pred(proj); ir_node *a = get_CopyB_dst(copyb); ir_node *b = get_CopyB_src(copyb); @@ -4747,7 +4831,8 @@ static ir_node *transform_node_Proj_CopyB(ir_node *proj) { /** * Optimize Bounds(idx, idx, upper) into idx. */ -static ir_node *transform_node_Proj_Bound(ir_node *proj) { +static ir_node *transform_node_Proj_Bound(ir_node *proj) +{ ir_node *oldn = proj; ir_node *bound = get_Proj_pred(proj); ir_node *idx = get_Bound_index(bound); @@ -4805,7 +4890,8 @@ static ir_node *transform_node_Proj_Bound(ir_node *proj) { * Does all optimizations on nodes that must be done on it's Proj's * because of creating new nodes. */ -static ir_node *transform_node_Proj(ir_node *proj) { +static ir_node *transform_node_Proj(ir_node *proj) +{ ir_node *n = get_Proj_pred(proj); if (n->op->ops.transform_node_Proj) @@ -4816,7 +4902,8 @@ static ir_node *transform_node_Proj(ir_node *proj) { /** * Move Confirms down through Phi nodes. */ -static ir_node *transform_node_Phi(ir_node *phi) { +static ir_node *transform_node_Phi(ir_node *phi) +{ int i, n; ir_mode *mode = get_irn_mode(phi); @@ -4860,7 +4947,8 @@ static ir_node *transform_node_Phi(ir_node *phi) { * Returns the operands of a commutative bin-op, if one operand is * a const, it is returned as the second one. */ -static void get_comm_Binop_Ops(ir_node *binop, ir_node **a, ir_node **c) { +static void get_comm_Binop_Ops(ir_node *binop, ir_node **a, ir_node **c) +{ ir_node *op_a = get_binop_left(binop); ir_node *op_b = get_binop_right(binop); @@ -4890,7 +4978,8 @@ static void get_comm_Binop_Ops(ir_node *binop, ir_node **a, ir_node **c) { * AND c1 ===> OR if (c1 | c2) == 0x111..11 * OR */ -static ir_node *transform_node_Or_bf_store(ir_node *or) { +static ir_node *transform_node_Or_bf_store(ir_node *or) +{ ir_node *and, *c1; ir_node *or_l, *c2; ir_node *and_l, *c3; @@ -4900,7 +4989,7 @@ static ir_node *transform_node_Or_bf_store(ir_node *or) { tarval *tv1, *tv2, *tv3, *tv4, *tv; - while (1) { + for (;;) { get_comm_Binop_Ops(or, &and, &c1); if (!is_Const(c1) || !is_And(and)) return or; @@ -4974,7 +5063,8 @@ static ir_node *transform_node_Or_bf_store(ir_node *or) { /** * Optimize an Or(shl(x, c), shr(x, bits - c)) into a Rotl */ -static ir_node *transform_node_Or_Rotl(ir_node *or) { +static ir_node *transform_node_Or_Rotl(ir_node *or) +{ ir_mode *mode = get_irn_mode(or); ir_node *shl, *shr, *block; ir_node *irn, *x, *c1, *c2, *v, *sub, *n, *rotval; @@ -5062,7 +5152,8 @@ static ir_node *transform_node_Or_Rotl(ir_node *or) { /** * Transform an Or. */ -static ir_node *transform_node_Or(ir_node *n) { +static ir_node *transform_node_Or(ir_node *n) +{ ir_node *c, *oldn = n; ir_node *a = get_Or_left(n); ir_node *b = get_Or_right(n); @@ -5087,13 +5178,12 @@ static ir_node *transform_node_Or(ir_node *n) { ir_node *pred_b = get_Proj_pred(b); if (pred_a == pred_b) { dbg_info *dbgi = get_irn_dbg_info(n); - ir_node *block = get_nodes_block(pred_a); pn_Cmp pn_a = get_Proj_proj(a); pn_Cmp pn_b = get_Proj_proj(b); /* yes, we can simply calculate with pncs */ pn_Cmp new_pnc = pn_a | pn_b; - return new_rd_Proj(dbgi, block, pred_a, mode_b, new_pnc); + return new_rd_Proj(dbgi, pred_a, mode_b, new_pnc); } } @@ -5119,7 +5209,8 @@ static ir_node *transform_node(ir_node *n); * * Should be moved to reassociation? */ -static ir_node *transform_node_shift(ir_node *n) { +static ir_node *transform_node_shift(ir_node *n) +{ ir_node *left, *right; ir_mode *mode; tarval *tv1, *tv2, *res; @@ -5189,7 +5280,8 @@ static ir_node *transform_node_shift(ir_node *n) { * - Shl, Shr, Shrs, rotl instead of >> * (with a special case for Or/Xor + Shrs) */ -static ir_node *transform_node_bitop_shift(ir_node *n) { +static ir_node *transform_node_bitop_shift(ir_node *n) +{ ir_node *left; ir_node *right = get_binop_right(n); ir_mode *mode = get_irn_mode(n); @@ -5237,10 +5329,10 @@ static ir_node *transform_node_bitop_shift(ir_node *n) { if (is_Shl(n)) { new_shift = new_rd_Shl(dbgi, block, bitop_left, right, mode); tv_shift = tarval_shl(tv1, tv2); - } else if(is_Shr(n)) { + } else if (is_Shr(n)) { new_shift = new_rd_Shr(dbgi, block, bitop_left, right, mode); tv_shift = tarval_shr(tv1, tv2); - } else if(is_Shrs(n)) { + } else if (is_Shrs(n)) { new_shift = new_rd_Shrs(dbgi, block, bitop_left, right, mode); tv_shift = tarval_shrs(tv1, tv2); } else { @@ -5254,7 +5346,7 @@ static ir_node *transform_node_bitop_shift(ir_node *n) { if (op_left == op_And) { new_bitop = new_rd_And(dbgi, block, new_shift, new_const, mode); - } else if(op_left == op_Or) { + } else if (op_left == op_Or) { new_bitop = new_rd_Or(dbgi, block, new_shift, new_const, mode); } else { assert(op_left == op_Eor); @@ -5271,11 +5363,11 @@ static ir_node *transform_node_bitop_shift(ir_node *n) { * (x >> c1) << c2 <=> x OP (c2-c1) & ((-1 >> c1) << c2) * (also with x >>s c1 when c1>=c2) */ -static ir_node *transform_node_shl_shr(ir_node *n) { +static ir_node *transform_node_shl_shr(ir_node *n) +{ ir_node *left; ir_node *right = get_binop_right(n); ir_node *x; - ir_graph *irg; ir_node *block; ir_mode *mode; dbg_info *dbgi; @@ -5317,7 +5409,7 @@ static ir_node *transform_node_shl_shr(ir_node *n) { tv_mask = tarval_shr(get_mode_all_one(mode), tv_shr); } tv_mask = tarval_shl(tv_mask, tv_shl); - } else if(is_Shr(n) && is_Shl(left)) { + } else if (is_Shr(n) && is_Shl(left)) { ir_node *shl_right = get_Shl_right(left); if (!is_Const(shl_right)) @@ -5340,7 +5432,6 @@ static ir_node *transform_node_shl_shr(ir_node *n) { assert(tv_mask != tarval_bad); assert(get_tarval_mode(tv_mask) == mode); - irg = get_irn_irg(n); block = get_nodes_block(n); dbgi = get_irn_dbg_info(n); @@ -5369,7 +5460,8 @@ static ir_node *transform_node_shl_shr(ir_node *n) { /** * Transform a Shr. */ -static ir_node *transform_node_Shr(ir_node *n) { +static ir_node *transform_node_Shr(ir_node *n) +{ ir_node *c, *oldn = n; ir_node *left = get_Shr_left(n); ir_node *right = get_Shr_right(n); @@ -5389,7 +5481,8 @@ static ir_node *transform_node_Shr(ir_node *n) { /** * Transform a Shrs. */ -static ir_node *transform_node_Shrs(ir_node *n) { +static ir_node *transform_node_Shrs(ir_node *n) +{ ir_node *c, *oldn = n; ir_node *a = get_Shrs_left(n); ir_node *b = get_Shrs_right(n); @@ -5407,7 +5500,8 @@ static ir_node *transform_node_Shrs(ir_node *n) { /** * Transform a Shl. */ -static ir_node *transform_node_Shl(ir_node *n) { +static ir_node *transform_node_Shl(ir_node *n) +{ ir_node *c, *oldn = n; ir_node *a = get_Shl_left(n); ir_node *b = get_Shl_right(n); @@ -5427,7 +5521,8 @@ static ir_node *transform_node_Shl(ir_node *n) { /** * Transform a Rotl. */ -static ir_node *transform_node_Rotl(ir_node *n) { +static ir_node *transform_node_Rotl(ir_node *n) +{ ir_node *c, *oldn = n; ir_node *a = get_Rotl_left(n); ir_node *b = get_Rotl_right(n); @@ -5445,7 +5540,8 @@ static ir_node *transform_node_Rotl(ir_node *n) { /** * Transform a Conv. */ -static ir_node *transform_node_Conv(ir_node *n) { +static ir_node *transform_node_Conv(ir_node *n) +{ ir_node *c, *oldn = n; ir_mode *mode = get_irn_mode(n); ir_node *a = get_Conv_op(n); @@ -5473,17 +5569,17 @@ static ir_node *transform_node_Conv(ir_node *n) { ir_node *r = get_Add_right(a); dbg_info *dbgi = get_irn_dbg_info(a); ir_node *block = get_nodes_block(n); - if(is_Conv(l)) { + if (is_Conv(l)) { ir_node *lop = get_Conv_op(l); - if(get_irn_mode(lop) == mode) { + if (get_irn_mode(lop) == mode) { /* ConvP(AddI(ConvI(P), x)) -> AddP(P, x) */ n = new_rd_Add(dbgi, block, lop, r, mode); return n; } } - if(is_Conv(r)) { + if (is_Conv(r)) { ir_node *rop = get_Conv_op(r); - if(get_irn_mode(rop) == mode) { + if (get_irn_mode(rop) == mode) { /* ConvP(AddI(x, ConvI(P))) -> AddP(x, P) */ n = new_rd_Add(dbgi, block, l, rop, mode); return n; @@ -5498,7 +5594,8 @@ static ir_node *transform_node_Conv(ir_node *n) { * Remove dead blocks and nodes in dead blocks * in keep alive list. We do not generate a new End node. */ -static ir_node *transform_node_End(ir_node *n) { +static ir_node *transform_node_End(ir_node *n) +{ int i, j, n_keepalives = get_End_n_keepalives(n); ir_node **in; @@ -5546,7 +5643,8 @@ bool is_negated_value(ir_node *a, ir_node *b) /** * Optimize a Mux into some simpler cases. */ -static ir_node *transform_node_Mux(ir_node *n) { +static ir_node *transform_node_Mux(ir_node *n) +{ ir_node *oldn = n, *sel = get_Mux_sel(n); ir_mode *mode = get_irn_mode(n); ir_node *t = get_Mux_true(n); @@ -5620,7 +5718,7 @@ static ir_node *transform_node_Mux(ir_node *n) { /* Mux(x, 0, y) => Mux(x, y, 0) */ pn_Cmp pnc = get_Proj_proj(sel); - sel = new_r_Proj(get_nodes_block(cmp), cmp, mode_b, + sel = new_r_Proj(cmp, mode_b, get_negated_pnc(pnc, get_irn_mode(get_Cmp_left(cmp)))); n = new_rd_Mux(get_irn_dbg_info(n), get_nodes_block(n), sel, t, f, mode); tmp = t; @@ -5824,7 +5922,8 @@ static ir_node *transform_node_Mux(ir_node *n) { * optimize Sync nodes that have other syncs as input we simply add the inputs * of the other sync to our own inputs */ -static ir_node *transform_node_Sync(ir_node *n) { +static ir_node *transform_node_Sync(ir_node *n) +{ int arity = get_Sync_n_preds(n); int i; @@ -5866,7 +5965,8 @@ static ir_node *transform_node_Sync(ir_node *n) { /** * optimize a trampoline Call into a direct Call */ -static ir_node *transform_node_Call(ir_node *call) { +static ir_node *transform_node_Call(ir_node *call) +{ ir_node *callee = get_Call_ptr(call); ir_node *adr, *mem, *res, *bl, **in; ir_type *ctp, *mtp, *tp; @@ -5939,7 +6039,8 @@ static ir_node *transform_node_Call(ir_node *call) { * transformations _do_ generate new nodes, and thus the old node must * not be freed even if the equivalent node isn't the old one. */ -static ir_node *transform_node(ir_node *n) { +static ir_node *transform_node(ir_node *n) +{ ir_node *oldn; /* @@ -6032,92 +6133,94 @@ static ir_op_ops *firm_set_default_transform_node(ir_opcode code, ir_op_ops *ops #define N_IR_NODES 512 /** Compares the attributes of two Const nodes. */ -static int node_cmp_attr_Const(ir_node *a, ir_node *b) { +static int node_cmp_attr_Const(ir_node *a, ir_node *b) +{ return (get_Const_tarval(a) != get_Const_tarval(b)) || (get_Const_type(a) != get_Const_type(b)); } /* node_cmp_attr_Const */ /** Compares the attributes of two Proj nodes. */ -static int node_cmp_attr_Proj(ir_node *a, ir_node *b) { - return get_irn_proj_attr(a) != get_irn_proj_attr(b); +static int node_cmp_attr_Proj(ir_node *a, ir_node *b) +{ + return a->attr.proj != b->attr.proj; } /* node_cmp_attr_Proj */ /** Compares the attributes of two Filter nodes. */ -static int node_cmp_attr_Filter(ir_node *a, ir_node *b) { +static int node_cmp_attr_Filter(ir_node *a, ir_node *b) +{ return get_Filter_proj(a) != get_Filter_proj(b); } /* node_cmp_attr_Filter */ /** Compares the attributes of two Alloc nodes. */ -static int node_cmp_attr_Alloc(ir_node *a, ir_node *b) { - const alloc_attr *pa = get_irn_alloc_attr(a); - const alloc_attr *pb = get_irn_alloc_attr(b); +static int node_cmp_attr_Alloc(ir_node *a, ir_node *b) +{ + const alloc_attr *pa = &a->attr.alloc; + const alloc_attr *pb = &b->attr.alloc; return (pa->where != pb->where) || (pa->type != pb->type); } /* node_cmp_attr_Alloc */ /** Compares the attributes of two Free nodes. */ -static int node_cmp_attr_Free(ir_node *a, ir_node *b) { - const free_attr *pa = get_irn_free_attr(a); - const free_attr *pb = get_irn_free_attr(b); +static int node_cmp_attr_Free(ir_node *a, ir_node *b) +{ + const free_attr *pa = &a->attr.free; + const free_attr *pb = &b->attr.free; return (pa->where != pb->where) || (pa->type != pb->type); } /* node_cmp_attr_Free */ /** Compares the attributes of two SymConst nodes. */ -static int node_cmp_attr_SymConst(ir_node *a, ir_node *b) { - const symconst_attr *pa = get_irn_symconst_attr(a); - const symconst_attr *pb = get_irn_symconst_attr(b); +static int node_cmp_attr_SymConst(ir_node *a, ir_node *b) +{ + const symconst_attr *pa = &a->attr.symc; + const symconst_attr *pb = &b->attr.symc; return (pa->kind != pb->kind) || (pa->sym.type_p != pb->sym.type_p) || (pa->tp != pb->tp); } /* node_cmp_attr_SymConst */ /** Compares the attributes of two Call nodes. */ -static int node_cmp_attr_Call(ir_node *a, ir_node *b) { - const call_attr *pa = get_irn_call_attr(a); - const call_attr *pb = get_irn_call_attr(b); +static int node_cmp_attr_Call(ir_node *a, ir_node *b) +{ + const call_attr *pa = &a->attr.call; + const call_attr *pb = &b->attr.call; return (pa->type != pb->type) || (pa->tail_call != pb->tail_call); } /* node_cmp_attr_Call */ /** Compares the attributes of two Sel nodes. */ -static int node_cmp_attr_Sel(ir_node *a, ir_node *b) { +static int node_cmp_attr_Sel(ir_node *a, ir_node *b) +{ const ir_entity *a_ent = get_Sel_entity(a); const ir_entity *b_ent = get_Sel_entity(b); -#if 0 - return - (a_ent->kind != b_ent->kind) || - (a_ent->name != b_ent->name) || - (a_ent->owner != b_ent->owner) || - (a_ent->ld_name != b_ent->ld_name) || - (a_ent->type != b_ent->type); -#endif - /* Matze: inlining of functions can produce 2 entities with same type, - * name, etc. */ return a_ent != b_ent; } /* node_cmp_attr_Sel */ /** Compares the attributes of two Phi nodes. */ -static int node_cmp_attr_Phi(ir_node *a, ir_node *b) { +static int node_cmp_attr_Phi(ir_node *a, ir_node *b) +{ /* we can only enter this function if both nodes have the same number of inputs, hence it is enough to check if one of them is a Phi0 */ if (is_Phi0(a)) { /* check the Phi0 pos attribute */ - return get_irn_phi_attr(a)->u.pos != get_irn_phi_attr(b)->u.pos; + return a->attr.phi.u.pos != b->attr.phi.u.pos; } return 0; } /* node_cmp_attr_Phi */ /** Compares the attributes of two Conv nodes. */ -static int node_cmp_attr_Conv(ir_node *a, ir_node *b) { +static int node_cmp_attr_Conv(ir_node *a, ir_node *b) +{ return get_Conv_strict(a) != get_Conv_strict(b); } /* node_cmp_attr_Conv */ /** Compares the attributes of two Cast nodes. */ -static int node_cmp_attr_Cast(ir_node *a, ir_node *b) { +static int node_cmp_attr_Cast(ir_node *a, ir_node *b) +{ return get_Cast_type(a) != get_Cast_type(b); } /* node_cmp_attr_Cast */ /** Compares the attributes of two Load nodes. */ -static int node_cmp_attr_Load(ir_node *a, ir_node *b) { +static int node_cmp_attr_Load(ir_node *a, ir_node *b) +{ if (get_Load_volatility(a) == volatility_is_volatile || get_Load_volatility(b) == volatility_is_volatile) /* NEVER do CSE on volatile Loads */ @@ -6130,7 +6233,8 @@ static int node_cmp_attr_Load(ir_node *a, ir_node *b) { } /* node_cmp_attr_Load */ /** Compares the attributes of two Store nodes. */ -static int node_cmp_attr_Store(ir_node *a, ir_node *b) { +static int node_cmp_attr_Store(ir_node *a, ir_node *b) +{ /* do not CSE Stores with different alignment. Be conservative. */ if (get_Store_align(a) != get_Store_align(b)) return 1; @@ -6141,9 +6245,10 @@ static int node_cmp_attr_Store(ir_node *a, ir_node *b) { } /* node_cmp_attr_Store */ /** Compares two exception attributes */ -static int node_cmp_exception(ir_node *a, ir_node *b) { - const except_attr *ea = get_irn_except_attr(a); - const except_attr *eb = get_irn_except_attr(b); +static int node_cmp_exception(ir_node *a, ir_node *b) +{ + const except_attr *ea = &a->attr.except; + const except_attr *eb = &b->attr.except; return ea->pin_state != eb->pin_state; } @@ -6151,55 +6256,53 @@ static int node_cmp_exception(ir_node *a, ir_node *b) { #define node_cmp_attr_Bound node_cmp_exception /** Compares the attributes of two Div nodes. */ -static int node_cmp_attr_Div(ir_node *a, ir_node *b) { - const divmod_attr *ma = get_irn_divmod_attr(a); - const divmod_attr *mb = get_irn_divmod_attr(b); +static int node_cmp_attr_Div(ir_node *a, ir_node *b) +{ + const divmod_attr *ma = &a->attr.divmod; + const divmod_attr *mb = &b->attr.divmod; return ma->exc.pin_state != mb->exc.pin_state || ma->resmode != mb->resmode || ma->no_remainder != mb->no_remainder; } /* node_cmp_attr_Div */ /** Compares the attributes of two DivMod nodes. */ -static int node_cmp_attr_DivMod(ir_node *a, ir_node *b) { - const divmod_attr *ma = get_irn_divmod_attr(a); - const divmod_attr *mb = get_irn_divmod_attr(b); +static int node_cmp_attr_DivMod(ir_node *a, ir_node *b) +{ + const divmod_attr *ma = &a->attr.divmod; + const divmod_attr *mb = &b->attr.divmod; return ma->exc.pin_state != mb->exc.pin_state || ma->resmode != mb->resmode; } /* node_cmp_attr_DivMod */ /** Compares the attributes of two Mod nodes. */ -static int node_cmp_attr_Mod(ir_node *a, ir_node *b) { - const divmod_attr *ma = get_irn_divmod_attr(a); - const divmod_attr *mb = get_irn_divmod_attr(b); - return ma->exc.pin_state != mb->exc.pin_state || - ma->resmode != mb->resmode; +static int node_cmp_attr_Mod(ir_node *a, ir_node *b) +{ + return node_cmp_attr_DivMod(a, b); } /* node_cmp_attr_Mod */ /** Compares the attributes of two Quot nodes. */ -static int node_cmp_attr_Quot(ir_node *a, ir_node *b) { - const divmod_attr *ma = get_irn_divmod_attr(a); - const divmod_attr *mb = get_irn_divmod_attr(b); - return ma->exc.pin_state != mb->exc.pin_state || - ma->resmode != mb->resmode; +static int node_cmp_attr_Quot(ir_node *a, ir_node *b) +{ + return node_cmp_attr_DivMod(a, b); } /* node_cmp_attr_Quot */ /** Compares the attributes of two Confirm nodes. */ -static int node_cmp_attr_Confirm(ir_node *a, ir_node *b) { +static int node_cmp_attr_Confirm(ir_node *a, ir_node *b) +{ /* no need to compare the bound, as this is a input */ return (get_Confirm_cmp(a) != get_Confirm_cmp(b)); } /* node_cmp_attr_Confirm */ /** Compares the attributes of two Builtin nodes. */ -static int node_cmp_attr_Builtin(ir_node *a, ir_node *b) { - const builtin_attr *ma = get_irn_builtin_attr(a); - const builtin_attr *mb = get_irn_builtin_attr(b); - +static int node_cmp_attr_Builtin(ir_node *a, ir_node *b) +{ /* no need to compare the type, equal kind means equal type */ - return ma->kind != mb->kind; + return get_Builtin_kind(a) != get_Builtin_kind(b); } /* node_cmp_attr_Builtin */ /** Compares the attributes of two ASM nodes. */ -static int node_cmp_attr_ASM(ir_node *a, ir_node *b) { +static int node_cmp_attr_ASM(ir_node *a, ir_node *b) +{ int i, n; const ir_asm_constraint *ca; const ir_asm_constraint *cb; @@ -6293,7 +6396,8 @@ static ir_op_ops *firm_set_default_node_cmp_attr(ir_opcode code, ir_op_ops *ops) CASE(Dummy); /* FIXME CopyB */ default: - /* leave NULL */; + /* leave NULL */ + break; } return ops; @@ -6304,7 +6408,8 @@ static ir_op_ops *firm_set_default_node_cmp_attr(ir_opcode code, ir_op_ops *ops) * Compare function for two nodes in the value table. Gets two * nodes as parameters. Returns 0 if the nodes are a Common Sub Expression. */ -int identities_cmp(const void *elt, const void *key) { +int identities_cmp(const void *elt, const void *key) +{ ir_node *a = (ir_node *)elt; ir_node *b = (ir_node *)key; int i, irn_arity_a; @@ -6355,22 +6460,26 @@ int identities_cmp(const void *elt, const void *key) { * * @param node The IR-node */ -unsigned ir_node_hash(const ir_node *node) { +unsigned ir_node_hash(const ir_node *node) +{ return node->op->ops.hash(node); } /* ir_node_hash */ -pset *new_identities(void) { +pset *new_identities(void) +{ return new_pset(identities_cmp, N_IR_NODES); } /* new_identities */ -void del_identities(pset *value_table) { +void del_identities(pset *value_table) +{ del_pset(value_table); } /* del_identities */ /* Normalize a node by putting constants (and operands with larger * node index) on the right (operator side). */ -void ir_normalize_node(ir_node *n) { +void ir_normalize_node(ir_node *n) +{ if (is_op_commutative(get_irn_op(n))) { ir_node *l = get_binop_left(n); ir_node *r = get_binop_right(n); @@ -6396,7 +6505,8 @@ void ir_normalize_node(ir_node *n) { * dominance info here: We known, that one block must dominate the other and * following the only block input will allow to find it. */ -static void update_known_irn(ir_node *known_irn, const ir_node *new_ir_node) { +static void update_known_irn(ir_node *known_irn, const ir_node *new_ir_node) +{ ir_node *known_blk, *new_block, *block, *mbh; if (get_opt_global_cse()) { @@ -6446,7 +6556,8 @@ static void update_known_irn(ir_node *known_irn, const ir_node *new_ir_node) { * @return a node that computes the same value as n or n if no such * node could be found */ -ir_node *identify_remember(pset *value_table, ir_node *n) { +ir_node *identify_remember(pset *value_table, ir_node *n) +{ ir_node *nn = NULL; if (!value_table) return n; @@ -6473,7 +6584,8 @@ ir_node *identify_remember(pset *value_table, ir_node *n) { * @param value_table The value table * @param n The node to lookup */ -static inline ir_node *identify_cons(pset *value_table, ir_node *n) { +static inline ir_node *identify_cons(pset *value_table, ir_node *n) +{ ir_node *old = n; n = identify_remember(value_table, n); @@ -6483,13 +6595,15 @@ static inline ir_node *identify_cons(pset *value_table, ir_node *n) { } /* identify_cons */ /* Add a node to the identities value table. */ -void add_identities(pset *value_table, ir_node *node) { +void add_identities(pset *value_table, ir_node *node) +{ if (get_opt_cse() && is_no_Block(node)) identify_remember(value_table, node); } /* add_identities */ /* Visit each node in the value table of a graph. */ -void visit_all_identities(ir_graph *irg, irg_walk_func visit, void *env) { +void visit_all_identities(ir_graph *irg, irg_walk_func visit, void *env) +{ ir_node *node; ir_graph *rem = current_ir_graph; @@ -6503,7 +6617,8 @@ void visit_all_identities(ir_graph *irg, irg_walk_func visit, void *env) { * Garbage in, garbage out. If a node has a dead input, i.e., the * Bad node is input to the node, return the Bad node. */ -static ir_node *gigo(ir_node *node) { +static ir_node *gigo(ir_node *node) +{ int i, irn_arity; ir_op *op = get_irn_op(node); @@ -6599,7 +6714,8 @@ static ir_node *gigo(ir_node *node) { * * current_ir_graph must be set to the graph of the node! */ -ir_node *optimize_node(ir_node *n) { +ir_node *optimize_node(ir_node *n) +{ tarval *tv; ir_node *oldn = n; ir_opcode iro = get_irn_opcode(n); @@ -6715,7 +6831,8 @@ ir_node *optimize_node(ir_node *n) { * nodes lying on the obstack. Remove these by a dead node elimination, * i.e., a copying garbage collection. */ -ir_node *optimize_in_place_2(ir_node *n) { +ir_node *optimize_in_place_2(ir_node *n) +{ tarval *tv; ir_node *oldn = n; ir_opcode iro = get_irn_opcode(n); @@ -6806,7 +6923,8 @@ ir_node *optimize_in_place_2(ir_node *n) { /** * Wrapper for external use, set proper status bits after optimization. */ -ir_node *optimize_in_place(ir_node *n) { +ir_node *optimize_in_place(ir_node *n) +{ /* Handle graph state */ assert(get_irg_phase_state(current_ir_graph) != phase_building); @@ -6824,11 +6942,12 @@ ir_node *optimize_in_place(ir_node *n) { /** * Calculate a hash value of a Const node. */ -static unsigned hash_Const(const ir_node *node) { +static unsigned hash_Const(const ir_node *node) +{ unsigned h; /* special value for const, as they only differ in their tarval. */ - h = HASH_PTR(node->attr.con.tv); + h = HASH_PTR(node->attr.con.tarval); return h; } /* hash_Const */ @@ -6836,7 +6955,8 @@ static unsigned hash_Const(const ir_node *node) { /** * Calculate a hash value of a SymConst node. */ -static unsigned hash_SymConst(const ir_node *node) { +static unsigned hash_SymConst(const ir_node *node) +{ unsigned h; /* all others are pointers */ @@ -6880,7 +7000,8 @@ static ir_op_ops *firm_set_default_hash(ir_opcode code, ir_op_ops *ops) /* * Sets the default operation for an ir_ops. */ -ir_op_ops *firm_set_default_operations(ir_opcode code, ir_op_ops *ops) { +ir_op_ops *firm_set_default_operations(ir_opcode code, ir_op_ops *ops) +{ ops = firm_set_default_hash(code, ops); ops = firm_set_default_computed_value(code, ops); ops = firm_set_default_equivalent_node(code, ops);