+/**
+ * Optimize a * 1 = 1 * a = a.
+ */
+static ir_node *equivalent_node_Mul(ir_node *n)
+{
+ ir_node *oldn = n;
+
+ ir_node *a = get_Mul_left(n);
+ ir_node *b = get_Mul_right(n);
+
+ /* Mul is commutative and has again an other neutral element. */
+ if (classify_tarval(value_of(a)) == TV_CLASSIFY_ONE) {
+ n = b;
+ DBG_OPT_ALGSIM1(oldn, a, b, n);
+ } else if (classify_tarval(value_of(b)) == TV_CLASSIFY_ONE) {
+ n = a;
+ DBG_OPT_ALGSIM1(oldn, a, b, n);
+ }
+ return n;
+}
+
+/**
+ * Optimize a / 1 = a.
+ */
+static ir_node *equivalent_node_Div(ir_node *n)
+{
+ ir_node *a = get_Div_left(n);
+ ir_node *b = get_Div_right(n);
+
+ /* Div is not commutative. */
+ if (classify_tarval(value_of(b)) == TV_CLASSIFY_ONE) { /* div(x, 1) == x */
+ /* Turn Div into a tuple (mem, bad, a) */
+ ir_node *mem = get_Div_mem(n);
+ turn_into_tuple(n, 3);
+ set_Tuple_pred(n, pn_Div_M, mem);
+ set_Tuple_pred(n, pn_Div_X_except, new_Bad()); /* no exception */
+ set_Tuple_pred(n, pn_Div_res, a);
+ }
+ return n;
+}
+
+/**
+ * Optimize a / 1 = a.
+ */
+static ir_node *equivalent_node_DivMod(ir_node *n)
+{
+ ir_node *a = get_DivMod_left(n);
+ ir_node *b = get_DivMod_right(n);
+
+ /* Div is not commutative. */
+ if (classify_tarval(value_of(b)) == TV_CLASSIFY_ONE) { /* div(x, 1) == x */
+ /* Turn DivMod into a tuple (mem, bad, a, 0) */
+ ir_node *mem = get_Div_mem(n);
+ ir_mode *mode = get_irn_mode(b);
+
+ turn_into_tuple(n, 4);
+ set_Tuple_pred(n, pn_DivMod_M, mem);
+ set_Tuple_pred(n, pn_DivMod_X_except, new_Bad()); /* no exception */
+ set_Tuple_pred(n, pn_DivMod_res_div, a);
+ set_Tuple_pred(n, pn_DivMod_res_mod, new_Const(mode, get_mode_null(mode)));
+ }
+ return n;
+}
+
+/**
+ * Optimize a & 0b1...1 = 0b1...1 & a = a & a = a.
+ */
+static ir_node *equivalent_node_And(ir_node *n)
+{
+ ir_node *oldn = n;
+
+ ir_node *a = get_And_left(n);
+ ir_node *b = get_And_right(n);
+
+ if (a == b) {
+ n = a; /* And has it's own neutral element */
+ } else if (classify_tarval(value_of(a)) == TV_CLASSIFY_ALL_ONE) {
+ n = b;
+ DBG_OPT_ALGSIM1(oldn, a, b, n);
+ } else if (classify_tarval(value_of(b)) == TV_CLASSIFY_ALL_ONE) {
+ n = a;
+ DBG_OPT_ALGSIM1(oldn, a, b, n);
+ }
+ return n;
+}
+
+/**
+ * Try to remove useless conv's:
+ */
+static ir_node *equivalent_node_Conv(ir_node *n)
+{
+ ir_node *oldn = n;
+ ir_node *a = get_Conv_op(n);
+ ir_node *b;
+
+ ir_mode *n_mode = get_irn_mode(n);
+ ir_mode *a_mode = get_irn_mode(a);
+
+ if (n_mode == a_mode) { /* No Conv necessary */
+ n = a;
+ DBG_OPT_ALGSIM3(oldn, a, n);
+ } else if (get_irn_op(a) == op_Conv) { /* Conv(Conv(b)) */
+ ir_mode *b_mode;
+
+ b = get_Conv_op(a);
+ n_mode = get_irn_mode(n);
+ b_mode = get_irn_mode(b);
+
+ if (n_mode == b_mode) {
+ if (n_mode == mode_b) {
+ n = b; /* Convb(Conv*(xxxb(...))) == xxxb(...) */
+ DBG_OPT_ALGSIM1(oldn, a, b, n);
+ }
+ else if (mode_is_int(n_mode) || mode_is_character(n_mode)) {
+ if (smaller_mode(b_mode, a_mode)){
+ n = b; /* ConvS(ConvL(xxxS(...))) == xxxS(...) */
+ DBG_OPT_ALGSIM1(oldn, a, b, n);
+ }
+ }
+ }
+ }
+ return n;
+}
+
+static ir_node *equivalent_node_Cast(ir_node *n) {
+ ir_node *pred = get_Cast_op(n);
+ if (get_irn_type(pred) == get_Cast_type(n))
+ n = pred;
+ return n;
+}
+
+static ir_node *equivalent_node_Phi(ir_node *n)
+{
+ /* Several optimizations:
+ - no Phi in start block.
+ - remove Id operators that are inputs to Phi
+ - fold Phi-nodes, iff they have only one predecessor except
+ themselves.
+ */
+ int i, n_preds;
+
+ ir_node *oldn = n;
+ ir_node *block = NULL; /* to shutup gcc */
+ ir_node *first_val = NULL; /* to shutup gcc */
+ ir_node *scnd_val = NULL; /* to shutup gcc */
+
+ if (!get_opt_normalize()) return n;
+
+ n_preds = get_Phi_n_preds(n);
+
+ block = get_nodes_block(n);
+ /* @@@ fliegt 'raus, sollte aber doch immer wahr sein!!!
+ assert(get_irn_arity(block) == n_preds && "phi in wrong block!"); */
+ if ((is_Bad(block)) || /* Control dead */
+ (block == current_ir_graph->start_block)) /* There should be no Phi nodes */
+ return new_Bad(); /* in the Start Block. */
+
+ if (n_preds == 0) return n; /* Phi of dead Region without predecessors. */
+
+#if 0
+ /* first we test for a special case: */
+ /* Confirm is a special node fixing additional information for a
+ value that is known at a certain point. This is useful for
+ dataflow analysis. */
+ if (n_preds == 2) {
+ ir_node *a = get_Phi_pred(n, 0);
+ ir_node *b = get_Phi_pred(n, 1);
+ if ( (get_irn_op(a) == op_Confirm)
+ && (get_irn_op(b) == op_Confirm)
+ && follow_Id (get_irn_n(a, 0) == get_irn_n(b, 0))
+ && (get_irn_n(a, 1) == get_irn_n (b, 1))
+ && (a->data.num == (~b->data.num & irpn_True) )) {
+ return get_irn_n(a, 0);
+ }
+ }
+#endif
+
+ /* If the Block has a Bad pred, we also have one. */
+ for (i = 0; i < n_preds; ++i)
+ if (is_Bad (get_Block_cfgpred(block, i)))
+ set_Phi_pred(n, i, new_Bad());
+
+ /* Find first non-self-referencing input */
+ for (i = 0; i < n_preds; ++i) {
+ first_val = get_Phi_pred(n, i);
+ if ( (first_val != n) /* not self pointer */
+#if 1
+ && (get_irn_op(first_val) != op_Bad)
+#endif
+ ) { /* value not dead */
+ break; /* then found first value. */
+ }
+ }
+
+ /* A totally Bad or self-referencing Phi (we didn't break the above loop) */
+ if (i >= n_preds) { return new_Bad(); }
+
+ scnd_val = NULL;
+
+ /* follow_Id () for rest of inputs, determine if any of these
+ are non-self-referencing */
+ while (++i < n_preds) {
+ scnd_val = get_Phi_pred(n, i);
+ if ( (scnd_val != n)
+ && (scnd_val != first_val)
+#if 1
+ && (get_irn_op(scnd_val) != op_Bad)
+#endif
+ ) {
+ break;
+ }
+ }