X-Git-Url: http://nsz.repo.hu/git/?a=blobdiff_plain;f=ir%2Fopt%2Fcfopt.c;h=d6f3c0ea715ec200c2023b9ecfce677639c05939;hb=762b472fc81c73cf7a1b0041b8cd286b7206d79d;hp=e3a14b48dcfda4af895d9c4e73faed5aeb5fd021;hpb=b597c7fd473086ca6374b2abbdf129f595c156d0;p=libfirm diff --git a/ir/opt/cfopt.c b/ir/opt/cfopt.c index e3a14b48d..d6f3c0ea7 100644 --- a/ir/opt/cfopt.c +++ b/ir/opt/cfopt.c @@ -22,16 +22,19 @@ * @brief Control flow optimizations. * @author Goetz Lindenmaier, Michael Beck, Sebastian Hack * @version $Id$ + * + * Removes Bad control flow predecessors and empty blocks. A block is empty + * if it contains only a Jmp node. Blocks can only be removed if they are not + * needed for the semantics of Phi nodes. Further, we NEVER remove labeled + * blocks (even if we could move the label). */ -#ifdef HAVE_CONFIG_H -# include "config.h" -#endif +#include "config.h" #include "iroptimize.h" #include +#include -#include "plist.h" #include "xmalloc.h" #include "irnode_t.h" #include "irgraph_t.h" @@ -42,192 +45,53 @@ #include "irgwalk.h" #include "irgmod.h" #include "irdump.h" -#include "irvrfy.h" +#include "irverify.h" #include "iredges.h" -#include "array.h" +#include "array_t.h" #include "irouts.h" #include "irbackedge_t.h" #include "irflag_t.h" #include "firmstat.h" +#include "irpass.h" +#include "irphase_t.h" #include "iropt_dbg.h" -/*------------------------------------------------------------------*/ -/* Control flow optimization. */ -/* */ -/* Removes Bad control flow predecessors and empty blocks. A block */ -/* is empty if it contains only a Jmp node. */ -/* Blocks can only be removed if they are not needed for the */ -/* semantics of Phi nodes. */ -/* Further, we NEVER remove labeled blocks (even if we could move */ -/* the label. */ -/*------------------------------------------------------------------*/ - -#define set_Block_removable(block) set_Block_mark(block, 1) -#define set_Block_non_removable(block) set_Block_mark(block, 0) -#define is_Block_removable(block) (get_Block_mark(block) != 0) - -/** - * Replace binary Conds that jumps twice into the same block - * by a simple Jmp. - * E.g. - * @verbatim - * Cond Jmp Bad - * / \ | / - * ProjX True ProjX False ==> | / - * \ / | / - * Block Block - * @endverbatim - * - * Such pattern are the result of if-conversion. - * - * Note that the simple case that Block has only these two - * predecessors are already handled in equivalent_node_Block(). - */ -static int remove_senseless_conds(ir_node *bl) { - int i, j; - int n = get_Block_n_cfgpreds(bl); - int changed = 0; - - for (i = 0; i < n; ++i) { - ir_node *pred_i = get_Block_cfgpred(bl, i); - ir_node *cond_i = skip_Proj(pred_i); - - /* binary Cond */ - if (is_Cond(cond_i) && get_irn_mode(get_Cond_selector(cond_i)) == mode_b) { - - for (j = i + 1; j < n; ++j) { - ir_node *pred_j = get_Block_cfgpred(bl, j); - ir_node *cond_j = skip_Proj(pred_j); - - if (cond_j == cond_i) { - ir_node *jmp = new_r_Jmp(current_ir_graph, get_nodes_block(cond_i)); - set_irn_n(bl, i, jmp); - set_irn_n(bl, j, new_Bad()); - - DBG_OPT_IFSIM2(cond_i, jmp); - changed = 1; - break; - } - } - } - } - return changed; -} - /** An environment for merge_blocks and collect nodes. */ -typedef struct _merge_env { - int changed; /**< Set if the graph was changed. */ - int phis_moved; /**< Set if Phi nodes were moved. */ - plist_t *list; /**< Helper list for all found Switch Conds. */ +typedef struct merge_env { + bool changed; /**< Set if the graph was changed. */ + bool phis_moved; /**< Set if Phi nodes were moved. */ } merge_env; -/** - * Removes Tuples from Block control flow predecessors. - * Optimizes blocks with equivalent_node(). This is tricky, - * as we want to avoid nodes that have as block predecessor Bads. - * Therefore we also optimize at control flow operations, depending - * how we first reach the Block. - */ -static void merge_blocks(ir_node *node, void *ctx) { - int i; - ir_node *new_block; - merge_env *env = ctx; - - /* clear the link field for ALL nodes first */ - set_irn_link(node, NULL); - - if (is_Block(node)) { - /* Remove Tuples */ - for (i = get_Block_n_cfgpreds(node) - 1; i >= 0; --i) { - ir_node *pred = get_Block_cfgpred(node, i); - ir_node *skipped = skip_Tuple(pred); - if (pred != skipped) { - set_Block_cfgpred(node, i, skipped); - env->changed = 1; - } - } - - /* see below */ - new_block = equivalent_node(node); - if (new_block != node && ! is_Block_dead(new_block)) { - exchange(node, new_block); - env->changed = 1; - } - - } else if (get_opt_optimize() && (get_irn_mode(node) == mode_X)) { - /* We will soon visit a block. Optimize it before visiting! */ - ir_node *b = get_nodes_block(skip_Proj(node)); - - if (!is_Block_dead(b)) { - new_block = equivalent_node(b); - - while (irn_not_visited(b) && (!is_Block_dead(new_block)) && (new_block != b)) { - /* We would have to run gigo() if new is bad, so we - promote it directly below. Nevertheless, we sometimes reach a block - the first time through a dataflow node. In this case we optimized the - block as such and have to promote the Bad here. */ - assert((get_opt_control_flow_straightening() || - get_opt_control_flow_weak_simplification()) && - ("strange flag setting")); - exchange(b, new_block); - env->changed = 1; - b = new_block; - new_block = equivalent_node(b); - } - - /* normally, we would create a Bad block here, but this must be - * prevented, so just set it's cf to Bad. - */ - if (is_Block_dead(new_block)) { - exchange(node, new_Bad()); - env->changed = 1; - } - } - } +/** set or reset the removable property of a block. */ +static void set_Block_removable(ir_node *block, bool removable) +{ + set_Block_mark(block, removable); } -/** - * Block walker removing control flow from dead block by - * inspecting dominance info. - * Do not replace blocks by Bad. This optimization shall - * ensure, that all Bad control flow predecessors are - * removed, and no new other Bads are introduced. - * Further removed useless Conds and clear the mark of all blocks. - * - * Must be run in the post walker. - */ -static void remove_unreachable_blocks_and_conds(ir_node *block, void *env) { - int i; - int *changed = env; - - /* Check block predecessors and turn control flow into bad. - Beware of Tuple, kill them. */ - for (i = get_Block_n_cfgpreds(block) - 1; i >= 0; --i) { - ir_node *pred_X = get_Block_cfgpred(block, i); - ir_node *skipped = skip_Tuple(pred_X); - - if (! is_Bad(skipped)) { - ir_node *pred_bl = get_nodes_block(skip_Proj(skipped)); - - if (is_Block_dead(pred_bl) || (get_Block_dom_depth(pred_bl) < 0)) { - set_Block_dead(pred_bl); - exchange(pred_X, new_Bad()); - *changed = 1; - } else if (skipped != pred_X) { - set_Block_cfgpred(block, i, skipped); - *changed = 1; - } - } - } +/** check if a block has the removable property set. */ +static bool is_Block_removable(ir_node *block) +{ + return get_Block_mark(block); +} - *changed |= remove_senseless_conds(block); +/** checks if a given Cond node is a switch Cond. */ +static bool is_switch_Cond(ir_node *cond) +{ + ir_node *sel = get_Cond_selector(cond); + return get_irn_mode(sel) != mode_b; +} - /* clear the block mark of all blocks */ - set_Block_removable(block); +/** Walker: clear link fields and mark all blocks as removable. */ +static void clear_link_and_mark_blocks_removable(ir_node *node, void *ctx) +{ + (void) ctx; + set_irn_link(node, NULL); + if (is_Block(node)) + set_Block_removable(node, true); } /** @@ -237,50 +101,48 @@ static void remove_unreachable_blocks_and_conds(ir_node *block, void *env) { * Links all Proj nodes to their predecessors. * Collects all switch-Conds in a list. */ -static void collect_nodes(ir_node *n, void *ctx) { - ir_opcode code = get_irn_opcode(n); - merge_env *env = ctx; - - if (code == iro_Block) { - /* mark the block as non-removable if it is labeled */ - if (has_Block_label(n)) - set_Block_non_removable(n); - } else { - ir_node *b = get_nodes_block(n); - - if (code == iro_Phi && get_irn_arity(n) > 0) { - /* Collect Phi nodes to compact ins along with block's ins. */ - set_irn_link(n, get_irn_link(b)); - set_irn_link(b, n); - } else if (code != iro_Jmp && !is_Bad(b)) { /* Check for non-empty block. */ - set_Block_non_removable(b); - - if (code == iro_Proj) { /* link Proj nodes */ - ir_node *pred = get_Proj_pred(n); - - set_irn_link(n, get_irn_link(pred)); - set_irn_link(pred, n); - } else if (code == iro_Cond) { - ir_node *sel = get_Cond_selector(n); - if (mode_is_int(get_irn_mode(sel))) { - /* found a switch-Cond, collect */ - plist_insert_back(env->list, n); - } - } +static void collect_nodes(ir_node *n, void *ctx) +{ + ir_node ***switch_conds = (ir_node***)ctx; + + if (is_Phi(n)) { + /* Collect Phi nodes to compact ins along with block's ins. */ + ir_node *block = get_nodes_block(n); + set_irn_link(n, get_irn_link(block)); + set_irn_link(block, n); + } else if (is_Block(n)) { + if (has_Block_entity(n)) { + /* block with a jump label attached cannot be removed. */ + set_Block_removable(n, false); + } + return; + } else if (!is_Jmp(n)) { /* Check for non-empty block. */ + ir_node *block = get_nodes_block(n); + set_Block_removable(block, false); + + if (is_Proj(n)) { + /* link Proj nodes */ + ir_node *pred = get_Proj_pred(n); + set_irn_link(n, get_irn_link(pred)); + set_irn_link(pred, n); + } else if (is_Cond(n) && is_switch_Cond(n)) { + /* found a switch-Cond, collect */ + ARR_APP1(ir_node*, *switch_conds, n); } } } -/** Returns true if pred is predecessor of block. */ -static int is_pred_of(ir_node *pred, ir_node *b) { +/** Returns true if pred is predecessor of block b. */ +static bool is_pred_of(ir_node *pred, ir_node *b) +{ int i; for (i = get_Block_n_cfgpreds(b) - 1; i >= 0; --i) { ir_node *b_pred = get_Block_cfgpred_block(b, i); if (b_pred == pred) - return 1; + return true; } - return 0; + return false; } /** Test whether we can optimize away pred block pos of b. @@ -310,61 +172,67 @@ static int is_pred_of(ir_node *pred, ir_node *b) { * To perform the test for pos, we must regard predecessors before pos * as already removed. **/ -static int test_whether_dispensable(ir_node *b, int pos) { - int i, j, n_preds = 1; - ir_node *pred = get_Block_cfgpred_block(b, pos); - - /* Bad blocks will be optimized away, so we don't need space for them */ - if (is_Block_dead(pred)) - return 0; - - if (is_Block_removable(pred)) { - if (!get_opt_optimize() || !get_opt_control_flow_strong_simplification()) { - /* Mark block so that is will not be removed: optimization is turned off. */ - set_Block_non_removable(pred); - return 1; - } +static unsigned test_whether_dispensable(ir_node *b, int pos) +{ + ir_node *pred = get_Block_cfgpred(b, pos); + ir_node *predb = get_nodes_block(pred); - /* Seems to be empty. At least we detected this in collect_nodes. */ - if (get_irn_link(b) == NULL) { - /* There are no Phi nodes ==> all predecessors are dispensable. */ - n_preds = get_Block_n_cfgpreds(pred); - } else { - /* b's pred blocks and pred's pred blocks must be pairwise disjunct. - Handle all pred blocks with preds < pos as if they were already removed. */ - for (i = 0; i < pos; i++) { - ir_node *b_pred = get_Block_cfgpred_block(b, i); - if (! is_Block_dead(b_pred) && is_Block_removable(b_pred)) { - for (j = get_Block_n_cfgpreds(b_pred) - 1; j >= 0; --j) { - ir_node *b_pred_pred = get_Block_cfgpred_block(b_pred, j); - if (is_pred_of(b_pred_pred, pred)) - goto non_dispensable; - } - } else { - if (is_pred_of(b_pred, pred)) + if (is_Bad(pred) || !is_Block_removable(predb)) + return 1; + + /* can't remove self-loops */ + if (predb == b) + goto non_dispensable; + if (is_unknown_jump(pred)) + goto non_dispensable; + + /* Seems to be empty. At least we detected this in collect_nodes. */ + if (get_irn_link(b) != NULL) { + int n_cfgpreds = get_Block_n_cfgpreds(b); + int i; + /* there are Phi nodes */ + + /* b's pred blocks and pred's pred blocks must be pairwise disjunct. + * Handle all pred blocks with preds < pos as if they were already + * removed. */ + for (i = 0; i < pos; i++) { + ir_node *other_pred = get_Block_cfgpred(b, i); + ir_node *other_predb = get_nodes_block(other_pred); + if (is_Bad(other_pred)) + continue; + if (is_Block_removable(other_predb) + && !Block_block_visited(other_predb)) { + int j; + for (j = get_Block_n_cfgpreds(other_predb) - 1; j >= 0; --j) { + ir_node *other_predpred + = get_Block_cfgpred_block(other_predb, j); + if (is_pred_of(other_predpred, predb)) goto non_dispensable; } + } else if (is_pred_of(other_predb, predb)) { + goto non_dispensable; } - for (i = pos +1; i < get_Block_n_cfgpreds(b); i++) { - ir_node *b_pred = get_Block_cfgpred_block(b, i); - if (is_pred_of(b_pred, pred)) - goto non_dispensable; - } - /* if we get here, the block is dispensable */ - n_preds = get_Block_n_cfgpreds(pred); + } + for (i = pos+1; i < n_cfgpreds; i++) { + ir_node *other_predb = get_Block_cfgpred_block(b, i); + if (is_pred_of(other_predb, predb)) + goto non_dispensable; } } - - return n_preds; + /* we will not dispense already visited blocks */ + if (Block_block_visited(predb)) + return 1; + /* if we get here, the block is dispensable, count useful preds */ + return get_irn_arity(predb); non_dispensable: - set_Block_non_removable(pred); + set_Block_removable(predb, false); return 1; } /** - * This method removed Bad cf predecessors from Blocks and Phis, and removes - * empty blocks. A block is empty if it only contains Phi and Jmp nodes. + * This method removes empty blocks. A block is empty if it only contains Phi + * and Jmp nodes. * * We first adapt Phi nodes, then Block nodes, as we need the old ins * of the Block to adapt the Phi nodes. We do this by computing new @@ -372,20 +240,22 @@ non_dispensable: * for all nodes, not regarding whether there is a possibility for optimization. * * For each predecessor p of a Block b there are three cases: - * -1. The predecessor p is a Bad node: just skip it. The in array of b shrinks by one. - * -2. The predecessor p is empty. Remove p. All predecessors of p are now - * predecessors of b. - * -3. The predecessor p is a block containing useful code. Just keep p as is. + * - The predecessor p is a Bad node: just skip it. The in array of b shrinks + * by one. + * - The predecessor p is empty. Remove p. All predecessors of p are now + * predecessors of b. + * - The predecessor p is a block containing useful code. Just keep p as is. * * For Phi nodes f we have to check the conditions at the Block of f. * For cases 1 and 3 we proceed as for Blocks. For case 2 we can have two * cases: - * -2a: The old predecessor of the Phi f is a Phi pred_f IN THE BLOCK REMOVED. In this - * case we proceed as for blocks. We remove pred_f. All - * predecessors of pred_f now are predecessors of f. - * -2b: The old predecessor of f is NOT in the block removed. It might be a Phi, too. - * We have to replicate f for each predecessor of the removed block. Or, with - * other words, the removed predecessor block has exactly one predecessor. + * -2a: The old predecessor of the Phi f is a Phi pred_f IN THE BLOCK REMOVED. + * In this case we proceed as for blocks. We remove pred_f. All + * predecessors of pred_f now are predecessors of f. + * -2b: The old predecessor of f is NOT in the block removed. It might be a Phi + * too. We have to replicate f for each predecessor of the removed block. + * Or, with other words, the removed predecessor block has exactly one + * predecessor. * * Further there is a special case for self referencing blocks: * @verbatim @@ -406,14 +276,18 @@ non_dispensable: * If there is a Phi in pred_b, but we remove pred_b, we have to generate a * Phi in loop_b, that has the ins of the Phi in pred_b and a self referencing * backedge. - * @@@ It is negotiable whether we should do this ... there might end up a copy - * from the Phi in the loop when removing the Phis. */ -static void optimize_blocks(ir_node *b, void *ctx) { +static void optimize_blocks(ir_node *b, void *ctx) +{ int i, j, k, n, max_preds, n_preds, p_preds = -1; - ir_node *pred, *phi; + ir_node *pred, *phi, *next; ir_node **in; - merge_env *env = ctx; + merge_env *env = (merge_env*)ctx; + + if (get_Block_dom_depth(b) < 0) { + /* ignore unreachable blocks */ + return; + } /* Count the number of predecessor if this block is merged with pred blocks that are empty. */ @@ -421,37 +295,42 @@ static void optimize_blocks(ir_node *b, void *ctx) { for (i = 0, k = get_Block_n_cfgpreds(b); i < k; ++i) { max_preds += test_whether_dispensable(b, i); } - in = xmalloc(max_preds * sizeof(*in)); + in = XMALLOCN(ir_node*, max_preds); /*- Fix the Phi nodes of the current block -*/ - for (phi = get_irn_link(b); phi; ) { - assert(get_irn_op(phi) == op_Phi); + for (phi = (ir_node*)get_irn_link(b); phi != NULL; phi = (ir_node*)next) { + assert(is_Phi(phi)); + next = (ir_node*)get_irn_link(phi); /* Find the new predecessors for the Phi */ p_preds = 0; for (i = 0, n = get_Block_n_cfgpreds(b); i < n; ++i) { + ir_graph *irg = get_irn_irg(b); pred = get_Block_cfgpred_block(b, i); - if (is_Block_dead(pred)) { - /* case Phi 1: Do nothing */ - } - else if (is_Block_removable(pred) && Block_not_block_visited(pred)) { + if (is_Bad(pred)) { + /* case Phi 1: maintain Bads, as somebody else is responsible to remove them */ + in[p_preds++] = new_r_Bad(irg, get_irn_mode(phi)); + } else if (is_Block_removable(pred) && !Block_block_visited(pred)) { /* case Phi 2: It's an empty block and not yet visited. */ ir_node *phi_pred = get_Phi_pred(phi, i); for (j = 0, k = get_Block_n_cfgpreds(pred); j < k; j++) { - /* because of breaking loops, not all predecessors are Bad-clean, - * so we must check this here again */ - if (! is_Bad(get_Block_cfgpred(pred, j))) { - if (get_nodes_block(phi_pred) == pred) { - /* case Phi 2a: */ - assert(get_irn_op(phi_pred) == op_Phi); /* Block is empty!! */ - - in[p_preds++] = get_Phi_pred(phi_pred, j); - } else { - /* case Phi 2b: */ - in[p_preds++] = phi_pred; - } + ir_node *pred_pred = get_Block_cfgpred(pred, j); + + if (is_Bad(pred_pred)) { + in[p_preds++] = new_r_Bad(irg, get_irn_mode(phi)); + continue; + } + + if (get_nodes_block(phi_pred) == pred) { + /* case Phi 2a: */ + assert(is_Phi(phi_pred)); /* Block is empty!! */ + + in[p_preds++] = get_Phi_pred(phi_pred, j); + } else { + /* case Phi 2b: */ + in[p_preds++] = phi_pred; } } } else { @@ -459,57 +338,66 @@ static void optimize_blocks(ir_node *b, void *ctx) { in[p_preds++] = get_Phi_pred(phi, i); } } - assert(p_preds <= max_preds); + assert(p_preds == max_preds); /* Fix the node */ if (p_preds == 1) - /* By removal of Bad ins the Phi might be degenerated. */ exchange(phi, in[0]); else set_irn_in(phi, p_preds, in); - env->changed = 1; - - phi = get_irn_link(phi); + env->changed = true; } /*- This happens only if merge between loop backedge and single loop entry. - Moreover, it is only needed if predb is the direct dominator of b, else there can be no uses - of the Phi's in predb ... -*/ + Moreover, it is only needed if predb is the direct dominator of b, + else there can be no uses of the Phi's in predb ... -*/ for (k = 0, n = get_Block_n_cfgpreds(b); k < n; ++k) { - ir_node *predb = get_nodes_block(get_Block_cfgpred(b, k)); + ir_node *pred = get_Block_cfgpred(b, k); + ir_node *predb = get_nodes_block(pred); + if (is_Bad(pred)) + continue; - if (is_Block_removable(predb) && Block_not_block_visited(predb)) { + if (is_Block_removable(predb) && !Block_block_visited(predb)) { ir_node *next_phi; /* we found a predecessor block at position k that will be removed */ - for (phi = get_irn_link(predb); phi; phi = next_phi) { + for (phi = (ir_node*)get_irn_link(predb); phi; phi = next_phi) { int q_preds = 0; - next_phi = get_irn_link(phi); + next_phi = (ir_node*)get_irn_link(phi); assert(is_Phi(phi)); if (get_Block_idom(b) != predb) { /* predb is not the dominator. There can't be uses of pred's Phi nodes, kill them .*/ - exchange(phi, new_Bad()); + ir_graph *irg = get_irn_irg(b); + ir_mode *mode = get_irn_mode(phi); + exchange(phi, new_r_Bad(irg, mode)); } else { /* predb is the direct dominator of b. There might be uses of the Phi nodes from predb in further block, so move this phi from the predecessor into the block b */ set_nodes_block(phi, b); set_irn_link(phi, get_irn_link(b)); set_irn_link(b, phi); - env->phis_moved = 1; + env->phis_moved = true; /* first, copy all 0..k-1 predecessors */ for (i = 0; i < k; i++) { pred = get_Block_cfgpred_block(b, i); - if (is_Block_dead(pred)) { - /* Do nothing */ - } else if (is_Block_removable(pred) && Block_not_block_visited(pred)) { + if (is_Bad(pred)) { + ir_graph *irg = get_irn_irg(b); + ir_mode *mode = get_irn_mode(phi); + in[q_preds++] = new_r_Bad(irg, mode); + } else if (is_Block_removable(pred) && !Block_block_visited(pred)) { /* It's an empty block and not yet visited. */ for (j = 0; j < get_Block_n_cfgpreds(pred); j++) { - if (! is_Bad(get_Block_cfgpred(pred, j))) + if (! is_Bad(get_Block_cfgpred(pred, j))) { in[q_preds++] = phi; + } else { + ir_graph *irg = get_irn_irg(b); + ir_mode *mode = get_irn_mode(phi); + in[q_preds++] = new_r_Bad(irg, mode); + } } } else { in[q_preds++] = phi; @@ -519,21 +407,27 @@ static void optimize_blocks(ir_node *b, void *ctx) { /* now we are at k, copy the phi predecessors */ pred = get_nodes_block(get_Block_cfgpred(b, k)); for (i = 0; i < get_Phi_n_preds(phi); i++) { - if (! is_Bad(get_Block_cfgpred(pred, i))) - in[q_preds++] = get_Phi_pred(phi, i); + in[q_preds++] = get_Phi_pred(phi, i); } /* and now all the rest */ for (i = k+1; i < get_Block_n_cfgpreds(b); i++) { pred = get_Block_cfgpred_block(b, i); - if (is_Block_dead(pred)) { - /* Do nothing */ - } else if (is_Block_removable(pred) && Block_not_block_visited(pred)) { + if (is_Bad(pred)) { + ir_graph *irg = get_irn_irg(b); + ir_mode *mode = get_irn_mode(phi); + in[q_preds++] = new_r_Bad(irg, mode); + } else if (is_Block_removable(pred) && !Block_block_visited(pred)) { /* It's an empty block and not yet visited. */ for (j = 0; j < get_Block_n_cfgpreds(pred); j++) { - if (! is_Bad(get_Block_cfgpred(pred, j))) + if (! is_Bad(get_Block_cfgpred(pred, j))) { in[q_preds++] = phi; + } else { + ir_graph *irg = get_irn_irg(b); + ir_mode *mode = get_irn_mode(phi); + in[q_preds++] = new_r_Bad(irg, mode); + } } } else { in[q_preds++] = phi; @@ -545,7 +439,7 @@ static void optimize_blocks(ir_node *b, void *ctx) { exchange(phi, in[0]); else set_irn_in(phi, q_preds, in); - env->changed = 1; + env->changed = true; assert(q_preds <= max_preds); // assert(p_preds == q_preds && "Wrong Phi Fix"); @@ -557,273 +451,408 @@ static void optimize_blocks(ir_node *b, void *ctx) { /*- Fix the block -*/ n_preds = 0; for (i = 0; i < get_Block_n_cfgpreds(b); i++) { - pred = get_Block_cfgpred_block(b, i); - - if (is_Block_dead(pred)) { - /* case 1: Do nothing */ - } else if (is_Block_removable(pred) && Block_not_block_visited(pred)) { + ir_node *pred = get_Block_cfgpred(b, i); + ir_node *predb = get_nodes_block(pred); + ir_graph *irg = get_irn_irg(pred); + + /* case 1: Bad predecessor */ + if (is_Bad(pred)) { + in[n_preds++] = new_r_Bad(irg, mode_X); + continue; + } + if (is_Block_removable(predb) && !Block_block_visited(predb)) { /* case 2: It's an empty block and not yet visited. */ - assert(get_Block_n_cfgpreds(b) > 1); - /* Else it should be optimized by equivalent_node. */ - for (j = 0; j < get_Block_n_cfgpreds(pred); j++) { - ir_node *pred_X = get_Block_cfgpred(pred, j); - - /* because of breaking loops, not all predecessors are Bad-clean, - * so we must check this here again */ - if (! is_Bad(pred_X)) - in[n_preds++] = pred_X; + for (j = 0; j < get_Block_n_cfgpreds(predb); j++) { + ir_node *predpred = get_Block_cfgpred(predb, j); + + if (is_Bad(predpred)) { + in[n_preds++] = new_r_Bad(irg, mode_X); + continue; + } + + in[n_preds++] = predpred; } - /* Remove block as it might be kept alive. */ - exchange(pred, b/*new_Bad()*/); + /* Remove block+jump as it might be kept alive. */ + exchange(pred, new_r_Bad(get_irn_irg(b), mode_X)); + exchange(predb, new_r_Bad(get_irn_irg(b), mode_BB)); } else { /* case 3: */ - in[n_preds++] = get_Block_cfgpred(b, i); + in[n_preds++] = pred; } } - assert(n_preds <= max_preds); + assert(n_preds == max_preds); set_irn_in(b, n_preds, in); - env->changed = 1; + env->changed = true; - assert(get_irn_link(b) == NULL || p_preds == -1 || (n_preds == p_preds && "Wrong Phi Fix")); + /* see if phi-fix was correct */ + assert(get_irn_link(b) == NULL || p_preds == -1 || (n_preds == p_preds)); xfree(in); } /** - * Block walker: optimize all blocks using the default optimizations. - * This removes Blocks that with only a Jmp predecessor. - */ -static void remove_simple_blocks(ir_node *block, void *ctx) { - ir_node *new_blk = equivalent_node(block); - merge_env *env = ctx; - - if (new_blk != block) { - exchange(block, new_blk); - env->changed = 1; - } -} - -/** - * Handle pre-optimized table switch Cond's. - * During iropt, all Projs from a switch-Cond are already removed except - * the defProj and maybe the taken one. - * The defProj cannot be removed WITHOUT looking backwards, so we do this here. + * Optimize table-switch Conds. * * @param cond the switch-Cond - * - * @return non-zero if a switch-Cond was optimized - * - * Expects all Proj's linked to the cond node + * @return true if the switch-Cond was optimized */ -static int handle_switch_cond(ir_node *cond) { - ir_node *sel = get_Cond_selector(cond); - - ir_node *proj1 = get_irn_link(cond); - ir_node *proj2 = get_irn_link(proj1); - ir_node *jmp, *blk; - - blk = get_nodes_block(cond); - +static bool handle_switch_cond(ir_node *cond) +{ + ir_node *sel = get_Cond_selector(cond); + ir_node *proj1 = (ir_node*)get_irn_link(cond); + ir_node *proj2 = (ir_node*)get_irn_link(proj1); + ir_node *blk = get_nodes_block(cond); + + /* exactly 1 Proj on the Cond node: must be the defaultProj */ if (proj2 == NULL) { - /* this Cond has only one Proj: must be the defProj */ - assert(get_Cond_defaultProj(cond) == get_Proj_proj(proj1)); + ir_node *jmp = new_r_Jmp(blk); + assert(get_Cond_default_proj(cond) == get_Proj_proj(proj1)); /* convert it into a Jmp */ - jmp = new_r_Jmp(current_ir_graph, blk); exchange(proj1, jmp); - return 1; - } else if (get_irn_link(proj2) == NULL) { - /* We have two Proj's here. Check if the Cond has - a constant argument */ - tarval *tv = value_of(sel); + return true; + } + + /* handle Cond nodes with constant argument. In this case the localopt rules + * should have killed all obviously impossible cases. + * So the only case left to handle here is 1 defaultProj + 1 case + * (this one case should be the one taken) */ + if (get_irn_link(proj2) == NULL) { + ir_tarval *tv = value_of(sel); if (tv != tarval_bad) { /* we have a constant switch */ - long num = get_tarval_long(tv); - long def_num = get_Cond_defaultProj(cond); + long num = get_tarval_long(tv); + long def_num = get_Cond_default_proj(cond); + ir_graph *irg = get_irn_irg(cond); + ir_node *bad = new_r_Bad(irg, mode_X); if (def_num == get_Proj_proj(proj1)) { /* first one is the defProj */ if (num == get_Proj_proj(proj2)) { - jmp = new_r_Jmp(current_ir_graph, blk); + ir_node *jmp = new_r_Jmp(blk); exchange(proj2, jmp); - exchange(proj1, new_Bad()); - return 1; + exchange(proj1, bad); + return true; } } else if (def_num == get_Proj_proj(proj2)) { /* second one is the defProj */ if (num == get_Proj_proj(proj1)) { - jmp = new_r_Jmp(current_ir_graph, blk); + ir_node *jmp = new_r_Jmp(blk); exchange(proj1, jmp); - exchange(proj2, new_Bad()); - return 1; + exchange(proj2, bad); + return true; } } else { /* neither: strange, Cond was not optimized so far */ if (num == get_Proj_proj(proj1)) { - jmp = new_r_Jmp(current_ir_graph, blk); + ir_node *jmp = new_r_Jmp(blk); exchange(proj1, jmp); - exchange(proj2, new_Bad()); - return 1; + exchange(proj2, bad); + return true; } else if (num == get_Proj_proj(proj2)) { - jmp = new_r_Jmp(current_ir_graph, blk); + ir_node *jmp = new_r_Jmp(blk); exchange(proj2, jmp); - exchange(proj1, new_Bad()); - return 1; + exchange(proj1, bad); + return true; } } } } - return 0; + return false; } -/* Optimizations of the control flow that also require changes of Phi nodes. - * - * This optimization performs two passes over the graph. - * - * The first pass collects all Phi nodes in a link list in the block - * nodes. Further it performs simple control flow optimizations. - * Finally it marks all blocks that do not contain useful - * computations, i.e., these blocks might be removed. - * - * The second pass performs the optimizations intended by this algorithm. - * It walks only over block nodes and adapts these and the Phi nodes in these blocks, - * which it finds in a linked list computed by the first pass. +/** + * Optimize boolean Conds, where true and false jump to the same block into a Jmp + * Block must contain no Phi nodes. * - * We use the mark flag to mark removable blocks in the first - * phase. + * Cond + * / \ + * projA projB => Jmp Bad + * \ / \ / + * block block + */ +static bool optimize_pred_cond(ir_node *block, int i, int j) +{ + ir_node *projA, *projB, *cond, *pred_block, *jmp, *bad; + assert(i != j); + + projA = get_Block_cfgpred(block, i); + if (!is_Proj(projA)) return false; + projB = get_Block_cfgpred(block, j); + if (!is_Proj(projB)) return false; + cond = get_Proj_pred(projA); + if (!is_Cond(cond)) return false; + + if (cond != get_Proj_pred(projB)) return false; + if (is_switch_Cond(cond)) return false; + + /* cond should actually be a Jmp */ + pred_block = get_nodes_block(cond); + jmp = new_r_Jmp(pred_block); + bad = new_r_Bad(get_irn_irg(block), mode_X); + + assert(projA != projB); + exchange(projA, jmp); + exchange(projB, bad); + return true; +} + +typedef enum block_flags_t { + BF_HAS_OPERATIONS = 1 << 0, + BF_HAS_PHIS = 1 << 1, + BF_IS_UNKNOWN_JUMP_TARGET = 1 << 2, +} block_flags_t; + +static bool get_phase_flag(ir_phase *block_info, ir_node *block, int flag) +{ + return PTR_TO_INT(phase_get_irn_data(block_info, block)) & flag; +} + +static void set_phase_flag(ir_phase *block_info, ir_node *block, + block_flags_t flag) +{ + int data = PTR_TO_INT(phase_get_irn_data(block_info, block)); + data |= flag; + phase_set_irn_data(block_info, block, INT_TO_PTR(data)); +} + +static bool has_operations(ir_phase *block_info, ir_node *block) +{ + return get_phase_flag(block_info, block, BF_HAS_OPERATIONS); +} + +static void set_has_operations(ir_phase *block_info, ir_node *block) +{ + set_phase_flag(block_info, block, BF_HAS_OPERATIONS); +} + +static bool has_phis(ir_phase *block_info, ir_node *block) +{ + return get_phase_flag(block_info, block, BF_HAS_PHIS); +} + +static void set_has_phis(ir_phase *block_info, ir_node *block) +{ + set_phase_flag(block_info, block, BF_HAS_PHIS); +} + +static bool is_unknown_jump_target(ir_phase *block_info, ir_node *block) +{ + return get_phase_flag(block_info, block, BF_IS_UNKNOWN_JUMP_TARGET); +} + +static void set_is_unknown_jump_target(ir_phase *block_info, ir_node *block) +{ + set_phase_flag(block_info, block, BF_IS_UNKNOWN_JUMP_TARGET); +} + +/** + * Walker: fill block info information. + */ +static void compute_block_info(ir_node *n, void *x) +{ + ir_phase *block_info = (ir_phase *)x; + + if (is_Block(n)) { + int i, max = get_Block_n_cfgpreds(n); + for (i=0; iphase, block)) + return; + + /* optimize Cond predecessors (might produce Bad predecessors) */ + for (i = 0; i < n_preds; ++i) { + for (j = i+1; j < n_preds; ++j) { + optimize_pred_cond(block, i, j); + } + } +} + +/** + * Pre-Block walker: remove empty blocks that are + * predecessors of the current block. */ -void optimize_cf(ir_graph *irg) { +static void remove_empty_blocks(ir_node *block, void *x) +{ + skip_env *env = (skip_env*)x; + int i; + int n_preds = get_Block_n_cfgpreds(block); + + for (i = 0; i < n_preds; ++i) { + ir_node *jmp, *jmp_block, *pred, *pred_block; + + jmp = get_Block_cfgpred(block, i); + if (!is_Jmp(jmp)) + continue; + jmp_block = get_nodes_block(jmp); + if (is_unknown_jump_target(env->phase, jmp_block)) + continue; + if (has_operations(env->phase,jmp_block)) + continue; + /* jmp_block is an empty block! */ + + if (get_Block_n_cfgpreds(jmp_block) != 1) + continue; + pred = get_Block_cfgpred(jmp_block, 0); + exchange(jmp, pred); + env->changed = true; + + /* cleanup: jmp_block might have a Keep edge! */ + pred_block = get_nodes_block(pred); + exchange(jmp_block, pred_block); + } +} + +/* + * Some cfg optimizations, which do not touch Phi nodes + */ +static void cfgopt_ignoring_phis(ir_graph *irg) +{ + ir_phase *block_info = new_phase(irg, NULL); + skip_env env = { false, block_info }; + + irg_walk_graph(irg, compute_block_info, NULL, block_info); + + for (;;) { + env.changed = false; + + /* optimize useless ifs: will not touch empty blocks */ + irg_block_walk_graph(irg, NULL, optimize_ifs, &env); + + /* Remove empty blocks */ + irg_block_walk_graph(irg, remove_empty_blocks, NULL, &env); + if (env.changed) { + set_irg_doms_inconsistent(irg); + /* Removing blocks might enable more useless-if optimizations */ + continue; + } else { + break; + } + } + + phase_free(block_info); +} + +/* Optimizations of the control flow that also require changes of Phi nodes. */ +void optimize_cf(ir_graph *irg) +{ int i, j, n; ir_node **in = NULL; - ir_node *cond, *end = get_irg_end(irg); - ir_graph *rem = current_ir_graph; - plist_element_t *el; + ir_node *end = get_irg_end(irg); + ir_node *new_end; merge_env env; + env.changed = false; + env.phis_moved = false; + assert(get_irg_phase_state(irg) != phase_building); /* if the graph is not pinned, we cannot determine empty blocks */ assert(get_irg_pinned(irg) != op_pin_state_floats && "Control flow optimization need a pinned graph"); - current_ir_graph = irg; - - /* FIXME: control flow opt destroys block edges. So edges are deactivated here. Fix the edges! */ + /* FIXME: control flow opt destroys block edges. So edges are deactivated + * here. Fix the edges! */ edges_deactivate(irg); + cfgopt_ignoring_phis(irg); + /* we use the mark flag to mark removable blocks */ - ir_reserve_resources(irg, IR_RESOURCE_BLOCK_MARK); -restart: - env.changed = 0; - env.phis_moved = 0; + ir_reserve_resources(irg, IR_RESOURCE_BLOCK_MARK | IR_RESOURCE_IRN_LINK); - /* ALWAYS kill unreachable control flow. Backend cannot handle it anyway. - Use dominator info to kill blocks. Also optimize useless Conds. */ - assure_doms(irg); - irg_block_walk_graph(irg, NULL, remove_unreachable_blocks_and_conds, &env.changed); + /* The switch Cond optimization might expose unreachable code, so we loop */ + for (;;) { + int length; + ir_node **switch_conds = NULL; + bool changed = false; - /* fix the keep-alives */ - for (i = 0, n = get_End_n_keepalives(end); i < n; ++i) { - ir_node *ka = get_End_keepalive(end, i); + assure_doms(irg); - if (is_Block(ka)) { - /* do NOT keep dead blocks */ - if (is_Block_dead(ka) || get_Block_dom_depth(ka) < 0) { - set_End_keepalive(end, i, new_Bad()); - env.changed = 1; - } - } else if (is_Block_dead(get_nodes_block(ka)) || - get_Block_dom_depth(get_nodes_block(ka)) < 0) { - /* do NOT keep nodes in dead blocks */ - set_End_keepalive(end, i, new_Bad()); - env.changed = 1; + /* + * This pass collects all Phi nodes in a link list in the block + * nodes. Further it performs simple control flow optimizations. + * Finally it marks all blocks that do not contain useful + * computations, i.e., these blocks might be removed. + */ + switch_conds = NEW_ARR_F(ir_node*, 0); + irg_walk(end, clear_link_and_mark_blocks_removable, collect_nodes, &switch_conds); + + /* handle all collected switch-Conds */ + length = ARR_LEN(switch_conds); + for (i = 0; i < length; ++i) { + ir_node *cond = switch_conds[i]; + changed |= handle_switch_cond(cond); } - } - - ir_reserve_resources(irg, IR_RESOURCE_IRN_LINK); + DEL_ARR_F(switch_conds); - env.list = plist_new(); - irg_walk(end, merge_blocks, collect_nodes, &env); + if (!changed) + break; - ir_free_resources(irg, IR_RESOURCE_IRN_LINK); - - if (env.changed) { - /* Handle graph state if was changed. */ - set_irg_outs_inconsistent(irg); set_irg_doms_inconsistent(irg); set_irg_extblk_inconsistent(irg); - set_irg_loopinfo_inconsistent(irg); - env.changed = 0; + set_irg_entity_usage_state(irg, ir_entity_usage_not_computed); } - /* handle all collected switch-Conds */ - foreach_plist(env.list, el) { - cond = plist_element_get_value(el); - env.changed |= handle_switch_cond(cond); - } - plist_free(env.list); + /* assert due to collect_nodes: + * 1. removable blocks are now marked as such + * 2. phi lists are up to date + */ - if (env.changed) { - /* The Cond optimization might generate unreachable code, so restart if - it happens. */ - goto restart; - } - - /* Optimize the standard code. */ - env.changed = 0; + /* Optimize the standard code. + * It walks only over block nodes and adapts these and the Phi nodes in these + * blocks, which it finds in a linked list computed before. + * */ assure_doms(irg); - irg_block_walk(get_irg_end_block(irg), optimize_blocks, remove_simple_blocks, &env); - - /* Walk all keep alives, optimize them if block, add to new in-array - for end if useful. */ - n = get_End_n_keepalives(end); - if (n > 0) - NEW_ARR_A(ir_node *, in, n); - - /* in rare cases a node may be kept alive more than once, use the visited flag to detect this */ - inc_irg_visited(irg); - ir_reserve_resources(irg, IR_RESOURCE_IRN_VISITED); - - /* fix the keep alive */ - for (i = j = 0; i < n; i++) { - ir_node *ka = get_End_keepalive(end, i); - - if (irn_not_visited(ka)) { - ir_op *op = get_irn_op(ka); - - if ((op == op_Block) && Block_not_block_visited(ka)) { - /* irg_block_walk() will increase the block visited flag, but we must visit only - these blocks that are not visited yet, so decrease it first. */ - set_irg_block_visited(irg, get_irg_block_visited(irg) - 1); - irg_block_walk(ka, optimize_blocks, remove_simple_blocks, &env.changed); - mark_irn_visited(ka); - in[j++] = ka; - } else { - mark_irn_visited(ka); - /* don't keep alive dead blocks */ - if (! is_Block_dead(get_nodes_block(ka))) - in[j++] = ka; - } - } - } - if (j != n) { - set_End_keepalives(end, j, in); - env.changed = 1; + irg_block_walk_graph(irg, optimize_blocks, NULL, &env); + + new_end = optimize_in_place(end); + if (new_end != end) { + set_irg_end(irg, new_end); + end = new_end; } + remove_End_Bads_and_doublets(end); - ir_free_resources(irg, IR_RESOURCE_BLOCK_MARK | IR_RESOURCE_IRN_VISITED); + ir_free_resources(irg, IR_RESOURCE_BLOCK_MARK | IR_RESOURCE_IRN_LINK); if (env.phis_moved) { /* Bad: when we moved Phi's, we might produce dead Phi nodes that are kept-alive. - Some other phases cannot copy with this, so will them. + Some other phases cannot copy with this, so kill them. */ n = get_End_n_keepalives(end); if (n > 0) { - if (env.changed) { - /* Handle graph state if was changed. */ - set_irg_outs_inconsistent(irg); - } + NEW_ARR_A(ir_node *, in, n); assure_irg_outs(irg); for (i = j = 0; i < n; ++i) { @@ -847,29 +876,21 @@ restart: } if (j != n) { set_End_keepalives(end, j, in); - env.changed = 1; + env.changed = true; } } } if (env.changed) { /* Handle graph state if was changed. */ - set_irg_outs_inconsistent(irg); set_irg_doms_inconsistent(irg); set_irg_extblk_inconsistent(irg); - set_irg_loopinfo_inconsistent(irg); - } - - - /* the verifier doesn't work yet with floating nodes */ - if (get_irg_pinned(irg) == op_pin_state_pinned) { - /* after optimize_cf(), only Bad data flow may remain. */ - if (irg_vrfy_bads(irg, BAD_DF | BAD_BLOCK | TUPLE)) { - dump_ir_block_graph(irg, "-vrfy-cf"); - dump_ir_graph(irg, "-vrfy-cf"); - fprintf(stderr, "VRFY_BAD in optimize_cf()\n"); - } + set_irg_entity_usage_state(irg, ir_entity_usage_not_computed); } +} - current_ir_graph = rem; +/* Creates an ir_graph pass for optimize_cf. */ +ir_graph_pass_t *optimize_cf_pass(const char *name) +{ + return def_graph_pass(name ? name : "optimize_cf", optimize_cf); }