X-Git-Url: http://nsz.repo.hu/git/?a=blobdiff_plain;f=ir%2Fopt%2Fcfopt.c;h=76160e58da0a081a3a2ae6965695759fc59ff7b5;hb=cf946e20e9cac9f8e42e7a1fc93886a75db7a80c;hp=d289c9906e4df74167f86c9a237fc63cfa4cc802;hpb=0756df288fc4e43a3c8fe2a6fe427a6f6d1040c4;p=libfirm diff --git a/ir/opt/cfopt.c b/ir/opt/cfopt.c index d289c9906..76160e58d 100644 --- a/ir/opt/cfopt.c +++ b/ir/opt/cfopt.c @@ -23,9 +23,7 @@ * @author Goetz Lindenmaier, Michael Beck, Sebastian Hack * @version $Id$ */ -#ifdef HAVE_CONFIG_H -# include "config.h" -#endif +#include "config.h" #include "iroptimize.h" @@ -45,13 +43,14 @@ #include "irvrfy.h" #include "iredges.h" -#include "array.h" +#include "array_t.h" #include "irouts.h" #include "irbackedge_t.h" #include "irflag_t.h" #include "firmstat.h" +#include "irpass.h" #include "iropt_dbg.h" @@ -104,7 +103,7 @@ static int remove_senseless_conds(ir_node *bl) { ir_node *cond_j = skip_Proj(pred_j); if (cond_j == cond_i) { - ir_node *jmp = new_r_Jmp(current_ir_graph, get_nodes_block(cond_i)); + ir_node *jmp = new_r_Jmp(get_nodes_block(cond_i)); set_irn_n(bl, i, jmp); set_irn_n(bl, j, new_Bad()); @@ -165,7 +164,7 @@ static void merge_blocks(ir_node *node, void *ctx) { if (!is_Block_dead(b)) { new_block = equivalent_node(b); - while (irn_not_visited(b) && (!is_Block_dead(new_block)) && (new_block != b)) { + while (!irn_visited(b) && !is_Block_dead(new_block) && new_block != b) { /* We would have to run gigo() if new is bad, so we promote it directly below. Nevertheless, we sometimes reach a block the first time through a dataflow node. In this case we optimized the @@ -226,8 +225,11 @@ static void remove_unreachable_blocks_and_conds(ir_node *block, void *env) { *changed |= remove_senseless_conds(block); - /* clear the block mark of all blocks */ - set_Block_removable(block); + /* clear the block mark of all non labeled blocks */ + if (has_Block_entity(block)) + set_Block_non_removable(block); + else + set_Block_removable(block); } /** @@ -243,7 +245,7 @@ static void collect_nodes(ir_node *n, void *ctx) { if (code == iro_Block) { /* mark the block as non-removable if it is labeled */ - if (has_Block_label(n)) + if (has_Block_entity(n)) set_Block_non_removable(n); } else { ir_node *b = get_nodes_block(n); @@ -411,7 +413,7 @@ non_dispensable: */ static void optimize_blocks(ir_node *b, void *ctx) { int i, j, k, n, max_preds, n_preds, p_preds = -1; - ir_node *pred, *phi; + ir_node *pred, *phi, *next; ir_node **in; merge_env *env = ctx; @@ -421,11 +423,12 @@ static void optimize_blocks(ir_node *b, void *ctx) { for (i = 0, k = get_Block_n_cfgpreds(b); i < k; ++i) { max_preds += test_whether_dispensable(b, i); } - in = xmalloc(max_preds * sizeof(*in)); + in = XMALLOCN(ir_node*, max_preds); /*- Fix the Phi nodes of the current block -*/ - for (phi = get_irn_link(b); phi; ) { - assert(get_irn_op(phi) == op_Phi); + for (phi = get_irn_link(b); phi != NULL; phi = next) { + assert(is_Phi(phi)); + next = get_irn_link(phi); /* Find the new predecessors for the Phi */ p_preds = 0; @@ -434,8 +437,7 @@ static void optimize_blocks(ir_node *b, void *ctx) { if (is_Block_dead(pred)) { /* case Phi 1: Do nothing */ - } - else if (is_Block_removable(pred) && Block_not_block_visited(pred)) { + } else if (is_Block_removable(pred) && !Block_block_visited(pred)) { /* case Phi 2: It's an empty block and not yet visited. */ ir_node *phi_pred = get_Phi_pred(phi, i); @@ -445,7 +447,7 @@ static void optimize_blocks(ir_node *b, void *ctx) { if (! is_Bad(get_Block_cfgpred(pred, j))) { if (get_nodes_block(phi_pred) == pred) { /* case Phi 2a: */ - assert(get_irn_op(phi_pred) == op_Phi); /* Block is empty!! */ + assert(is_Phi(phi_pred)); /* Block is empty!! */ in[p_preds++] = get_Phi_pred(phi_pred, j); } else { @@ -468,8 +470,6 @@ static void optimize_blocks(ir_node *b, void *ctx) { else set_irn_in(phi, p_preds, in); env->changed = 1; - - phi = get_irn_link(phi); } /*- This happens only if merge between loop backedge and single loop entry. @@ -478,7 +478,7 @@ static void optimize_blocks(ir_node *b, void *ctx) { for (k = 0, n = get_Block_n_cfgpreds(b); k < n; ++k) { ir_node *predb = get_nodes_block(get_Block_cfgpred(b, k)); - if (is_Block_removable(predb) && Block_not_block_visited(predb)) { + if (is_Block_removable(predb) && !Block_block_visited(predb)) { ir_node *next_phi; /* we found a predecessor block at position k that will be removed */ @@ -505,7 +505,7 @@ static void optimize_blocks(ir_node *b, void *ctx) { if (is_Block_dead(pred)) { /* Do nothing */ - } else if (is_Block_removable(pred) && Block_not_block_visited(pred)) { + } else if (is_Block_removable(pred) && !Block_block_visited(pred)) { /* It's an empty block and not yet visited. */ for (j = 0; j < get_Block_n_cfgpreds(pred); j++) { if (! is_Bad(get_Block_cfgpred(pred, j))) @@ -529,7 +529,7 @@ static void optimize_blocks(ir_node *b, void *ctx) { if (is_Block_dead(pred)) { /* Do nothing */ - } else if (is_Block_removable(pred) && Block_not_block_visited(pred)) { + } else if (is_Block_removable(pred) && !Block_block_visited(pred)) { /* It's an empty block and not yet visited. */ for (j = 0; j < get_Block_n_cfgpreds(pred); j++) { if (! is_Bad(get_Block_cfgpred(pred, j))) @@ -561,9 +561,9 @@ static void optimize_blocks(ir_node *b, void *ctx) { if (is_Block_dead(pred)) { /* case 1: Do nothing */ - } else if (is_Block_removable(pred) && Block_not_block_visited(pred)) { + } else if (is_Block_removable(pred) && !Block_block_visited(pred)) { /* case 2: It's an empty block and not yet visited. */ - assert(get_Block_n_cfgpreds(b) > 1); + assert(get_Block_n_cfgpreds(b) > 1 || has_Block_entity(b)); /* Else it should be optimized by equivalent_node. */ for (j = 0; j < get_Block_n_cfgpreds(pred); j++) { ir_node *pred_X = get_Block_cfgpred(pred, j); @@ -626,9 +626,9 @@ static int handle_switch_cond(ir_node *cond) { if (proj2 == NULL) { /* this Cond has only one Proj: must be the defProj */ - assert(get_Cond_defaultProj(cond) == get_Proj_proj(proj1)); + assert(get_Cond_default_proj(cond) == get_Proj_proj(proj1)); /* convert it into a Jmp */ - jmp = new_r_Jmp(current_ir_graph, blk); + jmp = new_r_Jmp(blk); exchange(proj1, jmp); return 1; } else if (get_irn_link(proj2) == NULL) { @@ -639,12 +639,12 @@ static int handle_switch_cond(ir_node *cond) { if (tv != tarval_bad) { /* we have a constant switch */ long num = get_tarval_long(tv); - long def_num = get_Cond_defaultProj(cond); + long def_num = get_Cond_default_proj(cond); if (def_num == get_Proj_proj(proj1)) { /* first one is the defProj */ if (num == get_Proj_proj(proj2)) { - jmp = new_r_Jmp(current_ir_graph, blk); + jmp = new_r_Jmp(blk); exchange(proj2, jmp); exchange(proj1, new_Bad()); return 1; @@ -652,7 +652,7 @@ static int handle_switch_cond(ir_node *cond) { } else if (def_num == get_Proj_proj(proj2)) { /* second one is the defProj */ if (num == get_Proj_proj(proj1)) { - jmp = new_r_Jmp(current_ir_graph, blk); + jmp = new_r_Jmp(blk); exchange(proj1, jmp); exchange(proj2, new_Bad()); return 1; @@ -660,12 +660,12 @@ static int handle_switch_cond(ir_node *cond) { } else { /* neither: strange, Cond was not optimized so far */ if (num == get_Proj_proj(proj1)) { - jmp = new_r_Jmp(current_ir_graph, blk); + jmp = new_r_Jmp(blk); exchange(proj1, jmp); exchange(proj2, new_Bad()); return 1; } else if (num == get_Proj_proj(proj2)) { - jmp = new_r_Jmp(current_ir_graph, blk); + jmp = new_r_Jmp(blk); exchange(proj2, jmp); exchange(proj1, new_Bad()); return 1; @@ -693,7 +693,7 @@ static int handle_switch_cond(ir_node *cond) { * phase. */ void optimize_cf(ir_graph *irg) { - int i, j, n; + int i, j, n, changed; ir_node **in = NULL; ir_node *cond, *end = get_irg_end(irg); ir_graph *rem = current_ir_graph; @@ -712,7 +712,7 @@ void optimize_cf(ir_graph *irg) { edges_deactivate(irg); /* we use the mark flag to mark removable blocks */ - set_using_block_mark(irg); + ir_reserve_resources(irg, IR_RESOURCE_BLOCK_MARK); restart: env.changed = 0; env.phis_moved = 0; @@ -723,6 +723,7 @@ restart: irg_block_walk_graph(irg, NULL, remove_unreachable_blocks_and_conds, &env.changed); /* fix the keep-alives */ + changed = 0; for (i = 0, n = get_End_n_keepalives(end); i < n; ++i) { ir_node *ka = get_End_keepalive(end, i); @@ -730,22 +731,26 @@ restart: /* do NOT keep dead blocks */ if (is_Block_dead(ka) || get_Block_dom_depth(ka) < 0) { set_End_keepalive(end, i, new_Bad()); - env.changed = 1; + changed = 1; + } + } else { + ir_node *block = get_nodes_block(ka); + + if (is_Bad(block) || is_Block_dead(block) || get_Block_dom_depth(block) < 0) { + /* do NOT keep nodes in dead blocks */ + set_End_keepalive(end, i, new_Bad()); + changed = 1; } - } else if (is_Block_dead(get_nodes_block(ka)) || - get_Block_dom_depth(get_nodes_block(ka)) < 0) { - /* do NOT keep nodes in dead blocks */ - set_End_keepalive(end, i, new_Bad()); - env.changed = 1; } } + env.changed |= changed; - set_using_irn_link(irg); + ir_reserve_resources(irg, IR_RESOURCE_IRN_LINK); env.list = plist_new(); irg_walk(end, merge_blocks, collect_nodes, &env); - clear_using_irn_link(irg); + ir_free_resources(irg, IR_RESOURCE_IRN_LINK); if (env.changed) { /* Handle graph state if was changed. */ @@ -753,6 +758,7 @@ restart: set_irg_doms_inconsistent(irg); set_irg_extblk_inconsistent(irg); set_irg_loopinfo_inconsistent(irg); + set_irg_entity_usage_state(irg, ir_entity_usage_not_computed); env.changed = 0; } @@ -772,51 +778,39 @@ restart: /* Optimize the standard code. */ env.changed = 0; assure_doms(irg); - irg_block_walk(get_irg_end_block(irg), optimize_blocks, remove_simple_blocks, &env); - - /* Walk all keep alives, optimize them if block, add to new in-array - for end if useful. */ - n = get_End_n_keepalives(end); - if (n > 0) - NEW_ARR_A(ir_node *, in, n); + irg_block_walk_graph(irg, optimize_blocks, remove_simple_blocks, &env); /* in rare cases a node may be kept alive more than once, use the visited flag to detect this */ + ir_reserve_resources(irg, IR_RESOURCE_IRN_VISITED); inc_irg_visited(irg); - set_using_irn_visited(irg); - /* fix the keep alive */ - for (i = j = 0; i < n; i++) { + /* fix the keep-alives again */ + changed = 0; + for (i = 0, n = get_End_n_keepalives(end); i < n; ++i) { ir_node *ka = get_End_keepalive(end, i); - if (irn_not_visited(ka)) { - ir_op *op = get_irn_op(ka); - - if ((op == op_Block) && Block_not_block_visited(ka)) { - /* irg_block_walk() will increase the block visited flag, but we must visit only - these blocks that are not visited yet, so decrease it first. */ - set_irg_block_visited(irg, get_irg_block_visited(irg) - 1); - irg_block_walk(ka, optimize_blocks, remove_simple_blocks, &env.changed); - mark_irn_visited(ka); - in[j++] = ka; - } else if (op == op_Phi) { - mark_irn_visited(ka); - /* don't keep alive dead blocks */ - if (! is_Block_dead(get_nodes_block(ka))) - in[j++] = ka; - } else if (is_op_keep(op)) { - mark_irn_visited(ka); - if (! is_Block_dead(get_nodes_block(ka))) - in[j++] = ka; + if (is_Block(ka)) { + /* do NOT keep dead blocks */ + if (is_Block_dead(ka) || get_Block_dom_depth(ka) < 0) { + set_End_keepalive(end, i, new_Bad()); + changed = 1; + } + } else { + ir_node *block = get_nodes_block(ka); + + if (is_Bad(block) || is_Block_dead(block) || get_Block_dom_depth(block) < 0) { + /* do NOT keep nodes in dead blocks */ + set_End_keepalive(end, i, new_Bad()); + changed = 1; } } } - if (j != n) { - set_End_keepalives(end, j, in); - env.changed = 1; - } + env.changed |= changed; + + remove_End_Bads_and_doublets(end); - clear_using_block_mark(irg); - clear_using_irn_visited(irg); + + ir_free_resources(irg, IR_RESOURCE_BLOCK_MARK | IR_RESOURCE_IRN_VISITED); if (env.phis_moved) { /* Bad: when we moved Phi's, we might produce dead Phi nodes @@ -825,6 +819,7 @@ restart: */ n = get_End_n_keepalives(end); if (n > 0) { + NEW_ARR_A(ir_node *, in, n); if (env.changed) { /* Handle graph state if was changed. */ set_irg_outs_inconsistent(irg); @@ -863,6 +858,7 @@ restart: set_irg_doms_inconsistent(irg); set_irg_extblk_inconsistent(irg); set_irg_loopinfo_inconsistent(irg); + set_irg_entity_usage_state(irg, ir_entity_usage_not_computed); } @@ -878,3 +874,9 @@ restart: current_ir_graph = rem; } + +/* Creates an ir_graph pass for optimize_cf. */ +ir_graph_pass_t *optimize_cf_pass(const char *name) +{ + return def_graph_pass(name ? name : "optimize_cf", optimize_cf); +} /* optimize_cf_pass */