X-Git-Url: http://nsz.repo.hu/git/?a=blobdiff_plain;ds=sidebyside;f=ir%2Fbe%2Fbechordal_main.c;h=67f9eaede3621673c2ba34b86f1921fa1b1f5560;hb=fef8dc3eff6468d7d65485af8e1ac9884c504e1a;hp=5e2f38f8587a47e01419ce743033bff25843b3fc;hpb=db82f30bbda008e296e8dfab17f899aa175fc469;p=libfirm diff --git a/ir/be/bechordal_main.c b/ir/be/bechordal_main.c index 5e2f38f85..67f9eaede 100644 --- a/ir/be/bechordal_main.c +++ b/ir/be/bechordal_main.c @@ -22,11 +22,8 @@ * @brief Driver for the chordal register allocator. * @author Sebastian Hack * @date 29.11.2005 - * @version $Id$ */ -#ifdef HAVE_CONFIG_H #include "config.h" -#endif #include #include @@ -35,8 +32,6 @@ #include "pset.h" #include "list.h" #include "bitset.h" -#include "iterator.h" -#include "firm_config.h" #include "lc_opts.h" #include "lc_opts_enum.h" @@ -44,81 +39,62 @@ #include "ircons_t.h" #include "irmode_t.h" #include "irgraph_t.h" -#include "irprintf_t.h" +#include "irprintf.h" #include "irgwalk.h" #include "ircons.h" #include "irdump.h" #include "irdom.h" #include "ircons.h" -#include "irbitset.h" #include "irnode.h" #include "ircons.h" +#include "irtools.h" #include "debug.h" -#include "xmalloc.h" #include "execfreq.h" #include "iredges_t.h" +#include "error.h" #include "bechordal_t.h" #include "beabi.h" -#include "bejavacoal.h" #include "beutil.h" #include "besched.h" -#include "besched_t.h" +#include "besched.h" #include "belive_t.h" -#include "bearch_t.h" -#include "beifg_t.h" -#include "beifg_impl.h" -#include "benode_t.h" -#include "bestatevent.h" +#include "bearch.h" +#include "beifg.h" +#include "benode.h" +#include "statev_t.h" #include "bestat.h" #include "bemodule.h" #include "be_t.h" #include "bera.h" -#include "beirg_t.h" +#include "beirg.h" +#include "bestack.h" #include "bespillslots.h" -#include "bespilloptions.h" +#include "bespill.h" #include "belower.h" -#ifdef WITH_ILP -#include "bespillremat.h" -#endif /* WITH_ILP */ - -#include "bejavacoal.h" #include "becopystat.h" #include "becopyopt.h" #include "bessadestr.h" #include "beverify.h" -#include "benode_t.h" +#include "benode.h" + +#include "bepbqpcoloring.h" static be_ra_chordal_opts_t options = { BE_CH_DUMP_NONE, BE_CH_LOWER_PERM_SWAP, - BE_CH_VRFY_WARN, - "", - "" + BE_CH_VRFY_WARN }; -typedef struct _post_spill_env_t { - be_chordal_env_t cenv; - be_irg_t *birg; - const arch_register_class_t *cls; - double pre_spill_cost; -} post_spill_env_t; - -static be_options_t *main_opts; - static const lc_opt_enum_int_items_t lower_perm_items[] = { { "copy", BE_CH_LOWER_PERM_COPY }, { "swap", BE_CH_LOWER_PERM_SWAP }, { NULL, 0 } }; -static const lc_opt_enum_int_items_t lower_perm_stat_items[] = { - { NULL, 0 } -}; - -static const lc_opt_enum_int_items_t dump_items[] = { +static const lc_opt_enum_mask_items_t dump_items[] = { { "none", BE_CH_DUMP_NONE }, { "spill", BE_CH_DUMP_SPILL }, { "live", BE_CH_DUMP_LIVE }, @@ -126,6 +102,7 @@ static const lc_opt_enum_int_items_t dump_items[] = { { "copymin", BE_CH_DUMP_COPYMIN }, { "ssadestr", BE_CH_DUMP_SSADESTR }, { "tree", BE_CH_DUMP_TREE_INTV }, + { "split", BE_CH_DUMP_SPLIT }, { "constr", BE_CH_DUMP_CONSTR }, { "lower", BE_CH_DUMP_LOWER }, { "spillslots", BE_CH_DUMP_SPILLSLOTS }, @@ -145,7 +122,7 @@ static lc_opt_enum_int_var_t lower_perm_var = { &options.lower_perm_opt, lower_perm_items }; -static lc_opt_enum_int_var_t dump_var = { +static lc_opt_enum_mask_var_t dump_var = { &options.dump_flags, dump_items }; @@ -154,283 +131,196 @@ static lc_opt_enum_int_var_t be_ch_vrfy_var = { }; static const lc_opt_table_entry_t be_chordal_options[] = { - LC_OPT_ENT_ENUM_PTR ("perm", "perm lowering options", &lower_perm_var), + LC_OPT_ENT_ENUM_INT ("perm", "perm lowering options", &lower_perm_var), LC_OPT_ENT_ENUM_MASK("dump", "select dump phases", &dump_var), - LC_OPT_ENT_ENUM_PTR ("vrfy", "verify options", &be_ch_vrfy_var), + LC_OPT_ENT_ENUM_INT ("verify", "verify options", &be_ch_vrfy_var), LC_OPT_LAST }; +static be_module_list_entry_t *colorings = NULL; +static const be_ra_chordal_coloring_t *selected_coloring = NULL; + +void be_register_chordal_coloring(const char *name, be_ra_chordal_coloring_t *coloring) +{ + if (selected_coloring == NULL) + selected_coloring = coloring; + + be_add_module_to_list(&colorings, name, coloring); +} + +static void be_ra_chordal_coloring(be_chordal_env_t *env) +{ + selected_coloring->allocate(env); +} + static void dump(unsigned mask, ir_graph *irg, const arch_register_class_t *cls, - const char *suffix, - void (*dump_func)(ir_graph *, const char *)) + const char *suffix) { - if((options.dump_flags & mask) == mask) { + if ((options.dump_flags & mask) == mask) { if (cls) { char buf[256]; - snprintf(buf, sizeof(buf), "-%s%s", cls->name, suffix); - be_dump(irg, buf, dump_func); + snprintf(buf, sizeof(buf), "%s-%s", cls->name, suffix); + dump_ir_graph(irg, buf); + } else { + dump_ir_graph(irg, suffix); } - else - be_dump(irg, suffix, dump_func); } } /** - * Checks for every reload if it's user can perform the load on itself. + * Post-Walker: Checks for the given reload if has only one user that can perform the + * reload as part of its address mode. + * Fold the reload into the user it that is possible. */ -static void memory_operand_walker(ir_node *irn, void *env) { - be_chordal_env_t *cenv = env; - const arch_env_t *aenv = cenv->birg->main_env->arch_env; - const ir_edge_t *edge, *ne; - ir_node *block; - ir_node *spill; +static void memory_operand_walker(ir_node *irn, void *env) +{ + ir_node *block; + ir_node *spill; + + (void)env; if (! be_is_Reload(irn)) return; /* only use memory operands, if the reload is only used by 1 node */ - if(get_irn_n_edges(irn) > 1) + if (get_irn_n_edges(irn) > 1) return; spill = be_get_Reload_mem(irn); block = get_nodes_block(irn); - foreach_out_edge_safe(irn, edge, ne) { + foreach_out_edge_safe(irn, edge) { ir_node *src = get_edge_src_irn(edge); int pos = get_edge_src_pos(edge); assert(src && "outedges broken!"); - if (get_nodes_block(src) == block && arch_possible_memory_operand(aenv, src, pos)) { - arch_perform_memory_operand(aenv, src, spill, pos); + if (get_nodes_block(src) == block && arch_possible_memory_operand(src, pos)) { + arch_perform_memory_operand(src, spill, pos); } } - /* kill the Reload */ + /* kill the Reload if it was folded */ if (get_irn_n_edges(irn) == 0) { + ir_graph *irg = get_irn_irg(irn); + ir_mode *frame_mode = get_irn_mode(get_irn_n(irn, n_be_Reload_frame)); sched_remove(irn); - set_irn_n(irn, be_pos_Reload_mem, new_Bad()); - set_irn_n(irn, be_pos_Reload_frame, new_Bad()); + set_irn_n(irn, n_be_Reload_mem, new_r_Bad(irg, mode_X)); + set_irn_n(irn, n_be_Reload_frame, new_r_Bad(irg, frame_mode)); } } /** * Starts a walk for memory operands if supported by the backend. */ -static INLINE void check_for_memory_operands(be_chordal_env_t *chordal_env) { - irg_walk_graph(chordal_env->irg, NULL, memory_operand_walker, chordal_env); -} - -/** - * Sorry for doing stats again... - */ -typedef struct _node_stat_t { - unsigned int n_phis; /**< Phis of the current register class. */ - unsigned int n_mem_phis; /**< Memory Phis (Phis with spill operands). */ - unsigned int n_copies; /**< Copies */ - unsigned int n_perms; /**< Perms */ - unsigned int n_spills; /**< Spill nodes */ - unsigned int n_reloads; /**< Reloads */ -} node_stat_t; - -struct node_stat_walker { - node_stat_t *stat; - const arch_env_t *arch_env; -}; - -static void node_stat_walker(ir_node *irn, void *data) -{ - struct node_stat_walker *env = data; - const arch_env_t *aenv = env->arch_env; - - /* if the node is a normal phi */ - if(is_Phi(irn)) { - if (get_irn_mode(irn) == mode_M) { - env->stat->n_mem_phis++; - } else { - env->stat->n_phis++; - } - } else { - arch_irn_class_t classify = arch_irn_classify(aenv, irn); - - if(classify & arch_irn_class_spill) - ++env->stat->n_spills; - if(classify & arch_irn_class_reload) - ++env->stat->n_reloads; - if(classify & arch_irn_class_copy) - ++env->stat->n_copies; - if(classify & arch_irn_class_perm) - ++env->stat->n_perms; - } -} - -static void node_stats(be_irg_t *birg, node_stat_t *stat) +void check_for_memory_operands(ir_graph *irg) { - struct node_stat_walker env; - - memset(stat, 0, sizeof(*stat)); - env.arch_env = birg->main_env->arch_env; - env.stat = stat; - irg_walk_graph(birg->irg, NULL, node_stat_walker, &env); + irg_walk_graph(irg, NULL, memory_operand_walker, NULL); } -static void insn_count_walker(ir_node *irn, void *data) -{ - unsigned long *cnt = data; - - switch(get_irn_opcode(irn)) { - case iro_Proj: - case iro_Phi: - case iro_Start: - case iro_End: - break; - default: - (*cnt)++; - } -} -static unsigned long count_insns(ir_graph *irg) -{ - unsigned long cnt = 0; - irg_walk_graph(irg, insn_count_walker, NULL, &cnt); - return cnt; -} - -static void block_count_walker(ir_node *node, void *data) -{ - unsigned long *cnt = data; - if (node == get_irg_end_block(current_ir_graph)) - return; - (*cnt)++; -} - -static unsigned long count_blocks(ir_graph *irg) -{ - unsigned long cnt = 0; - irg_block_walk_graph(irg, block_count_walker, NULL, &cnt); - return cnt; -} - -static node_stat_t last_node_stat; +static be_node_stats_t last_node_stats; /** * Perform things which need to be done per register class before spilling. */ -static void pre_spill(post_spill_env_t *pse, const arch_register_class_t *cls) +static void pre_spill(be_chordal_env_t *const chordal_env, arch_register_class_t const *const cls, ir_graph *const irg) { - be_chordal_env_t *chordal_env = &pse->cenv; - be_irg_t *birg = pse->birg; - ir_graph *irg = be_get_birg_irg(birg); - const be_main_env_t *main_env = birg->main_env; - - pse->cls = cls; - chordal_env->cls = cls; - chordal_env->border_heads = pmap_create(); - chordal_env->ignore_colors = bitset_malloc(chordal_env->cls->n_regs); - - be_assure_liveness(birg); - be_liveness_assure_chk(be_get_birg_liveness(birg)); + chordal_env->cls = cls; + chordal_env->border_heads = pmap_create(); + chordal_env->allocatable_regs = bitset_malloc(chordal_env->cls->n_regs); - stat_ev_do(pse->pre_spill_cost = be_estimate_irg_costs(irg, main_env->arch_env, birg->exec_freq)); + be_assure_live_chk(irg); /* put all ignore registers into the ignore register set. */ - be_put_ignore_regs(birg, pse->cls, chordal_env->ignore_colors); + be_put_allocatable_regs(irg, cls, chordal_env->allocatable_regs); - BE_TIMER_PUSH(t_ra_constr); - be_pre_spill_prepare_constr(chordal_env); - BE_TIMER_POP(t_ra_constr); + be_timer_push(T_RA_CONSTR); + be_pre_spill_prepare_constr(irg, chordal_env->cls); + be_timer_pop(T_RA_CONSTR); - dump(BE_CH_DUMP_CONSTR, birg->irg, pse->cls, "-constr-pre", dump_ir_block_graph_sched); + dump(BE_CH_DUMP_CONSTR, irg, cls, "constr-pre"); } /** * Perform things which need to be done per register class after spilling. */ -static void post_spill(post_spill_env_t *pse, int iteration) { - be_chordal_env_t *chordal_env = &pse->cenv; - be_irg_t *birg = pse->birg; - ir_graph *irg = birg->irg; - const be_main_env_t *main_env = birg->main_env; - int colors_n = arch_register_class_n_regs(chordal_env->cls); - int allocatable_regs = colors_n - be_put_ignore_regs(birg, chordal_env->cls, NULL); - +static void post_spill(be_chordal_env_t *const chordal_env, ir_graph *const irg) +{ /* some special classes contain only ignore regs, no work to be done */ + int const allocatable_regs = be_get_n_allocatable_regs(irg, chordal_env->cls); if (allocatable_regs > 0) { - stat_ev_dbl("bechordal_spillcosts", be_estimate_irg_costs(irg, main_env->arch_env, birg->exec_freq) - pse->pre_spill_cost); - /* If we have a backend provided spiller, post spill is called in a loop after spilling for each register class. But we only need to fix stack nodes once in this case. */ - BE_TIMER_PUSH(t_ra_spill); - check_for_memory_operands(chordal_env); - if (iteration == 0) { - be_abi_fix_stack_nodes(birg->abi); - } - BE_TIMER_POP(t_ra_spill); + be_timer_push(T_RA_SPILL_APPLY); + check_for_memory_operands(irg); + be_abi_fix_stack_nodes(irg); + be_timer_pop(T_RA_SPILL_APPLY); - BE_TIMER_PUSH(t_verify); /* verify schedule and register pressure */ + be_timer_push(T_VERIFY); if (chordal_env->opts->vrfy_option == BE_CH_VRFY_WARN) { - be_verify_schedule(birg); - be_verify_register_pressure(birg, pse->cls, irg); + be_verify_schedule(irg); + be_verify_register_pressure(irg, chordal_env->cls); } else if (chordal_env->opts->vrfy_option == BE_CH_VRFY_ASSERT) { - assert(be_verify_schedule(birg) && "Schedule verification failed"); - assert(be_verify_register_pressure(birg, pse->cls, irg) + assert(be_verify_schedule(irg) && "Schedule verification failed"); + assert(be_verify_register_pressure(irg, chordal_env->cls) && "Register pressure verification failed"); } - BE_TIMER_POP(t_verify); + be_timer_pop(T_VERIFY); /* Color the graph. */ - BE_TIMER_PUSH(t_ra_color); - be_ra_chordal_color(chordal_env); - BE_TIMER_POP(t_ra_color); + be_timer_push(T_RA_COLOR); + be_ra_chordal_coloring(chordal_env); + be_timer_pop(T_RA_COLOR); - dump(BE_CH_DUMP_CONSTR, irg, pse->cls, "-color", dump_ir_block_graph_sched); + dump(BE_CH_DUMP_CONSTR, irg, chordal_env->cls, "color"); /* Create the ifg with the selected flavor */ - BE_TIMER_PUSH(t_ra_ifg); + be_timer_push(T_RA_IFG); chordal_env->ifg = be_create_ifg(chordal_env); - BE_TIMER_POP(t_ra_ifg); + be_timer_pop(T_RA_IFG); - stat_ev_if { - be_ifg_stat_t stat; - node_stat_t node_stat; + if (stat_ev_enabled) { + be_ifg_stat_t stat; + be_node_stats_t node_stats; - be_ifg_stat(birg, chordal_env->ifg, &stat); + be_ifg_stat(irg, chordal_env->ifg, &stat); stat_ev_dbl("bechordal_ifg_nodes", stat.n_nodes); stat_ev_dbl("bechordal_ifg_edges", stat.n_edges); stat_ev_dbl("bechordal_ifg_comps", stat.n_comps); - node_stats(birg, &node_stat); + be_collect_node_stats(&node_stats, irg); + be_subtract_node_stats(&node_stats, &last_node_stats); + stat_ev_dbl("bechordal_perms_before_coal", - node_stat.n_perms - last_node_stat.n_perms); + node_stats[BE_STAT_PERMS]); stat_ev_dbl("bechordal_copies_before_coal", - node_stat.n_copies - last_node_stat.n_copies); + node_stats[BE_STAT_COPIES]); } - /* copy minimization */ - BE_TIMER_PUSH(t_ra_copymin); + be_timer_push(T_RA_COPYMIN); co_driver(chordal_env); - BE_TIMER_POP(t_ra_copymin); - - dump(BE_CH_DUMP_COPYMIN, irg, pse->cls, "-copymin", dump_ir_block_graph_sched); + be_timer_pop(T_RA_COPYMIN); + dump(BE_CH_DUMP_COPYMIN, irg, chordal_env->cls, "copymin"); /* ssa destruction */ - BE_TIMER_PUSH(t_ra_ssa); + be_timer_push(T_RA_SSA); be_ssa_destruction(chordal_env); - BE_TIMER_POP(t_ra_ssa); + be_timer_pop(T_RA_SSA); - dump(BE_CH_DUMP_SSADESTR, irg, pse->cls, "-ssadestr", dump_ir_block_graph_sched); + dump(BE_CH_DUMP_SSADESTR, irg, chordal_env->cls, "ssadestr"); if (chordal_env->opts->vrfy_option != BE_CH_VRFY_OFF) { - BE_TIMER_PUSH(t_verify); + be_timer_push(T_VERIFY); be_ssa_destruction_check(chordal_env); - BE_TIMER_POP(t_verify); + be_timer_pop(T_VERIFY); } /* the ifg exists only if there are allocatable regs */ @@ -439,155 +329,111 @@ static void post_spill(post_spill_env_t *pse, int iteration) { /* free some always allocated data structures */ pmap_destroy(chordal_env->border_heads); - bitset_free(chordal_env->ignore_colors); + bitset_free(chordal_env->allocatable_regs); } /** * Performs chordal register allocation for each register class on given irg. * - * @param birg Backend irg object - * @return Structure containing timer for the single phases or NULL if no timing requested. + * @param irg the graph + * @return Structure containing timer for the single phases or NULL if no + * timing requested. */ -static void be_ra_chordal_main(be_irg_t *birg) +static void be_ra_chordal_main(ir_graph *irg) { - const be_main_env_t *main_env = birg->main_env; - const arch_env_t *arch_env = main_env->arch_env; - ir_graph *irg = birg->irg; - int j, m; - be_chordal_env_t chordal_env; - struct obstack obst; - - main_opts = main_env->options; - - BE_TIMER_PUSH(t_ra_other); - - BE_TIMER_PUSH(t_ra_prolog); - - be_assure_liveness(birg); - - chordal_env.obst = &obst; - chordal_env.opts = &options; - chordal_env.irg = irg; - chordal_env.birg = birg; - chordal_env.border_heads = NULL; - chordal_env.ifg = NULL; - chordal_env.ignore_colors = NULL; - - obstack_init(&obst); - - BE_TIMER_POP(t_ra_prolog); - - stat_ev_if { - be_stat_ev("bechordal_insns_before", count_insns(irg)); - be_stat_ev("bechordal_blocks_before", count_blocks(irg)); - node_stats(birg, &last_node_stat); + const arch_env_t *arch_env = be_get_irg_arch_env(irg); + int j; + int m; + + be_timer_push(T_RA_OTHER); + + be_chordal_env_t chordal_env; + obstack_init(&chordal_env.obst); + chordal_env.opts = &options; + chordal_env.irg = irg; + chordal_env.border_heads = NULL; + chordal_env.ifg = NULL; + chordal_env.allocatable_regs = NULL; + + if (stat_ev_enabled) { + be_collect_node_stats(&last_node_stats, irg); } - if (! arch_code_generator_has_spiller(birg->cg)) { - /* use one of the generic spiller */ - - /* Perform the following for each register class. */ - for (j = 0, m = arch_env_get_n_reg_class(arch_env); j < m; ++j) { - post_spill_env_t pse; - const arch_register_class_t *cls - = arch_env_get_reg_class(arch_env, j); - - if(arch_register_class_flags(cls) & arch_register_class_flag_manual_ra) - continue; - + /* use one of the generic spiller */ - stat_ev_ctx_push_str("bechordal_cls", cls->name); + /* Perform the following for each register class. */ + for (j = 0, m = arch_env->n_register_classes; j < m; ++j) { + const arch_register_class_t *cls = &arch_env->register_classes[j]; - memcpy(&pse.cenv, &chordal_env, sizeof(chordal_env)); - pse.birg = birg; - pre_spill(&pse, cls); + if (arch_register_class_flags(cls) & arch_register_class_flag_manual_ra) + continue; - BE_TIMER_PUSH(t_ra_spill); - be_do_spill(birg, cls); - BE_TIMER_POP(t_ra_spill); - dump(BE_CH_DUMP_SPILL, irg, pse.cls, "-spill", - dump_ir_block_graph_sched); + stat_ev_ctx_push_str("bechordal_cls", cls->name); - post_spill(&pse, 0); + double pre_spill_cost = 0; + if (stat_ev_enabled) { + be_do_stat_reg_pressure(irg, cls); + pre_spill_cost = be_estimate_irg_costs(irg); + } - stat_ev_if { - node_stat_t node_stat; + pre_spill(&chordal_env, cls, irg); - node_stats(birg, &node_stat); - stat_ev_dbl("bechordal_phis", - node_stat.n_phis - last_node_stat.n_phis); - stat_ev_dbl("bechordal_mem_phis", - node_stat.n_mem_phis - last_node_stat.n_mem_phis); - stat_ev_dbl("bechordal_reloads", - node_stat.n_reloads - last_node_stat.n_reloads); - stat_ev_dbl("bechordal_spills", - node_stat.n_spills - last_node_stat.n_spills); - stat_ev_dbl("bechordal_perms_after_coal", - node_stat.n_perms - last_node_stat.n_perms); - stat_ev_dbl("bechordal_copies_after_coal", - node_stat.n_copies - last_node_stat.n_copies); + be_timer_push(T_RA_SPILL); + be_do_spill(irg, cls); + be_timer_pop(T_RA_SPILL); - last_node_stat = node_stat; - stat_ev_ctx_pop("bechordal_cls"); - } - } - } else { - post_spill_env_t *pse; + dump(BE_CH_DUMP_SPILL, irg, cls, "spill"); - /* the backend has its own spiller */ - m = arch_env_get_n_reg_class(arch_env); + stat_ev_dbl("bechordal_spillcosts", be_estimate_irg_costs(irg) - pre_spill_cost); - pse = alloca(m * sizeof(pse[0])); + post_spill(&chordal_env, irg); - for (j = 0; j < m; ++j) { - memcpy(&pse[j].cenv, &chordal_env, sizeof(chordal_env)); - pse[j].birg = birg; - pre_spill(&pse[j], pse[j].cls); - } + if (stat_ev_enabled) { + be_node_stats_t node_stats; - BE_TIMER_PUSH(t_ra_spill); - arch_code_generator_spill(birg->cg, birg); - BE_TIMER_POP(t_ra_spill); - dump(BE_CH_DUMP_SPILL, irg, NULL, "-spill", dump_ir_block_graph_sched); + be_collect_node_stats(&node_stats, irg); + be_subtract_node_stats(&node_stats, &last_node_stats); + be_emit_node_stats(&node_stats, "bechordal_"); - for (j = 0; j < m; ++j) { - post_spill(&pse[j], j); + be_copy_node_stats(&last_node_stats, &node_stats); + stat_ev_ctx_pop("bechordal_cls"); } } - BE_TIMER_PUSH(t_verify); - be_verify_register_allocation(birg); - BE_TIMER_POP(t_verify); - - BE_TIMER_PUSH(t_ra_epilog); - lower_nodes_after_ra(birg, options.lower_perm_opt & BE_CH_LOWER_PERM_COPY ? 1 : 0); - dump(BE_CH_DUMP_LOWER, irg, NULL, "-belower-after-ra", dump_ir_block_graph_sched); + be_timer_push(T_VERIFY); + if (chordal_env.opts->vrfy_option == BE_CH_VRFY_WARN) { + be_verify_register_allocation(irg); + } else if (chordal_env.opts->vrfy_option == BE_CH_VRFY_ASSERT) { + assert(be_verify_register_allocation(irg) + && "Register allocation invalid"); + } + be_timer_pop(T_VERIFY); - obstack_free(&obst, NULL); - be_liveness_invalidate(be_get_birg_liveness(birg)); - BE_TIMER_POP(t_ra_epilog); + be_timer_push(T_RA_EPILOG); + lower_nodes_after_ra(irg, options.lower_perm_opt == BE_CH_LOWER_PERM_COPY); + dump(BE_CH_DUMP_LOWER, irg, NULL, "belower-after-ra"); - BE_TIMER_POP(t_ra_other); + obstack_free(&chordal_env.obst, NULL); + be_invalidate_live_sets(irg); + be_timer_pop(T_RA_EPILOG); - stat_ev_if { - be_stat_ev("bechordal_insns_after", count_insns(irg)); - } + be_timer_pop(T_RA_OTHER); } -static be_ra_t be_ra_chordal_allocator = { - be_ra_chordal_main, -}; - +BE_REGISTER_MODULE_CONSTRUCTOR(be_init_chordal_main) void be_init_chordal_main(void) { + static be_ra_t be_ra_chordal_allocator = { + be_ra_chordal_main, + }; + lc_opt_entry_t *be_grp = lc_opt_get_grp(firm_opt_get_root(), "be"); lc_opt_entry_t *ra_grp = lc_opt_get_grp(be_grp, "ra"); lc_opt_entry_t *chordal_grp = lc_opt_get_grp(ra_grp, "chordal"); - lc_opt_add_table(chordal_grp, be_chordal_options); - be_register_allocator("chordal", &be_ra_chordal_allocator); -} -BE_REGISTER_MODULE_CONSTRUCTOR(be_init_chordal_main); + lc_opt_add_table(chordal_grp, be_chordal_options); + be_add_module_list_opt(chordal_grp, "coloring", "select coloring method", &colorings, (void**) &selected_coloring); +}