X-Git-Url: http://nsz.repo.hu/git/?a=blobdiff_plain;f=ir%2Fbe%2Fbechordal_main.c;h=5e2f38f8587a47e01419ce743033bff25843b3fc;hb=db82f30bbda008e296e8dfab17f899aa175fc469;hp=9fb4870323e4cce5ab47f6c12217f08a06c8416b;hpb=57fb9fc8605ae538d91ca191301e0d3c111cc3ba;p=libfirm diff --git a/ir/be/bechordal_main.c b/ir/be/bechordal_main.c index 9fb487032..5e2f38f85 100644 --- a/ir/be/bechordal_main.c +++ b/ir/be/bechordal_main.c @@ -38,9 +38,8 @@ #include "iterator.h" #include "firm_config.h" -#include -#include -#include +#include "lc_opts.h" +#include "lc_opts_enum.h" #include "ircons_t.h" #include "irmode_t.h" @@ -238,8 +237,6 @@ typedef struct _node_stat_t { struct node_stat_walker { node_stat_t *stat; const arch_env_t *arch_env; - bitset_t *mem_phis; - const arch_register_class_t *cls; }; static void node_stat_walker(ir_node *irn, void *data) @@ -247,57 +244,40 @@ static void node_stat_walker(ir_node *irn, void *data) struct node_stat_walker *env = data; const arch_env_t *aenv = env->arch_env; - if (arch_irn_consider_in_reg_alloc(aenv, env->cls, irn)) { - - /* if the node is a normal phi */ - if(is_Phi(irn)) + /* if the node is a normal phi */ + if(is_Phi(irn)) { + if (get_irn_mode(irn) == mode_M) { + env->stat->n_mem_phis++; + } else { env->stat->n_phis++; + } + } else { + arch_irn_class_t classify = arch_irn_classify(aenv, irn); - else if(arch_irn_classify(aenv, irn) & arch_irn_class_spill) + if(classify & arch_irn_class_spill) ++env->stat->n_spills; - - else if(arch_irn_classify(aenv, irn) & arch_irn_class_reload) + if(classify & arch_irn_class_reload) ++env->stat->n_reloads; - - else if(arch_irn_classify(aenv, irn) & arch_irn_class_copy) + if(classify & arch_irn_class_copy) ++env->stat->n_copies; - - else if(arch_irn_classify(aenv, irn) & arch_irn_class_perm) + if(classify & arch_irn_class_perm) ++env->stat->n_perms; } - - /* a mem phi is a PhiM with a mem phi operand or a Spill operand */ - else if(is_Phi(irn) && get_irn_mode(irn) == mode_M) { - int i; - - for(i = get_irn_arity(irn) - 1; i >= 0; --i) { - ir_node *op = get_irn_n(irn, i); - - if((is_Phi(op) && bitset_contains_irn(env->mem_phis, op)) || (arch_irn_classify(aenv, op) & arch_irn_class_spill)) { - bitset_add_irn(env->mem_phis, irn); - env->stat->n_mem_phis++; - break; - } - } - } } -static void node_stats(be_irg_t *birg, const arch_register_class_t *cls, node_stat_t *stat) +static void node_stats(be_irg_t *birg, node_stat_t *stat) { struct node_stat_walker env; - memset(stat, 0, sizeof(stat[0])); + memset(stat, 0, sizeof(*stat)); env.arch_env = birg->main_env->arch_env; - env.mem_phis = bitset_irg_malloc(birg->irg); env.stat = stat; - env.cls = cls; irg_walk_graph(birg->irg, NULL, node_stat_walker, &env); - bitset_free(env.mem_phis); } static void insn_count_walker(ir_node *irn, void *data) { - int *cnt = data; + unsigned long *cnt = data; switch(get_irn_opcode(irn)) { case iro_Proj: @@ -310,13 +290,30 @@ static void insn_count_walker(ir_node *irn, void *data) } } -static unsigned int count_insns(ir_graph *irg) +static unsigned long count_insns(ir_graph *irg) { - int cnt = 0; + unsigned long cnt = 0; irg_walk_graph(irg, insn_count_walker, NULL, &cnt); return cnt; } +static void block_count_walker(ir_node *node, void *data) +{ + unsigned long *cnt = data; + if (node == get_irg_end_block(current_ir_graph)) + return; + (*cnt)++; +} + +static unsigned long count_blocks(ir_graph *irg) +{ + unsigned long cnt = 0; + irg_block_walk_graph(irg, block_count_walker, NULL, &cnt); + return cnt; +} + +static node_stat_t last_node_stat; + /** * Perform things which need to be done per register class before spilling. */ @@ -326,7 +323,6 @@ static void pre_spill(post_spill_env_t *pse, const arch_register_class_t *cls) be_irg_t *birg = pse->birg; ir_graph *irg = be_get_birg_irg(birg); const be_main_env_t *main_env = birg->main_env; - node_stat_t node_stat; pse->cls = cls; chordal_env->cls = cls; @@ -336,10 +332,7 @@ static void pre_spill(post_spill_env_t *pse, const arch_register_class_t *cls) be_assure_liveness(birg); be_liveness_assure_chk(be_get_birg_liveness(birg)); - stat_ev_ctx_push_str("bechordal_cls", pse->cls->name); - stat_ev_do(node_stats(birg, pse->cls, &node_stat)); stat_ev_do(pse->pre_spill_cost = be_estimate_irg_costs(irg, main_env->arch_env, birg->exec_freq)); - stat_ev_dbl("phis_before_spill", node_stat.n_phis); /* put all ignore registers into the ignore register set. */ be_put_ignore_regs(birg, pse->cls, chordal_env->ignore_colors); @@ -349,8 +342,6 @@ static void pre_spill(post_spill_env_t *pse, const arch_register_class_t *cls) BE_TIMER_POP(t_ra_constr); dump(BE_CH_DUMP_CONSTR, birg->irg, pse->cls, "-constr-pre", dump_ir_block_graph_sched); - - stat_ev_ctx_pop("bechordal_cls"); } /** @@ -361,20 +352,12 @@ static void post_spill(post_spill_env_t *pse, int iteration) { be_irg_t *birg = pse->birg; ir_graph *irg = birg->irg; const be_main_env_t *main_env = birg->main_env; - node_stat_t node_stat; int colors_n = arch_register_class_n_regs(chordal_env->cls); int allocatable_regs = colors_n - be_put_ignore_regs(birg, chordal_env->cls, NULL); /* some special classes contain only ignore regs, no work to be done */ if (allocatable_regs > 0) { - - stat_ev_ctx_push_str("bechordal_cls", pse->cls->name); - stat_ev_do(node_stats(birg, pse->cls, &node_stat)); - stat_ev_dbl("phis_after_spill", node_stat.n_phis); - stat_ev_dbl("mem_phis", node_stat.n_mem_phis); - stat_ev_dbl("reloads", node_stat.n_reloads); - stat_ev_dbl("spills", node_stat.n_spills); - stat_ev_dbl("spillcosts", be_estimate_irg_costs(irg, main_env->arch_env, birg->exec_freq) - pse->pre_spill_cost); + stat_ev_dbl("bechordal_spillcosts", be_estimate_irg_costs(irg, main_env->arch_env, birg->exec_freq) - pse->pre_spill_cost); /* If we have a backend provided spiller, post spill is @@ -394,8 +377,7 @@ static void post_spill(post_spill_env_t *pse, int iteration) { if (chordal_env->opts->vrfy_option == BE_CH_VRFY_WARN) { be_verify_schedule(birg); be_verify_register_pressure(birg, pse->cls, irg); - } - else if (chordal_env->opts->vrfy_option == BE_CH_VRFY_ASSERT) { + } else if (chordal_env->opts->vrfy_option == BE_CH_VRFY_ASSERT) { assert(be_verify_schedule(birg) && "Schedule verification failed"); assert(be_verify_register_pressure(birg, pse->cls, irg) && "Register pressure verification failed"); @@ -414,17 +396,20 @@ static void post_spill(post_spill_env_t *pse, int iteration) { chordal_env->ifg = be_create_ifg(chordal_env); BE_TIMER_POP(t_ra_ifg); - { + stat_ev_if { be_ifg_stat_t stat; - - stat_ev_do(be_ifg_stat(birg, chordal_env->ifg, &stat)); - stat_ev_dbl("ifg_nodes", stat.n_nodes); - stat_ev_dbl("ifg_edges", stat.n_edges); - stat_ev_dbl("ifg_comps", stat.n_comps); - - stat_ev_do(node_stats(birg, pse->cls, &node_stat)); - stat_ev_dbl("perms_before_coal", node_stat.n_perms); - stat_ev_dbl("copies_before_coal", node_stat.n_copies); + node_stat_t node_stat; + + be_ifg_stat(birg, chordal_env->ifg, &stat); + stat_ev_dbl("bechordal_ifg_nodes", stat.n_nodes); + stat_ev_dbl("bechordal_ifg_edges", stat.n_edges); + stat_ev_dbl("bechordal_ifg_comps", stat.n_comps); + + node_stats(birg, &node_stat); + stat_ev_dbl("bechordal_perms_before_coal", + node_stat.n_perms - last_node_stat.n_perms); + stat_ev_dbl("bechordal_copies_before_coal", + node_stat.n_copies - last_node_stat.n_copies); } /* copy minimization */ @@ -437,9 +422,7 @@ static void post_spill(post_spill_env_t *pse, int iteration) { /* ssa destruction */ BE_TIMER_PUSH(t_ra_ssa); - stat_ev_ctx_push_str("berachordal_phase", "ssadestr"); be_ssa_destruction(chordal_env); - stat_ev_ctx_pop("berachordal_phase"); BE_TIMER_POP(t_ra_ssa); dump(BE_CH_DUMP_SSADESTR, irg, pse->cls, "-ssadestr", dump_ir_block_graph_sched); @@ -450,11 +433,6 @@ static void post_spill(post_spill_env_t *pse, int iteration) { BE_TIMER_POP(t_verify); } - stat_ev_do(node_stats(birg, pse->cls, &node_stat)); - stat_ev_dbl("perms_after_coal", node_stat.n_perms); - stat_ev_dbl("copies_after_coal", node_stat.n_copies); - stat_ev_ctx_pop("bechordal_cls"); - /* the ifg exists only if there are allocatable regs */ be_ifg_free(chordal_env->ifg); } @@ -472,9 +450,9 @@ static void post_spill(post_spill_env_t *pse, int iteration) { */ static void be_ra_chordal_main(be_irg_t *birg) { - const be_main_env_t *main_env = birg->main_env; - const arch_isa_t *isa = arch_env_get_isa(main_env->arch_env); - ir_graph *irg = birg->irg; + const be_main_env_t *main_env = birg->main_env; + const arch_env_t *arch_env = main_env->arch_env; + ir_graph *irg = birg->irg; int j, m; be_chordal_env_t chordal_env; struct obstack obst; @@ -485,7 +463,6 @@ static void be_ra_chordal_main(be_irg_t *birg) BE_TIMER_PUSH(t_ra_prolog); - be_assure_dom_front(birg); be_assure_liveness(birg); chordal_env.obst = &obst; @@ -501,22 +478,26 @@ static void be_ra_chordal_main(be_irg_t *birg) BE_TIMER_POP(t_ra_prolog); stat_ev_if { - be_stat_ev("insns_before", count_insns(irg)); + be_stat_ev("bechordal_insns_before", count_insns(irg)); + be_stat_ev("bechordal_blocks_before", count_blocks(irg)); + node_stats(birg, &last_node_stat); } if (! arch_code_generator_has_spiller(birg->cg)) { /* use one of the generic spiller */ /* Perform the following for each register class. */ - for (j = 0, m = arch_isa_get_n_reg_class(isa); j < m; ++j) { + for (j = 0, m = arch_env_get_n_reg_class(arch_env); j < m; ++j) { post_spill_env_t pse; const arch_register_class_t *cls - = arch_isa_get_reg_class(isa, j); + = arch_env_get_reg_class(arch_env, j); if(arch_register_class_flags(cls) & arch_register_class_flag_manual_ra) continue; + stat_ev_ctx_push_str("bechordal_cls", cls->name); + memcpy(&pse.cenv, &chordal_env, sizeof(chordal_env)); pse.birg = birg; pre_spill(&pse, cls); @@ -529,12 +510,33 @@ static void be_ra_chordal_main(be_irg_t *birg) dump_ir_block_graph_sched); post_spill(&pse, 0); + + stat_ev_if { + node_stat_t node_stat; + + node_stats(birg, &node_stat); + stat_ev_dbl("bechordal_phis", + node_stat.n_phis - last_node_stat.n_phis); + stat_ev_dbl("bechordal_mem_phis", + node_stat.n_mem_phis - last_node_stat.n_mem_phis); + stat_ev_dbl("bechordal_reloads", + node_stat.n_reloads - last_node_stat.n_reloads); + stat_ev_dbl("bechordal_spills", + node_stat.n_spills - last_node_stat.n_spills); + stat_ev_dbl("bechordal_perms_after_coal", + node_stat.n_perms - last_node_stat.n_perms); + stat_ev_dbl("bechordal_copies_after_coal", + node_stat.n_copies - last_node_stat.n_copies); + + last_node_stat = node_stat; + stat_ev_ctx_pop("bechordal_cls"); + } } } else { post_spill_env_t *pse; - /* the backend has it's own spiller */ - m = arch_isa_get_n_reg_class(isa); + /* the backend has its own spiller */ + m = arch_env_get_n_reg_class(arch_env); pse = alloca(m * sizeof(pse[0])); @@ -569,10 +571,8 @@ static void be_ra_chordal_main(be_irg_t *birg) BE_TIMER_POP(t_ra_other); stat_ev_if { - be_stat_ev("insns_after", count_insns(irg)); + be_stat_ev("bechordal_insns_after", count_insns(irg)); } - - return; } static be_ra_t be_ra_chordal_allocator = {