X-Git-Url: http://nsz.repo.hu/git/?a=blobdiff_plain;f=ir%2Fbe%2Fbeblocksched.c;h=a647e3484796e17858573327573a12a21e131e20;hb=efdb09dd64658e15cd9fd2599884bfe72bba7fd2;hp=e97b183d93bed2a1d4f6940007074e618a6df347;hpb=0f068c12c1abcb8cdbf4165998d7401141e0e0d6;p=libfirm diff --git a/ir/be/beblocksched.c b/ir/be/beblocksched.c index e97b183d9..a647e3484 100644 --- a/ir/be/beblocksched.c +++ b/ir/be/beblocksched.c @@ -1,13 +1,40 @@ /* - * Author: Matthias Braun, Christoph Mallon - * Date: 27.09.2006 - * Copyright: (c) Universitaet Karlsruhe - * License: This file is protected by GPL - GNU GENERAL PUBLIC LICENSE. - * CVS-Id: $Id$ + * Copyright (C) 1995-2008 University of Karlsruhe. All right reserved. + * + * This file is part of libFirm. + * + * This file may be distributed and/or modified under the terms of the + * GNU General Public License version 2 as published by the Free Software + * Foundation and appearing in the file LICENSE.GPL included in the + * packaging of this file. + * + * Licensees holding valid libFirm Professional Edition licenses may use + * this file in accordance with the libFirm Commercial License. + * Agreement provided with the Software. + * + * This file is provided AS IS with NO WARRANTY OF ANY KIND, INCLUDING THE + * WARRANTY OF DESIGN, MERCHANTABILITY AND FITNESS FOR A PARTICULAR + * PURPOSE. + */ + +/** + * @file + * @brief Block-scheduling strategies. + * @author Matthias Braun, Christoph Mallon + * @date 27.09.2006 + * @version $Id$ + * + * The goals of the greedy (and ILP) algorithm here works by assuming that + * we want to change as many jumps to fallthroughs as possible (executed jumps + * actually, we have to look at the execution frequencies). The algorithms + * do this by collecting execution frequencies of all branches (which is easily + * possible when all critical edges are split) then removes critical edges where + * possible as we don't need and want them anymore now. The algorithms then try + * to change as many edges to fallthroughs as possible, this is done by setting + * a next and prev pointers on blocks. The greedy algorithm sorts the edges by + * execution frequencies and tries to transform them to fallthroughs in this order */ -#ifdef HAVE_CONFIG_H #include "config.h" -#endif /* HAVE_CONFIG_H */ #include "beblocksched.h" @@ -18,35 +45,37 @@ #include "iredges.h" #include "irgwalk.h" +#include "irnode_t.h" #include "irgraph_t.h" #include "irloop.h" #include "irprintf.h" +#include "execfreq.h" #include "irdump_t.h" #include "irtools.h" #include "debug.h" #include "beirgmod.h" +#include "bemodule.h" +#include "be.h" +#include "error.h" -#ifdef WITH_LIBCORE -#include -#include -#include -#endif /* WITH_LIBCORE */ +#include "lc_opts.h" +#include "lc_opts_enum.h" #ifdef WITH_ILP #include #include #endif /* WITH_ILP */ +DEBUG_ONLY(static firm_dbg_module_t *dbg = NULL;) + typedef enum _blocksched_algos_t { - BLOCKSCHED_NAIV, BLOCKSCHED_EXTBB, BLOCKSCHED_GREEDY, BLOCKSCHED_ILP + BLOCKSCHED_NAIV, BLOCKSCHED_GREEDY, BLOCKSCHED_ILP } blocksched_algos_t; static int algo = BLOCKSCHED_GREEDY; -#ifdef WITH_LIBCORE static const lc_opt_enum_int_items_t blockschedalgo_items[] = { { "naiv", BLOCKSCHED_NAIV }, - { "extbb", BLOCKSCHED_EXTBB }, { "greedy", BLOCKSCHED_GREEDY }, #ifdef WITH_ILP { "ilp", BLOCKSCHED_ILP }, @@ -60,9 +89,8 @@ static lc_opt_enum_int_var_t algo_var = { static const lc_opt_table_entry_t be_blocksched_options[] = { LC_OPT_ENT_ENUM_INT ("algo", "the block scheduling algorithm", &algo_var), - { NULL } + LC_OPT_LAST }; -#endif /* * ____ _ @@ -73,29 +101,35 @@ static const lc_opt_table_entry_t be_blocksched_options[] = { * |___/ */ -typedef struct _blocksched_entry_t { - ir_node *block; - struct _blocksched_entry_t *next; - struct _blocksched_entry_t *prev; -} blocksched_entry_t; +typedef struct blocksched_entry_t blocksched_entry_t; +struct blocksched_entry_t { + ir_node *block; + blocksched_entry_t *next; + blocksched_entry_t *prev; +}; -typedef struct _edge_t { +typedef struct edge_t edge_t; +struct edge_t { ir_node *block; /**< source block */ int pos; /**< number of cfg predecessor (target) */ double execfreq; /**< the frequency */ - int highest_execfreq; /**< flag that indicates wether this edge is the edge with the highest - execfreq pointing away from this block */ -} edge_t; + double outedge_penalty_freq; /**< for edges leaving the loop this is the + penality when we make them a + fallthrough. */ + int highest_execfreq; /**< flag that indicates whether this edge is + the edge with the highest execfreq pointing + away from this block */ +}; -typedef struct _blocksched_env_t { +typedef struct blocksched_env_t blocksched_env_t; +struct blocksched_env_t { ir_graph *irg; struct obstack *obst; ir_exec_freq *execfreqs; edge_t *edges; pdeq *worklist; int blockcount; - DEBUG_ONLY(firm_dbg_module_t *dbg;) -} blocksched_env_t; +}; /** * Collect cfg frequencies of all edges between blocks. @@ -103,33 +137,48 @@ typedef struct _blocksched_env_t { */ static void collect_egde_frequency(ir_node *block, void *data) { - blocksched_env_t *env = data; - ir_graph *irg = env->irg; - ir_node *startblock = get_irg_start_block(irg); + blocksched_env_t *env = data; int arity; edge_t edge; blocksched_entry_t *entry; + ir_loop *loop; - entry = obstack_alloc(env->obst, sizeof(entry[0])); + memset(&edge, 0, sizeof(edge)); + + entry = OALLOCZ(env->obst, blocksched_entry_t); entry->block = block; - entry->next = NULL; - entry->prev = NULL; set_irn_link(block, entry); - if (block == startblock) - return; + loop = get_irn_loop(block); - arity = get_irn_arity(block); + arity = get_Block_n_cfgpreds(block); + + if (arity == 0) { + /* must be the start block (or end-block for endless loops), + * everything else is dead code and should be removed by now */ + assert(block == get_irg_start_block(env->irg) + || block == get_irg_end_block(env->irg)); + /* nothing to do here */ + return; + } else if (arity == 1) { + ir_node *pred_block = get_Block_cfgpred_block(block, 0); + ir_loop *pred_loop = get_irn_loop(pred_block); + float freq = (float)get_block_execfreq(env->execfreqs, block); + + /* is it an edge leaving a loop */ + if (get_loop_depth(pred_loop) > get_loop_depth(loop)) { + float pred_freq = (float)get_block_execfreq(env->execfreqs, pred_block); + edge.outedge_penalty_freq = -(pred_freq - freq); + } - if (arity == 1) { edge.block = block; edge.pos = 0; - edge.execfreq = get_block_execfreq(env->execfreqs, block); + edge.execfreq = freq; edge.highest_execfreq = 1; ARR_APP1(edge_t, env->edges, edge); } else { int i; - double highest_execfreq = -1; + double highest_execfreq = -1.0; int highest_edge_num = -1; edge.block = block; @@ -150,7 +199,7 @@ static void collect_egde_frequency(ir_node *block, void *data) } } - if(highest_edge_num >= 0) + if (highest_edge_num >= 0) env->edges[highest_edge_num].highest_execfreq = 1; } } @@ -163,27 +212,54 @@ static int cmp_edges(const void *d1, const void *d2) return QSORT_CMP(e2->execfreq, e1->execfreq); } +static int cmp_edges_outedge_penalty(const void *d1, const void *d2) +{ + const edge_t *e1 = d1; + const edge_t *e2 = d2; + /* reverse sorting as penalties are negative */ + return QSORT_CMP(e1->outedge_penalty_freq, e2->outedge_penalty_freq); +} + +static void clear_loop_links(ir_loop *loop) +{ + int i, n; + + set_loop_link(loop, NULL); + n = get_loop_n_elements(loop); + for (i = 0; i < n; ++i) { + loop_element elem = get_loop_element(loop, i); + if (*elem.kind == k_ir_loop) { + clear_loop_links(elem.son); + } + } +} + static void coalesce_blocks(blocksched_env_t *env) { int i; int edge_count = ARR_LEN(env->edges); + edge_t *edges = env->edges; + + /* sort interblock edges by execution frequency */ + qsort(edges, ARR_LEN(edges), sizeof(edges[0]), cmp_edges); /* run1: only look at jumps */ for (i = 0; i < edge_count; ++i) { - const edge_t *edge = &env->edges[i]; + const edge_t *edge = &edges[i]; ir_node *block = edge->block; + int pos = edge->pos; ir_node *pred_block; blocksched_entry_t *entry, *pred_entry; - /* the block might have been removed already... */ - if (is_Bad(get_Block_cfgpred(block, 0))) - continue; - /* only check edge with highest frequency */ if (! edge->highest_execfreq) continue; - pred_block = get_Block_cfgpred_block(block, edge->pos); + /* the block might have been removed already... */ + if (is_Bad(get_Block_cfgpred(block, 0))) + continue; + + pred_block = get_Block_cfgpred_block(block, pos); entry = get_irn_link(block); pred_entry = get_irn_link(pred_block); @@ -195,24 +271,76 @@ static void coalesce_blocks(blocksched_env_t *env) continue; /* schedule the 2 blocks behind each other */ - DBG((env->dbg, LEVEL_1, "Coalesce (Jump) %+F -> %+F (%.3g)\n", + DB((dbg, LEVEL_1, "Coalesce (Jump) %+F -> %+F (%.3g)\n", pred_entry->block, entry->block, edge->execfreq)); pred_entry->next = entry; entry->prev = pred_entry; } - /* run2: remaining edges */ + /* run2: pick loop fallthroughs */ + clear_loop_links(get_irg_loop(env->irg)); + + qsort(edges, ARR_LEN(edges), sizeof(edges[0]), cmp_edges_outedge_penalty); for (i = 0; i < edge_count; ++i) { - const edge_t *edge = &env->edges[i]; + const edge_t *edge = &edges[i]; ir_node *block = edge->block; + int pos = edge->pos; ir_node *pred_block; blocksched_entry_t *entry, *pred_entry; + ir_loop *loop; + ir_loop *outer_loop; + + /* already seen all loop outedges? */ + if (edge->outedge_penalty_freq == 0) + break; /* the block might have been removed already... */ - if (is_Bad(get_Block_cfgpred(block, 0))) + if (is_Bad(get_Block_cfgpred(block, pos))) continue; - pred_block = get_Block_cfgpred_block(block, edge->pos); + pred_block = get_Block_cfgpred_block(block, pos); + entry = get_irn_link(block); + pred_entry = get_irn_link(pred_block); + + if (pred_entry->next != NULL || entry->prev != NULL) + continue; + + /* we want at most 1 outedge fallthrough per loop */ + loop = get_irn_loop(pred_block); + if (get_loop_link(loop) != NULL) + continue; + + /* schedule the 2 blocks behind each other */ + DB((dbg, LEVEL_1, "Coalesce (Loop Outedge) %+F -> %+F (%.3g)\n", + pred_entry->block, entry->block, edge->execfreq)); + pred_entry->next = entry; + entry->prev = pred_entry; + + /* all loops left have an outedge now */ + outer_loop = get_irn_loop(block); + do { + /* we set loop link to loop to mark it */ + set_loop_link(loop, loop); + loop = get_loop_outer_loop(loop); + } while (loop != outer_loop); + } + + /* sort interblock edges by execution frequency */ + qsort(edges, ARR_LEN(edges), sizeof(edges[0]), cmp_edges); + + /* run3: remaining edges */ + for (i = 0; i < edge_count; ++i) { + const edge_t *edge = &edges[i]; + ir_node *block = edge->block; + int pos = edge->pos; + ir_node *pred_block; + blocksched_entry_t *entry, *pred_entry; + + /* the block might have been removed already... */ + if (is_Bad(get_Block_cfgpred(block, pos))) + continue; + + pred_block = get_Block_cfgpred_block(block, pos); entry = get_irn_link(block); pred_entry = get_irn_link(pred_block); @@ -221,7 +349,7 @@ static void coalesce_blocks(blocksched_env_t *env) continue; /* schedule the 2 blocks behind each other */ - DBG((env->dbg, LEVEL_1, "Coalesce (CondJump) %+F -> %+F (%.3g)\n", + DB((dbg, LEVEL_1, "Coalesce (CondJump) %+F -> %+F (%.3g)\n", pred_entry->block, entry->block, edge->execfreq)); pred_entry->next = entry; entry->prev = pred_entry; @@ -236,13 +364,12 @@ static void pick_block_successor(blocksched_entry_t *entry, blocksched_env_t *en const ir_edge_t *edge; double best_succ_execfreq; - if (irn_visited(block)) + if (irn_visited_else_mark(block)) return; env->blockcount++; - mark_irn_visited(block); - DBG((env->dbg, LEVEL_1, "Pick succ of %+F\n", block)); + DB((dbg, LEVEL_1, "Pick succ of %+F\n", block)); /* put all successors into the worklist */ foreach_block_succ(block, edge) { @@ -267,7 +394,7 @@ static void pick_block_successor(blocksched_entry_t *entry, blocksched_env_t *en if (irn_visited(succ_entry->block)) continue; - DBG((env->dbg, LEVEL_1, "Put %+F into worklist\n", succ_entry->block)); + DB((dbg, LEVEL_1, "Put %+F into worklist\n", succ_entry->block)); pdeq_putr(env->worklist, succ_entry->block); } @@ -276,7 +403,7 @@ static void pick_block_successor(blocksched_entry_t *entry, blocksched_env_t *en return; } - DBG((env->dbg, LEVEL_1, "deciding...\n")); + DB((dbg, LEVEL_1, "deciding...\n")); best_succ_execfreq = -1; /* no successor yet: pick the successor block with the highest execution @@ -301,11 +428,11 @@ static void pick_block_successor(blocksched_entry_t *entry, blocksched_env_t *en } if (succ == NULL) { - DBG((env->dbg, LEVEL_1, "pick from worklist\n")); + DB((dbg, LEVEL_1, "pick from worklist\n")); do { if (pdeq_empty(env->worklist)) { - DBG((env->dbg, LEVEL_1, "worklist empty\n")); + DB((dbg, LEVEL_1, "worklist empty\n")); return; } succ = pdeq_getl(env->worklist); @@ -325,6 +452,7 @@ static blocksched_entry_t *finish_block_schedule(blocksched_env_t *env) ir_node *startblock = get_irg_start_block(irg); blocksched_entry_t *entry = get_irn_link(startblock); + ir_reserve_resources(irg, IR_RESOURCE_IRN_VISITED); inc_irg_visited(irg); env->worklist = new_pdeq(); @@ -332,6 +460,8 @@ static blocksched_entry_t *finish_block_schedule(blocksched_env_t *env) assert(pdeq_empty(env->worklist)); del_pdeq(env->worklist); + ir_free_resources(irg, IR_RESOURCE_IRN_VISITED); + return entry; } @@ -341,14 +471,15 @@ static ir_node **create_blocksched_array(blocksched_env_t *env, blocksched_entry int i = 0; ir_node **block_list; blocksched_entry_t *entry; + (void) env; block_list = NEW_ARR_D(ir_node *, obst, count); - DBG((env->dbg, LEVEL_1, "Blockschedule:\n")); + DB((dbg, LEVEL_1, "Blockschedule:\n")); for (entry = first; entry != NULL; entry = entry->next) { assert(i < count); block_list[i++] = entry->block; - DBG((env->dbg, LEVEL_1, "\t%+F\n", entry->block)); + DB((dbg, LEVEL_1, "\t%+F\n", entry->block)); } assert(i == count); @@ -370,21 +501,23 @@ static ir_node **create_block_schedule_greedy(ir_graph *irg, ir_exec_freq *execf env.edges = NEW_ARR_F(edge_t, 0); env.worklist = NULL; env.blockcount = 0; - FIRM_DBG_REGISTER(env.dbg, "firm.be.blocksched"); + + /* make sure loopinfo is up-to-date */ + if (! (get_irg_loopinfo_state(irg) & loopinfo_cf_consistent)) { + construct_cf_backedges(irg); + } // collect edge execution frequencies irg_block_walk_graph(irg, collect_egde_frequency, NULL, &env); - // sort interblock edges by execution frequency - qsort(env.edges, ARR_LEN(env.edges), sizeof(env.edges[0]), cmp_edges); - - be_remove_empty_blocks(irg); + (void)be_remove_empty_blocks(irg); if (algo != BLOCKSCHED_NAIV) coalesce_blocks(&env); start_entry = finish_block_schedule(&env); - block_list = create_blocksched_array(&env, start_entry, env.blockcount, get_irg_obstack(irg)); + block_list = create_blocksched_array(&env, start_entry, env.blockcount, + be_get_be_obst(irg)); DEL_ARR_F(env.edges); obstack_free(&obst, NULL); @@ -452,7 +585,7 @@ static void collect_egde_frequency_ilp(ir_node *block, void *data) snprintf(name, sizeof(name), "block_out_constr_%ld", get_irn_node_nr(block)); out_count = get_irn_n_edges_kind(block, EDGE_KIND_BLOCK); - entry = obstack_alloc(env->env.obst, sizeof(entry[0])); + entry = OALLOC(env->env.obst, blocksched_ilp_entry_t); entry->block = block; entry->next = NULL; entry->prev = NULL; @@ -469,7 +602,6 @@ static void collect_egde_frequency_ilp(ir_node *block, void *data) } else { int i; - int *edgenums = alloca(sizeof(edgenums[0]) * arity); snprintf(name, sizeof(name), "block_in_constr_%ld", get_irn_node_nr(block)); cst = lpp_add_cst_uniq(env->lpp, name, lpp_greater, arity - 1); @@ -491,11 +623,12 @@ static void collect_egde_frequency_ilp(ir_node *block, void *data) static void coalesce_blocks_ilp(blocksched_ilp_env_t *env) { - int i; - int edge_count = ARR_LEN(env->ilpedges); + int edge_count = ARR_LEN(env->ilpedges); + be_options_t *options = be_get_irg_options(env->env.irg); + int i; /* complete out constraints */ - for(i = 0; i < edge_count; ++i) { + for (i = 0; i < edge_count; ++i) { const ilp_edge_t *edge = &env->ilpedges[i]; ir_node *block = edge->block; ir_node *pred; @@ -508,25 +641,12 @@ static void coalesce_blocks_ilp(blocksched_ilp_env_t *env) pred = get_Block_cfgpred_block(block, edge->pos); entry = get_irn_link(pred); - DBG((env->env.dbg, LEVEL_1, "Adding out cst to %+F from %+F,%d\n", + DB((dbg, LEVEL_1, "Adding out cst to %+F from %+F,%d\n", pred, block, edge->pos)); lpp_set_factor_fast(env->lpp, entry->out_cst, edge->ilpvar, 1.0); } -#if 0 - { - FILE *f; - char fname[256]; - lpp_dump(env->lpp, "lpp.out"); - snprintf(fname, sizeof(fname), "lpp_%s.plain", get_irg_dump_name(env->env.irg)); - f = fopen(fname, "w"); - lpp_dump_plain(env->lpp, f); - fclose(f); - } -#endif - - //lpp_solve_net(env->lpp, main_env->options->ilp_server, main_env->options->ilp_solver); - lpp_solve_net(env->lpp, "i44pc52", "cplex"); + lpp_solve_net(env->lpp, options->ilp_server, options->ilp_solver); assert(lpp_is_sol_valid(env->lpp)); /* Apply results to edges */ @@ -542,7 +662,7 @@ static void coalesce_blocks_ilp(blocksched_ilp_env_t *env) if (is_Bad(get_Block_cfgpred(block, 0))) continue; - is_jump = lpp_get_var_sol(env->lpp, edge->ilpvar); + is_jump = (int)lpp_get_var_sol(env->lpp, edge->ilpvar); if (is_jump) continue; @@ -571,7 +691,6 @@ static ir_node **create_block_schedule_ilp(ir_graph *irg, ir_exec_freq *execfreq env.env.worklist = NULL; env.env.blockcount = 0; env.ilpedges = NEW_ARR_F(ilp_edge_t, 0); - FIRM_DBG_REGISTER(env.env.dbg, "firm.be.blocksched"); env.lpp = new_lpp("blockschedule", lpp_minimize); lpp_set_time_limit(env.lpp, 20); @@ -579,11 +698,13 @@ static ir_node **create_block_schedule_ilp(ir_graph *irg, ir_exec_freq *execfreq irg_block_walk_graph(irg, collect_egde_frequency_ilp, NULL, &env); - be_remove_empty_blocks(irg); + (void)be_remove_empty_blocks(irg); coalesce_blocks_ilp(&env); start_entry = finish_block_schedule(&env.env); - block_list = create_blocksched_array(&env.env, start_entry, env.env.blockcount, get_irg_obstack(irg)); + block_list = create_blocksched_array(&env.env, start_entry, + env.env.blockcount, + be_get_be_obst(irg)); DEL_ARR_F(env.ilpedges); free_lpp(env.lpp); @@ -593,101 +714,6 @@ static ir_node **create_block_schedule_ilp(ir_graph *irg, ir_exec_freq *execfreq } #endif /* WITH_ILP */ -/* - * _____ _ ____ ____ - * | ____|_ _| |_| __ )| __ ) - * | _| \ \/ / __| _ \| _ \ - * | |___ > <| |_| |_) | |_) | - * |_____/_/\_\\__|____/|____/ - * - */ - -/** A simple forward single linked list. */ -typedef struct { - ir_node *start; /**< start of the list */ - ir_node *end; /**< last block in the list */ - unsigned n_blks; /**< number of blocks in the list */ -} anchor; - -static void add_block(anchor *list, ir_node *block) { - if (list->start == NULL) { - list->start = block; - list->end = block; - } - else { - set_irn_link(list->end, block); - list->end = block; - } - - list->n_blks++; -} - -static void create_block_list(ir_node *leader_block, anchor *list) { - int i; - const ir_edge_t *edge; - ir_node *block = NULL; - ir_extblk *extbb = get_Block_extbb(leader_block); - - if (extbb_visited(extbb)) - return; - mark_extbb_visited(extbb); - - for (i = 0; i < get_extbb_n_blocks(extbb); ++i) { - block = get_extbb_block(extbb, i); - add_block(list, block); - } - - assert(block != NULL); - - /* pick successor extbbs */ - foreach_block_succ(block, edge) { - ir_node *succ = get_edge_src_irn(edge); - create_block_list(succ, list); - } - - for (i = 0; i < get_extbb_n_blocks(extbb) - 1; ++i) { - block = get_extbb_block(extbb, i); - - foreach_block_succ(block, edge) { - ir_node *succ = get_edge_src_irn(edge); - create_block_list(succ, list); - } - } -} - -void compute_extbb_execfreqs(ir_graph *irg, ir_exec_freq *execfreqs); - -/* - * Calculates a block schedule. The schedule is stored as a linked - * list starting at the start_block of the irg. - */ -static ir_node **create_extbb_block_schedule(ir_graph *irg, ir_exec_freq *execfreqs) -{ - anchor list; - ir_node **blk_list, *b, *n; - unsigned i; - - /* schedule extended basic blocks */ - compute_extbb_execfreqs(irg, execfreqs); - //compute_extbb(irg); - - list.start = NULL; - list.end = NULL; - list.n_blks = 0; - inc_irg_block_visited(irg); - create_block_list(get_irg_start_block(irg), &list); - - /** create an array, so we can go forward and backward */ - blk_list = NEW_ARR_D(ir_node *, irg->obst,list.n_blks); - - for (i = 0, b = list.start; b; b = n, ++i) { - n = get_irn_link(b); - blk_list[i] = b; - } - - return blk_list; -} - /* * __ __ _ * | \/ | __ _(_)_ __ @@ -696,37 +722,30 @@ static ir_node **create_extbb_block_schedule(ir_graph *irg, ir_exec_freq *execfr * |_| |_|\__,_|_|_| |_| * */ - -#ifdef WITH_LIBCORE -void be_block_schedule_register_options(lc_opt_entry_t *grp) +BE_REGISTER_MODULE_CONSTRUCTOR(be_init_blocksched); +void be_init_blocksched(void) { - static int run_once = 0; - lc_opt_entry_t *blocksched_grp; - - if (run_once) - return; - - run_once = 1; - blocksched_grp = lc_opt_get_grp(grp, "blocksched"); + lc_opt_entry_t *be_grp = lc_opt_get_grp(firm_opt_get_root(), "be"); + lc_opt_entry_t *blocksched_grp = lc_opt_get_grp(be_grp, "blocksched"); lc_opt_add_table(blocksched_grp, be_blocksched_options); + + FIRM_DBG_REGISTER(dbg, "firm.be.blocksched"); } -#endif /* WITH_LIBCORE */ -ir_node **be_create_block_schedule(ir_graph *irg, ir_exec_freq *execfreqs) +ir_node **be_create_block_schedule(ir_graph *irg) { - switch(algo) { + ir_exec_freq *execfreqs = be_get_irg_exec_freq(irg); + + switch (algo) { case BLOCKSCHED_GREEDY: case BLOCKSCHED_NAIV: return create_block_schedule_greedy(irg, execfreqs); - case BLOCKSCHED_EXTBB: - return create_extbb_block_schedule(irg, execfreqs); #ifdef WITH_ILP case BLOCKSCHED_ILP: return create_block_schedule_ilp(irg, execfreqs); #endif /* WITH_ILP */ } - assert(0 && "unknown blocksched algo"); - return NULL; + panic("unknown blocksched algo"); }