* @date 29.11.2005
* @version $Id$
*/
-#ifdef HAVE_CONFIG_H
#include "config.h"
-#endif
#include <stdlib.h>
#include <time.h>
#include "list.h"
#include "bitset.h"
#include "iterator.h"
-#include "firm_config.h"
#include "lc_opts.h"
#include "lc_opts_enum.h"
#include "irnode.h"
#include "ircons.h"
#include "debug.h"
-#include "xmalloc.h"
#include "execfreq.h"
#include "iredges_t.h"
#include "bechordal_t.h"
#include "beabi.h"
-#include "bejavacoal.h"
#include "beutil.h"
#include "besched.h"
#include "besched_t.h"
#include "bespilloptions.h"
#include "belower.h"
-#ifdef WITH_ILP
-#include "bespillremat.h"
-#endif /* WITH_ILP */
-
-#include "bejavacoal.h"
#include "becopystat.h"
#include "becopyopt.h"
#include "bessadestr.h"
double pre_spill_cost;
} post_spill_env_t;
-static be_options_t *main_opts;
-
static const lc_opt_enum_int_items_t lower_perm_items[] = {
{ "copy", BE_CH_LOWER_PERM_COPY },
{ "swap", BE_CH_LOWER_PERM_SWAP },
static const lc_opt_table_entry_t be_chordal_options[] = {
LC_OPT_ENT_ENUM_PTR ("perm", "perm lowering options", &lower_perm_var),
LC_OPT_ENT_ENUM_MASK("dump", "select dump phases", &dump_var),
- LC_OPT_ENT_ENUM_PTR ("vrfy", "verify options", &be_ch_vrfy_var),
+ LC_OPT_ENT_ENUM_PTR ("verify", "verify options", &be_ch_vrfy_var),
LC_OPT_LAST
};
}
/**
- * Checks for every reload if it's user can perform the load on itself.
+ * Checks for every reload if its user can perform the load on itself.
*/
-static void memory_operand_walker(ir_node *irn, void *env) {
- be_chordal_env_t *cenv = env;
- const arch_env_t *aenv = cenv->birg->main_env->arch_env;
+static void memory_operand_walker(ir_node *irn, void *env)
+{
const ir_edge_t *edge, *ne;
ir_node *block;
ir_node *spill;
+ (void)env;
+
if (! be_is_Reload(irn))
return;
assert(src && "outedges broken!");
- if (get_nodes_block(src) == block && arch_possible_memory_operand(aenv, src, pos)) {
- arch_perform_memory_operand(aenv, src, spill, pos);
+ if (get_nodes_block(src) == block && arch_possible_memory_operand(src, pos)) {
+ arch_perform_memory_operand(src, spill, pos);
}
}
/**
* Starts a walk for memory operands if supported by the backend.
*/
-static INLINE void check_for_memory_operands(be_chordal_env_t *chordal_env) {
- irg_walk_graph(chordal_env->irg, NULL, memory_operand_walker, chordal_env);
+static inline void check_for_memory_operands(ir_graph *irg)
+{
+ irg_walk_graph(irg, NULL, memory_operand_walker, NULL);
}
be_chordal_env_t *chordal_env = &pse->cenv;
be_irg_t *birg = pse->birg;
ir_graph *irg = be_get_birg_irg(birg);
- const be_main_env_t *main_env = birg->main_env;
pse->cls = cls;
chordal_env->cls = cls;
be_assure_liveness(birg);
be_liveness_assure_chk(be_get_birg_liveness(birg));
- stat_ev_do(pse->pre_spill_cost = be_estimate_irg_costs(irg, main_env->arch_env, birg->exec_freq));
+ stat_ev_do(pse->pre_spill_cost = be_estimate_irg_costs(irg, birg->exec_freq));
/* put all ignore registers into the ignore register set. */
be_put_ignore_regs(birg, pse->cls, chordal_env->ignore_colors);
be_chordal_env_t *chordal_env = &pse->cenv;
be_irg_t *birg = pse->birg;
ir_graph *irg = birg->irg;
- const be_main_env_t *main_env = birg->main_env;
- int colors_n = arch_register_class_n_regs(chordal_env->cls);
+ int colors_n = arch_register_class_n_regs(chordal_env->cls);
int allocatable_regs = colors_n - be_put_ignore_regs(birg, chordal_env->cls, NULL);
/* some special classes contain only ignore regs, no work to be done */
if (allocatable_regs > 0) {
- stat_ev_dbl("bechordal_spillcosts", be_estimate_irg_costs(irg, main_env->arch_env, birg->exec_freq) - pse->pre_spill_cost);
+ stat_ev_dbl("bechordal_spillcosts", be_estimate_irg_costs(irg, birg->exec_freq) - pse->pre_spill_cost);
/*
If we have a backend provided spiller, post spill is
But we only need to fix stack nodes once in this case.
*/
BE_TIMER_PUSH(t_ra_spill_apply);
- check_for_memory_operands(chordal_env);
+ check_for_memory_operands(irg);
if (iteration == 0) {
be_abi_fix_stack_nodes(birg->abi);
}
BE_TIMER_POP(t_ra_spill_apply);
- BE_TIMER_PUSH(t_verify);
/* verify schedule and register pressure */
+ BE_TIMER_PUSH(t_verify);
if (chordal_env->opts->vrfy_option == BE_CH_VRFY_WARN) {
be_verify_schedule(birg);
be_verify_register_pressure(birg, pse->cls, irg);
*/
static void be_ra_chordal_main(be_irg_t *birg)
{
- const be_main_env_t *main_env = birg->main_env;
- const arch_env_t *arch_env = main_env->arch_env;
- ir_graph *irg = birg->irg;
- int j, m;
- be_chordal_env_t chordal_env;
- struct obstack obst;
-
- main_opts = main_env->options;
+ const arch_env_t *arch_env = birg->main_env->arch_env;
+ ir_graph *irg = birg->irg;
+ int j;
+ int m;
+ be_chordal_env_t chordal_env;
+ struct obstack obst;
BE_TIMER_PUSH(t_ra_other);
/* the backend has its own spiller */
m = arch_env_get_n_reg_class(arch_env);
- pse = alloca(m * sizeof(pse[0]));
+ pse = ALLOCAN(post_spill_env_t, m);
for (j = 0; j < m; ++j) {
memcpy(&pse[j].cenv, &chordal_env, sizeof(chordal_env));
}
BE_TIMER_PUSH(t_verify);
- be_verify_register_allocation(birg);
+ if (chordal_env.opts->vrfy_option == BE_CH_VRFY_WARN) {
+ be_verify_register_allocation(birg);
+ } else if(chordal_env.opts->vrfy_option == BE_CH_VRFY_ASSERT) {
+ assert(be_verify_register_allocation(birg)
+ && "Register allocation invalid");
+ }
BE_TIMER_POP(t_verify);
BE_TIMER_PUSH(t_ra_epilog);