DEBUG_ONLY(static firm_dbg_module_t *dbg = NULL;)
+/** Defines an invalid register index. */
+#define NO_COLOR (-1)
+
#define SEARCH_FREE_COLORS
#define SLOTS_PINNED_GLOBAL 64
static int set_cmp_conflict_t(const void *x, const void *y, size_t size)
{
- const conflict_t *xx = x;
- const conflict_t *yy = y;
+ const conflict_t *xx = (const conflict_t*)x;
+ const conflict_t *yy = (const conflict_t*)y;
(void) size;
return xx->n1 != yy->n1 || xx->n2 != yy->n2;
{
node_stat_t find;
find.irn = irn;
- return set_find(qn->changed_nodes, &find, sizeof(find), hash_irn(irn));
+ return (const node_stat_t*)set_find(qn->changed_nodes, &find, sizeof(find), hash_irn(irn));
}
/**
find.irn = irn;
find.new_color = NO_COLOR;
find.pinned_local = 0;
- return set_insert(qn->changed_nodes, &find, sizeof(find), hash_irn(irn));
+ return (node_stat_t*)set_insert(qn->changed_nodes, &find, sizeof(find), hash_irn(irn));
}
/**
ir_node *sub_res, *curr;
be_ifg_t *ifg = chordal_env->ifg;
neighbours_iter_t iter;
-
+ const arch_register_req_t *req;
DBG((dbg, LEVEL_3, "\t %+F \tcaused col(%+F) \t%2d --> %2d\n", trigger, irn, irn_col, col));
return irn;
}
+ req = arch_get_irn_register_req(irn);
#ifdef SEARCH_FREE_COLORS
/* If we resolve conflicts (recursive calls) we can use any unused color.
* In case of the first call @p col must be used.
*/
if (irn != trigger) {
bitset_t *free_cols = bitset_alloca(cls->n_regs);
- const arch_register_req_t *req;
ir_node *curr;
int free_col;
bitset_copy(free_cols, co->cenv->allocatable_regs);
/* Exclude colors not assignable to the irn */
- req = arch_get_register_req_out(irn);
if (arch_register_req_is(req, limited)) {
bitset_t *limited = bitset_alloca(cls->n_regs);
rbitset_copy_to_bitset(req->limited, limited);
#endif /* SEARCH_FREE_COLORS */
/* If target color is not allocatable changing color is impossible */
- if (!arch_reg_out_is_allocatable(irn, arch_register_for_index(cls, col))) {
+ if (!arch_reg_is_allocatable(req, arch_register_for_index(cls, col))) {
DBG((dbg, LEVEL_3, "\t %+F impossible\n", irn));
return CHANGE_IMPOSSIBLE;
}
ir_node **safe, **unsafe;
int i, o, safe_count, safe_costs, unsafe_count, *unsafe_costs;
bitset_t *curr, *best;
- unsigned pos;
+ size_t pos;
int next, curr_weight, best_weight = 0;
/* assign the nodes into two groups.
goto no_stable_set;
/* if we arrive here, we have a stable set */
- /* compute the weigth of the stable set*/
+ /* compute the weight of the stable set*/
curr_weight = 0;
bitset_foreach(curr, pos)
curr_weight += unsafe_costs[pos];
/* init queue */
INIT_LIST_HEAD(&ou->queue);
- req = arch_get_register_req_out(ou->nodes[0]);
+ req = arch_get_irn_register_req(ou->nodes[0]);
allocatable_regs = ou->co->cenv->allocatable_regs;
n_regs = req->cls->n_regs;
if (arch_register_req_is(req, limited)) {
}
/* set color of all changed nodes */
- for (ns = set_first(curr->changed_nodes); ns; ns = set_next(curr->changed_nodes)) {
+ for (ns = (node_stat_t*)set_first(curr->changed_nodes); ns != NULL;
+ ns = (node_stat_t*)set_next(curr->changed_nodes)) {
/* NO_COLOR is possible, if we had an undo */
if (ns->new_color != NO_COLOR) {
DBG((dbg, LEVEL_1, "\t color(%+F) := %d\n", ns->irn, ns->new_color));
return 0;
}
-BE_REGISTER_MODULE_CONSTRUCTOR(be_init_copyheur);
+BE_REGISTER_MODULE_CONSTRUCTOR(be_init_copyheur)
void be_init_copyheur(void)
{
static co_algo_info copyheur = {