projects
/
libfirm
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
merge kaps
[libfirm]
/
ir
/
be
/
becopyopt.c
diff --git
a/ir/be/becopyopt.c
b/ir/be/becopyopt.c
index
cde32cd
..
e9e0e9c
100644
(file)
--- a/
ir/be/becopyopt.c
+++ b/
ir/be/becopyopt.c
@@
-194,7
+194,7
@@
DEBUG_ONLY(static firm_dbg_module_t *dbg = NULL;)
copy_opt_t *new_copy_opt(be_chordal_env_t *chordal_env, cost_fct_t get_costs)
{
const char *s1, *s2, *s3;
copy_opt_t *new_copy_opt(be_chordal_env_t *chordal_env, cost_fct_t get_costs)
{
const char *s1, *s2, *s3;
-
in
t len;
+
size_
t len;
copy_opt_t *co;
FIRM_DBG_REGISTER(dbg, "ir.be.copyopt");
copy_opt_t *co;
FIRM_DBG_REGISTER(dbg, "ir.be.copyopt");
@@
-228,15
+228,10
@@
void free_copy_opt(copy_opt_t *co)
static int co_is_optimizable_root(ir_node *irn)
{
const arch_register_req_t *req;
static int co_is_optimizable_root(ir_node *irn)
{
const arch_register_req_t *req;
- const arch_register_t *reg;
if (arch_irn_is_ignore(irn))
return 0;
if (arch_irn_is_ignore(irn))
return 0;
- reg = arch_get_irn_register(irn);
- if (arch_register_type_is(reg, ignore))
- return 0;
-
if (is_Reg_Phi(irn) || is_Perm_Proj(irn))
return 1;
if (is_Reg_Phi(irn) || is_Perm_Proj(irn))
return 1;
@@
-313,7
+308,7
@@
static int ou_max_ind_set_costs(unit_t *ou)
ir_node **safe, **unsafe;
int i, o, safe_count, safe_costs, unsafe_count, *unsafe_costs;
bitset_t *curr;
ir_node **safe, **unsafe;
int i, o, safe_count, safe_costs, unsafe_count, *unsafe_costs;
bitset_t *curr;
-
unsigned
pos;
+
size_t
pos;
int curr_weight, best_weight = 0;
/* assign the nodes into two groups.
int curr_weight, best_weight = 0;
/* assign the nodes into two groups.
@@
-349,7
+344,7
@@
static int ou_max_ind_set_costs(unit_t *ou)
/* now compute the best set out of the unsafe nodes*/
if (unsafe_count > MIS_HEUR_TRIGGER) {
bitset_t *best = bitset_alloca(unsafe_count);
/* now compute the best set out of the unsafe nodes*/
if (unsafe_count > MIS_HEUR_TRIGGER) {
bitset_t *best = bitset_alloca(unsafe_count);
- /* Heuristi
k
: Greedy trial and error form index 0 to unsafe_count-1 */
+ /* Heuristi
c
: Greedy trial and error form index 0 to unsafe_count-1 */
for (i=0; i<unsafe_count; ++i) {
bitset_set(best, i);
/* check if it is a stable set */
for (i=0; i<unsafe_count; ++i) {
bitset_set(best, i);
/* check if it is a stable set */
@@
-374,7
+369,7
@@
static int ou_max_ind_set_costs(unit_t *ou)
goto no_stable_set;
/* if we arrive here, we have a stable set */
goto no_stable_set;
/* if we arrive here, we have a stable set */
- /* compute the weig
th
of the stable set*/
+ /* compute the weig
ht
of the stable set*/
curr_weight = 0;
bitset_foreach(curr, pos)
curr_weight += unsafe_costs[pos];
curr_weight = 0;
bitset_foreach(curr, pos)
curr_weight += unsafe_costs[pos];
@@
-395,7
+390,7
@@
static int ou_max_ind_set_costs(unit_t *ou)
static void co_collect_units(ir_node *irn, void *env)
{
const arch_register_req_t *req;
static void co_collect_units(ir_node *irn, void *env)
{
const arch_register_req_t *req;
- copy_opt_t *co = env;
+ copy_opt_t *co =
(copy_opt_t*)
env;
unit_t *unit;
if (get_irn_mode(irn) == mode_T)
unit_t *unit;
if (get_irn_mode(irn) == mode_T)
@@
-753,8
+748,8
@@
void co_complete_stats(const copy_opt_t *co, co_complete_stats_t *stat)
static int compare_affinity_node_t(const void *k1, const void *k2, size_t size)
{
static int compare_affinity_node_t(const void *k1, const void *k2, size_t size)
{
- const affinity_node_t *n1 = k1;
- const affinity_node_t *n2 = k2;
+ const affinity_node_t *n1 =
(const affinity_node_t*)
k1;
+ const affinity_node_t *n2 =
(const affinity_node_t*)
k2;
(void) size;
return (n1->irn != n2->irn);
(void) size;
return (n1->irn != n2->irn);
@@
-769,7
+764,7
@@
static void add_edge(copy_opt_t *co, ir_node *n1, ir_node *n2, int costs)
new_node.irn = n1;
new_node.degree = 0;
new_node.neighbours = NULL;
new_node.irn = n1;
new_node.degree = 0;
new_node.neighbours = NULL;
- node = set_insert(co->nodes, &new_node, sizeof(new_node), hash_irn(new_node.irn));
+ node =
(affinity_node_t*)
set_insert(co->nodes, &new_node, sizeof(new_node), hash_irn(new_node.irn));
for (nbr = node->neighbours; nbr; nbr = nbr->next)
if (nbr->irn == n2) {
for (nbr = node->neighbours; nbr; nbr = nbr->next)
if (nbr->irn == n2) {
@@
-803,9
+798,8
@@
static inline void add_edges(copy_opt_t *co, ir_node *n1, ir_node *n2, int costs
static void build_graph_walker(ir_node *irn, void *env)
{
const arch_register_req_t *req;
static void build_graph_walker(ir_node *irn, void *env)
{
const arch_register_req_t *req;
- copy_opt_t *co = env;
+ copy_opt_t *co =
(copy_opt_t*)
env;
int pos, max;
int pos, max;
- const arch_register_t *reg;
if (get_irn_mode(irn) == mode_T)
return;
if (get_irn_mode(irn) == mode_T)
return;
@@
-813,10
+807,6
@@
static void build_graph_walker(ir_node *irn, void *env)
if (req->cls != co->cls || arch_irn_is_ignore(irn))
return;
if (req->cls != co->cls || arch_irn_is_ignore(irn))
return;
- reg = arch_get_irn_register(irn);
- if (arch_register_type_is(reg, ignore))
- return;
-
if (is_Reg_Phi(irn)) { /* Phis */
for (pos=0, max=get_irn_arity(irn); pos<max; ++pos) {
ir_node *arg = get_irn_n(irn, pos);
if (is_Reg_Phi(irn)) { /* Phis */
for (pos=0, max=get_irn_arity(irn); pos<max; ++pos) {
ir_node *arg = get_irn_n(irn, pos);
@@
-865,7
+855,7
@@
int co_gs_is_optimizable(copy_opt_t *co, ir_node *irn)
ASSERT_GS_AVAIL(co);
new_node.irn = irn;
ASSERT_GS_AVAIL(co);
new_node.irn = irn;
- n = set_find(co->nodes, &new_node, sizeof(new_node), hash_irn(new_node.irn));
+ n =
(affinity_node_t*)
set_find(co->nodes, &new_node, sizeof(new_node), hash_irn(new_node.irn));
if (n) {
return (n->degree > 0);
} else
if (n) {
return (n->degree > 0);
} else
@@
-898,6
+888,8
@@
void co_dump_appel_graph(const copy_opt_t *co, FILE *f)
be_ifg_t *ifg = co->cenv->ifg;
int *color_map = ALLOCAN(int, co->cls->n_regs);
int *node_map = XMALLOCN(int, get_irg_last_idx(co->irg) + 1);
be_ifg_t *ifg = co->cenv->ifg;
int *color_map = ALLOCAN(int, co->cls->n_regs);
int *node_map = XMALLOCN(int, get_irg_last_idx(co->irg) + 1);
+ ir_graph *irg = co->irg;
+ be_irg_t *birg = be_birg_from_irg(irg);
ir_node *irn;
nodes_iter_t it;
ir_node *irn;
nodes_iter_t it;
@@
-908,7
+900,11
@@
void co_dump_appel_graph(const copy_opt_t *co, FILE *f)
n_regs = 0;
for (i = 0; i < co->cls->n_regs; ++i) {
const arch_register_t *reg = &co->cls->regs[i];
n_regs = 0;
for (i = 0; i < co->cls->n_regs; ++i) {
const arch_register_t *reg = &co->cls->regs[i];
- color_map[i] = arch_register_type_is(reg, ignore) ? -1 : n_regs++;
+ if (rbitset_is_set(birg->allocatable_regs, reg->global_index)) {
+ color_map[i] = n_regs++;
+ } else {
+ color_map[i] = -1;
+ }
}
/*
}
/*
@@
-1032,7
+1028,7
@@
static int ifg_is_dump_node(void *self, ir_node *irn)
static void ifg_dump_node_attr(FILE *f, void *self, ir_node *irn)
{
static void ifg_dump_node_attr(FILE *f, void *self, ir_node *irn)
{
- co_ifg_dump_t *env = self;
+ co_ifg_dump_t *env =
(co_ifg_dump_t*)
self;
const arch_register_t *reg = arch_get_irn_register(irn);
const arch_register_req_t *req = arch_get_register_req_out(irn);
int limited = arch_register_req_is(req, limited);
const arch_register_t *reg = arch_get_irn_register(irn);
const arch_register_req_t *req = arch_get_register_req_out(irn);
int limited = arch_register_req_is(req, limited);
@@
-1059,7
+1055,7
@@
static void ifg_dump_node_attr(FILE *f, void *self, ir_node *irn)
static void ifg_dump_at_end(FILE *file, void *self)
{
static void ifg_dump_at_end(FILE *file, void *self)
{
- co_ifg_dump_t *env = self;
+ co_ifg_dump_t *env =
(co_ifg_dump_t*)
self;
affinity_node_t *a;
co_gs_foreach_aff_node(env->co, a) {
affinity_node_t *a;
co_gs_foreach_aff_node(env->co, a) {