projects
/
libfirm
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
Use SPARC_STACK_ALIGNMENT and round_up2() instead of magic numbers and calculations.
[libfirm]
/
ir
/
be
/
becopyopt.c
diff --git
a/ir/be/becopyopt.c
b/ir/be/becopyopt.c
index
7b7affc
..
f90ec70
100644
(file)
--- a/
ir/be/becopyopt.c
+++ b/
ir/be/becopyopt.c
@@
-50,7
+50,7
@@
#include "bearch.h"
#include "benode.h"
#include "beutil.h"
#include "bearch.h"
#include "benode.h"
#include "beutil.h"
-#include "beifg
_t
.h"
+#include "beifg.h"
#include "beintlive_t.h"
#include "becopyopt_t.h"
#include "becopystat.h"
#include "beintlive_t.h"
#include "becopyopt_t.h"
#include "becopystat.h"
@@
-75,7
+75,7
@@
static unsigned dump_flags = 0;
static unsigned style_flags = 0;
static unsigned dump_flags = 0;
static unsigned style_flags = 0;
-static
unsigned
do_stats = 0;
+static
int
do_stats = 0;
static cost_fct_t cost_func = co_get_costs_exec_freq;
static int improve = 1;
static cost_fct_t cost_func = co_get_costs_exec_freq;
static int improve = 1;
@@
-136,6
+136,7
@@
void be_register_copyopt(const char *name, co_algo_info *copyopt)
be_add_module_to_list(©opts, name, copyopt);
}
be_add_module_to_list(©opts, name, copyopt);
}
+BE_REGISTER_MODULE_CONSTRUCTOR(be_init_copyopt)
void be_init_copyopt(void)
{
lc_opt_entry_t *be_grp = lc_opt_get_grp(firm_opt_get_root(), "be");
void be_init_copyopt(void)
{
lc_opt_entry_t *be_grp = lc_opt_get_grp(firm_opt_get_root(), "be");
@@
-148,14
+149,13
@@
void be_init_copyopt(void)
©opts, (void**) &selected_copyopt);
}
©opts, (void**) &selected_copyopt);
}
-BE_REGISTER_MODULE_CONSTRUCTOR(be_init_copyopt);
-
static int void_algo(copy_opt_t *co)
{
(void) co;
return 0;
}
static int void_algo(copy_opt_t *co)
{
(void) co;
return 0;
}
+BE_REGISTER_MODULE_CONSTRUCTOR(be_init_copynone)
void be_init_copynone(void)
{
static co_algo_info copyheur = {
void be_init_copynone(void)
{
static co_algo_info copyheur = {
@@
-165,16
+165,16
@@
void be_init_copynone(void)
be_register_copyopt("none", ©heur);
}
be_register_copyopt("none", ©heur);
}
-BE_REGISTER_MODULE_CONSTRUCTOR(be_init_copynone);
-
#undef QUICK_AND_DIRTY_HACK
static int nodes_interfere(const be_chordal_env_t *env, const ir_node *a, const ir_node *b)
{
if (env->ifg)
return be_ifg_connected(env->ifg, a, b);
#undef QUICK_AND_DIRTY_HACK
static int nodes_interfere(const be_chordal_env_t *env, const ir_node *a, const ir_node *b)
{
if (env->ifg)
return be_ifg_connected(env->ifg, a, b);
- else
- return be_values_interfere(env->birg->lv, a, b);
+ else {
+ be_lv_t *lv = be_get_irg_liveness(env->irg);
+ return be_values_interfere(lv, a, b);
+ }
}
}
@@
-194,7
+194,7
@@
DEBUG_ONLY(static firm_dbg_module_t *dbg = NULL;)
copy_opt_t *new_copy_opt(be_chordal_env_t *chordal_env, cost_fct_t get_costs)
{
const char *s1, *s2, *s3;
copy_opt_t *new_copy_opt(be_chordal_env_t *chordal_env, cost_fct_t get_costs)
{
const char *s1, *s2, *s3;
-
in
t len;
+
size_
t len;
copy_opt_t *co;
FIRM_DBG_REGISTER(dbg, "ir.be.copyopt");
copy_opt_t *co;
FIRM_DBG_REGISTER(dbg, "ir.be.copyopt");
@@
-215,7
+215,8
@@
copy_opt_t *new_copy_opt(be_chordal_env_t *chordal_env, cost_fct_t get_costs)
return co;
}
return co;
}
-void free_copy_opt(copy_opt_t *co) {
+void free_copy_opt(copy_opt_t *co)
+{
xfree(co->name);
free(co);
}
xfree(co->name);
free(co);
}
@@
-227,26
+228,22
@@
void free_copy_opt(copy_opt_t *co) {
static int co_is_optimizable_root(ir_node *irn)
{
const arch_register_req_t *req;
static int co_is_optimizable_root(ir_node *irn)
{
const arch_register_req_t *req;
- const arch_register_t *reg;
if (arch_irn_is_ignore(irn))
return 0;
if (arch_irn_is_ignore(irn))
return 0;
- reg = arch_get_irn_register(irn);
- if (arch_register_type_is(reg, ignore))
- return 0;
-
if (is_Reg_Phi(irn) || is_Perm_Proj(irn))
return 1;
if (is_Reg_Phi(irn) || is_Perm_Proj(irn))
return 1;
- req = arch_get_
register_req_out
(irn);
+ req = arch_get_
irn_register_req
(irn);
if (is_2addr_code(req))
return 1;
return 0;
}
if (is_2addr_code(req))
return 1;
return 0;
}
-int co_get_costs_loop_depth(const copy_opt_t *co, ir_node *root, ir_node* arg, int pos) {
+int co_get_costs_loop_depth(const copy_opt_t *co, ir_node *root, ir_node* arg, int pos)
+{
int cost = 0;
ir_loop *loop;
ir_node *root_block = get_nodes_block(root);
int cost = 0;
ir_loop *loop;
ir_node *root_block = get_nodes_block(root);
@@
-267,19
+264,22
@@
int co_get_costs_loop_depth(const copy_opt_t *co, ir_node *root, ir_node* arg, i
return 1+cost;
}
return 1+cost;
}
-int co_get_costs_exec_freq(const copy_opt_t *co, ir_node *root, ir_node* arg, int pos) {
+int co_get_costs_exec_freq(const copy_opt_t *co, ir_node *root, ir_node* arg, int pos)
+{
int res;
ir_node *root_bl = get_nodes_block(root);
ir_node *copy_bl = is_Phi(root) ? get_Block_cfgpred_block(root_bl, pos) : root_bl;
int res;
ir_node *root_bl = get_nodes_block(root);
ir_node *copy_bl = is_Phi(root) ? get_Block_cfgpred_block(root_bl, pos) : root_bl;
+ ir_exec_freq *exec_freq = be_get_irg_exec_freq(co->cenv->irg);
(void) arg;
(void) arg;
- res = get_block_execfreq_ulong(
co->cenv->birg->
exec_freq, copy_bl);
+ res = get_block_execfreq_ulong(exec_freq, copy_bl);
/* don't allow values smaller than one. */
return res < 1 ? 1 : res;
}
/* don't allow values smaller than one. */
return res < 1 ? 1 : res;
}
-int co_get_costs_all_one(const copy_opt_t *co, ir_node *root, ir_node *arg, int pos) {
+int co_get_costs_all_one(const copy_opt_t *co, ir_node *root, ir_node *arg, int pos)
+{
(void) co;
(void) root;
(void) arg;
(void) co;
(void) root;
(void) arg;
@@
-302,13
+302,14
@@
int co_get_costs_all_one(const copy_opt_t *co, ir_node *root, ir_node *arg, int
* Determines a maximum weighted independent set with respect to
* the interference and conflict edges of all nodes in a qnode.
*/
* Determines a maximum weighted independent set with respect to
* the interference and conflict edges of all nodes in a qnode.
*/
-static int ou_max_ind_set_costs(unit_t *ou) {
+static int ou_max_ind_set_costs(unit_t *ou)
+{
be_chordal_env_t *chordal_env = ou->co->cenv;
ir_node **safe, **unsafe;
int i, o, safe_count, safe_costs, unsafe_count, *unsafe_costs;
bitset_t *curr;
be_chordal_env_t *chordal_env = ou->co->cenv;
ir_node **safe, **unsafe;
int i, o, safe_count, safe_costs, unsafe_count, *unsafe_costs;
bitset_t *curr;
-
bitset_pos_t
pos;
- int
max,
curr_weight, best_weight = 0;
+
size_t
pos;
+ int curr_weight, best_weight = 0;
/* assign the nodes into two groups.
* safe: node has no interference, hence it is in every max stable set.
/* assign the nodes into two groups.
* safe: node has no interference, hence it is in every max stable set.
@@
-320,9
+321,9
@@
static int ou_max_ind_set_costs(unit_t *ou) {
unsafe = ALLOCAN(ir_node*, ou->node_count - 1);
unsafe_costs = ALLOCAN(int, ou->node_count - 1);
unsafe_count = 0;
unsafe = ALLOCAN(ir_node*, ou->node_count - 1);
unsafe_costs = ALLOCAN(int, ou->node_count - 1);
unsafe_count = 0;
- for(i=1; i<ou->node_count; ++i) {
+ for
(i=1; i<ou->node_count; ++i) {
int is_safe = 1;
int is_safe = 1;
- for(o=1; o<ou->node_count; ++o) {
+ for
(o=1; o<ou->node_count; ++o) {
if (i==o)
continue;
if (nodes_interfere(chordal_env, ou->nodes[i], ou->nodes[o])) {
if (i==o)
continue;
if (nodes_interfere(chordal_env, ou->nodes[i], ou->nodes[o])) {
@@
-343,7
+344,7
@@
static int ou_max_ind_set_costs(unit_t *ou) {
/* now compute the best set out of the unsafe nodes*/
if (unsafe_count > MIS_HEUR_TRIGGER) {
bitset_t *best = bitset_alloca(unsafe_count);
/* now compute the best set out of the unsafe nodes*/
if (unsafe_count > MIS_HEUR_TRIGGER) {
bitset_t *best = bitset_alloca(unsafe_count);
- /* Heuristi
k
: Greedy trial and error form index 0 to unsafe_count-1 */
+ /* Heuristi
c
: Greedy trial and error form index 0 to unsafe_count-1 */
for (i=0; i<unsafe_count; ++i) {
bitset_set(best, i);
/* check if it is a stable set */
for (i=0; i<unsafe_count; ++i) {
bitset_set(best, i);
/* check if it is a stable set */
@@
-360,7
+361,7
@@
static int ou_max_ind_set_costs(unit_t *ou) {
/* Exact Algorithm: Brute force */
curr = bitset_alloca(unsafe_count);
bitset_set_all(curr);
/* Exact Algorithm: Brute force */
curr = bitset_alloca(unsafe_count);
bitset_set_all(curr);
- while (
(max = bitset_popcnt(curr)
) != 0) {
+ while (
bitset_popcount(curr
) != 0) {
/* check if curr is a stable set */
for (i=bitset_next_set(curr, 0); i!=-1; i=bitset_next_set(curr, i+1))
for (o=bitset_next_set(curr, i+1); o!=-1; o=bitset_next_set(curr, o+1)) /* !!!!! difference to qnode_max_ind_set(): NOT (curr, i) */
/* check if curr is a stable set */
for (i=bitset_next_set(curr, 0); i!=-1; i=bitset_next_set(curr, i+1))
for (o=bitset_next_set(curr, i+1); o!=-1; o=bitset_next_set(curr, o+1)) /* !!!!! difference to qnode_max_ind_set(): NOT (curr, i) */
@@
-368,7
+369,7
@@
static int ou_max_ind_set_costs(unit_t *ou) {
goto no_stable_set;
/* if we arrive here, we have a stable set */
goto no_stable_set;
/* if we arrive here, we have a stable set */
- /* compute the weig
th
of the stable set*/
+ /* compute the weig
ht
of the stable set*/
curr_weight = 0;
bitset_foreach(curr, pos)
curr_weight += unsafe_costs[pos];
curr_weight = 0;
bitset_foreach(curr, pos)
curr_weight += unsafe_costs[pos];
@@
-388,10
+389,13
@@
static int ou_max_ind_set_costs(unit_t *ou) {
static void co_collect_units(ir_node *irn, void *env)
{
static void co_collect_units(ir_node *irn, void *env)
{
- const arch_register_req_t *req
= arch_get_register_req_out(irn)
;
- copy_opt_t *co = env;
+ const arch_register_req_t *req;
+ copy_opt_t *co =
(copy_opt_t*)
env;
unit_t *unit;
unit_t *unit;
+ if (get_irn_mode(irn) == mode_T)
+ return;
+ req = arch_get_irn_register_req(irn);
if (req->cls != co->cls)
return;
if (!co_is_optimizable_root(irn))
if (req->cls != co->cls)
return;
if (!co_is_optimizable_root(irn))
@@
-418,7
+422,7
@@
static void co_collect_units(ir_node *irn, void *env)
int o, arg_pos;
ir_node *arg = get_irn_n(irn, i);
int o, arg_pos;
ir_node *arg = get_irn_n(irn, i);
- assert(arch_get_irn_reg_class
_out
(arg) == co->cls && "Argument not in same register class.");
+ assert(arch_get_irn_reg_class(arg) == co->cls && "Argument not in same register class.");
if (arg == irn)
continue;
if (nodes_interfere(co->cenv, irn, arg)) {
if (arg == irn)
continue;
if (nodes_interfere(co->cenv, irn, arg)) {
@@
-511,7
+515,7
@@
static void co_collect_units(ir_node *irn, void *env)
struct list_head *tmp;
/* Determine the maximum costs this unit can cause: all_nodes_cost */
struct list_head *tmp;
/* Determine the maximum costs this unit can cause: all_nodes_cost */
- for(i=1; i<unit->node_count; ++i) {
+ for
(i=1; i<unit->node_count; ++i) {
unit->sort_key = MAX(unit->sort_key, unit->costs[i]);
unit->all_nodes_costs += unit->costs[i];
}
unit->sort_key = MAX(unit->sort_key, unit->costs[i]);
unit->all_nodes_costs += unit->costs[i];
}
@@
-530,7
+534,8
@@
static void co_collect_units(ir_node *irn, void *env)
#ifdef QUICK_AND_DIRTY_HACK
#ifdef QUICK_AND_DIRTY_HACK
-static int compare_ous(const void *k1, const void *k2) {
+static int compare_ous(const void *k1, const void *k2)
+{
const unit_t *u1 = *((const unit_t **) k1);
const unit_t *u2 = *((const unit_t **) k2);
int i, o, u1_has_constr, u2_has_constr;
const unit_t *u1 = *((const unit_t **) k1);
const unit_t *u2 = *((const unit_t **) k2);
int i, o, u1_has_constr, u2_has_constr;
@@
-539,7
+544,7
@@
static int compare_ous(const void *k1, const void *k2) {
/* Units with constraints come first */
u1_has_constr = 0;
for (i=0; i<u1->node_count; ++i) {
/* Units with constraints come first */
u1_has_constr = 0;
for (i=0; i<u1->node_count; ++i) {
- arch_get_
register_req_out
(&req, u1->nodes[i]);
+ arch_get_
irn_register_req
(&req, u1->nodes[i]);
if (arch_register_req_is(&req, limited)) {
u1_has_constr = 1;
break;
if (arch_register_req_is(&req, limited)) {
u1_has_constr = 1;
break;
@@
-548,7
+553,7
@@
static int compare_ous(const void *k1, const void *k2) {
u2_has_constr = 0;
for (i=0; i<u2->node_count; ++i) {
u2_has_constr = 0;
for (i=0; i<u2->node_count; ++i) {
- arch_get_
register_req_out
(&req, u2->nodes[i]);
+ arch_get_
irn_register_req
(&req, u2->nodes[i]);
if (arch_register_req_is(&req, limited)) {
u2_has_constr = 1;
break;
if (arch_register_req_is(&req, limited)) {
u2_has_constr = 1;
break;
@@
-574,7
+579,8
@@
static int compare_ous(const void *k1, const void *k2) {
/**
* Sort the ou's according to constraints and their sort_key
*/
/**
* Sort the ou's according to constraints and their sort_key
*/
-static void co_sort_units(copy_opt_t *co) {
+static void co_sort_units(copy_opt_t *co)
+{
int i, count = 0, costs;
unit_t *ou, **ous;
int i, count = 0, costs;
unit_t *ou, **ous;
@@
-610,7
+616,8
@@
static void co_sort_units(copy_opt_t *co) {
}
#endif
}
#endif
-void co_build_ou_structure(copy_opt_t *co) {
+void co_build_ou_structure(copy_opt_t *co)
+{
DBG((dbg, LEVEL_1, "\tCollecting optimization units\n"));
INIT_LIST_HEAD(&co->units);
irg_walk_graph(co->irg, co_collect_units, NULL, co);
DBG((dbg, LEVEL_1, "\tCollecting optimization units\n"));
INIT_LIST_HEAD(&co->units);
irg_walk_graph(co->irg, co_collect_units, NULL, co);
@@
-619,7
+626,8
@@
void co_build_ou_structure(copy_opt_t *co) {
#endif
}
#endif
}
-void co_free_ou_structure(copy_opt_t *co) {
+void co_free_ou_structure(copy_opt_t *co)
+{
unit_t *curr, *tmp;
ASSERT_OU_AVAIL(co);
list_for_each_entry_safe(unit_t, curr, tmp, &co->units, units) {
unit_t *curr, *tmp;
ASSERT_OU_AVAIL(co);
list_for_each_entry_safe(unit_t, curr, tmp, &co->units, units) {
@@
-632,7
+640,8
@@
void co_free_ou_structure(copy_opt_t *co) {
/* co_solve_heuristic() is implemented in becopyheur.c */
/* co_solve_heuristic() is implemented in becopyheur.c */
-int co_get_max_copy_costs(const copy_opt_t *co) {
+int co_get_max_copy_costs(const copy_opt_t *co)
+{
int i, res = 0;
unit_t *curr;
int i, res = 0;
unit_t *curr;
@@
-646,7
+655,8
@@
int co_get_max_copy_costs(const copy_opt_t *co) {
return res;
}
return res;
}
-int co_get_inevit_copy_costs(const copy_opt_t *co) {
+int co_get_inevit_copy_costs(const copy_opt_t *co)
+{
int res = 0;
unit_t *curr;
int res = 0;
unit_t *curr;
@@
-657,7
+667,8
@@
int co_get_inevit_copy_costs(const copy_opt_t *co) {
return res;
}
return res;
}
-int co_get_copy_costs(const copy_opt_t *co) {
+int co_get_copy_costs(const copy_opt_t *co)
+{
int i, res = 0;
unit_t *curr;
int i, res = 0;
unit_t *curr;
@@
-678,7
+689,8
@@
int co_get_copy_costs(const copy_opt_t *co) {
return res;
}
return res;
}
-int co_get_lower_bound(const copy_opt_t *co) {
+int co_get_lower_bound(const copy_opt_t *co)
+{
int res = 0;
unit_t *curr;
int res = 0;
unit_t *curr;
@@
-702,7
+714,7
@@
void co_complete_stats(const copy_opt_t *co, co_complete_stats_t *stat)
stat->aff_nodes += 1;
bitset_add_irn(seen, an->irn);
co_gs_foreach_neighb(an, neigh) {
stat->aff_nodes += 1;
bitset_add_irn(seen, an->irn);
co_gs_foreach_neighb(an, neigh) {
- if(!bitset_contains_irn(seen, neigh->irn)) {
+ if
(!bitset_contains_irn(seen, neigh->irn)) {
stat->aff_edges += 1;
stat->max_costs += neigh->costs;
stat->aff_edges += 1;
stat->max_costs += neigh->costs;
@@
-711,7
+723,7
@@
void co_complete_stats(const copy_opt_t *co, co_complete_stats_t *stat)
stat->unsatisfied_edges += 1;
}
stat->unsatisfied_edges += 1;
}
- if(nodes_interfere(co->cenv, an->irn, neigh->irn)) {
+ if
(nodes_interfere(co->cenv, an->irn, neigh->irn)) {
stat->aff_int += 1;
stat->inevit_costs += neigh->costs;
}
stat->aff_int += 1;
stat->inevit_costs += neigh->costs;
}
@@
-734,15
+746,17
@@
void co_complete_stats(const copy_opt_t *co, co_complete_stats_t *stat)
|_| |___/
******************************************************************************/
|_| |___/
******************************************************************************/
-static int compare_affinity_node_t(const void *k1, const void *k2, size_t size) {
- const affinity_node_t *n1 = k1;
- const affinity_node_t *n2 = k2;
+static int compare_affinity_node_t(const void *k1, const void *k2, size_t size)
+{
+ const affinity_node_t *n1 = (const affinity_node_t*)k1;
+ const affinity_node_t *n2 = (const affinity_node_t*)k2;
(void) size;
return (n1->irn != n2->irn);
}
(void) size;
return (n1->irn != n2->irn);
}
-static void add_edge(copy_opt_t *co, ir_node *n1, ir_node *n2, int costs) {
+static void add_edge(copy_opt_t *co, ir_node *n1, ir_node *n2, int costs)
+{
affinity_node_t new_node, *node;
neighb_t *nbr;
int allocnew = 1;
affinity_node_t new_node, *node;
neighb_t *nbr;
int allocnew = 1;
@@
-750,7
+764,7
@@
static void add_edge(copy_opt_t *co, ir_node *n1, ir_node *n2, int costs) {
new_node.irn = n1;
new_node.degree = 0;
new_node.neighbours = NULL;
new_node.irn = n1;
new_node.degree = 0;
new_node.neighbours = NULL;
- node = set_insert(co->nodes, &new_node, sizeof(new_node), hash_irn(new_node.irn));
+ node =
(affinity_node_t*)
set_insert(co->nodes, &new_node, sizeof(new_node), hash_irn(new_node.irn));
for (nbr = node->neighbours; nbr; nbr = nbr->next)
if (nbr->irn == n2) {
for (nbr = node->neighbours; nbr; nbr = nbr->next)
if (nbr->irn == n2) {
@@
-773,24
+787,24
@@
static void add_edge(copy_opt_t *co, ir_node *n1, ir_node *n2, int costs) {
nbr->costs += costs;
}
nbr->costs += costs;
}
-static inline void add_edges(copy_opt_t *co, ir_node *n1, ir_node *n2, int costs) {
+static inline void add_edges(copy_opt_t *co, ir_node *n1, ir_node *n2, int costs)
+{
if (! be_ifg_connected(co->cenv->ifg, n1, n2)) {
add_edge(co, n1, n2, costs);
add_edge(co, n2, n1, costs);
}
}
if (! be_ifg_connected(co->cenv->ifg, n1, n2)) {
add_edge(co, n1, n2, costs);
add_edge(co, n2, n1, costs);
}
}
-static void build_graph_walker(ir_node *irn, void *env) {
- const arch_register_req_t *req = arch_get_register_req_out(irn);
- copy_opt_t *co = env;
+static void build_graph_walker(ir_node *irn, void *env)
+{
+ const arch_register_req_t *req;
+ copy_opt_t *co = (copy_opt_t*)env;
int pos, max;
int pos, max;
- const arch_register_t *reg;
- if (
req->cls != co->cls || arch_irn_is_ignore(irn)
)
+ if (
get_irn_mode(irn) == mode_T
)
return;
return;
-
- reg = arch_get_irn_register(irn);
- if (arch_register_type_is(reg, ignore))
+ req = arch_get_irn_register_req(irn);
+ if (req->cls != co->cls || arch_irn_is_ignore(irn))
return;
if (is_Reg_Phi(irn)) { /* Phis */
return;
if (is_Reg_Phi(irn)) { /* Phis */
@@
-817,14
+831,16
@@
static void build_graph_walker(ir_node *irn, void *env) {
}
}
}
}
-void co_build_graph_structure(copy_opt_t *co) {
+void co_build_graph_structure(copy_opt_t *co)
+{
obstack_init(&co->obst);
co->nodes = new_set(compare_affinity_node_t, 32);
irg_walk_graph(co->irg, build_graph_walker, NULL, co);
}
obstack_init(&co->obst);
co->nodes = new_set(compare_affinity_node_t, 32);
irg_walk_graph(co->irg, build_graph_walker, NULL, co);
}
-void co_free_graph_structure(copy_opt_t *co) {
+void co_free_graph_structure(copy_opt_t *co)
+{
ASSERT_GS_AVAIL(co);
del_set(co->nodes);
ASSERT_GS_AVAIL(co);
del_set(co->nodes);
@@
-832,15
+848,14
@@
void co_free_graph_structure(copy_opt_t *co) {
co->nodes = NULL;
}
co->nodes = NULL;
}
-/* co_solve_ilp1() co_solve_ilp2() are implemented in becopyilpX.c */
-
-int co_gs_is_optimizable(copy_opt_t *co, ir_node *irn) {
+int co_gs_is_optimizable(copy_opt_t *co, ir_node *irn)
+{
affinity_node_t new_node, *n;
ASSERT_GS_AVAIL(co);
new_node.irn = irn;
affinity_node_t new_node, *n;
ASSERT_GS_AVAIL(co);
new_node.irn = irn;
- n = set_find(co->nodes, &new_node, sizeof(new_node), hash_irn(new_node.irn));
+ n =
(affinity_node_t*)
set_find(co->nodes, &new_node, sizeof(new_node), hash_irn(new_node.irn));
if (n) {
return (n->degree > 0);
} else
if (n) {
return (n->degree > 0);
} else
@@
-857,8
+872,8
@@
static int co_dump_appel_disjoint_constraints(const copy_opt_t *co, ir_node *a,
constr[1] = bitset_alloca(co->cls->n_regs);
for (j = 0; j < 2; ++j) {
constr[1] = bitset_alloca(co->cls->n_regs);
for (j = 0; j < 2; ++j) {
- const arch_register_req_t *req = arch_get_
register_req_out
(nodes[j]);
- if(arch_register_req_is(req, limited))
+ const arch_register_req_t *req = arch_get_
irn_register_req
(nodes[j]);
+ if
(arch_register_req_is(req, limited))
rbitset_copy_to_bitset(req->limited, constr[j]);
else
bitset_set_all(constr[j]);
rbitset_copy_to_bitset(req->limited, constr[j]);
else
bitset_set_all(constr[j]);
@@
-873,16
+888,23
@@
void co_dump_appel_graph(const copy_opt_t *co, FILE *f)
be_ifg_t *ifg = co->cenv->ifg;
int *color_map = ALLOCAN(int, co->cls->n_regs);
int *node_map = XMALLOCN(int, get_irg_last_idx(co->irg) + 1);
be_ifg_t *ifg = co->cenv->ifg;
int *color_map = ALLOCAN(int, co->cls->n_regs);
int *node_map = XMALLOCN(int, get_irg_last_idx(co->irg) + 1);
+ ir_graph *irg = co->irg;
+ be_irg_t *birg = be_birg_from_irg(irg);
ir_node *irn;
ir_node *irn;
- void *it, *nit;
+ nodes_iter_t it;
+ neighbours_iter_t nit;
int n, n_regs;
unsigned i;
n_regs = 0;
int n, n_regs;
unsigned i;
n_regs = 0;
- for(i = 0; i < co->cls->n_regs; ++i) {
+ for
(i = 0; i < co->cls->n_regs; ++i) {
const arch_register_t *reg = &co->cls->regs[i];
const arch_register_t *reg = &co->cls->regs[i];
- color_map[i] = arch_register_type_is(reg, ignore) ? -1 : n_regs++;
+ if (rbitset_is_set(birg->allocatable_regs, reg->global_index)) {
+ color_map[i] = n_regs++;
+ } else {
+ color_map[i] = -1;
+ }
}
/*
}
/*
@@
-890,11
+912,8
@@
void co_dump_appel_graph(const copy_opt_t *co, FILE *f)
* the values below n are the pre-colored register nodes
*/
* the values below n are the pre-colored register nodes
*/
- it = be_ifg_nodes_iter_alloca(ifg);
- nit = be_ifg_neighbours_iter_alloca(ifg);
-
n = n_regs;
n = n_regs;
- be_ifg_foreach_node(ifg, it, irn) {
+ be_ifg_foreach_node(ifg,
&
it, irn) {
if (arch_irn_is_ignore(irn))
continue;
node_map[get_irn_idx(irn)] = n++;
if (arch_irn_is_ignore(irn))
continue;
node_map[get_irn_idx(irn)] = n++;
@@
-902,36
+921,36
@@
void co_dump_appel_graph(const copy_opt_t *co, FILE *f)
fprintf(f, "%d %d\n", n, n_regs);
fprintf(f, "%d %d\n", n, n_regs);
- be_ifg_foreach_node(ifg, it, irn) {
+ be_ifg_foreach_node(ifg,
&
it, irn) {
if (!arch_irn_is_ignore(irn)) {
int idx = node_map[get_irn_idx(irn)];
affinity_node_t *a = get_affinity_info(co, irn);
if (!arch_irn_is_ignore(irn)) {
int idx = node_map[get_irn_idx(irn)];
affinity_node_t *a = get_affinity_info(co, irn);
- const arch_register_req_t *req = arch_get_
register_req_out
(irn);
+ const arch_register_req_t *req = arch_get_
irn_register_req
(irn);
ir_node *adj;
ir_node *adj;
- if(arch_register_req_is(req, limited)) {
- for(i = 0; i < co->cls->n_regs; ++i) {
- if(!rbitset_is_set(req->limited, i) && color_map[i] >= 0)
+ if
(arch_register_req_is(req, limited)) {
+ for
(i = 0; i < co->cls->n_regs; ++i) {
+ if
(!rbitset_is_set(req->limited, i) && color_map[i] >= 0)
fprintf(f, "%d %d -1\n", color_map[i], idx);
}
}
fprintf(f, "%d %d -1\n", color_map[i], idx);
}
}
- be_ifg_foreach_neighbour(ifg, nit, irn, adj) {
+ be_ifg_foreach_neighbour(ifg,
&
nit, irn, adj) {
if (!arch_irn_is_ignore(adj) &&
!co_dump_appel_disjoint_constraints(co, irn, adj)) {
int adj_idx = node_map[get_irn_idx(adj)];
if (!arch_irn_is_ignore(adj) &&
!co_dump_appel_disjoint_constraints(co, irn, adj)) {
int adj_idx = node_map[get_irn_idx(adj)];
- if(idx < adj_idx)
+ if
(idx < adj_idx)
fprintf(f, "%d %d -1\n", idx, adj_idx);
}
}
fprintf(f, "%d %d -1\n", idx, adj_idx);
}
}
- if(a) {
+ if
(a) {
neighb_t *n;
co_gs_foreach_neighb(a, n) {
if (!arch_irn_is_ignore(n->irn)) {
int n_idx = node_map[get_irn_idx(n->irn)];
neighb_t *n;
co_gs_foreach_neighb(a, n) {
if (!arch_irn_is_ignore(n->irn)) {
int n_idx = node_map[get_irn_idx(n->irn)];
- if(idx < n_idx)
+ if
(idx < n_idx)
fprintf(f, "%d %d %d\n", idx, n_idx, (int) n->costs);
}
}
fprintf(f, "%d %d %d\n", idx, n_idx, (int) n->costs);
}
}
@@
-990,7
+1009,7
@@
static const char *get_dot_color_name(size_t col)
return col < sizeof(names)/sizeof(names[0]) ? names[col] : "white";
}
return col < sizeof(names)/sizeof(names[0]) ? names[col] : "white";
}
-typedef struct
_
co_ifg_dump_t {
+typedef struct co_ifg_dump_t {
const copy_opt_t *co;
unsigned flags;
} co_ifg_dump_t;
const copy_opt_t *co;
unsigned flags;
} co_ifg_dump_t;
@@
-1009,15
+1028,15
@@
static int ifg_is_dump_node(void *self, ir_node *irn)
static void ifg_dump_node_attr(FILE *f, void *self, ir_node *irn)
{
static void ifg_dump_node_attr(FILE *f, void *self, ir_node *irn)
{
- co_ifg_dump_t *env = self;
+ co_ifg_dump_t *env =
(co_ifg_dump_t*)
self;
const arch_register_t *reg = arch_get_irn_register(irn);
const arch_register_t *reg = arch_get_irn_register(irn);
- const arch_register_req_t *req = arch_get_
register_req_out
(irn);
+ const arch_register_req_t *req = arch_get_
irn_register_req
(irn);
int limited = arch_register_req_is(req, limited);
int limited = arch_register_req_is(req, limited);
- if(env->flags & CO_IFG_DUMP_LABELS) {
+ if
(env->flags & CO_IFG_DUMP_LABELS) {
ir_fprintf(f, "label=\"%+F", irn);
ir_fprintf(f, "label=\"%+F", irn);
- if((env->flags & CO_IFG_DUMP_CONSTR) && limited) {
+ if
((env->flags & CO_IFG_DUMP_CONSTR) && limited) {
bitset_t *bs = bitset_alloca(env->co->cls->n_regs);
rbitset_copy_to_bitset(req->limited, bs);
ir_fprintf(f, "\\n%B", bs);
bitset_t *bs = bitset_alloca(env->co->cls->n_regs);
rbitset_copy_to_bitset(req->limited, bs);
ir_fprintf(f, "\\n%B", bs);
@@
-1027,16
+1046,16
@@
static void ifg_dump_node_attr(FILE *f, void *self, ir_node *irn)
fprintf(f, "label=\"\" shape=point " );
}
fprintf(f, "label=\"\" shape=point " );
}
- if(env->flags & CO_IFG_DUMP_SHAPE)
+ if
(env->flags & CO_IFG_DUMP_SHAPE)
fprintf(f, "shape=%s ", limited ? "diamond" : "ellipse");
fprintf(f, "shape=%s ", limited ? "diamond" : "ellipse");
- if(env->flags & CO_IFG_DUMP_COLORS)
+ if
(env->flags & CO_IFG_DUMP_COLORS)
fprintf(f, "style=filled color=%s ", get_dot_color_name(reg->index));
}
static void ifg_dump_at_end(FILE *file, void *self)
{
fprintf(f, "style=filled color=%s ", get_dot_color_name(reg->index));
}
static void ifg_dump_at_end(FILE *file, void *self)
{
- co_ifg_dump_t *env = self;
+ co_ifg_dump_t *env =
(co_ifg_dump_t*)
self;
affinity_node_t *a;
co_gs_foreach_aff_node(env->co, a) {
affinity_node_t *a;
co_gs_foreach_aff_node(env->co, a) {
@@
-1048,12
+1067,12
@@
static void ifg_dump_at_end(FILE *file, void *self)
const arch_register_t *nr = arch_get_irn_register(n->irn);
unsigned nidx = get_irn_idx(n->irn);
const arch_register_t *nr = arch_get_irn_register(n->irn);
unsigned nidx = get_irn_idx(n->irn);
- if(aidx < nidx) {
+ if
(aidx < nidx) {
const char *color = nr == ar ? "blue" : "red";
const char *color = nr == ar ? "blue" : "red";
- fprintf(file, "\tn%
d -- n%d
[weight=0.01 ", aidx, nidx);
- if(env->flags & CO_IFG_DUMP_LABELS)
+ fprintf(file, "\tn%
u -- n%u
[weight=0.01 ", aidx, nidx);
+ if
(env->flags & CO_IFG_DUMP_LABELS)
fprintf(file, "label=\"%d\" ", n->costs);
fprintf(file, "label=\"%d\" ", n->costs);
- if(env->flags & CO_IFG_DUMP_COLORS)
+ if
(env->flags & CO_IFG_DUMP_COLORS)
fprintf(file, "color=%s ", color);
else
fprintf(file, "style=dotted");
fprintf(file, "color=%s ", color);
else
fprintf(file, "style=dotted");
@@
-1105,10
+1124,11
@@
static FILE *my_open(const be_chordal_env_t *env, const char *prefix, const char
char buf[1024];
size_t i, n;
char *tu_name;
char buf[1024];
size_t i, n;
char *tu_name;
+ const char *cup_name = be_get_irg_main_env(env->irg)->cup_name;
- n = strlen(
env->birg->main_env->
cup_name);
+ n = strlen(cup_name);
tu_name = XMALLOCN(char, n + 1);
tu_name = XMALLOCN(char, n + 1);
- strcpy(tu_name,
env->birg->main_env->
cup_name);
+ strcpy(tu_name, cup_name);
for (i = 0; i < n; ++i)
if (tu_name[i] == '.')
tu_name[i] = '_';
for (i = 0; i < n; ++i)
if (tu_name[i] == '.')
tu_name[i] = '_';
@@
-1117,7
+1137,7
@@
static FILE *my_open(const be_chordal_env_t *env, const char *prefix, const char
ir_snprintf(buf, sizeof(buf), "%s%s_%F_%s%s", prefix, tu_name, env->irg, env->cls->name, suffix);
xfree(tu_name);
result = fopen(buf, "wt");
ir_snprintf(buf, sizeof(buf), "%s%s_%F_%s%s", prefix, tu_name, env->irg, env->cls->name, suffix);
xfree(tu_name);
result = fopen(buf, "wt");
- if(result == NULL) {
+ if
(result == NULL) {
panic("Couldn't open '%s' for writing.", buf);
}
panic("Couldn't open '%s' for writing.", buf);
}
@@
-1134,10
+1154,10
@@
void co_driver(be_chordal_env_t *cenv)
assert(selected_copyopt);
/* skip copymin if algo is 'none' */
assert(selected_copyopt);
/* skip copymin if algo is 'none' */
- if(selected_copyopt->copyopt == void_algo)
+ if
(selected_copyopt->copyopt == void_algo)
return;
return;
- be_liveness_assure_chk(be_get_
birg_liveness(cenv->b
irg));
+ be_liveness_assure_chk(be_get_
irg_liveness(cenv->
irg));
co = new_copy_opt(cenv, cost_func);
co_build_ou_structure(co);
co = new_copy_opt(cenv, cost_func);
co_build_ou_structure(co);
@@
-1196,7
+1216,7
@@
void co_driver(be_chordal_env_t *cenv)
ir_printf("%30F ", cenv->irg);
printf("%10s %10" ULL_FMT "%10" ULL_FMT "%10" ULL_FMT, cenv->cls->name, after.max_costs, before.costs, after.inevit_costs);
ir_printf("%30F ", cenv->irg);
printf("%10s %10" ULL_FMT "%10" ULL_FMT "%10" ULL_FMT, cenv->cls->name, after.max_costs, before.costs, after.inevit_costs);
- if(optimizable_costs > 0)
+ if
(optimizable_costs > 0)
printf("%10" ULL_FMT " %5.2f\n", after.costs, (evitable * 100.0) / optimizable_costs);
else
printf("%10" ULL_FMT " %5s\n", after.costs, "-");
printf("%10" ULL_FMT " %5.2f\n", after.costs, (evitable * 100.0) / optimizable_costs);
else
printf("%10" ULL_FMT " %5s\n", after.costs, "-");
@@
-1205,7
+1225,7
@@
void co_driver(be_chordal_env_t *cenv)
/* Dump the interference graph in Appel's format. */
if (dump_flags & DUMP_APPEL) {
FILE *f = my_open(cenv, "", ".apl");
/* Dump the interference graph in Appel's format. */
if (dump_flags & DUMP_APPEL) {
FILE *f = my_open(cenv, "", ".apl");
- fprintf(f, "# %ll
d %lld
\n", after.costs, after.unsatisfied_edges);
+ fprintf(f, "# %ll
u %llu
\n", after.costs, after.unsatisfied_edges);
co_dump_appel_graph(co, f);
fclose(f);
}
co_dump_appel_graph(co, f);
fclose(f);
}