/*
- * Copyright (C) 1995-2011 University of Karlsruhe. All right reserved.
- *
* This file is part of libFirm.
- *
- * This file may be distributed and/or modified under the terms of the
- * GNU General Public License version 2 as published by the Free Software
- * Foundation and appearing in the file LICENSE.GPL included in the
- * packaging of this file.
- *
- * Licensees holding valid libFirm Professional Edition licenses may use
- * this file in accordance with the libFirm Commercial License.
- * Agreement provided with the Software.
- *
- * This file is provided AS IS with NO WARRANTY OF ANY KIND, INCLUDING THE
- * WARRANTY OF DESIGN, MERCHANTABILITY AND FITNESS FOR A PARTICULAR
- * PURPOSE.
+ * Copyright (C) 2012 University of Karlsruhe.
*/
/**
#include "array_t.h"
#include "bearch.h"
+#include "beirg.h"
#include "belower.h"
#include "benode.h"
#include "besched.h"
typedef enum perm_type_t {
PERM_CYCLE,
PERM_CHAIN,
- PERM_SWAP,
- PERM_COPY
} perm_type_t;
/** Structure to represent cycles or chains in a Perm. */
* IN_2 = in node with register i + 1
* OUT_1 = out node with register i + 1
* OUT_2 = out node with register i */
+ ir_node *cpyxchg;
if (cycle.type == PERM_CYCLE && !do_copy) {
ir_node *in[2];
- ir_node *cpyxchg;
in[0] = arg1;
in[1] = arg2;
arch_set_irn_register(res2, cycle.elems[i + 1]);
arch_set_irn_register(res1, cycle.elems[i]);
- /* insert the copy/exchange node in schedule after the magic schedule node (see above) */
- sched_add_after(skip_Proj(sched_point), cpyxchg);
-
DB((dbg, LEVEL_1, "replacing %+F with %+F, placed new node after %+F\n", irn, cpyxchg, sched_point));
-
- /* set the new scheduling point */
- sched_point = res1;
} else {
- ir_node *cpyxchg;
-
DB((dbg, LEVEL_1, "%+F creating copy node (%+F, %s) -> (%+F, %s)\n",
irn, arg1, cycle.elems[i]->name, res2, cycle.elems[i + 1]->name));
/* exchange copy node and proj */
exchange(res2, cpyxchg);
+ }
- /* insert the copy/exchange node in schedule after the magic schedule node (see above) */
- sched_add_after(skip_Proj(sched_point), cpyxchg);
+ /* insert the copy/exchange node in schedule after the magic schedule node (see above) */
+ sched_add_after(sched_point, cpyxchg);
- /* set the new scheduling point */
- sched_point = cpyxchg;
- }
+ /* set the new scheduling point */
+ sched_point = cpyxchg;
}
}
ir_node *keep, *cpy;
op_copy_assoc_t *entry;
- if (arch_irn_is_ignore(other_different) ||
+ arch_register_req_t const *const req = arch_get_irn_register_req(other_different);
+ if (arch_register_req_is(req, ignore) ||
!mode_is_datab(get_irn_mode(other_different))) {
DB((dbg_constr, LEVEL_1, "ignore constraint for %+F because other_irn is ignore or not a datab node\n", irn));
return;
op_set = &env->op_set;
block = get_nodes_block(irn);
- cls = arch_get_irn_reg_class(other_different);
+ cls = req->cls;
/* Make a not spillable copy of the different node */
/* this is needed because the different irn could be */
constraint_env_t *env = (constraint_env_t*)walk_env;
sched_foreach_reverse(block, irn) {
- ir_mode *mode = get_irn_mode(irn);
-
- if (mode == mode_T) {
- foreach_out_edge(irn, edge) {
- ir_node *proj = get_edge_src_irn(edge);
-
- mode = get_irn_mode(proj);
- if (mode_is_datab(mode))
- assure_different_constraints(proj, irn, env);
- }
- } else if (mode_is_datab(mode)) {
- assure_different_constraints(irn, irn, env);
- }
+ be_foreach_value(irn, value,
+ if (mode_is_datab(get_irn_mode(value)))
+ assure_different_constraints(value, irn, env);
+ );
}
}
if (ck_arr[idx]) {
int j, n_melt;
ir_node **new_ck_in;
- ir_node *new_ck;
ir_node *sched_pt = NULL;
n_melt = 1;
kill_node(melt_arr[j]);
}
+ ir_node *const new_ck = be_new_CopyKeep(get_nodes_block(ref), be_get_CopyKeep_op(ref), n_melt, new_ck_in);
#ifdef KEEP_ALIVE_COPYKEEP_HACK
- new_ck = be_new_CopyKeep(get_nodes_block(ref), be_get_CopyKeep_op(ref), n_melt, new_ck_in);
keep_alive(new_ck);
-#else
- new_ck = be_new_CopyKeep(get_nodes_block(ref), be_get_CopyKeep_op(ref), n_melt, new_ck_in);
#endif /* KEEP_ALIVE_COPYKEEP_HACK */
/* set register class for all kept inputs */
ir_node *keep;
keep = be_new_Keep(get_nodes_block(cp), n, get_irn_in(cp) + 1);
- sched_add_before(cp, keep);
+ sched_replace(cp, keep);
/* Set all ins (including the block) of the CopyKeep BAD to keep the verifier happy. */
- sched_remove(cp);
kill_node(cp);
}
}
int new_size;
ir_node *frontier = bl;
int i, n;
+ be_lv_t *lv = be_get_irg_liveness(irg);
/* get some Proj and find out the register class of that Proj. */
ir_node *one_proj = get_edge_src_irn(get_irn_out_edge_first_kind(perm, EDGE_KIND_NORMAL));
* the former dead operand would be live now at the point of
* the Perm, increasing the register pressure by one.
*/
- sched_foreach_reverse_from(sched_prev(perm), irn) {
- for (i = get_irn_arity(irn) - 1; i >= 0; --i) {
- ir_node *op = get_irn_n(irn, i);
- be_lv_t *lv = be_get_irg_liveness(irg);
- if (arch_irn_consider_in_reg_alloc(cls, op) &&
- !be_values_interfere(lv, op, one_proj)) {
+ sched_foreach_reverse_before(perm, irn) {
+ be_foreach_use(irn, cls, in_req_, op, op_req_,
+ if (!be_values_interfere(lv, op, one_proj)) {
frontier = irn;
goto found_front;
}
- }
+ );
}
found_front:
n_moved = 0;
for (;;) {
ir_node *const node = sched_prev(perm);
- if (sched_is_begin(node))
+ if (node == frontier)
break;
const arch_register_req_t *req;
/* it wasn't an input to the perm, we can't do anything more */
if (input < 0)
break;
- if (!sched_comes_after(frontier, node))
- break;
if (arch_irn_is(node, modify_flags))
break;
req = arch_get_irn_register_req(node);