+/*
+ * Copyright (C) 1995-2007 University of Karlsruhe. All right reserved.
+ *
+ * This file is part of libFirm.
+ *
+ * This file may be distributed and/or modified under the terms of the
+ * GNU General Public License version 2 as published by the Free Software
+ * Foundation and appearing in the file LICENSE.GPL included in the
+ * packaging of this file.
+ *
+ * Licensees holding valid libFirm Professional Edition licenses may use
+ * this file in accordance with the libFirm Commercial License.
+ * Agreement provided with the Software.
+ *
+ * This file is provided AS IS with NO WARRANTY OF ANY KIND, INCLUDING THE
+ * WARRANTY OF DESIGN, MERCHANTABILITY AND FITNESS FOR A PARTICULAR
+ * PURPOSE.
+ */
+
/**
- * Scheduling algorithms.
+ * @file
+ * @brief ILP based instruction scheduling.
+ * @author Christian Wuerdig
+ * @date 22.10.2006
+ * @version $Id$
+ *
* An ILP scheduler based on
* "ILP-based Instruction Scheduling for IA-64"
* by Daniel Kaestner and Sebastian Winkel
- *
- * @date 22.10.2005
- * @author Christian Wuerdig
- * @cvs-id $Id$
+ * extended with register pressure constraints by Christian Wuerdig
*/
#ifdef HAVE_CONFIG_H
#include "config.h"
#include "irgwalk.h"
#include "irbitset.h"
#include "irphase_t.h"
+#include "height.h"
#include "iredges.h"
-#include "debug.h"
#include "pdeq.h"
+#include "debug.h"
#include "irtools.h"
#include "irdump.h"
+#include "irprintf.h"
+#include "plist.h"
+#include "irprintf.h"
#include <lpp/lpp.h>
#include <lpp/lpp_net.h>
-#ifdef WITH_LIBCORE
#include <libcore/lc_opts.h>
#include <libcore/lc_opts_enum.h>
#include <libcore/lc_timing.h>
-#endif /* WITH_LIBCORE */
#include "be.h"
#include "benode_t.h"
#include "besched_t.h"
+#include "beilpsched.h"
+#include "beutil.h"
+#include "bestat.h"
+#include "beirg_t.h"
+#include "benodesets.h"
typedef struct _ilpsched_options_t {
+ unsigned regpress;
unsigned time_limit;
char log_file[1024];
} ilpsched_options_t;
const be_execution_unit_type_t *tp;
} unit_type_info_t;
+/**
+ * holding the ILP variables of the different types
+ */
+typedef struct _ilp_var_types_t {
+ int *x; /* x_{nt}^k variables */
+ int *a; /* a_{nt}^k variables */
+ int *y; /* y_{nt}^k variables */
+} ilp_var_types_t;
+
+/**
+ * Holds alive variables for a node live-in to a block.
+ */
+typedef struct _ilp_livein_node_t {
+ ir_node *irn;
+ unsigned max_alive_steps;
+ int *a;
+} ilp_livein_node_t;
+
/* attributes for a node */
typedef struct _ilpsched_node_attr_t {
unsigned asap; /**< The ASAP scheduling control step */
unsigned alap; /**< The ALAP scheduling control step */
+ unsigned latency; /**< Latency of this node (needed for sorting) */
unsigned sched_point; /**< the step in which the node is finally scheduled */
unsigned visit_idx; /**< Index of the node having visited this node last */
- unsigned block_idx : 31; /**< A unique per block index */
- unsigned enqueued : 1; /**< Node is already enqueued for ALAP calculation */
+ unsigned consumer_idx; /**< Index of the node having counted this node as consumer last */
+ unsigned n_consumer; /**< Number of consumers */
+ ir_node **block_consumer; /**< List of consumer being in the same block */
+ waitq *projkeeps; /**< A List of Projs and Keeps belonging to this node */
+ unsigned block_idx : 30; /**< A unique per block index */
+ unsigned alap_changed : 1; /**< the current ALAP has changed, revisit preds */
+ unsigned is_dummy_node : 1; /**< this node is assigned to DUMMY unit */
bitset_t *transitive_block_nodes; /**< Set of transitive block nodes (predecessors
for ASAP, successors for ALAP */
- unsigned n_unit_types; /**< number of allowed execution unit types */
+ unsigned n_unit_types; /**< number of allowed execution unit types */
unit_type_info_t *type_info; /**< list of allowed execution unit types */
- int *ilp_vars; /**< the binary ilp variables x_{nt}^k assigned to this node
- (== 1 iff: node n is executed at step t on execunit k
- So we have: |ASAP(n) ... ALAP(n)| * |execunits| variables
- */
+ ilp_var_types_t ilp_vars; /**< the different ILP variables */
} ilpsched_node_attr_t;
/* attributes for a block */
typedef struct _ilpsched_block_attr_t {
unsigned block_last_idx; /**< The highest node index in block so far */
unsigned n_interesting_nodes; /**< The number of nodes interesting for scheduling */
- waitq *root_nodes; /**< A queue of nodes having no user in current block */
+ unsigned max_steps; /**< Upper bound for block execution */
+ plist_t *root_nodes; /**< A list of nodes having no user in current block */
ir_node *head_ilp_nodes; /**< A linked list of nodes which will contribute to ILP */
+ pset *livein_nodes; /**< A set of nodes which are live-in to this block */
} ilpsched_block_attr_t;
typedef union _ilpsched_attr_ {
/* The ILP scheduling environment */
typedef struct {
- phase_t ph; /**< The phase */
- ir_graph *irg; /**< The current irg */
- waitq *alap_queue; /**< An queue of nodes waiting for final ALAP calculation */
- const arch_env_t *arch_env;
- const be_main_env_t *main_env;
- const be_machine_t *cpu; /**< the current abstract machine */
- ilpsched_options_t *opts; /**< the ilp options for current irg */
+ ir_phase ph; /**< The phase */
+ ir_graph *irg; /**< The current irg */
+ heights_t *height; /**< The heights object of the irg */
+ void *irg_env; /**< An environment for the irg scheduling, provided by the backend */
+ void *block_env; /**< An environment for scheduling a block, provided by the backend */
+ const arch_env_t *arch_env;
+ const arch_isa_t *isa; /**< The ISA */
+ const be_main_env_t *main_env;
+ const be_machine_t *cpu; /**< the current abstract machine */
+ ilpsched_options_t *opts; /**< the ilp options for current irg */
+ const be_irg_t *birg; /**< The birg object */
+ be_options_t *be_opts; /**< backend options */
+ const ilp_sched_selector_t *sel; /**< The ILP sched selector provided by the backend */
DEBUG_ONLY(firm_dbg_module_t *dbg);
} be_ilpsched_env_t;
#define get_ilpsched_block_attr(block) (&(block)->attr.block_attr)
#define get_ilpsched_node_attr(node) (&(node)->attr.node_attr)
-/* iterate over a list of ir_nodes linked by link field */
-#define foreach_linked_irns(head, iter) for ((iter) = (head); (iter); (iter) = get_irn_link((iter)))
-
/* check if node is considered for ILP scheduling */
-#define consider_for_sched(irn) \
- (! (is_Block(irn) || \
- is_Proj(irn) || \
- is_Phi(irn) || \
- be_is_Keep(irn) || \
- is_NoMem(irn) || \
- is_Jmp(irn) || \
- is_End(irn) \
+#define consider_for_sched(isa, irn) \
+ (! (is_Block(irn) || \
+ is_normal_Proj(isa, irn) || \
+ is_Phi(irn) || \
+ is_NoMem(irn) || \
+ is_Unknown(irn) || \
+ is_End(irn) \
))
/* gives the valid scheduling time step interval for a node */
#define VALID_SCHED_INTERVAL(na) ((na)->alap - (na)->asap + 1)
+/* gives the valid interval where a node can die */
+#define VALID_KILL_INTERVAL(ba, na) ((ba)->max_steps - (na)->asap + 1)
+
/* gives the corresponding ILP variable for given node, unit and time step */
#define ILPVAR_IDX(na, unit, control_step) \
((unit) * VALID_SCHED_INTERVAL((na)) + (control_step) - (na)->asap + 1)
+/* gives the corresponding dead nodes ILP variable for given node, unit and time step */
+#define ILPVAR_IDX_DEAD(ba, na, unit, control_step) \
+ ((unit) * VALID_KILL_INTERVAL((ba), (na)) + (control_step) - (na)->asap + 1)
+
/* check if a double value is within an epsilon environment of 0 */
#define LPP_VALUE_IS_0(dbl) (fabs((dbl)) <= 1e-10)
+#define ilp_timer_push(t) lc_timer_push((t))
+#define ilp_timer_pop() lc_timer_pop()
+#define ilp_timer_elapsed_usec(t) lc_timer_elapsed_usec((t))
+
/* option variable */
static ilpsched_options_t ilp_opts = {
- 120, /* 120 sec per block time limit */
+ 1, /* default is with register pressure constraints */
+ 300, /* 300 sec per block time limit */
"" /* no log file */
};
-#ifdef WITH_LIBCORE
/* ILP options */
static const lc_opt_table_entry_t ilpsched_option_table[] = {
+ LC_OPT_ENT_BOOL("regpress", "Use register pressure constraints", &ilp_opts.regpress),
LC_OPT_ENT_INT("time_limit", "ILP time limit per block", &ilp_opts.time_limit),
LC_OPT_ENT_STR("lpp_log", "LPP logfile (stderr and stdout are supported)", ilp_opts.log_file, sizeof(ilp_opts.log_file)),
- { NULL }
+ LC_OPT_LAST
};
-#endif /* WITH_LIBCORE */
+
+/*
+ We need this global variable as we compare nodes dependent on heights,
+ but we cannot pass any information to the qsort compare function.
+*/
+static heights_t *glob_heights;
+
+/**
+ * Check if irn is a Proj, which has no execution units assigned.
+ * @return 1 if irn is a Proj having no execution units assigned, 0 otherwise
+ */
+static INLINE int is_normal_Proj(const arch_isa_t *isa, const ir_node *irn) {
+ return is_Proj(irn) && (arch_isa_get_allowed_execution_units(isa, irn) == NULL);
+}
+
+/**
+ * Skips normal Projs.
+ * @return predecessor if irn is a normal Proj, otherwise irn.
+ */
+static INLINE ir_node *skip_normal_Proj(const arch_isa_t *isa, ir_node *irn) {
+ if (is_normal_Proj(isa, irn))
+ return get_Proj_pred(irn);
+ return irn;
+}
+
+static INLINE int fixed_latency(const ilp_sched_selector_t *sel, ir_node *irn, void *env) {
+ unsigned lat = be_ilp_sched_latency(sel, irn, env);
+ if (lat == 0 && ! is_Proj(irn) && ! be_is_Keep(irn))
+ lat = 1;
+ return lat;
+}
+
+static int cmp_live_in_nodes(const void *a, const void *b) {
+ const ilp_livein_node_t *n1 = a;
+ const ilp_livein_node_t *n2 = b;
+
+ return n1->irn != n2->irn;
+}
/**
* Compare scheduling time steps of two be_ilpsched_irn's.
ilpsched_node_attr_t *n1_a = get_ilpsched_node_attr(n1);
ilpsched_node_attr_t *n2_a = get_ilpsched_node_attr(n2);
- return QSORT_CMP(n1_a->sched_point, n2_a->sched_point);
+ if (n1_a->sched_point == n2_a->sched_point) {
+ ir_node *irn_a = n1->irn;
+ ir_node *irn_b = n2->irn;
+
+ if (heights_reachable_in_block(glob_heights, irn_a, irn_b))
+ return 1;
+ if (heights_reachable_in_block(glob_heights, irn_b, irn_a))
+ return -1;
+
+ /*
+ Ok, timestep is equal and the nodes are parallel,
+ so check latency and schedule high latency first.
+ */
+ return QSORT_CMP(n2_a->latency, n1_a->latency);
+ }
+ else
+ return QSORT_CMP(n1_a->sched_point, n2_a->sched_point);
}
/**
* In case there is no phase information for irn, initialize it.
*/
-static void *init_ilpsched_irn(phase_t *ph, ir_node *irn, void *old) {
+static void *init_ilpsched_irn(ir_phase *ph, ir_node *irn, void *old) {
be_ilpsched_irn_t *res = old ? old : phase_alloc(ph, sizeof(res[0]));
if (res == old) {
be_ilpsched_irn_t *block_node = phase_get_or_set_irn_data(ph, block);
ilpsched_block_attr_t *ba = get_ilpsched_block_attr(block_node);
- /* we are called after the block indicees have been build: create bitset */
+ /* we are called after the block indices have been build: create bitset */
na->transitive_block_nodes = bitset_obstack_alloc(phase_obst(ph), ba->block_last_idx);
}
else {
/* we are called from reinit block data: clear the bitset */
bitset_clear_all(na->transitive_block_nodes);
- na->visit_idx = 0;
+ na->visit_idx = 0;
+ na->alap_changed = 1;
}
}
return old;
ba->n_interesting_nodes = 0;
ba->block_last_idx = 0;
- ba->root_nodes = new_waitq();
+ ba->root_nodes = plist_new();
ba->head_ilp_nodes = NULL;
+ ba->livein_nodes = new_pset(cmp_live_in_nodes, 16);
+ ba->max_steps = 0;
}
else {
ilpsched_node_attr_t *na = get_ilpsched_node_attr(res);
ilpsched_node_attr_t *na;
ilpsched_block_attr_t *ba;
- if (! consider_for_sched(irn))
+ set_irn_link(irn, NULL);
+ if (! consider_for_sched(env->arch_env->isa, irn))
return;
node = get_ilpsched_irn(env, irn);
static void collect_alap_root_nodes(ir_node *irn, void *walk_env) {
ir_node *block;
const ir_edge_t *edge;
- be_ilpsched_irn_t *block_node;
+ be_ilpsched_irn_t *block_node, *node;
ilpsched_block_attr_t *ba;
+ ilpsched_node_attr_t *na;
+ int i, j;
be_ilpsched_env_t *env = walk_env;
int has_block_user = 0;
+ unsigned n_consumer = 0;
ir_edge_kind_t ekind[2] = { EDGE_KIND_NORMAL, EDGE_KIND_DEP };
- int i, j;
+ ir_node **consumer;
+ unsigned idx;
- if (! consider_for_sched(irn))
+ if (! consider_for_sched(env->arch_env->isa, irn))
return;
- block = get_nodes_block(irn);
+ block = get_nodes_block(irn);
+ idx = get_irn_idx(irn);
+ consumer = NEW_ARR_F(ir_node *, 0);
DBG((env->dbg, LEVEL_3, "%+F (%+F) is interesting, examining ... ", irn, block));
- for (i = 0; i < 2; ++i) {
+ /* check data and dependency out edges */
+ for (i = 0; i < 2 && ! has_block_user; ++i) {
foreach_out_edge_kind(irn, edge, ekind[i]) {
ir_node *user = get_edge_src_irn(edge);
- if (is_Proj(user)) {
+ if (is_normal_Proj(env->arch_env->isa, user)) {
const ir_edge_t *user_edge;
if (get_irn_mode(user) == mode_X)
foreach_out_edge_kind(user, user_edge, ekind[j]) {
ir_node *real_user = get_edge_src_irn(user_edge);
- if (! is_Phi(real_user) && ! be_is_Keep(real_user) && get_nodes_block(real_user) == block) {
- has_block_user = 1;
- break;
+ if (! is_Phi(real_user) && ! is_Block(real_user)) {
+ be_ilpsched_irn_t *node = get_ilpsched_irn(env, real_user);
+ ilpsched_node_attr_t *ua = get_ilpsched_node_attr(node);
+
+ /* skip already visited nodes */
+ if (ua->consumer_idx == idx)
+ continue;
+
+ /* check if node has user in this block and collect the user if it's a data user */
+ if (get_nodes_block(real_user) == block) {
+ if (i == 0 && j == 0)
+ ARR_APP1(ir_node *, consumer, real_user);
+ has_block_user = 1;
+ }
+
+ /* only count data consumer */
+ if (i == 0)
+ n_consumer++;
+
+ /* mark user as visited by this node */
+ ua->consumer_idx = idx;
}
}
}
-
- if (has_block_user)
- break;
}
else if (is_Block(user)) {
continue;
}
- else if (! is_Phi(user) && ! be_is_Keep(user) && get_nodes_block(user) == block) {
- has_block_user = 1;
- break;
+ else if (! is_Phi(user)) {
+ be_ilpsched_irn_t *node = get_ilpsched_irn(env, user);
+ ilpsched_node_attr_t *ua = get_ilpsched_node_attr(node);
+
+ /* skip already visited nodes */
+ if (ua->consumer_idx == idx)
+ continue;
+
+ /* check if node has user in this block and collect the user if it's a data user */
+ if (get_nodes_block(user) == block) {
+ if (i == 0)
+ ARR_APP1(ir_node *, consumer, user);
+ has_block_user = 1;
+ }
+
+ /* only count data consumer */
+ if (i == 0)
+ n_consumer++;
+
+ /* mark user visited by this node */
+ ua->consumer_idx = idx;
+ }
+ else if (get_nodes_block(user) != block) {
+ n_consumer++;
}
}
}
/* current irn has no user inside this block, add to queue */
if (! has_block_user) {
DB((env->dbg, LEVEL_3, "root node\n"));
- waitq_put(ba->root_nodes, irn);
+ plist_insert_back(ba->root_nodes, irn);
}
else {
DB((env->dbg, LEVEL_3, "normal node\n"));
}
+
+ /* record number of all consumer and the consumer within the same block */
+ node = get_ilpsched_irn(env, irn);
+ na = get_ilpsched_node_attr(node);
+ na->n_consumer = n_consumer;
+ na->block_consumer = NEW_ARR_D(ir_node *, phase_obst(&env->ph), ARR_LEN(consumer));
+ memcpy(na->block_consumer, consumer, ARR_LEN(consumer) * sizeof(na->block_consumer[0]));
+ DEL_ARR_F(consumer);
}
/**
* Calculate the ASAP scheduling step for current irn.
*/
static void calculate_irn_asap(ir_node *irn, void *walk_env) {
- be_ilpsched_irn_t *node;
- be_ilpsched_env_t *env = walk_env;
- int i;
- ir_node *block;
- ilpsched_node_attr_t *na;
+ be_ilpsched_env_t *env = walk_env;
+ int i;
+ ir_node *block;
+ be_ilpsched_irn_t *node, *block_node;
+ ilpsched_node_attr_t *na;
+ ilpsched_block_attr_t *ba;
/* These nodes are handled separate */
- if (! consider_for_sched(irn))
+ if (! consider_for_sched(env->arch_env->isa, irn))
return;
- DBG((env->dbg, LEVEL_2, "Calculating ASAP of node %+F\n", irn));
+ DBG((env->dbg, LEVEL_2, "Calculating ASAP of node %+F ... ", irn));
- node = get_ilpsched_irn(env, irn);
- block = get_nodes_block(irn);
- na = get_ilpsched_node_attr(node);
+ block = get_nodes_block(irn);
+ node = get_ilpsched_irn(env, irn);
+ na = get_ilpsched_node_attr(node);
+ na->asap = 1;
- /* accumulate all transitive predecessors of current node */
for (i = get_irn_ins_or_deps(irn) - 1; i >= 0; --i) {
- ir_node *pred = skip_Proj(get_irn_in_or_dep(irn, i));
- be_ilpsched_irn_t *pred_node;
- ilpsched_node_attr_t *pna;
- unsigned idx;
+ ir_node *pred = skip_normal_Proj(env->arch_env->isa, get_irn_in_or_dep(irn, i));
- if (be_is_Keep(pred))
- pred = skip_Proj(get_irn_n(pred, 0));
+ /* check for greatest distance to top */
+ if (! is_Phi(pred) && ! is_NoMem(pred) && get_nodes_block(pred) == block) {
+ be_ilpsched_irn_t *pred_node = get_ilpsched_irn(env, pred);
+ ilpsched_node_attr_t *pna = get_ilpsched_node_attr(pred_node);
+ unsigned lat;
- if (is_Phi(pred) || block != get_nodes_block(pred) || is_NoMem(pred))
- continue;
-
- pred_node = get_ilpsched_irn(env, pred);
- pna = get_ilpsched_node_attr(pred_node);
- idx = get_irn_idx(irn);
-
- assert(pna->asap && "missing ASAP of predecessor");
-
- /*
- We have not already visited this predecessor
- -> accumulate it's predecessors
- */
- if (pna->visit_idx != idx) {
- pna->visit_idx = idx;
- na->transitive_block_nodes = bitset_or(na->transitive_block_nodes, pna->transitive_block_nodes);
- DBG((env->dbg, LEVEL_3, "\taccumulating preds of %+F\n", pred));
+ lat = fixed_latency(env->sel, pred, env->block_env);
+ na->latency = lat;
+ na->asap = MAX(na->asap, pna->asap + lat);
}
}
- /* every node is it's own transitive predecessor in block */
- bitset_set(na->transitive_block_nodes, na->block_idx);
+ /* add node to ILP node list and update max_steps */
+ block_node = get_ilpsched_irn(env, block);
+ ba = get_ilpsched_block_attr(block_node);
- /* asap = number of transitive predecessors in this block */
- na->asap = bitset_popcnt(na->transitive_block_nodes);
+ set_irn_link(irn, ba->head_ilp_nodes);
+ ba->head_ilp_nodes = irn;
+ ba->max_steps += fixed_latency(env->sel, irn, env->block_env);
- DBG((env->dbg, LEVEL_2, "\tcalculated ASAP is %u\n", na->asap));
+ DB((env->dbg, LEVEL_2, "%u\n", na->asap));
}
/**
- * Accumulate the successors of all nodes from irn on upwards.
+ * Calculate the ALAP scheduling step of all irns in current block.
+ * Depends on max_steps being calculated.
*/
-static void accumulate_succs(be_ilpsched_env_t *env, ir_node *irn) {
- unsigned i, n;
- be_ilpsched_irn_t *node = get_ilpsched_irn(env, irn);
- ilpsched_node_attr_t *na = get_ilpsched_node_attr(node);
- ir_node *block = get_nodes_block(irn);
- waitq *wq = new_waitq();
-
- DBG((env->dbg, LEVEL_3, "\taccumulating succs of %+F\n", irn));
-
- /* enqueue node for final alap calculation */
- if (! na->enqueued) {
- be_ilpsched_irn_t *block_node = get_ilpsched_irn(env, block);
- ilpsched_block_attr_t *ba = get_ilpsched_block_attr(block_node);
-
- na->enqueued = 1;
- na->alap = ba->n_interesting_nodes;
- waitq_put(env->alap_queue, node);
-
- set_irn_link(irn, ba->head_ilp_nodes);
- ba->head_ilp_nodes = irn;
- DBG((env->dbg, LEVEL_5, "\t\tlinked %+F to ilp nodes of %+F, attr %p\n", irn, block, ba));
- DBG((env->dbg, LEVEL_4, "\t\tenqueueing %+F for final ALAP calculation\n", irn));
+static void calculate_block_alap(ir_node *block, void *walk_env) {
+ be_ilpsched_env_t *env = walk_env;
+ be_ilpsched_irn_t *block_node = get_ilpsched_irn(env, block);
+ ilpsched_block_attr_t *ba = get_ilpsched_block_attr(block_node);
+ waitq *cur_queue = new_waitq();
+ plist_element_t *el;
+
+ assert(is_Block(block));
+
+ DBG((env->dbg, LEVEL_2, "Calculating ALAP for nodes in %+F (%u nodes, %u max steps)\n",
+ block, ba->n_interesting_nodes, ba->max_steps));
+
+ /* TODO: Might be faster to use out edges and call phase_reinit_single_irn_data */
+ //phase_reinit_block_irn_data(&env->ph, block);
+
+ /* init start queue */
+ foreach_plist(ba->root_nodes, el) {
+ waitq_put(cur_queue, plist_element_get_value(el));
}
- for (i = 0, n = get_irn_ins_or_deps(irn); i < n; ++i) {
- ir_node *pred = skip_Proj(get_irn_in_or_dep(irn, i));
- unsigned idx;
- be_ilpsched_irn_t *pred_node;
- ilpsched_node_attr_t *pna;
+ /* repeat until all nodes are processed */
+ while (! waitq_empty(cur_queue)) {
+ waitq *next_queue = new_waitq();
+
+ /* process all nodes in current step */
+ while (! waitq_empty(cur_queue)) {
+ ir_node *cur_irn = waitq_get(cur_queue);
+ be_ilpsched_irn_t *node = get_ilpsched_irn(env, cur_irn);
+ ilpsched_node_attr_t *na = get_ilpsched_node_attr(node);
+ int i;
+
+ /* cur_node has no alap set -> it's a root node, set to max alap */
+ if (na->alap == 0) {
+ na->alap = ba->max_steps;
+ DBG((env->dbg, LEVEL_2, "setting ALAP of node %+F to %u, handling preds:\n",
+ cur_irn, na->alap));
+ }
+ else {
+ DBG((env->dbg, LEVEL_2, "ALAP of node %+F is %u, handling preds:\n",
+ cur_irn, na->alap));
+ }
- if (be_is_Keep(pred))
- pred = skip_Proj(get_irn_n(pred, 0));
+ /* set the alap's of all predecessors */
+ for (i = get_irn_ins_or_deps(cur_irn) - 1; i >= 0; --i) {
+ ir_node *pred = skip_normal_Proj(env->arch_env->isa, get_irn_in_or_dep(cur_irn, i));
- if (is_Phi(pred) || block != get_nodes_block(pred) || is_NoMem(pred))
- continue;
+ /* check for greatest distance to bottom */
+ if (! is_Phi(pred) && ! is_NoMem(pred) && get_nodes_block(pred) == block) {
+ be_ilpsched_irn_t *pred_node = get_ilpsched_irn(env, pred);
+ ilpsched_node_attr_t *pna = get_ilpsched_node_attr(pred_node);
+ unsigned lat;
- pred_node = get_ilpsched_irn(env, pred);
- pna = get_ilpsched_node_attr(pred_node);
- idx = get_irn_idx(irn);
+ /* mark the predecessor as visited by current irn */
+ if (pna->visit_idx == get_irn_idx(cur_irn) && ! na->alap_changed)
+ continue;
+ pna->visit_idx = get_irn_idx(cur_irn);
- /* accumulate the successors */
- if (pna->visit_idx != idx) {
- pna->visit_idx = idx;
- pna->transitive_block_nodes = bitset_or(pna->transitive_block_nodes, na->transitive_block_nodes);
+ lat = fixed_latency(env->sel, pred, env->block_env);
- /* set current node as successor */
- bitset_set(pna->transitive_block_nodes, na->block_idx);
- waitq_put(wq, pred);
+ /* set ALAP of current pred */
+ if (pna->alap == 0) {
+ /* current ALAP is 0: node has not yet been visited */
+ pna->alap_changed = 1;
+ pna->alap = na->alap - lat;
+ }
+ else if (pna->alap > na->alap - lat) {
+ /* we found a longer path to root node: change ALAP */
+ pna->alap = na->alap - lat;
+ pna->alap_changed = 1;
+ }
+ else {
+ /* current ALAP is best found so far: keep it */
+ pna->alap_changed = 0;
+ }
+
+ DBG((env->dbg, LEVEL_2, "\tsetting ALAP of node %+F to %u\n", pred, pna->alap));
- DBG((env->dbg, LEVEL_3, "\taccumulating succs of %+F to %+F\n", irn, pred));
+ /* enqueue node for next iteration */
+ if (get_irn_ins_or_deps(pred) > 0)
+ waitq_put(next_queue, pred);
+ }
+ }
}
- }
- /* process all predecessors */
- while (! waitq_empty(wq)) {
- accumulate_succs(env, waitq_get(wq));
+ /* prepare for next iteration */
+ del_waitq(cur_queue);
+ cur_queue = next_queue;
}
-
- del_waitq(wq);
}
/**
- * Calculate the ALAP scheduling step of all irns in current block.
- * Depends on ASAP beeing calculated.
+ * Free list of root nodes and the set of live-in nodes.
*/
-static void calculate_block_alap(ir_node *block, void *walk_env) {
+static void clear_unwanted_data(ir_node *block, void *walk_env) {
be_ilpsched_env_t *env = walk_env;
be_ilpsched_irn_t *block_node = get_ilpsched_irn(env, block);
ilpsched_block_attr_t *ba = get_ilpsched_block_attr(block_node);
- assert(is_Block(block));
-
- DBG((env->dbg, LEVEL_2, "Calculating ALAP for nodes in %+F (%u nodes)\n", block, ba->n_interesting_nodes));
-
- /* TODO: Might be faster to use out edges and call phase_reinit_single_irn_data */
- phase_reinit_block_irn_data(&env->ph, block);
-
- /* calculate the alap of all nodes, starting at collected roots upwards */
- while (! waitq_empty(ba->root_nodes)) {
- accumulate_succs(env, waitq_get(ba->root_nodes));
- }
-
- /* we don't need it anymore */
- del_waitq(ba->root_nodes);
+ plist_free(ba->root_nodes);
ba->root_nodes = NULL;
+ del_pset(ba->livein_nodes);
+ ba->livein_nodes = NULL;
+}
- /* all interesting nodes should have their successors accumulated now */
- while (! waitq_empty(env->alap_queue)) {
- be_ilpsched_irn_t *node = waitq_get(env->alap_queue);
- ilpsched_node_attr_t *na = get_ilpsched_node_attr(node);
+/**
+ * Refine the {ASAP(n), ALAP(n)} interval for the nodes.
+ * Set the ASAP/ALAP times of Projs and Keeps to their ancestor ones.
+ */
+static void refine_asap_alap_times(ir_node *irn, void *walk_env) {
+ be_ilpsched_env_t *env = walk_env;
+ ir_node *pred = irn;
+ be_ilpsched_irn_t *node, *pred_node;
+ ilpsched_node_attr_t *na, *pna;
- /* control flow ops must always be scheduled last */
- if (is_cfop(node->irn) && ! is_Start(node->irn) && get_irn_opcode(node->irn) != iro_End)
- na->asap = na->alap;
- else
- na->alap -= bitset_popcnt(na->transitive_block_nodes);
- DBG((env->dbg, LEVEL_2, "\tALAP of %+F is %u (%u succs)\n", node->irn, na->alap, bitset_popcnt(na->transitive_block_nodes)));
- }
-}
+ if (! consider_for_sched(env->arch_env->isa, irn))
+ return;
+ if (! is_Proj(irn) && ! be_is_Keep(irn))
+ return;
-/*************************
- * _____ _ _____
- * |_ _| | | __ \
- * | | | | | |__) |
- * | | | | | ___/
- * _| |_| |____| |
- * |_____|______|_|
- *
- *************************/
+ /* go to the ancestor */
+ if (be_is_Keep(irn))
+ pred = get_irn_n(irn, 0);
+ pred = skip_Proj(pred);
-/**
- * Check if node can be executed on given unit type.
- */
-static INLINE int is_valid_unit_type_for_node(const be_execution_unit_type_t *tp, be_ilpsched_irn_t *node) {
- int i;
- ilpsched_node_attr_t *na = get_ilpsched_node_attr(node);
+ node = get_ilpsched_irn(env, irn);
+ pred_node = get_ilpsched_irn(env, pred);
+ na = get_ilpsched_node_attr(node);
+ pna = get_ilpsched_node_attr(pred_node);
- for (i = na->n_unit_types - 1; i >= 0; --i) {
- if (na->type_info[i].tp == tp)
- return i;
- }
+ na->asap = pna->asap;
+ na->alap = pna->alap;
- return -1;
+ /* record all Projs and Keeps for this node */
+ if (! pna->projkeeps)
+ pna->projkeeps = new_waitq();
+ waitq_put(pna->projkeeps, irn);
+
+ DBG((env->dbg, LEVEL_2, "fixing ASAP/ALAP of %+F to %u/%u\n", irn, na->asap, na->alap));
}
+/*******************************************
+ * _ _ _
+ * | | | | | |
+ * ___ ___| |__ ___ __| |_ _| | ___
+ * / __|/ __| '_ \ / _ \/ _` | | | | |/ _ \
+ * \__ \ (__| | | | __/ (_| | |_| | | __/
+ * |___/\___|_| |_|\___|\__,_|\__,_|_|\___|
+ *
+ *******************************************/
+
static INLINE void check_for_keeps(waitq *keeps, ir_node *block, ir_node *irn) {
const ir_edge_t *edge;
}
}
+/**
+ * Inserts @p irn before @p before into schedule and notifies backend.
+ */
+static INLINE void notified_sched_add_before(be_ilpsched_env_t *env,
+ ir_node *before, ir_node *irn, unsigned cycle)
+{
+ be_ilp_sched_node_scheduled(env->sel, irn, cycle, env->block_env);
+ sched_add_before(before, irn);
+}
+
/**
* Adds a node, it's Projs (in case of mode_T nodes) and
* it's Keeps to schedule.
*/
-static void add_to_sched(ir_node *block, ir_node *irn) {
+static void add_to_sched(be_ilpsched_env_t *env, ir_node *block, ir_node *irn, unsigned cycle) {
const ir_edge_t *edge;
waitq *keeps = new_waitq();
+ /* mode_M nodes are not scheduled */
if (get_irn_mode(irn) == mode_M)
return;
- sched_add_before(block, irn);
+ if (! sched_is_scheduled(irn))
+ notified_sched_add_before(env, block, irn, cycle);
/* add Projs */
if (get_irn_mode(irn) == mode_T) {
foreach_out_edge(irn, edge) {
ir_node *user = get_edge_src_irn(edge);
- assert(is_Proj(user) && "User of mode_T node must be a Proj");
-
- if (to_appear_in_schedule(user))
- sched_add_before(block, user);
+ if ((to_appear_in_schedule(user) || get_irn_mode(user) == mode_b) &&
+ get_irn_n_edges(user) > 0)
+ {
+ notified_sched_add_before(env, block, user, cycle);
+ }
check_for_keeps(keeps, block, user);
}
while (! waitq_empty(keeps)) {
ir_node *keep = waitq_get(keeps);
if (! sched_is_scheduled(keep))
- sched_add_before(block, keep);
+ notified_sched_add_before(env, block, keep, cycle);
}
del_waitq(keeps);
ARR_APP1(be_ilpsched_irn_t *, sched_nodes, node);
}
else {
+ /* check all nodes for their positive solution */
foreach_linked_irns(ba->head_ilp_nodes, irn) {
be_ilpsched_irn_t *node;
ilpsched_node_attr_t *na;
cur_var = 0;
found = 0;
+ if (! na->is_dummy_node) {
+ for (tp_idx = na->n_unit_types - 1; ! found && tp_idx >= 0; --tp_idx) {
+ for (t = na->asap - 1; ! found && t <= na->alap - 1; ++t) {
+ double cost = lpp_get_var_sol(lpp, na->ilp_vars.y[cur_var]);
+
+ if (! LPP_VALUE_IS_0(cost)) {
+ DBG((env->dbg, LEVEL_3, "%+F has additional regpressure costs of %f\n", irn, cost));
+ found = 1;
+ }
+ }
+ }
+ }
+
+ found = 0;
+ /* go over all variables of a node until the non-zero one is found */
for (tp_idx = na->n_unit_types - 1; ! found && tp_idx >= 0; --tp_idx) {
for (t = na->asap - 1; ! found && t <= na->alap - 1; ++t) {
- double val = lpp_get_var_sol(lpp, na->ilp_vars[cur_var++]);
+ double val = lpp_get_var_sol(lpp, na->ilp_vars.x[cur_var++]);
+
+ /* check, if variable is set to one (it's not zero then :) */
if (! LPP_VALUE_IS_0(val)) {
na->sched_point = t;
ARR_APP1(be_ilpsched_irn_t *, sched_nodes, node);
- DBG((env->dbg, LEVEL_1, "Schedpoint of %+F is %u at unit type %s\n",
+ DBG((env->dbg, LEVEL_2, "Schedpoint of %+F is %u at unit type %s\n",
irn, t, na->type_info[tp_idx].tp->name));
found = 1;
}
}
}
+ glob_heights = env->height;
/* sort nodes ascending by scheduling time step */
qsort(sched_nodes, ARR_LEN(sched_nodes), sizeof(sched_nodes[0]), cmp_ilpsched_irn);
}
switch (get_irn_opcode(irn)) {
case iro_Phi:
- add_to_sched(block, irn);
+ add_to_sched(env, block, irn, 0);
break;
case iro_Start:
case iro_End:
case iro_Proj:
case iro_Bad:
+ case iro_Unknown:
break;
default:
if (is_cfop(irn)) {
- assert(cfop == NULL && "Highlander!");
+ assert(cfop == NULL && "Highlander - there can be only one");
cfop = irn;
}
break;
/* add all nodes from list */
for (i = 0, l = ARR_LEN(sched_nodes); i < l; ++i) {
- add_to_sched(block, sched_nodes[i]->irn);
+ ilpsched_node_attr_t *na = get_ilpsched_node_attr(sched_nodes[i]);
+ if (sched_nodes[i]->irn != cfop)
+ add_to_sched(env, block, sched_nodes[i]->irn, na->sched_point);
}
/* schedule control flow node if not already done */
if (cfop && ! sched_is_scheduled(cfop))
- add_to_sched(block, cfop);
+ add_to_sched(env, block, cfop, 0);
DEL_ARR_F(sched_nodes);
}
+/***************************************************************
+ * _____ _ _____ _____ _ _
+ * |_ _| | | __ \ / ____| | | (_)
+ * | | | | | |__) | | (___ ___ ___| |_ _ ___ _ __
+ * | | | | | ___/ \___ \ / _ \/ __| __| |/ _ \| '_ \
+ * _| |_| |____| | ____) | __/ (__| |_| | (_) | | | |
+ * |_____|______|_| |_____/ \___|\___|\__|_|\___/|_| |_|
+ *
+ ***************************************************************/
+
/**
- * Create the ilp (add variables, build constraints, solve, build schedule from solution).
+ * Check if node can be executed on given unit type.
*/
-static void create_ilp(ir_node *block, void *walk_env) {
- be_ilpsched_env_t *env = walk_env;
- be_ilpsched_irn_t *block_node = get_ilpsched_irn(env, block);
- ilpsched_block_attr_t *ba = get_ilpsched_block_attr(block_node);
- unsigned num_block_var = 0;
- unsigned num_nodes = 0;
- unsigned n_instr_max = env->cpu->bundle_size * env->cpu->bundels_per_cycle;
- bitset_t *bs_block_irns = bitset_alloca(ba->block_last_idx);
- FILE *logfile = NULL;
- lpp_t *lpp = NULL;
- unsigned num_cst_assign, num_cst_prec, num_cst_resrc, num_cst_bundle;
- unsigned t;
- int glob_type_idx;
- ir_node *irn;
- char buf[1024];
- struct obstack var_obst;
- lc_timer_t *t_var;
- lc_timer_t *t_cst_assign;
- lc_timer_t *t_cst_prec;
- lc_timer_t *t_cst_rsrc;
- lc_timer_t *t_cst_bundle;
- double fact_var, fact_cst;
-
-#ifdef WITH_LIBCORE
- t_var = lc_timer_register("beilpsched_var", "create ilp variables");
- t_cst_assign = lc_timer_register("beilpsched_cst_assign", "create assignment constraints");
- t_cst_prec = lc_timer_register("beilpsched_cst_prec", "create precedence constraints");
- t_cst_rsrc = lc_timer_register("beilpsched_cst_rsrc", "create resource constraints");
- t_cst_bundle = lc_timer_register("beilpsched_cst_bundle", "create bundle constraints");
- LC_STOP_AND_RESET_TIMER(t_var);
- LC_STOP_AND_RESET_TIMER(t_cst_assign);
- LC_STOP_AND_RESET_TIMER(t_cst_prec);
- LC_STOP_AND_RESET_TIMER(t_cst_rsrc);
- LC_STOP_AND_RESET_TIMER(t_cst_bundle);
-#endif /* WITH_LIBCORE */
+static INLINE int is_valid_unit_type_for_node(const be_execution_unit_type_t *tp, be_ilpsched_irn_t *node) {
+ int i;
+ ilpsched_node_attr_t *na = get_ilpsched_node_attr(node);
- DBG((env->dbg, 255, "\n\n\n=========================================\n"));
- DBG((env->dbg, 255, " ILP Scheduling for %+F\n", block));
- DBG((env->dbg, 255, "=========================================\n\n"));
+ for (i = na->n_unit_types - 1; i >= 0; --i) {
+ if (na->type_info[i].tp == tp)
+ return i;
+ }
- DBG((env->dbg, LEVEL_1, "Creating ILP Variables for nodes in %+F (%u interesting nodes)\n",
- block, ba->n_interesting_nodes));
+ return -1;
+}
- if (ba->n_interesting_nodes > 1) {
- fact_var = ba->n_interesting_nodes < 25 ? 1.0 : 0.6;
- fact_cst = ba->n_interesting_nodes < 25 ? 0.8 : 0.6;
+/************************************************
+ * _ _ _
+ * (_) | | | |
+ * __ ____ _ _ __ _ __ _| |__ | | ___ ___
+ * \ \ / / _` | '__| |/ _` | '_ \| |/ _ \/ __|
+ * \ V / (_| | | | | (_| | |_) | | __/\__ \
+ * \_/ \__,_|_| |_|\__,_|_.__/|_|\___||___/
+ *
+ ************************************************/
+
+static int be_ilpsched_set_type_info(be_ilpsched_env_t *env, ir_node *irn, struct obstack *obst) {
+ const be_execution_unit_t ***execunits = arch_isa_get_allowed_execution_units(env->arch_env->isa, irn);
+ unsigned n_unit_types = 0;
+ be_ilpsched_irn_t *node;
+ ilpsched_node_attr_t *na;
+ unsigned unit_idx, tp_idx;
+
+ /* count number of available unit types for this node */
+ for (n_unit_types = 0; execunits[n_unit_types]; ++n_unit_types)
+ /* just count */ ;
+
+ node = get_ilpsched_irn(env, irn);
+ na = get_ilpsched_node_attr(node);
+
+ if (! na->type_info) {
+ na->n_unit_types = n_unit_types;
+ na->type_info = NEW_ARR_D(unit_type_info_t, obst, n_unit_types);
+
+ /* fill the type info array */
+ for (tp_idx = 0; tp_idx < n_unit_types; ++tp_idx) {
+ for (unit_idx = 0; execunits[tp_idx][unit_idx]; ++unit_idx) {
+ /* beware: we also count number of available units here */
+ if (be_machine_is_dummy_unit(execunits[tp_idx][unit_idx]))
+ na->is_dummy_node = 1;
+ }
- lpp = new_lpp_userdef(
- "be ilp scheduling",
- lpp_minimize,
- (int)((double)(ba->n_interesting_nodes * ba->n_interesting_nodes) * fact_var) + 1, /* num vars */
- (int)((double)(ba->n_interesting_nodes * ba->n_interesting_nodes) * fact_cst) + 20, /* num cst */
- 1.2); /* grow factor */
- obstack_init(&var_obst);
+ na->type_info[tp_idx].tp = execunits[tp_idx][0]->tp;
+ na->type_info[tp_idx].n_units = unit_idx;
+ }
+ }
- lc_timer_push(t_var);
- foreach_linked_irns(ba->head_ilp_nodes, irn) {
- const be_execution_unit_t ***execunits = arch_isa_get_allowed_execution_units(env->arch_env->isa, irn);
- be_ilpsched_irn_t *node;
- ilpsched_node_attr_t *na;
- unsigned n_unit_types, tp_idx, unit_idx, cur_var, n_var, cur_unit;
+ return n_unit_types;
+}
- /* count number of available unit types for this node */
- for (n_unit_types = 0; execunits[n_unit_types]; ++n_unit_types)
- /* just count */ ;
+/**
+ * Returns the largest alap time of a user of @p irn.
+ * The user must be in block @p block.
+ */
+static unsigned be_ilpsched_get_max_alap_user(be_ilpsched_env_t *env, ir_node *irn, ir_node *block) {
+ const ir_edge_t *edge;
+ unsigned max_alap = 0;
- node = get_ilpsched_irn(env, irn);
- na = get_ilpsched_node_attr(node);
+ foreach_out_edge(irn, edge) {
+ ir_node *user = get_edge_src_irn(edge);
+
+ if (get_nodes_block(user) == block) {
+ be_ilpsched_irn_t *node = get_ilpsched_irn(env, user);
+ ilpsched_node_attr_t *na = get_ilpsched_node_attr(node);
- na->n_unit_types = n_unit_types;
- na->type_info = NEW_ARR_D(unit_type_info_t, &var_obst, n_unit_types);
+ max_alap = MAX(max_alap, na->alap);
+ }
+ }
- /* fill the type info array */
- for (tp_idx = 0; tp_idx < n_unit_types; ++tp_idx) {
- for (unit_idx = 0; execunits[tp_idx][unit_idx]; ++unit_idx)
- /* just count units of this type */;
+ assert(max_alap > 0);
+ return max_alap;
+}
- na->type_info[tp_idx].tp = execunits[tp_idx][0]->tp;
- na->type_info[tp_idx].n_units = unit_idx;
- }
+/**
+ * Create the following variables:
+ * - x_{nt}^k binary weigthed with: t
+ * node n is scheduled at time step t to unit type k
+ * ==>> These variables represent the schedule
+ *
+ * - a_{nt}^k binary weighted with num_nodes
+ * node n is alive at time step t on unit type k
+ *
+ * - y_{nt}^k continuous weighted with: num_nodes^2
+ * register pressure over limit for unit type k
+ * ==>> These variables represent the register pressure
+ *
+ */
+static void create_variables(be_ilpsched_env_t *env, lpp_t *lpp, be_ilpsched_irn_t *block_node, struct obstack *var_obst) {
+ char buf[1024];
+ ir_node *irn;
+ unsigned num_block_var, num_nodes;
+ ilp_livein_node_t *livein;
+ ilpsched_block_attr_t *ba = get_ilpsched_block_attr(block_node);
+ unsigned weigth_y = ba->n_interesting_nodes * ba->n_interesting_nodes;
+ lc_timer_t *t_var = lc_timer_register("beilpsched_var", "create ilp variables");
+
+ ilp_timer_push(t_var);
+ num_block_var = num_nodes = 0;
+ foreach_linked_irns(ba->head_ilp_nodes, irn) {
+ be_ilpsched_irn_t *node;
+ ilpsched_node_attr_t *na;
+ unsigned n_unit_types, tp_idx, n_var, cur_unit;
+ unsigned cur_var_ad, cur_var_x, cur_var_y, num_ad;
+ int i;
+
+ node = get_ilpsched_irn(env, irn);
+ na = get_ilpsched_node_attr(node);
+ n_unit_types = be_ilpsched_set_type_info(env, irn, var_obst);
+
+ /* allocate space for ilp variables */
+ na->ilp_vars.x = NEW_ARR_D(int, var_obst, n_unit_types * VALID_SCHED_INTERVAL(na));
+ memset(na->ilp_vars.x, -1, ARR_LEN(na->ilp_vars.x) * sizeof(na->ilp_vars.x[0]));
+
+ /* we need these variables only for "real" nodes */
+ if (! na->is_dummy_node) {
+ na->ilp_vars.y = NEW_ARR_D(int, var_obst, n_unit_types * VALID_SCHED_INTERVAL(na));
+ memset(na->ilp_vars.y, -1, ARR_LEN(na->ilp_vars.y) * sizeof(na->ilp_vars.y[0]));
+
+ num_ad = ba->max_steps - na->asap + 1;
+ na->ilp_vars.a = NEW_ARR_D(int, var_obst, n_unit_types * num_ad);
+ memset(na->ilp_vars.a, -1, ARR_LEN(na->ilp_vars.a) * sizeof(na->ilp_vars.a[0]));
+ }
- /* allocate space for ilp variables */
- na->ilp_vars = NEW_ARR_D(int, &var_obst, n_unit_types * VALID_SCHED_INTERVAL(na));
- memset(na->ilp_vars, -1, n_unit_types * VALID_SCHED_INTERVAL(na) * sizeof(na->ilp_vars[0]));
+ DBG((env->dbg, LEVEL_3, "\thandling %+F (asap %u, alap %u, unit types %u):\n",
+ irn, na->asap, na->alap, na->n_unit_types));
- DBG((env->dbg, LEVEL_3, "\thandling %+F (asap %u, alap %u, unit types %u):\n",
- irn, na->asap, na->alap, na->n_unit_types));
+ cur_var_x = cur_var_ad = cur_var_y = cur_unit = n_var = 0;
+ /* create variables */
+ for (tp_idx = 0; tp_idx < n_unit_types; ++tp_idx) {
+ unsigned t;
- cur_var = cur_unit = n_var = 0;
- /* create variables */
- for (tp_idx = 0; tp_idx < n_unit_types; ++tp_idx) {
- for (t = na->asap - 1; t <= na->alap - 1; ++t) {
- snprintf(buf, sizeof(buf), "n%u_%s_%u",
+ for (t = na->asap - 1; t <= na->alap - 1; ++t) {
+ /* x_{nt}^k variables */
+ snprintf(buf, sizeof(buf), "x_n%u_%s_%u",
+ get_irn_idx(irn), na->type_info[tp_idx].tp->name, t);
+ na->ilp_vars.x[cur_var_x++] = lpp_add_var(lpp, buf, lpp_binary, (double)(t + 1));
+ DBG((env->dbg, LEVEL_4, "\t\tcreated ILP variable %s\n", buf));
+ /* variable counter */
+ n_var++;
+ num_block_var++;
+
+ if (! na->is_dummy_node) {
+ /* y_{nt}^k variables */
+ snprintf(buf, sizeof(buf), "y_n%u_%s_%u",
get_irn_idx(irn), na->type_info[tp_idx].tp->name, t);
- na->ilp_vars[cur_var++] = lpp_add_var(lpp, buf, lpp_binary, (double)(t + 1));
+ na->ilp_vars.y[cur_var_y++] = lpp_add_var(lpp, buf, lpp_continous, (double)(weigth_y));
+ DBG((env->dbg, LEVEL_4, "\t\tcreated ILP variable %s\n", buf));
+
+ /* variable counter */
n_var++;
num_block_var++;
+ }
+ }
+
+ /* a node can die at any step t: asap(n) <= t <= U */
+ if (! na->is_dummy_node) {
+ for (t = na->asap - 1; t <= ba->max_steps; ++t) {
+
+ /* a_{nt}^k variables */
+ snprintf(buf, sizeof(buf), "a_n%u_%s_%u",
+ get_irn_idx(irn), na->type_info[tp_idx].tp->name, t);
+ na->ilp_vars.a[cur_var_ad++] = lpp_add_var(lpp, buf, lpp_binary, (double)(ba->n_interesting_nodes));
DBG((env->dbg, LEVEL_4, "\t\tcreated ILP variable %s\n", buf));
+
+ /* variable counter */
+ n_var++;
+ num_block_var++;
}
}
- DB((env->dbg, LEVEL_3, "%u variables created\n", n_var));
- num_nodes++;
+ /* collect live-in nodes */
+ for (i = get_irn_arity(irn) - 1; i >= 0; --i) {
+ ir_node *pred = get_irn_n(irn, i);
+
+ if (get_nodes_block(pred) != block_node->irn && consider_for_sched(env->arch_env->isa, pred)) {
+ be_ilpsched_set_type_info(env, pred, var_obst);
+ if (! na->is_dummy_node) {
+ ilp_livein_node_t *entry = obstack_alloc(var_obst, sizeof(*entry));
+ entry->irn = pred;
+ entry->a = NULL;
+ pset_insert(ba->livein_nodes, entry, (unsigned)get_irn_idx(pred));
+ }
+ }
+ }
}
- lc_timer_pop();
- DBG((env->dbg, LEVEL_1, "... %u variables for %u nodes created (%g sec)\n",
- num_block_var, num_nodes, lc_timer_elapsed_usec(t_var) / 1000000.0));
- /*
- 1st:
- - the assignment constraints:
- assure each node is executed once by exactly one (allowed) execution unit
- - the precedence constraints:
- assure that no data dependencies are violated
- */
- num_cst_assign = num_cst_prec = num_cst_resrc = num_cst_bundle = 0;
- DBG((env->dbg, LEVEL_1, "Creating constraints for nodes in %+F:\n", block));
- foreach_linked_irns(ba->head_ilp_nodes, irn) {
- int cst, tp_idx, i;
- unsigned cur_var;
- be_ilpsched_irn_t *node;
- ilpsched_node_attr_t *na;
+ DB((env->dbg, LEVEL_3, "%u variables created\n", n_var));
+ num_nodes++;
+ }
- /* the assignment constraint */
- lc_timer_push(t_cst_assign);
- snprintf(buf, sizeof(buf), "assignment_cst_n%u", get_irn_idx(irn));
- cst = lpp_add_cst_uniq(lpp, buf, lpp_equal, 1.0);
- DBG((env->dbg, LEVEL_2, "added constraint %s\n", buf));
- num_cst_assign++;
+ /* create alive variables a_{nt}^k for live-ins */
+ foreach_pset(ba->livein_nodes, livein) {
+ be_ilpsched_irn_t *node;
+ ilpsched_node_attr_t *na;
+ unsigned tp_idx, var_idx;
+ ir_node *irn;
- node = get_ilpsched_irn(env, irn);
- na = get_ilpsched_node_attr(node);
- cur_var = 0;
+ irn = livein->irn;
+ node = get_ilpsched_irn(env, irn);
+ na = get_ilpsched_node_attr(node);
- lpp_set_factor_fast_bulk(lpp, cst, na->ilp_vars, na->n_unit_types * VALID_SCHED_INTERVAL(na), 1.0);
- lc_timer_pop();
+ livein->max_alive_steps = be_ilpsched_get_max_alap_user(env, irn, block_node->irn);
- /* the precedence constraints */
- lc_timer_push(t_cst_prec);
- bs_block_irns = bitset_clear_all(bs_block_irns);
- for (i = get_irn_ins_or_deps(irn) - 1; i >= 0; --i) {
- ir_node *pred = skip_Proj(get_irn_in_or_dep(irn, i));
- unsigned t_low, t_high;
- be_ilpsched_irn_t *pred_node;
- ilpsched_node_attr_t *pna;
+ livein->a = NEW_ARR_D(int, var_obst, na->n_unit_types * livein->max_alive_steps);
+ var_idx = 0;
- if (be_is_Keep(pred))
- pred = skip_Proj(get_irn_n(pred, 0));
+ /* create variables */
+ for (tp_idx = 0; tp_idx < na->n_unit_types; ++tp_idx) {
+ unsigned t;
- if (is_Phi(pred) || block != get_nodes_block(pred) || is_NoMem(pred))
- continue;
+ for (t = 0; t < livein->max_alive_steps; ++t) {
+ /* a_{nt}^k variables */
+ snprintf(buf, sizeof(buf), "al_n%u_%s_%u",
+ get_irn_idx(irn), na->type_info[tp_idx].tp->name, t);
+ livein->a[var_idx++] = lpp_add_var(lpp, buf, lpp_binary, (double)(ba->n_interesting_nodes));
+ DBG((env->dbg, LEVEL_4, "\t\tcreated ILP variable %s\n", buf));
+ num_block_var++;
+ }
+ }
+ }
- pred_node = get_ilpsched_irn(env, pred);
- pna = get_ilpsched_node_attr(pred_node);
+ ilp_timer_pop();
+ DBG((env->dbg, LEVEL_1, "... %u variables for %u nodes created (%g sec)\n",
+ num_block_var, num_nodes, ilp_timer_elapsed_usec(t_var) / 1000000.0));
+}
- assert(pna->asap > 0 && pna->alap >= pna->asap && "Invalid scheduling interval.");
+/*******************************************************
+ * _ _ _
+ * | | (_) | |
+ * ___ ___ _ __ ___| |_ _ __ __ _ _ _ __ | |_ ___
+ * / __/ _ \| '_ \/ __| __| '__/ _` | | '_ \| __/ __|
+ * | (_| (_) | | | \__ \ |_| | | (_| | | | | | |_\__ \
+ * \___\___/|_| |_|___/\__|_| \__,_|_|_| |_|\__|___/
+ *
+ *******************************************************/
- if (! bitset_is_set(bs_block_irns, pna->block_idx))
- bitset_set(bs_block_irns, pna->block_idx);
- else
- continue;
+/**
+ * Collect all operands and nodes @p irn depends on.
+ * If there is a Proj within the dependencies, all other Projs of the parent node are added as well.
+ */
+static nodeset *sta_collect_in_deps(ir_node *irn, nodeset *deps) {
+ int i;
- /* irn = n, pred = m */
- t_low = MAX(na->asap, pna->asap);
- t_high = MIN(na->alap, pna->alap);
- for (t = t_low - 1; t <= t_high - 1; ++t) {
- int tn, tm, cur_idx;
- int int_na = (t >= na->asap - 1) ? t - na->asap + 2 : 0;
- int int_pna = (t < pna->alap) ? pna->alap - t : 0;
- int *tmp_var_idx = NEW_ARR_F(int, int_na * na->n_unit_types + int_pna * pna->n_unit_types);
-
- snprintf(buf, sizeof(buf), "precedence_n%u_n%u_%u", get_irn_idx(pred), get_irn_idx(irn), t);
- cst = lpp_add_cst_uniq(lpp, buf, lpp_less, 1.0);
- DBG((env->dbg, LEVEL_2, "added constraint %s\n", buf));
- num_cst_prec++;
-
- cur_idx = 0;
-
- /* lpp_set_factor_fast_bulk needs variables sorted ascending by index */
- if (na->ilp_vars[0] < pna->ilp_vars[0]) {
- /* node variables have smaller index than pred variables */
- for (tp_idx = na->n_unit_types - 1; tp_idx >= 0; --tp_idx) {
- for (tn = na->asap - 1; tn <= t; ++tn) {
- unsigned idx = ILPVAR_IDX(na, tp_idx, tn);
- tmp_var_idx[cur_idx++] = na->ilp_vars[idx];
- }
+ for (i = get_irn_ins_or_deps(irn) - 1; i >= 0; --i) {
+ ir_node *p = get_irn_in_or_dep(irn, i);
+
+ if (is_Proj(p)) {
+ const ir_edge_t *edge;
+
+ p = get_Proj_pred(p);
+ foreach_out_edge(p, edge) {
+ ir_node *src = get_edge_src_irn(edge);
+ nodeset_insert(deps, src);
+ }
+ }
+ else {
+ nodeset_insert(deps, p);
+ }
+ }
+
+ return deps;
+}
+
+/**
+ * Create following ILP constraints:
+ * - the assignment constraints:
+ * assure each node is executed once by exactly one (allowed) execution unit
+ * - the dead node assignment constraints:
+ * assure a node can only die at most once
+ * - the precedence constraints:
+ * assure that no data dependencies are violated
+ */
+static void create_assignment_and_precedence_constraints(be_ilpsched_env_t *env, lpp_t *lpp, be_ilpsched_irn_t *block_node) {
+ unsigned num_cst_assign, num_cst_prec, num_cst_dead;
+ char buf[1024];
+ ir_node *irn;
+ ilpsched_block_attr_t *ba = get_ilpsched_block_attr(block_node);
+ bitset_t *bs_block_irns = bitset_alloca(ba->block_last_idx);
+ lc_timer_t *t_cst_assign = lc_timer_register("beilpsched_cst_assign", "create assignment constraints");
+ lc_timer_t *t_cst_prec = lc_timer_register("beilpsched_cst_prec", "create precedence constraints");
+
+ num_cst_assign = num_cst_prec = num_cst_dead = 0;
+ foreach_linked_irns(ba->head_ilp_nodes, irn) {
+ int cst, tp_idx;
+ unsigned cur_var;
+ be_ilpsched_irn_t *node;
+ ilpsched_node_attr_t *na;
+ ir_node *pred;
+ nodeset *deps = new_nodeset(16);
+
+ node = get_ilpsched_irn(env, irn);
+ na = get_ilpsched_node_attr(node);
+ cur_var = 0;
+
+ /* the assignment constraint */
+ ilp_timer_push(t_cst_assign);
+ snprintf(buf, sizeof(buf), "assignment_cst_n%u", get_irn_idx(irn));
+ cst = lpp_add_cst_uniq(lpp, buf, lpp_equal, 1.0);
+ DBG((env->dbg, LEVEL_2, "added constraint %s\n", buf));
+ num_cst_assign++;
+
+ lpp_set_factor_fast_bulk(lpp, cst, na->ilp_vars.x, ARR_LEN(na->ilp_vars.x), 1.0);
+ ilp_timer_pop();
+
+ /* We have separate constraints for Projs and Keeps */
+ // ILP becomes infeasible ?!?
+// if (is_Proj(irn) || be_is_Keep(irn))
+// continue;
+
+ /* the precedence constraints */
+ ilp_timer_push(t_cst_prec);
+ bs_block_irns = bitset_clear_all(bs_block_irns);
+
+ deps = sta_collect_in_deps(irn, deps);
+ foreach_nodeset(deps, pred) {
+ unsigned t_low, t_high, t;
+ be_ilpsched_irn_t *pred_node;
+ ilpsched_node_attr_t *pna;
+ unsigned delay;
+
+ pred = skip_normal_Proj(env->arch_env->isa, pred);
+ if (is_Phi(pred) || block_node->irn != get_nodes_block(pred) || is_NoMem(pred))
+ continue;
+
+ pred_node = get_ilpsched_irn(env, pred);
+ pna = get_ilpsched_node_attr(pred_node);
+
+ assert(pna->asap > 0 && pna->alap >= pna->asap && "Invalid scheduling interval.");
+
+ if (! bitset_is_set(bs_block_irns, pna->block_idx))
+ bitset_set(bs_block_irns, pna->block_idx);
+ else
+ continue;
+
+ /* irn = n, pred = m */
+ delay = fixed_latency(env->sel, pred, env->block_env);
+ t_low = MAX(na->asap, pna->asap + delay - 1);
+ t_high = MIN(na->alap, pna->alap + delay - 1);
+ for (t = t_low - 1; t <= t_high - 1; ++t) {
+ unsigned tn, tm;
+ int *tmp_var_idx = NEW_ARR_F(int, 0);
+
+ snprintf(buf, sizeof(buf), "precedence_n%u_n%u_%u", get_irn_idx(pred), get_irn_idx(irn), t);
+ cst = lpp_add_cst_uniq(lpp, buf, lpp_less, 1.0);
+ DBG((env->dbg, LEVEL_2, "added constraint %s\n", buf));
+ num_cst_prec++;
+
+ /* lpp_set_factor_fast_bulk needs variables sorted ascending by index */
+ if (na->ilp_vars.x[0] < pna->ilp_vars.x[0]) {
+ /* node variables have smaller index than pred variables */
+ for (tp_idx = na->n_unit_types - 1; tp_idx >= 0; --tp_idx) {
+ for (tn = na->asap - 1; tn <= t; ++tn) {
+ unsigned idx = ILPVAR_IDX(na, tp_idx, tn);
+ ARR_APP1(int, tmp_var_idx, na->ilp_vars.x[idx]);
}
+ }
- for (tp_idx = pna->n_unit_types - 1; tp_idx >= 0; --tp_idx) {
- for (tm = t; tm < pna->alap; ++tm) {
- unsigned idx = ILPVAR_IDX(pna, tp_idx, tm);
- tmp_var_idx[cur_idx++] = pna->ilp_vars[idx];
- }
+ for (tp_idx = pna->n_unit_types - 1; tp_idx >= 0; --tp_idx) {
+ for (tm = t - delay + 1; tm < pna->alap; ++tm) {
+ unsigned idx = ILPVAR_IDX(pna, tp_idx, tm);
+ ARR_APP1(int, tmp_var_idx, pna->ilp_vars.x[idx]);
}
}
- else {
- /* pred variables have smaller index than node variables */
- for (tp_idx = pna->n_unit_types - 1; tp_idx >= 0; --tp_idx) {
- for (tm = t; tm < pna->alap; ++tm) {
- unsigned idx = ILPVAR_IDX(pna, tp_idx, tm);
- tmp_var_idx[cur_idx++] = pna->ilp_vars[idx];
- }
+ }
+ else {
+ /* pred variables have smaller index than node variables */
+ for (tp_idx = pna->n_unit_types - 1; tp_idx >= 0; --tp_idx) {
+ for (tm = t - delay + 1; tm < pna->alap; ++tm) {
+ unsigned idx = ILPVAR_IDX(pna, tp_idx, tm);
+ ARR_APP1(int, tmp_var_idx, pna->ilp_vars.x[idx]);
}
+ }
- for (tp_idx = na->n_unit_types - 1; tp_idx >= 0; --tp_idx) {
- for (tn = na->asap - 1; tn <= t; ++tn) {
- unsigned idx = ILPVAR_IDX(na, tp_idx, tn);
- tmp_var_idx[cur_idx++] = na->ilp_vars[idx];
- }
+ for (tp_idx = na->n_unit_types - 1; tp_idx >= 0; --tp_idx) {
+ for (tn = na->asap - 1; tn <= t; ++tn) {
+ unsigned idx = ILPVAR_IDX(na, tp_idx, tn);
+ ARR_APP1(int, tmp_var_idx, na->ilp_vars.x[idx]);
}
}
+ }
+
+ if (ARR_LEN(tmp_var_idx) > 0)
+ lpp_set_factor_fast_bulk(lpp, cst, tmp_var_idx, ARR_LEN(tmp_var_idx), 1.0);
+
+ DEL_ARR_F(tmp_var_idx);
+ }
+ }
+ del_nodeset(deps);
+ ilp_timer_pop();
+ }
+ DBG((env->dbg, LEVEL_1, "\t%u assignement constraints (%g sec)\n",
+ num_cst_assign, ilp_timer_elapsed_usec(t_cst_assign) / 1000000.0));
+ DBG((env->dbg, LEVEL_1, "\t%u precedence constraints (%g sec)\n",
+ num_cst_prec, ilp_timer_elapsed_usec(t_cst_prec) / 1000000.0));
+}
+
+/**
+ * Create ILP resource constraints:
+ * - assure that for each time step not more instructions are scheduled
+ * to the same unit types as units of this type are available
+ */
+static void create_ressource_constraints(be_ilpsched_env_t *env, lpp_t *lpp, be_ilpsched_irn_t *block_node) {
+ int glob_type_idx;
+ char buf[1024];
+ unsigned num_cst_resrc = 0;
+ ilpsched_block_attr_t *ba = get_ilpsched_block_attr(block_node);
+ lc_timer_t *t_cst_rsrc = lc_timer_register("beilpsched_cst_rsrc", "create resource constraints");
+
+ ilp_timer_push(t_cst_rsrc);
+ for (glob_type_idx = env->cpu->n_unit_types - 1; glob_type_idx >= 0; --glob_type_idx) {
+ unsigned t;
+ be_execution_unit_type_t *cur_tp = &env->cpu->unit_types[glob_type_idx];
+
+ /* BEWARE: the DUMMY unit type is not in CPU, so it's skipped automatically */
+
+ /* check each time step */
+ for (t = 0; t < ba->max_steps; ++t) {
+ ir_node *irn;
+ int cst;
+ int *tmp_var_idx = NEW_ARR_F(int, 0);
+
+ snprintf(buf, sizeof(buf), "resource_cst_%s_%u", cur_tp->name, t);
+ cst = lpp_add_cst_uniq(lpp, buf, lpp_less, (double)cur_tp->n_units);
+ DBG((env->dbg, LEVEL_2, "added constraint %s\n", buf));
+ num_cst_resrc++;
- lpp_set_factor_fast_bulk(lpp, cst, tmp_var_idx, cur_idx, 1.0);
+ foreach_linked_irns(ba->head_ilp_nodes, irn) {
+ be_ilpsched_irn_t *node = get_ilpsched_irn(env, irn);
+ ilpsched_node_attr_t *na = get_ilpsched_node_attr(node);
+ int tp_idx;
- DEL_ARR_F(tmp_var_idx);
+ tp_idx = is_valid_unit_type_for_node(cur_tp, node);
+
+ if (tp_idx >= 0 && t >= na->asap - 1 && t <= na->alap - 1) {
+ int cur_var = ILPVAR_IDX(na, tp_idx, t);
+ ARR_APP1(int, tmp_var_idx, na->ilp_vars.x[cur_var]);
}
}
- lc_timer_pop();
+
+ /* set constraints if we have some */
+ if (ARR_LEN(tmp_var_idx) > 0)
+ lpp_set_factor_fast_bulk(lpp, cst, tmp_var_idx, ARR_LEN(tmp_var_idx), 1.0);
+
+ DEL_ARR_F(tmp_var_idx);
}
- DBG((env->dbg, LEVEL_1, "\t%u assignement constraints (%g sec)\n",
- num_cst_assign, lc_timer_elapsed_usec(t_cst_assign) / 1000000.0));
- DBG((env->dbg, LEVEL_1, "\t%u precedence constraints (%g sec)\n",
- num_cst_prec, lc_timer_elapsed_usec(t_cst_prec) / 1000000.0));
+ }
+ ilp_timer_pop();
+ DBG((env->dbg, LEVEL_1, "\t%u resource constraints (%g sec)\n",
+ num_cst_resrc, ilp_timer_elapsed_usec(t_cst_rsrc) / 1000000.0));
+}
- /*
- 2nd: the resource constraints:
- assure that for each time step not more instructions are scheduled
- to the same unit types as units of this type are available
- */
- lc_timer_push(t_cst_rsrc);
+/**
+ * Create ILP bundle constraints:
+ * - assure, at most bundle_size * bundles_per_cycle instructions
+ * can be started at a certain point.
+ */
+static void create_bundle_constraints(be_ilpsched_env_t *env, lpp_t *lpp, be_ilpsched_irn_t *block_node) {
+ char buf[1024];
+ unsigned t;
+ unsigned num_cst_bundle = 0;
+ unsigned n_instr_max = env->cpu->bundle_size * env->cpu->bundels_per_cycle;
+ ilpsched_block_attr_t *ba = get_ilpsched_block_attr(block_node);
+ lc_timer_t *t_cst_bundle = lc_timer_register("beilpsched_cst_bundle", "create bundle constraints");
+
+ ilp_timer_push(t_cst_bundle);
+ for (t = 0; t < ba->max_steps; ++t) {
+ ir_node *irn;
+ int cst;
+ int *tmp_var_idx = NEW_ARR_F(int, 0);
+
+ snprintf(buf, sizeof(buf), "bundle_cst_%u", t);
+ cst = lpp_add_cst_uniq(lpp, buf, lpp_less, (double)n_instr_max);
+ DBG((env->dbg, LEVEL_2, "added constraint %s\n", buf));
+ num_cst_bundle++;
+
+ foreach_linked_irns(ba->head_ilp_nodes, irn) {
+ be_ilpsched_irn_t *node;
+ ilpsched_node_attr_t *na;
+ int tp_idx;
+
+ /* Projs and Keeps do not contribute to bundle size */
+ if (is_Proj(irn) || be_is_Keep(irn))
+ continue;
+
+ node = get_ilpsched_irn(env, irn);
+ na = get_ilpsched_node_attr(node);
+
+ /* nodes assigned to DUMMY unit do not contribute to bundle size */
+ if (na->is_dummy_node)
+ continue;
+
+ if (t >= na->asap - 1 && t <= na->alap - 1) {
+ for (tp_idx = na->n_unit_types - 1; tp_idx >= 0; --tp_idx) {
+ int idx = ILPVAR_IDX(na, tp_idx, t);
+ ARR_APP1(int, tmp_var_idx, na->ilp_vars.x[idx]);
+ }
+ }
+ }
+
+ if (ARR_LEN(tmp_var_idx) > 0)
+ lpp_set_factor_fast_bulk(lpp, cst, tmp_var_idx, ARR_LEN(tmp_var_idx), 1.0);
+
+ DEL_ARR_F(tmp_var_idx);
+ }
+ ilp_timer_pop();
+ DBG((env->dbg, LEVEL_1, "\t%u bundle constraints (%g sec)\n",
+ num_cst_bundle, ilp_timer_elapsed_usec(t_cst_bundle) / 1000000.0));
+}
+
+/**
+ * Create ILP alive nodes constraints:
+ * - set variable a_{nt}^k to 1 if nodes n is alive at step t on unit k
+ */
+static void create_alive_nodes_constraint(be_ilpsched_env_t *env, lpp_t *lpp, be_ilpsched_irn_t *block_node) {
+ char buf[1024];
+ ir_node *irn;
+ unsigned num_cst = 0;
+ ilpsched_block_attr_t *ba = get_ilpsched_block_attr(block_node);
+ lc_timer_t *t_cst = lc_timer_register("beilpsched_cst_alive_nodes", "create alive nodes constraints");
+
+ ilp_timer_push(t_cst);
+ /* for each node */
+ foreach_linked_irns(ba->head_ilp_nodes, irn) {
+ be_ilpsched_irn_t *node = get_ilpsched_irn(env, irn);
+ ilpsched_node_attr_t *na = get_ilpsched_node_attr(node);
+ unsigned t;
+
+ /* we ignore nodes assigned to dummy unit here */
+ if (na->is_dummy_node)
+ continue;
+
+ /* check check all time steps: asap(n) <= t <= U */
+ for (t = na->asap - 1; t < ba->max_steps; ++t) {
+ int node_tp_idx;
+
+ /* for all unit types available for this node */
+ for (node_tp_idx = na->n_unit_types - 1; node_tp_idx >= 0; --node_tp_idx) {
+ unsigned tn, tn_max, idx;
+ int cst, i;
+ int *tmp_var_idx_n = NEW_ARR_F(int, 0);
+ int *tmp_var_idx_m = NEW_ARR_F(int, 0);
+
+ snprintf(buf, sizeof(buf), "alive_node_cst_%u_n%u_%s",
+ t, get_irn_idx(irn), na->type_info[node_tp_idx].tp->name);
+ cst = lpp_add_cst_uniq(lpp, buf, lpp_less, 0.0);
+ DBG((env->dbg, LEVEL_2, "added constraint %s\n", buf));
+ num_cst++;
+
+ tn_max = MIN(na->alap - 1, t);
+ /* check if the node has been scheduled so far */
+ for (tn = na->asap - 1; tn <= tn_max; ++tn) {
+ int idx = ILPVAR_IDX(na, node_tp_idx, tn);
+ ARR_APP1(int, tmp_var_idx_n, na->ilp_vars.x[idx]);
+ }
+
+ if (ARR_LEN(tmp_var_idx_n) > 0)
+ lpp_set_factor_fast_bulk(lpp, cst, tmp_var_idx_n, ARR_LEN(tmp_var_idx_n), (double)(na->n_consumer));
+ DEL_ARR_F(tmp_var_idx_n);
+
+ /* subtract the number of consumer scheduled so far */
+ for (i = ARR_LEN(na->block_consumer) - 1; i >= 0; --i) {
+ be_ilpsched_irn_t *cons = get_ilpsched_irn(env, na->block_consumer[i]);
+ ilpsched_node_attr_t *ca = get_ilpsched_node_attr(cons);
+ int tp_idx;
+ unsigned tm, tm_max;
+
+ tm_max = MIN(ca->alap - 1, t);
+ for (tp_idx = ca->n_unit_types - 1; tp_idx >= 0; --tp_idx) {
+ for (tm = ca->asap - 1; tm <= tm_max; ++tm) {
+ int idx = ILPVAR_IDX(ca, tp_idx, tm);
+ ARR_APP1(int, tmp_var_idx_m, ca->ilp_vars.x[idx]);
+ }
+ }
+ }
+
+ if (ARR_LEN(tmp_var_idx_m) > 0)
+ lpp_set_factor_fast_bulk(lpp, cst, tmp_var_idx_m, ARR_LEN(tmp_var_idx_m), -1.0);
+ DEL_ARR_F(tmp_var_idx_m);
+
+ /* -c * a_{nt}^k */
+ idx = ILPVAR_IDX_DEAD(ba, na, node_tp_idx, t);
+ lpp_set_factor_fast(lpp, cst, na->ilp_vars.a[idx], 0.0 - (double)(na->n_consumer));
+
+ }
+ }
+ }
+ ilp_timer_pop();
+ DBG((env->dbg, LEVEL_1, "\t%u alive nodes constraints (%g sec)\n",
+ num_cst, ilp_timer_elapsed_usec(t_cst) / 1000000.0));
+}
+
+/**
+ * Create ILP alive nodes constraints for live-in nodes:
+ * - set variable a_{nt}^k to 1 if nodes n is alive at step t on unit k
+ */
+static void create_alive_livein_nodes_constraint(be_ilpsched_env_t *env, lpp_t *lpp, be_ilpsched_irn_t *block_node) {
+ char buf[1024];
+ ilp_livein_node_t *livein;
+ unsigned num_cst = 0;
+ ilpsched_block_attr_t *ba = get_ilpsched_block_attr(block_node);
+ lc_timer_t *t_cst = lc_timer_register("beilpsched_cst_alive_livein_nodes", "create alive livein nodes constraints");
+
+ ilp_timer_push(t_cst);
+ /* for each node */
+ foreach_pset(ba->livein_nodes, livein) {
+ ir_node *irn = livein->irn;
+ be_ilpsched_irn_t *node = get_ilpsched_irn(env, irn);
+ ilpsched_node_attr_t *na = get_ilpsched_node_attr(node);
+ unsigned t;
+
+ /* check check all time steps: 0 <= t < max_alive_steps */
+ for (t = 0; t < livein->max_alive_steps; ++t) {
+ int node_tp_idx;
+
+ /* for all unit types available for this node */
+ for (node_tp_idx = na->n_unit_types - 1; node_tp_idx >= 0; --node_tp_idx) {
+ const ir_edge_t *edge;
+ unsigned idx;
+ int cst, num_block_user;
+ int *tmp_var_idx_m = NEW_ARR_F(int, 0);
+
+ /* check the number of consumer scheduled so far */
+ num_block_user = 0;
+ foreach_out_edge(irn, edge) {
+ ir_node *user = get_edge_src_irn(edge);
+ be_ilpsched_irn_t *cons;
+ ilpsched_node_attr_t *ca;
+ int tp_idx;
+ unsigned tm, tm_max;
+
+ /* check only users within current block */
+ if (get_nodes_block(user) != block_node->irn)
+ continue;
+
+ num_block_user++;
+ cons = get_ilpsched_irn(env, user);
+ ca = get_ilpsched_node_attr(cons);
+
+ tm_max = MIN(ca->alap - 1, t);
+ for (tp_idx = ca->n_unit_types - 1; tp_idx >= 0; --tp_idx) {
+ for (tm = ca->asap - 1; tm <= tm_max; ++tm) {
+ int idx = ILPVAR_IDX(ca, tp_idx, tm);
+ ARR_APP1(int, tmp_var_idx_m, ca->ilp_vars.x[idx]);
+ }
+ }
+ }
+
+ snprintf(buf, sizeof(buf), "alive_livein_node_cst_%u_n%u_%s",
+ t, get_irn_idx(irn), na->type_info[node_tp_idx].tp->name);
+ cst = lpp_add_cst_uniq(lpp, buf, lpp_greater, (double)num_block_user);
+ DBG((env->dbg, LEVEL_2, "added constraint %s\n", buf));
+ num_cst++;
+
+ /* sum(scheduled users) */
+ if (ARR_LEN(tmp_var_idx_m) > 0)
+ lpp_set_factor_fast_bulk(lpp, cst, tmp_var_idx_m, ARR_LEN(tmp_var_idx_m), 1.0);
+ DEL_ARR_F(tmp_var_idx_m);
+
+ /* + c * a_{nt}^k */
+ idx = node_tp_idx * livein->max_alive_steps + t;
+ lpp_set_factor_fast(lpp, cst, livein->a[idx], (double)(num_block_user));
+ }
+ }
+ }
+ ilp_timer_pop();
+ DBG((env->dbg, LEVEL_1, "\t%u alive livein nodes constraints (%g sec)\n",
+ num_cst, ilp_timer_elapsed_usec(t_cst) / 1000000.0));
+}
+
+/**
+ * Create ILP pressure constraints, based on alive nodes:
+ * - add additional costs to objective function if a node is scheduled
+ * on a unit although all units of this type are currently occupied
+ */
+static void create_pressure_alive_constraint(be_ilpsched_env_t *env, lpp_t *lpp, be_ilpsched_irn_t *block_node) {
+ char buf[1024];
+ ir_node *cur_irn;
+ unsigned num_cst = 0;
+ ilpsched_block_attr_t *ba = get_ilpsched_block_attr(block_node);
+ lc_timer_t *t_cst = lc_timer_register("beilpsched_cst_pressure", "create pressure constraints");
+
+ ilp_timer_push(t_cst);
+ /* y_{nt}^k is set for each node and timestep and unit type */
+ foreach_linked_irns(ba->head_ilp_nodes, cur_irn) {
+ unsigned cur_idx = get_irn_idx(cur_irn);
+ be_ilpsched_irn_t *cur_node = get_ilpsched_irn(env, cur_irn);
+ ilpsched_node_attr_t *cur_na = get_ilpsched_node_attr(cur_node);
+ int glob_type_idx;
+
+ /* we ignore nodes assigned to DUMMY unit here */
+ if (cur_na->is_dummy_node)
+ continue;
+
+ /* for all types */
for (glob_type_idx = env->cpu->n_unit_types - 1; glob_type_idx >= 0; --glob_type_idx) {
- unsigned t;
- be_execution_unit_type_t *cur_tp = &env->cpu->unit_types[glob_type_idx];
+ be_execution_unit_type_t *cur_tp = &env->cpu->unit_types[glob_type_idx];
+ int cur_tp_idx;
+ unsigned t;
+
+ /* BEWARE: the DUMMY unit types is not in CPU, so it's skipped automatically */
+
+ /* check if node can be executed on this unit type */
+ cur_tp_idx = is_valid_unit_type_for_node(cur_tp, cur_node);
+ if (cur_tp_idx < 0)
+ continue;
- for (t = 0; t < ba->n_interesting_nodes; ++t) {
- int cst;
- int *tmp_var_idx = NEW_ARR_F(int, ba->n_interesting_nodes);
- int cur_idx = 0;
+ /* check all time_steps at which the current node can be scheduled */
+ for (t = cur_na->asap - 1; t <= cur_na->alap - 1; ++t) {
+ int cst, y_idx;
+ ir_node *irn;
+ int *tmp_var_idx = NEW_ARR_F(int, 0);
+ ilp_livein_node_t *livein;
- snprintf(buf, sizeof(buf), "resource_cst_%s_%u", cur_tp->name, t);
- cst = lpp_add_cst_uniq(lpp, buf, lpp_less, (double)cur_tp->n_units);
+ snprintf(buf, sizeof(buf), "pressure_cst_n%u_%u_%s", cur_idx, t, cur_tp->name);
+ cst = lpp_add_cst_uniq(lpp, buf, lpp_less, (double)(cur_tp->n_units - 1));
DBG((env->dbg, LEVEL_2, "added constraint %s\n", buf));
- num_cst_resrc++;
+ num_cst++;
+ /* - accumulate all nodes alive at point t on unit type k */
foreach_linked_irns(ba->head_ilp_nodes, irn) {
be_ilpsched_irn_t *node = get_ilpsched_irn(env, irn);
ilpsched_node_attr_t *na = get_ilpsched_node_attr(node);
- int tp_idx;
+ int a_idx, tp_idx;
+
+ /* check if node can be alive here */
+ if (t < na->asap - 1)
+ continue;
tp_idx = is_valid_unit_type_for_node(cur_tp, node);
- if (tp_idx >= 0 && t >= na->asap - 1 && t <= na->alap - 1) {
- int cur_var = ILPVAR_IDX(na, tp_idx, t);
- tmp_var_idx[cur_idx++] = na->ilp_vars[cur_var];
- }
+ /* current type is not suitable */
+ if (tp_idx < 0)
+ continue;
+
+ a_idx = ILPVAR_IDX_DEAD(ba, na, tp_idx, t);
+ ARR_APP1(int, tmp_var_idx, na->ilp_vars.a[a_idx]);
}
+ /* do the same for livein nodes */
+ foreach_pset(ba->livein_nodes, livein) {
+ ir_node *irn = livein->irn;
+ be_ilpsched_irn_t *node = get_ilpsched_irn(env, irn);
+ int a_idx, tp_idx;
+
+ /* check if node can be alive here */
+ if (t >= livein->max_alive_steps)
+ continue;
+
+ tp_idx = is_valid_unit_type_for_node(cur_tp, node);
- if (cur_idx)
- lpp_set_factor_fast_bulk(lpp, cst, tmp_var_idx, cur_idx, 1.0);
+ /* current type is not suitable */
+ if (tp_idx < 0)
+ continue;
+ a_idx = tp_idx * livein->max_alive_steps + t;
+ ARR_APP1(int, tmp_var_idx, livein->a[a_idx]);
+ }
+
+ if (ARR_LEN(tmp_var_idx) > 0)
+ lpp_set_factor_fast_bulk(lpp, cst, tmp_var_idx, ARR_LEN(tmp_var_idx), 1.0);
DEL_ARR_F(tmp_var_idx);
+
+ /* - num_nodes * y_{nt}^k */
+ y_idx = ILPVAR_IDX(cur_na, cur_tp_idx, t);
+ lpp_set_factor_fast(lpp, cst, cur_na->ilp_vars.y[y_idx], -1.0);
}
}
- lc_timer_pop();
- DBG((env->dbg, LEVEL_1, "\t%u resource constraints (%g sec)\n",
- num_cst_resrc, lc_timer_elapsed_usec(t_cst_rsrc) / 1000000.0));
+ }
+ ilp_timer_pop();
+ DBG((env->dbg, LEVEL_1, "\t%u pressure constraints (%g sec)\n",
+ num_cst, ilp_timer_elapsed_usec(t_cst) / 1000000.0));
+}
- /*
- 3rd: bundle constraints:
- assure, at most bundle_size * bundles_per_cycle instructions
- can be started at a certain point.
- */
- lc_timer_push(t_cst_bundle);
- for (t = 0; t < ba->n_interesting_nodes; ++t) {
+/**
+ * Create ILP branch constraints:
+ * Assure, alle nodes are scheduled prior to cfg op.
+ */
+static void create_branch_constraint(be_ilpsched_env_t *env, lpp_t *lpp, be_ilpsched_irn_t *block_node) {
+ char buf[1024];
+ ir_node *cur_irn, *cfop;
+ unsigned num_cst = 0;
+ unsigned num_non_branches = 0;
+ ilpsched_block_attr_t *ba = get_ilpsched_block_attr(block_node);
+ lc_timer_t *t_cst = lc_timer_register("beilpsched_cst_branch", "create branch constraints");
+
+ ilp_timer_push(t_cst);
+ cfop = NULL;
+ /* determine number of non-branch nodes and the one and only branch node */
+ foreach_linked_irns(ba->head_ilp_nodes, cur_irn) {
+ switch (get_irn_opcode(cur_irn)) {
+ case iro_Phi:
+ case iro_Start:
+ case iro_End:
+ case iro_Proj:
+ case iro_Bad:
+ case iro_Unknown:
+ num_non_branches++;
+ break;
+ default:
+ if (is_cfop(cur_irn)) {
+ assert(cfop == NULL && "Highlander - there can be only one to be constrained");
+ cfop = cur_irn;
+ }
+ else {
+ num_non_branches++;
+ }
+ break;
+ }
+ }
+
+ if (cfop) {
+ be_ilpsched_irn_t *cf_node = get_ilpsched_irn(env, cfop);
+ ilpsched_node_attr_t *cf_na = get_ilpsched_node_attr(cf_node);
+ unsigned t;
+
+ /* for each time step */
+ for (t = cf_na->asap - 1; t <= cf_na->alap - 1; ++t) {
+ int *non_branch_vars, *branch_vars;
int cst;
- int *tmp_var_idx = NEW_ARR_F(int, 0);
- snprintf(buf, sizeof(buf), "bundle_cst_%u", t);
- cst = lpp_add_cst_uniq(lpp, buf, lpp_less, (double)n_instr_max);
+ snprintf(buf, sizeof(buf), "branch_cst_%u_n%u", t, get_irn_idx(cfop));
+ cst = lpp_add_cst_uniq(lpp, buf, lpp_greater, 0.0);
DBG((env->dbg, LEVEL_2, "added constraint %s\n", buf));
- num_cst_bundle++;
-
- foreach_linked_irns(ba->head_ilp_nodes, irn) {
- be_ilpsched_irn_t *node;
- ilpsched_node_attr_t *na;
+ num_cst++;
+
+ /* sum(overall non branches: n)x_{nt}^k - sum(overall branches: b)(num_non_branches * x_{bt}^k >= 0) */
+ non_branch_vars = NEW_ARR_F(int, 0);
+ branch_vars = NEW_ARR_F(int, 0);
+ foreach_linked_irns(ba->head_ilp_nodes, cur_irn) {
+ be_ilpsched_irn_t *node = get_ilpsched_irn(env, cur_irn);
+ ilpsched_node_attr_t *na = get_ilpsched_node_attr(node);
int tp_idx;
- node = get_ilpsched_irn(env, irn);
- na = get_ilpsched_node_attr(node);
-
- if (t >= na->asap - 1 && t <= na->alap - 1) {
+ if (cur_irn == cfop) {
+ /* for all unit types available for this node */
+ for (tp_idx = na->n_unit_types - 1; tp_idx >= 0; --tp_idx) {
+ unsigned idx = ILPVAR_IDX(na, tp_idx, t);
+ ARR_APP1(int, branch_vars, na->ilp_vars.x[idx]);
+ }
+ }
+ else {
+ /* sum up all possible schedule points for this node upto current timestep */
for (tp_idx = na->n_unit_types - 1; tp_idx >= 0; --tp_idx) {
- int idx = ILPVAR_IDX(na, tp_idx, t);
- ARR_APP1(int, tmp_var_idx, na->ilp_vars[idx]);
+ unsigned tn;
+ unsigned tmax = MIN(t, na->alap - 1);
+
+ for (tn = na->asap - 1; tn <= tmax; ++tn) {
+ unsigned idx = ILPVAR_IDX(na, tp_idx, tn);
+ ARR_APP1(int, non_branch_vars, na->ilp_vars.x[idx]);
+ }
}
}
+
}
- if (ARR_LEN(tmp_var_idx) > 0)
- lpp_set_factor_fast_bulk(lpp, cst, tmp_var_idx, ARR_LEN(tmp_var_idx), 1.0);
+ if (ARR_LEN(non_branch_vars) > 0)
+ lpp_set_factor_fast_bulk(lpp, cst, non_branch_vars, ARR_LEN(non_branch_vars), 1.0);
+ if (ARR_LEN(branch_vars) > 0)
+ lpp_set_factor_fast_bulk(lpp, cst, branch_vars, ARR_LEN(branch_vars), 0.0 - (double)num_non_branches);
- DEL_ARR_F(tmp_var_idx);
+ DEL_ARR_F(branch_vars);
+ DEL_ARR_F(non_branch_vars);
+ }
+ }
+ ilp_timer_pop();
+ DBG((env->dbg, LEVEL_1, "\t%u branch constraints (%g sec)\n",
+ num_cst, ilp_timer_elapsed_usec(t_cst) / 1000000.0));
+}
+
+#if 0
+static void create_proj_keep_constraints(be_ilpsched_env_t *env, lpp_t *lpp, be_ilpsched_irn_t *block_node) {
+ char buf[1024];
+ ir_node *irn;
+ unsigned num_cst = 0;
+ ilpsched_block_attr_t *ba = get_ilpsched_block_attr(block_node);
+ lc_timer_t *t_cst = lc_timer_register("beilpsched_cst_projkeep", "create proj and keep constraints");
+
+ ilp_timer_push(t_cst);
+ /* check all nodes */
+ foreach_linked_irns(ba->head_ilp_nodes, irn) {
+ be_ilpsched_irn_t *node;
+ ilpsched_node_attr_t *na;
+ unsigned t;
+ ir_node **pk;
+
+ /* only mode_T nodes can have Projs and Keeps assigned */
+ if (get_irn_mode(irn) != mode_T)
+ continue;
+
+ node = get_ilpsched_irn(env, irn);
+ na = get_ilpsched_node_attr(node);
+
+ /* check if has some Projs and Keeps assigned */
+ if (! na->projkeeps)
+ continue;
+
+ /* we can run only once over the queue, so preserve the nodes */
+ pk = NEW_ARR_F(ir_node *, 0);
+ while (! waitq_empty(na->projkeeps))
+ ARR_APP1(ir_node *, pk, waitq_get(na->projkeeps));
+ del_waitq(na->projkeeps);
+ na->projkeeps = NULL;
+
+ /* for all time steps at which this node can be scheduled */
+ for (t = na->asap - 1; t <= na->alap - 1; ++t) {
+ int cst, tp_idx, i;
+ int *tmp_var_idx_n = NEW_ARR_F(int, 0);
+
+ /* add the constraint, assure, that a node is always scheduled along with it's Projs and Keeps */
+ snprintf(buf, sizeof(buf), "projkeep_cst_n%u_%u", get_irn_idx(irn), t);
+ cst = lpp_add_cst_uniq(lpp, buf, lpp_equal, 0.0);
+ DBG((env->dbg, LEVEL_2, "added constraint %s\n", buf));
+ num_cst++;
+
+ /* sum up scheduling variables for this time step */
+ for (tp_idx = na->n_unit_types - 1; tp_idx >= 0; --tp_idx) {
+ int idx = ILPVAR_IDX(na, tp_idx, t);
+ ARR_APP1(int, tmp_var_idx_n, na->ilp_vars.x[idx]);
+ }
+
+ if (ARR_LEN(tmp_var_idx_n) > 0)
+ lpp_set_factor_fast_bulk(lpp, cst, tmp_var_idx_n, ARR_LEN(tmp_var_idx_n), (double)(ARR_LEN(pk)));
+ DEL_ARR_F(tmp_var_idx_n);
+
+ /* subtract all Proj and Keep variables for this step */
+ for (i = ARR_LEN(pk) - 1; i >= 0; --i) {
+ be_ilpsched_irn_t *pk_node = get_ilpsched_irn(env, pk[i]);
+ ilpsched_node_attr_t *pk_na = get_ilpsched_node_attr(pk_node);
+ int pk_tp_idx;
+
+ for (pk_tp_idx = pk_na->n_unit_types - 1; pk_tp_idx >= 0; --pk_tp_idx) {
+ int idx = ILPVAR_IDX(pk_na, pk_tp_idx, t);
+ lpp_set_factor_fast(lpp, cst, pk_na->ilp_vars.x[idx], -1.0);
+ }
+ }
+ }
+ }
+ ilp_timer_pop();
+ DBG((env->dbg, LEVEL_1, "\t%u Proj and Keep constraints (%g sec)\n",
+ num_cst, ilp_timer_elapsed_usec(t_cst) / 1000000.0));
+}
+#endif /* if 0 */
+
+/***************************************************
+ * _____ _ _____ _
+ * |_ _| | | __ \ (_)
+ * | | | | | |__) | _ __ ___ __ _ _ _ __
+ * | | | | | ___/ | '_ ` _ \ / _` | | '_ \
+ * _| |_| |____| | | | | | | | (_| | | | | |
+ * |_____|______|_| |_| |_| |_|\__,_|_|_| |_|
+ *
+ ***************************************************/
+
+/**
+ * Create the ilp (add variables, build constraints, solve, build schedule from solution).
+ */
+static void create_ilp(ir_node *block, void *walk_env) {
+ be_ilpsched_env_t *env = walk_env;
+ be_ilpsched_irn_t *block_node = get_ilpsched_irn(env, block);
+ ilpsched_block_attr_t *ba = get_ilpsched_block_attr(block_node);
+ FILE *logfile = NULL;
+ lpp_t *lpp = NULL;
+ int need_heur = 0;
+ struct obstack var_obst;
+ char name[1024];
+
+ DBG((env->dbg, 255, "\n\n\n=========================================\n"));
+ DBG((env->dbg, 255, " ILP Scheduling for %+F\n", block));
+ DBG((env->dbg, 255, "=========================================\n\n"));
+
+ DBG((env->dbg, LEVEL_1, "Creating ILP Variables for nodes in %+F (%u interesting nodes, %u max steps)\n",
+ block, ba->n_interesting_nodes, ba->max_steps));
+
+ /* notify backend and get block environment */
+ env->block_env = be_ilp_sched_init_block_ilp_schedule(env->sel, block);
+
+ /* if we have less than two interesting nodes, there is no need to create the ILP */
+ if (ba->n_interesting_nodes > 1) {
+ double fact_var = ba->n_interesting_nodes > 25 ? 2.3 : 3;
+ double fact_cst = ba->n_interesting_nodes > 25 ? 3 : 4.5;
+ int base_num = ba->n_interesting_nodes * ba->n_interesting_nodes;
+ int estimated_n_var = (int)((double)base_num * fact_var);
+ int estimated_n_cst = (int)((double)base_num * fact_cst);
+
+ DBG((env->dbg, LEVEL_1, "Creating LPP with estimated numbers: %d vars, %d cst\n",
+ estimated_n_var, estimated_n_cst));
+
+ /* set up the LPP object */
+ snprintf(name, sizeof(name), "ilp scheduling IRG %s", get_entity_ld_name(get_irg_entity(env->irg)));
+
+ lpp = new_lpp_userdef(
+ (const char *)name,
+ lpp_minimize,
+ estimated_n_cst, /* num vars */
+ estimated_n_cst + 1, /* num cst */
+ 1.3); /* grow factor */
+ obstack_init(&var_obst);
+
+ /* create ILP variables */
+ create_variables(env, lpp, block_node, &var_obst);
+
+ /* create ILP constraints */
+ DBG((env->dbg, LEVEL_1, "Creating constraints for nodes in %+F:\n", block));
+ create_assignment_and_precedence_constraints(env, lpp, block_node);
+ create_ressource_constraints(env, lpp, block_node);
+ create_bundle_constraints(env, lpp, block_node);
+ create_branch_constraint(env, lpp, block_node);
+ //create_proj_keep_constraints(env, lpp, block_node);
+
+ if (env->opts->regpress) {
+ create_alive_nodes_constraint(env, lpp, block_node);
+ create_alive_livein_nodes_constraint(env, lpp, block_node);
+ create_pressure_alive_constraint(env, lpp, block_node);
}
- lc_timer_pop();
- DBG((env->dbg, LEVEL_1, "\t%u bundle constraints (%g sec)\n",
- num_cst_bundle, lc_timer_elapsed_usec(t_cst_bundle) / 1000000.0));
DBG((env->dbg, LEVEL_1, "ILP to solve: %u variables, %u constraints\n", lpp->var_next, lpp->cst_next));
/* debug stuff, dump lpp when debugging is on */
- DEBUG_ONLY(
- if (firm_dbg_get_mask(env->dbg) > 0) {
+ DEBUG_ONLY({
+ if (firm_dbg_get_mask(env->dbg) > 1) {
+ char buf[1024];
FILE *f;
snprintf(buf, sizeof(buf), "lpp_block_%lu.txt", get_irn_node_nr(block));
lpp_dump_plain(lpp, f);
fclose(f);
snprintf(buf, sizeof(buf), "lpp_block_%lu.mps", get_irn_node_nr(block));
- lpp_dump(lpp, "buf");
+ lpp_dump(lpp, buf);
}
- );
+ })
/* set solve time limit */
lpp_set_time_limit(lpp, env->opts->time_limit);
/* check for valid solution */
if (! lpp_is_sol_valid(lpp)) {
+ char buf[1024];
FILE *f;
- snprintf(buf, sizeof(buf), "lpp_block_%lu.assert.txt", get_irn_node_nr(block));
- f = fopen(buf, "w");
- lpp_dump_plain(lpp, f);
- fclose(f);
- snprintf(buf, sizeof(buf), "lpp_block_%lu.assert.mps", get_irn_node_nr(block));
- lpp_dump(lpp, buf);
- dump_ir_block_graph(env->irg, "-assert");
+ DEBUG_ONLY({
+ if (firm_dbg_get_mask(env->dbg) >= 2) {
+ snprintf(buf, sizeof(buf), "lpp_block_%lu.infeasible.txt", get_irn_node_nr(block));
+ f = fopen(buf, "w");
+ lpp_dump_plain(lpp, f);
+ fclose(f);
+ snprintf(buf, sizeof(buf), "lpp_block_%lu.infeasible.mps", get_irn_node_nr(block));
+ lpp_dump(lpp, buf);
+ dump_ir_block_graph(env->irg, "-infeasible");
+ }
+ })
- assert(0 && "ILP solution is not feasible!");
+ ir_fprintf(stderr, "ILP found no solution within time (%+F, %+F), falling back to heuristics.\n", block, env->irg);
+ need_heur = 1;
}
DBG((env->dbg, LEVEL_1, "\nSolution:\n"));
DBG((env->dbg, LEVEL_1, "\tsend time: %g sec\n", lpp->send_time / 1000000.0));
DBG((env->dbg, LEVEL_1, "\treceive time: %g sec\n", lpp->recv_time / 1000000.0));
+ DBG((env->dbg, LEVEL_1, "\tmatrix: %u elements, density %.2f%%, size %.2fMB\n", lpp->n_elems, lpp->density, (double)lpp->matrix_mem / 1024.0 / 1024.0));
DBG((env->dbg, LEVEL_1, "\titerations: %d\n", lpp->iterations));
DBG((env->dbg, LEVEL_1, "\tsolution time: %g\n", lpp->sol_time));
DBG((env->dbg, LEVEL_1, "\tobjective function: %g\n", LPP_VALUE_IS_0(lpp->objval) ? 0.0 : lpp->objval));
DBG((env->dbg, LEVEL_1, "\tbest bound: %g\n", LPP_VALUE_IS_0(lpp->best_bound) ? 0.0 : lpp->best_bound));
- obstack_free(&var_obst, NULL);
+ DBG((env->dbg, LEVEL_1, "variables used %u bytes\n", obstack_memory_used(&var_obst)));
}
/* apply solution */
- apply_solution(env, lpp, block);
+ be_stat_ev("nodes", ba->block_last_idx);
+ be_stat_ev("vars", lpp ? lpp->var_next : 0);
+ be_stat_ev("csts", lpp ? lpp->cst_next : 0);
+ if (need_heur) {
+ be_stat_ev("time", -1);
+ be_stat_ev_dbl("opt", 0.0);
+ list_sched_single_block(env->birg, block, env->be_opts);
+ }
+ else {
+ if (lpp) {
+ double opt = lpp->sol_state == lpp_optimal ? 100.0 : 100.0 * lpp->best_bound / lpp->objval;
+ be_stat_ev_dbl("time", lpp->sol_time);
+ be_stat_ev_dbl("opt", opt);
+ }
+ else {
+ be_stat_ev_dbl("time", 0.0);
+ be_stat_ev_dbl("opt", 100.0);
+ }
+ apply_solution(env, lpp, block);
+ }
if (lpp)
free_lpp(lpp);
+
+ /* notify backend */
+ be_ilp_sched_finish_block_ilp_schedule(env->sel, block, env->block_env);
}
/**
* Perform ILP scheduling on the given irg.
*/
-void be_ilp_sched(const be_irg_t *birg) {
- be_ilpsched_env_t env;
- const char *name = "be ilp scheduling";
+void be_ilp_sched(const be_irg_t *birg, be_options_t *be_opts) {
+ be_ilpsched_env_t env;
+ const char *name = "be ilp scheduling";
+ ir_graph *irg = be_get_birg_irg(birg);
+ const arch_env_t *arch_env = be_get_birg_arch_env(birg);
+ const arch_isa_t *isa = arch_env->isa;
+ const ilp_sched_selector_t *sel = isa->impl->get_ilp_sched_selector(isa);
FIRM_DBG_REGISTER(env.dbg, "firm.be.sched.ilp");
- //firm_dbg_set_mask(env.dbg, 31);
+ stat_ev_ctx_push("phase", "ilpsched");
+
+// firm_dbg_set_mask(env.dbg, 1);
- env.irg = birg->irg;
+ env.irg_env = be_ilp_sched_init_irg_ilp_schedule(sel, irg);
+ env.sel = sel;
+ env.irg = irg;
+ env.height = heights_new(irg);
env.main_env = birg->main_env;
- env.alap_queue = new_waitq();
- env.arch_env = birg->main_env->arch_env;
- env.cpu = arch_isa_get_machine(birg->main_env->arch_env->isa);
+ env.arch_env = arch_env;
+ env.cpu = arch_isa_get_machine(arch_env->isa);
env.opts = &ilp_opts;
- phase_init(&env.ph, name, env.irg, PHASE_DEFAULT_GROWTH, init_ilpsched_irn);
+ env.birg = birg;
+ env.be_opts = be_opts;
+ phase_init(&env.ph, name, env.irg, PHASE_DEFAULT_GROWTH, init_ilpsched_irn, NULL);
+ /* assign a unique per block number to all interesting nodes */
irg_walk_in_or_dep_graph(env.irg, NULL, build_block_idx, &env);
/*
- The block indicees are completely build after the walk,
- now we can allocate the bitsets (size depends on block indicees)
+ The block indices are completely build after the walk,
+ now we can allocate the bitsets (size depends on block indices)
for all nodes.
*/
phase_reinit_irn_data(&env.ph);
/* Collect all root nodes (having no user in their block) and calculate ASAP. */
irg_walk_in_or_dep_blkwise_graph(env.irg, collect_alap_root_nodes, calculate_irn_asap, &env);
- /* calculate ALAP and create variables */
- irg_block_walk_graph(env.irg, calculate_block_alap, create_ilp, &env);
+ /* Calculate ALAP of all irns */
+ irg_block_walk_graph(env.irg, NULL, calculate_block_alap, &env);
+
+ /* We refine the {ASAP(n), ALAP(n)} interval and fix the time steps for Projs and Keeps */
+ irg_walk_in_or_dep_blkwise_graph(env.irg, NULL, refine_asap_alap_times, &env);
+
+ /* perform ILP scheduling */
+ irg_block_walk_graph(env.irg, NULL, create_ilp, &env);
DEBUG_ONLY(
if (firm_dbg_get_mask(env.dbg)) {
}
);
+ /* free data allocated dynamically */
+ irg_block_walk_graph(env.irg, NULL, clear_unwanted_data, &env);
+
/* free all allocated object */
- del_waitq(env.alap_queue);
phase_free(&env.ph);
+ heights_free(env.height);
+
+ /* notify backend */
+ be_ilp_sched_finish_irg_ilp_schedule(sel, birg->irg, env.irg_env);
+
+ stat_ev_ctx_pop();
}
-#ifdef WITH_LIBCORE
/**
* Register ILP scheduler options.
*/
-void ilpsched_register_options(lc_opt_entry_t *grp) {
- static int run_once = 0;
- lc_opt_entry_t *sched_grp;
-
- if (! run_once) {
- run_once = 1;
- sched_grp = lc_opt_get_grp(grp, "ilpsched");
+void be_init_ilpsched(void)
+{
+ lc_opt_entry_t *be_grp = lc_opt_get_grp(firm_opt_get_root(), "be");
+ lc_opt_entry_t *sched_grp = lc_opt_get_grp(be_grp, "ilpsched");
- lc_opt_add_table(sched_grp, ilpsched_option_table);
- }
+ lc_opt_add_table(sched_grp, ilpsched_option_table);
}
-#endif /* WITH_LIBCORE */
#else /* WITH_ILP */
-static int some_picky_compiler_do_not_allow_empty_files;
+static INLINE void some_picky_compiler_do_not_allow_empty_files(void)
+{}
#endif /* WITH_ILP */