#ifdef COMPARE_AGAINST_GAUSSJORDAN
{
- double *nw = xmalloc(size * size * sizeof(*nw));
- double *nx = xmalloc(size * sizeof(*nx));
+ double *nw = XMALLOCN(double, size * size);
+ double *nx = XMALLOCNZ(double, size);
- memset(nx, 0, size * sizeof(*nx));
gs_matrix_export(mat, nw, size);
stat_ev_tim_push();
ir_exec_freq *create_execfreq(ir_graph *irg)
{
- ir_exec_freq *execfreq = xmalloc(sizeof(execfreq[0]));
- memset(execfreq, 0, sizeof(execfreq[0]));
+ ir_exec_freq *execfreq = XMALLOCZ(ir_exec_freq);
execfreq->set = new_set(cmp_freq, 32);
memset(&execfreq->hook, 0, sizeof(execfreq->hook));
* => they can "flow" from start to end.
*/
dfs = dfs_new(&absgraph_irg_cfg_succ, irg);
- ef = xmalloc(sizeof(ef[0]));
- memset(ef, 0, sizeof(ef[0]));
+ ef = XMALLOCZ(ir_exec_freq);
ef->min_non_zero = HUGE_VAL; /* initialize with a reasonable large number. */
freqs = ef->set = new_set(cmp_freq, dfs_get_n_nodes(dfs));
irg_block_walk_graph(irg, collect_blocks, NULL, freqs);
construct_cf_backedges(irg);
- /* TODO: edges are corrupt for EDGE_KIND_BLOCK after the local optimize
- graph phase merges blocks in the x86 backend */
- edges_deactivate(irg);
- edges_activate(irg);
- /* edges_assure(irg); */
+ edges_assure(irg);
size = dfs_get_n_nodes(dfs);
mat = gs_new_matrix(size, size);
- x = xmalloc(size*sizeof(*x));
+ x = XMALLOCN(double, size);
for (idx = dfs_get_n_nodes(dfs) - 1; idx >= 0; --idx) {
ir_node *bb = (ir_node *) dfs_get_post_num_node(dfs, size - idx - 1);
freq = set_insert_freq(freqs, bb);
freq->idx = idx;
- gs_matrix_set(mat, idx, idx, -1.0);
for(i = get_Block_n_cfgpreds(bb) - 1; i >= 0; --i) {
ir_node *pred = get_Block_cfgpred_block(bb, i);
int pred_idx = size - dfs_get_post_num(dfs, pred) - 1;
gs_matrix_set(mat, idx, pred_idx, get_cf_probability(bb, i, loop_weight));
}
+ gs_matrix_set(mat, idx, idx, -1.0);
}
dfs_free(dfs);