}
}
-int arch_get_op_estimated_cost(const arch_env_t *env, const ir_node *irn)
+int arch_get_op_estimated_cost(const ir_node *irn)
{
const arch_irn_ops_t *ops = get_irn_ops(irn);
- (void)env; // TODO remove parameter
if(ops->get_op_estimated_cost) {
return ops->get_op_estimated_cost(irn);
void arch_set_frame_entity(ir_node *irn, ir_entity *ent);
int arch_get_sp_bias(ir_node *irn);
-extern int arch_get_op_estimated_cost(const arch_env_t *env, const ir_node *irn);
+int arch_get_op_estimated_cost(const ir_node *irn);
arch_inverse_t *arch_get_inverse(const ir_node *irn, int i, arch_inverse_t *inverse, struct obstack *obstack);
int arch_possible_memory_operand(const ir_node *irn, unsigned int i);
void arch_perform_memory_operand(ir_node *irn, ir_node *spill, unsigned int i);
if(be_is_Reload(spilled)) {
costs += 2;
} else {
- costs += arch_get_op_estimated_cost(env->arch_env, spilled);
+ costs += arch_get_op_estimated_cost(spilled);
}
if(parentcosts + costs >= env->reload_cost + env->spill_cost) {
return REMAT_COST_INFINITE;
double costs = 0.0;
sched_foreach(block, node) {
- costs += arch_get_op_estimated_cost(env->arch_env, node);
+ costs += arch_get_op_estimated_cost(node);
}
env->costs += costs * get_block_execfreq(env->execfreqs, block);