+void arch_perform_memory_operand(const arch_env_t *env, ir_node *irn, ir_node *spill, unsigned int i) {
+ const arch_irn_ops_t *ops = get_irn_ops(env, irn);
+
+ if(ops->impl->perform_memory_operand) {
+ ops->impl->perform_memory_operand(ops, irn, spill, i);
+ } else {
+ return;
+ }
+}
+
+int arch_get_op_estimated_cost(const arch_env_t *env, const ir_node *irn)
+{
+ const arch_irn_ops_t *ops = get_irn_ops(env, irn);
+
+ if(ops->impl->get_op_estimated_cost) {
+ return ops->impl->get_op_estimated_cost(ops, irn);
+ } else {
+ return 1;
+ }
+}
+
+int arch_is_possible_memory_operand(const arch_env_t *env, const ir_node *irn, int i)
+{
+ const arch_irn_ops_t *ops = get_irn_ops(env, irn);
+
+ if(ops->impl->possible_memory_operand) {
+ return ops->impl->possible_memory_operand(ops, irn, i);
+ } else {
+ return 0;
+ }
+}
+
+int arch_get_allocatable_regs(const arch_env_t *env, const ir_node *irn, int pos, bitset_t *bs)
+{
+ const arch_irn_ops_t *ops = get_irn_ops(env, irn);
+ const arch_register_req_t *req = ops->impl->get_irn_reg_req(ops, irn, pos);
+
+ if(req->type == arch_register_req_type_none) {
+ bitset_clear_all(bs);
+ return 0;
+ }
+
+ if(arch_register_req_is(req, limited)) {
+ rbitset_copy_to_bitset(req->limited, bs);
+ return bitset_popcnt(bs);
+ }
+
+ arch_register_class_put(req->cls, bs);
+ return req->cls->n_regs;
+}
+
+void arch_put_non_ignore_regs(const arch_env_t *env,
+ const arch_register_class_t *cls, bitset_t *bs)
+{
+ int i;
+ (void) env;
+
+ for(i = 0; i < cls->n_regs; ++i) {
+ if(!arch_register_type_is(&cls->regs[i], ignore))
+ bitset_set(bs, i);
+ }
+}
+
+int arch_count_non_ignore_regs(const arch_env_t *env,
+ const arch_register_class_t *cls)
+{
+ int i;
+ int result = 0;
+ (void) env;
+
+ for(i = 0; i < cls->n_regs; ++i) {
+ if(!arch_register_type_is(&cls->regs[i], ignore))
+ result++;
+ }
+
+ return result;