Use edges_reroute_except() instead of edges_reroute() + set_irn_n().
[libfirm] / ir / be / sparc / sparc_finish.c
index f94f916..4c34288 100644 (file)
@@ -21,7 +21,6 @@
  * @file
  * @brief    Peephole optimization and legalization of a sparc function
  * @author   Matthias Braun
- * @version  $Id$
  *
  * A note on sparc stackpointer (sp) behaviour:
  * The ABI expects SPARC_MIN_STACKSIZE bytes to be available at the
 #include "bearch_sparc_t.h"
 #include "gen_sparc_regalloc_if.h"
 #include "sparc_new_nodes.h"
+#include "sparc_transform.h"
 #include "irprog.h"
 #include "irgmod.h"
 #include "ircons.h"
 #include "irgwalk.h"
+#include "heights.h"
 
-#include "../bepeephole.h"
-#include "../benode.h"
-#include "../besched.h"
-#include "../bespillslots.h"
-#include "../bestack.h"
-#include "../beirgmod.h"
+#include "bepeephole.h"
+#include "benode.h"
+#include "besched.h"
+#include "bespillslots.h"
+#include "bestack.h"
+#include "beirgmod.h"
+
+static ir_heights_t *heights;
 
 static void kill_unused_stacknodes(ir_node *node)
 {
@@ -88,8 +91,10 @@ static void introduce_epilog(ir_node *ret)
 
        if (!layout->sp_relative) {
                const arch_register_t *fp_reg = &sparc_registers[REG_FRAME_POINTER];
+               const arch_register_t *sp_reg = &sparc_registers[REG_SP];
                ir_node *fp      = be_get_initial_reg_value(irg, fp_reg);
-               ir_node *restore = new_bd_sparc_RestoreZero(NULL, block, fp);
+               ir_node *sp      = be_get_initial_reg_value(irg, sp_reg);
+               ir_node *restore = new_bd_sparc_RestoreZero(NULL, block, sp, fp);
                sched_add_before(ret, restore);
                arch_set_irn_register(restore, sp_reg);
                set_irn_n(ret, sp_idx, restore);
@@ -109,7 +114,6 @@ void sparc_introduce_prolog_epilog(ir_graph *irg)
        be_stack_layout_t     *layout     = be_get_irg_stack_layout(irg);
        ir_node               *block      = get_nodes_block(start);
        ir_node               *initial_sp = be_get_initial_reg_value(irg, sp_reg);
-       ir_node               *sp         = initial_sp;
        ir_node               *schedpoint = start;
        ir_type               *frame_type = get_irg_frame_type(irg);
        unsigned               frame_size = get_type_size_bytes(frame_type);
@@ -131,189 +135,286 @@ void sparc_introduce_prolog_epilog(ir_graph *irg)
                schedpoint = sched_next(schedpoint);
 
        if (!layout->sp_relative) {
-               ir_node *incsp;
-               ir_node *save = new_bd_sparc_Save_imm(NULL, block, sp, NULL,
-                                                     -SPARC_MIN_STACKSIZE);
+               ir_node *const save = new_bd_sparc_Save_imm(NULL, block, initial_sp, NULL, -(SPARC_MIN_STACKSIZE + frame_size));
                arch_set_irn_register(save, sp_reg);
                sched_add_after(schedpoint, save);
                schedpoint = save;
 
-               incsp = be_new_IncSP(sp_reg, block, save, frame_size, 0);
-               edges_reroute(initial_sp, incsp);
-               set_irn_n(save, n_sparc_Save_stack, initial_sp);
-               sched_add_after(schedpoint, incsp);
-               schedpoint = incsp;
+               edges_reroute_except(initial_sp, save, save);
 
-               /* we still need the IncSP even if noone is explicitely using the
+               /* we still need the Save even if noone is explicitely using the
                 * value. (TODO: this isn't 100% correct yet, something at the end of
-                * the function should hold the IncSP, even if we use a restore
+                * the function should hold the Save, even if we use a restore
                 * which just overrides it instead of using the value)
                 */
-               if (get_irn_n_edges(incsp) == 0) {
-                       ir_node *in[] = { incsp };
+               if (get_irn_n_edges(save) == 0) {
+                       ir_node *in[] = { save };
                        ir_node *keep = be_new_Keep(block, 1, in);
                        sched_add_after(schedpoint, keep);
                }
        } else {
-               ir_node *incsp = be_new_IncSP(sp_reg, block, sp, frame_size, 0);
-               edges_reroute(initial_sp, incsp);
-               be_set_IncSP_pred(incsp, sp);
+               ir_node *const incsp = be_new_IncSP(sp_reg, block, initial_sp, frame_size, 0);
+               edges_reroute_except(initial_sp, incsp, incsp);
                sched_add_after(schedpoint, incsp);
        }
 }
 
+/**
+ * Creates a constant from an immediate value.
+ */
+static ir_node *create_constant_from_immediate(ir_node *node, int offset)
+{
+       dbg_info *dbgi  = get_irn_dbg_info(node);
+       ir_node  *block = get_nodes_block(node);
+       ir_node  *high  = new_bd_sparc_SetHi(dbgi, block, NULL, offset);
+
+       sched_add_before(node, high);
+       arch_set_irn_register(high, &sparc_registers[REG_G4]);
+
+       if ((offset & 0x3ff) != 0) {
+               ir_node *low = new_bd_sparc_Or_imm(dbgi, block, high, NULL, offset & 0x3ff);
+
+               sched_add_before(node, low);
+               arch_set_irn_register(low, &sparc_registers[REG_G4]);
+
+               return low;
+       }
+
+       return high;
+}
+
 static void finish_sparc_Save(ir_node *node)
 {
        sparc_attr_t *attr = get_sparc_attr(node);
        int offset = attr->immediate_value;
-       ir_node  *schedpoint = node;
-       dbg_info *dbgi;
-       ir_node  *block;
-       ir_node  *new_save;
-       ir_node  *stack;
-       ir_entity *entity;
-
-       if (sparc_is_value_imm_encodeable(offset))
-               return;
 
-       /* uhh only works for the imm variant yet */
-       assert(get_irn_arity(node) == 1);
-
-       block = get_nodes_block(node);
-       dbgi = get_irn_dbg_info(node);
-       stack = get_irn_n(node, n_sparc_Save_stack);
-       entity = attr->immediate_value_entity;
-       new_save = new_bd_sparc_Save_imm(dbgi, block, stack, entity, 0);
-       arch_set_irn_register(new_save, &sparc_registers[REG_SP]);
-       stack = new_save;
-
-       sched_add_after(node, new_save);
-       schedpoint = new_save;
-       while (offset > SPARC_IMMEDIATE_MAX || offset < SPARC_IMMEDIATE_MIN) {
-               if (offset > 0) {
-                       stack = be_new_IncSP(&sparc_registers[REG_SP], block, stack,
-                                            SPARC_IMMEDIATE_MIN, 0);
-                       offset -= -SPARC_IMMEDIATE_MIN;
-               } else {
-                       stack = be_new_IncSP(&sparc_registers[REG_SP], block, stack,
-                                            -SPARC_IMMEDIATE_MIN, 0);
-                       offset -= SPARC_IMMEDIATE_MIN;
-               }
-               sched_add_after(schedpoint, stack);
-               schedpoint = stack;
+       if (! sparc_is_value_imm_encodeable(offset)) {
+               ir_node               *base     = get_irn_n(node, n_sparc_Save_stack);
+               dbg_info              *dbgi     = get_irn_dbg_info(node);
+               ir_node               *block    = get_nodes_block(node);
+               ir_node               *constant = create_constant_from_immediate(node, offset);
+               ir_node               *new_save = new_bd_sparc_Save_reg(dbgi, block, base, constant);
+               const arch_register_t *reg      = arch_get_irn_register(node);
+
+               /* we have a Save with immediate */
+               assert(get_irn_arity(node) == 1);
+
+               sched_add_before(node, new_save);
+               arch_set_irn_register(new_save, reg);
+               be_peephole_exchange(node, new_save);
        }
-       attr = get_sparc_attr(new_save);
-       attr->immediate_value = offset;
-       be_peephole_exchange(node, stack);
 }
 
 /**
- * sparc immediates are limited. Split IncSP with bigger immediates if
+ * SPARC immediates are limited. Split IncSP with bigger immediates if
  * necessary.
  */
 static void finish_be_IncSP(ir_node *node)
 {
-       int      sign   = 1;
-       int      offset = be_get_IncSP_offset(node);
-       ir_node *sp     = be_get_IncSP_pred(node);
-       ir_node *block;
-
-       /* we might have to break the IncSP apart if the constant has become too
-        * big */
-       if (offset < 0) {
-               offset = -offset;
-               sign   = -1;
-       }
-
-       if (sparc_is_value_imm_encodeable(-offset))
-               return;
-
-       /* split incsp into multiple instructions */
-       block = get_nodes_block(node);
-       while (offset > -SPARC_IMMEDIATE_MIN) {
-               sp = be_new_IncSP(&sparc_registers[REG_SP], block, sp,
-                                 sign * -SPARC_IMMEDIATE_MIN, 0);
-               sched_add_before(node, sp);
-               offset -= -SPARC_IMMEDIATE_MIN;
+       int offset = be_get_IncSP_offset(node);
+
+       /* we might have to break the IncSP apart if the constant has become too big */
+       if (! sparc_is_value_imm_encodeable(offset) && ! sparc_is_value_imm_encodeable(-offset)) {
+               ir_node               *sp       = be_get_IncSP_pred(node);
+               dbg_info              *dbgi     = get_irn_dbg_info(node);
+               ir_node               *block    = get_nodes_block(node);
+               ir_node               *constant = create_constant_from_immediate(node, offset);
+               ir_node               *sub      = new_bd_sparc_Sub_reg(dbgi, block, sp, constant);
+
+               sched_add_before(node, sub);
+               arch_set_irn_register(sub, &sparc_registers[REG_SP]);
+               be_peephole_exchange(node, sub);
        }
-
-       be_set_IncSP_pred(node, sp);
-       be_set_IncSP_offset(node, sign*offset);
 }
 
 /**
- * adjust sp-relative offsets. Split into multiple instructions if offset
- * exceeds sparc immediate range.
+ * Adjust sp-relative offsets.
+ *
+ * Split into multiple instructions if offset exceeds SPARC immediate range.
  */
 static void finish_sparc_FrameAddr(ir_node *node)
 {
-       /* adapt to sparc stack magic */
        sparc_attr_t *attr   = get_sparc_attr(node);
        int           offset = attr->immediate_value;
-       ir_node      *base   = get_irn_n(node, n_sparc_FrameAddr_base);
-       dbg_info     *dbgi   = get_irn_dbg_info(node);
-       ir_node      *block  = get_nodes_block(node);
-       int           sign   = 1;
-       bool          sp_relative
-               = arch_get_irn_register(base) == &sparc_registers[REG_SP];
-       if (sp_relative) {
-               offset += SPARC_MIN_STACKSIZE;
+
+       if (! sparc_is_value_imm_encodeable(offset)) {
+               ir_node               *base          = get_irn_n(node, n_sparc_FrameAddr_base);
+               dbg_info              *dbgi          = get_irn_dbg_info(node);
+               ir_node               *block         = get_nodes_block(node);
+               ir_node               *constant      = create_constant_from_immediate(node, offset);
+               ir_node               *new_frameaddr = new_bd_sparc_Add_reg(dbgi, block, base, constant);
+               const arch_register_t *reg           = arch_get_irn_register(node);
+
+               sched_add_before(node, new_frameaddr);
+               arch_set_irn_register(new_frameaddr, reg);
+               be_peephole_exchange(node, new_frameaddr);
        }
+}
+
+static void finish_sparc_Ld(ir_node *node)
+{
+       sparc_attr_t                  *attr            = get_sparc_attr(node);
+       int                            offset          = attr->immediate_value;
+       const sparc_load_store_attr_t *load_store_attr = get_sparc_load_store_attr_const(node);
+
+       if (! load_store_attr->is_frame_entity)
+               return;
 
-       if (offset < 0) {
-               sign   = -1;
-               offset = -offset;
+       if (! sparc_is_value_imm_encodeable(offset)) {
+               ir_node                 *ptr             = get_irn_n(node, n_sparc_Ld_ptr);
+               dbg_info                *dbgi            = get_irn_dbg_info(node);
+               ir_node                 *block           = get_nodes_block(node);
+               ir_node                 *mem             = get_irn_n(node, n_sparc_Ld_mem);
+               ir_mode                 *load_store_mode = load_store_attr->load_store_mode;
+               ir_node                 *constant        = create_constant_from_immediate(node, offset);
+               ir_node                 *new_load        = new_bd_sparc_Ld_reg(dbgi, block, ptr, constant, mem, load_store_mode);
+               sparc_load_store_attr_t *new_load_attr   = get_sparc_load_store_attr(new_load);
+               unsigned                 n_outs          = arch_get_irn_n_outs(node);
+               unsigned                 i;
+
+               new_load_attr->is_frame_entity = load_store_attr->is_frame_entity;
+               new_load_attr->is_reg_reg      = load_store_attr->is_reg_reg;
+
+               sched_add_before(node, new_load);
+               for (i = 0; i < n_outs; i++) {
+                       arch_set_irn_register_out(new_load, i, arch_get_irn_register_out(node, i));
+               }
+               be_peephole_exchange(node, new_load);
        }
 
-       if (offset > -SPARC_IMMEDIATE_MIN) {
-               ir_entity *entity = attr->immediate_value_entity;
-               ir_node   *new_frameaddr
-                       = new_bd_sparc_FrameAddr(dbgi, block, base, entity, 0);
-               ir_node   *schedpoint = node;
-               const arch_register_t *reg = arch_get_irn_register(node);
+}
 
-               sched_add_after(schedpoint, new_frameaddr);
-               schedpoint = new_frameaddr;
-               arch_set_irn_register(new_frameaddr, reg);
-               base = new_frameaddr;
+static void split_sparc_ldf(ir_node *node)
+{
+       sparc_load_store_attr_t *attr = get_sparc_load_store_attr(node);
+       unsigned                 bits = get_mode_size_bits(attr->load_store_mode);
+       /* split 128bit loads into 2 64bit loads */
+       if (bits == 128) {
+               dbg_info *dbgi  = get_irn_dbg_info(node);
+               ir_node  *block = get_nodes_block(node);
+               ir_node  *ptr   = get_irn_n(node, n_sparc_Ldf_ptr);
+               ir_node  *mem   = get_irn_n(node, n_sparc_Ldf_mem);
+               ir_node  *new_load
+                       = new_bd_sparc_Ldf_d(dbgi, block, ptr, mem, mode_D,
+                                            attr->base.immediate_value_entity,
+                                            attr->base.immediate_value + 8,
+                                            attr->is_frame_entity);
+               ir_node  *new_mem = new_r_Proj(new_load, mode_M, pn_sparc_Ldf_M);
+
+               const arch_register_t *reg
+                       = arch_get_irn_register_out(node, pn_sparc_Ldf_res);
+               unsigned reg_index = reg->global_index;
+
+               arch_set_irn_register_out(new_load, pn_sparc_Ldf_res,
+                                         &sparc_registers[reg_index+2]);
+
+               attr->load_store_mode = mode_D;
+               set_irn_n(node, n_sparc_Ldf_mem, new_mem);
+               sched_add_before(node, new_load);
+       }
+}
 
-               while (offset > -SPARC_IMMEDIATE_MIN) {
-                       if (sign > 0) {
-                               base = new_bd_sparc_Sub_imm(dbgi, block, base, NULL,
-                                                                                       SPARC_IMMEDIATE_MIN);
-                       } else {
-                               base = new_bd_sparc_Add_imm(dbgi, block, base, NULL,
-                                                                                       SPARC_IMMEDIATE_MIN);
-                       }
-                       arch_set_irn_register(base, reg);
-                       sched_add_after(schedpoint, base);
-                       schedpoint = base;
+static void finish_sparc_Ldf(ir_node *node)
+{
+       sparc_attr_t                  *attr            = get_sparc_attr(node);
+       int                            offset          = attr->immediate_value;
+       const sparc_load_store_attr_t *load_store_attr = get_sparc_load_store_attr_const(node);
+
+       if (! load_store_attr->is_frame_entity)
+               return;
 
-                       offset -= -SPARC_IMMEDIATE_MIN;
+       if (! sparc_is_value_imm_encodeable(offset)) {
+               ir_node                 *ptr             = get_irn_n(node, n_sparc_Ldf_ptr);
+               dbg_info                *dbgi            = get_irn_dbg_info(node);
+               ir_node                 *block           = get_nodes_block(node);
+               ir_node                 *mem             = get_irn_n(node, n_sparc_Ldf_mem);
+               ir_mode                 *load_store_mode = load_store_attr->load_store_mode;
+               ir_node                 *constant        = create_constant_from_immediate(node, offset);
+               ir_node                 *new_ptr         = new_bd_sparc_Add_reg(dbgi, block, ptr, constant);
+               ir_node                 *new_load        = new_bd_sparc_Ldf_s(dbgi, block, new_ptr, mem, load_store_mode, NULL, 0, true);
+               sparc_load_store_attr_t *new_load_attr   = get_sparc_load_store_attr(new_load);
+               unsigned                 n_outs          = arch_get_irn_n_outs(node);
+               unsigned                 i;
+
+               new_load_attr->is_frame_entity = load_store_attr->is_frame_entity;
+               new_load_attr->is_reg_reg      = load_store_attr->is_reg_reg;
+
+               sched_add_before(node, new_load);
+               for (i = 0; i < n_outs; i++) {
+                       arch_set_irn_register_out(new_load, i, arch_get_irn_register_out(node, i));
                }
+               be_peephole_exchange(node, new_load);
+       }
 
-               be_peephole_exchange(node, base);
-               attr = get_sparc_attr(new_frameaddr);
+}
+
+static void finish_sparc_St(ir_node *node)
+{
+       sparc_attr_t                  *attr            = get_sparc_attr(node);
+       int                            offset          = attr->immediate_value;
+       const sparc_load_store_attr_t *load_store_attr = get_sparc_load_store_attr_const(node);
+
+       if (! load_store_attr->is_frame_entity)
+               return;
+
+       if (! sparc_is_value_imm_encodeable(offset)) {
+               ir_node                 *ptr             = get_irn_n(node, n_sparc_St_ptr);
+               dbg_info                *dbgi            = get_irn_dbg_info(node);
+               ir_node                 *block           = get_nodes_block(node);
+               ir_node                 *mem             = get_irn_n(node, n_sparc_St_mem);
+               ir_node                 *value           = get_irn_n(node, n_sparc_St_val);
+               ir_mode                 *load_store_mode = load_store_attr->load_store_mode;
+               ir_node                 *constant        = create_constant_from_immediate(node, offset);
+               ir_node                 *new_load        = new_bd_sparc_St_reg(dbgi, block, value, ptr, constant, mem, load_store_mode);
+               sparc_load_store_attr_t *new_load_attr   = get_sparc_load_store_attr(new_load);
+               unsigned                 n_outs          = arch_get_irn_n_outs(node);
+               unsigned                 i;
+
+               new_load_attr->is_frame_entity = load_store_attr->is_frame_entity;
+               new_load_attr->is_reg_reg      = load_store_attr->is_reg_reg;
+
+               sched_add_before(node, new_load);
+               for (i = 0; i < n_outs; i++) {
+                       arch_set_irn_register_out(new_load, i, arch_get_irn_register_out(node, i));
+               }
+               be_peephole_exchange(node, new_load);
        }
-       attr->immediate_value = sign*offset;
+
 }
 
-static void finish_sparc_LdSt(ir_node *node)
+static void finish_sparc_Stf(ir_node *node)
 {
-       sparc_load_store_attr_t *attr = get_sparc_load_store_attr(node);
-       if (attr->is_frame_entity) {
-               ir_node *base;
-               bool     sp_relative;
-               if (is_sparc_Ld(node) || is_sparc_Ldf(node)) {
-                       base = get_irn_n(node, n_sparc_Ld_ptr);
-               } else {
-                       assert(is_sparc_St(node) || is_sparc_Stf(node));
-                       base = get_irn_n(node, n_sparc_St_ptr);
+       sparc_attr_t                  *attr            = get_sparc_attr(node);
+       int                            offset          = attr->immediate_value;
+       const sparc_load_store_attr_t *load_store_attr = get_sparc_load_store_attr_const(node);
+
+       if (! load_store_attr->is_frame_entity)
+               return;
+
+       if (! sparc_is_value_imm_encodeable(offset)) {
+               ir_node                 *ptr             = get_irn_n(node, n_sparc_Stf_ptr);
+               dbg_info                *dbgi            = get_irn_dbg_info(node);
+               ir_node                 *block           = get_nodes_block(node);
+               ir_node                 *mem             = get_irn_n(node, n_sparc_Stf_mem);
+               ir_node                 *value           = get_irn_n(node, n_sparc_Stf_val);
+               ir_mode                 *load_store_mode = load_store_attr->load_store_mode;
+               ir_node                 *constant        = create_constant_from_immediate(node, offset);
+               ir_node                 *new_ptr         = new_bd_sparc_Add_reg(dbgi, block, ptr, constant);
+               ir_node                 *new_load        = new_bd_sparc_Stf_s(dbgi, block, value, new_ptr, mem, load_store_mode, NULL, 0, true);
+               sparc_load_store_attr_t *new_load_attr   = get_sparc_load_store_attr(new_load);
+               unsigned                 n_outs          = arch_get_irn_n_outs(node);
+               unsigned                 i;
+
+               new_load_attr->is_frame_entity = load_store_attr->is_frame_entity;
+               new_load_attr->is_reg_reg      = load_store_attr->is_reg_reg;
+
+               sched_add_before(node, new_load);
+               for (i = 0; i < n_outs; i++) {
+                       arch_set_irn_register_out(new_load, i, arch_get_irn_register_out(node, i));
                }
-               sp_relative = arch_get_irn_register(base) == &sparc_registers[REG_SP];
-               if (sp_relative)
-                       attr->base.immediate_value += SPARC_MIN_STACKSIZE;
+               be_peephole_exchange(node, new_load);
        }
+
 }
 
 static void peephole_be_IncSP(ir_node *node)
@@ -346,6 +447,143 @@ static void peephole_sparc_FrameAddr(ir_node *node)
        (void) node;
 }
 
+/* output must not be local, or out reg. Since the destination of the restore
+ * is the rotated register-file where only the old in-registers are still
+ * visible (as out-registers) */
+static bool is_restorezeroopt_reg(const arch_register_t *reg)
+{
+       unsigned index = reg->global_index;
+       return (index >= REG_G0 && index <= REG_G7)
+           || (index >= REG_I0 && index <= REG_I7);
+}
+
+static void replace_with_restore_reg(ir_node *node, ir_node *replaced,
+                                                                        ir_node *op0, ir_node *op1)
+{
+       dbg_info *dbgi     = get_irn_dbg_info(node);
+       ir_node  *stack_in = get_irn_n(node, n_sparc_RestoreZero_stack);
+       ir_node  *fp       = get_irn_n(node, n_sparc_RestoreZero_frame_pointer);
+       ir_node  *block    = get_nodes_block(node);
+       ir_mode  *mode     = get_irn_mode(node);
+       ir_node  *new_node = new_bd_sparc_Restore_reg(dbgi, block, stack_in, fp,
+                                                     op0, op1);
+       ir_node  *stack    = new_r_Proj(new_node, mode, pn_sparc_Restore_stack);
+       ir_node  *res      = new_r_Proj(new_node, mode, pn_sparc_Restore_res);
+       const arch_register_t *reg = arch_get_irn_register(replaced);
+       const arch_register_t *sp  = &sparc_registers[REG_SP];
+       arch_set_irn_register_out(new_node, pn_sparc_Restore_stack, sp);
+       arch_set_irn_register_out(new_node, pn_sparc_Restore_res, reg);
+
+       sched_add_before(node, new_node);
+       be_peephole_exchange(node, stack);
+       be_peephole_exchange(replaced, res);
+}
+
+static void replace_with_restore_imm(ir_node *node, ir_node *replaced,
+                                                                        ir_node *op, ir_entity *imm_entity,
+                                                                        int32_t immediate)
+{
+       dbg_info *dbgi     = get_irn_dbg_info(node);
+       ir_node  *stack_in = get_irn_n(node, n_sparc_RestoreZero_stack);
+       ir_node  *fp       = get_irn_n(node, n_sparc_RestoreZero_frame_pointer);
+       ir_node  *block    = get_nodes_block(node);
+       ir_mode  *mode     = get_irn_mode(node);
+       ir_node  *new_node = new_bd_sparc_Restore_imm(dbgi, block, stack_in, fp,
+                                                     op, imm_entity, immediate);
+       ir_node  *stack    = new_r_Proj(new_node, mode, pn_sparc_Restore_stack);
+       ir_node  *res      = new_r_Proj(new_node, mode, pn_sparc_Restore_res);
+       const arch_register_t *reg = arch_get_irn_register(replaced);
+       const arch_register_t *sp  = &sparc_registers[REG_SP];
+       arch_set_irn_register_out(new_node, pn_sparc_Restore_stack, sp);
+       arch_set_irn_register_out(new_node, pn_sparc_Restore_res, reg);
+
+       sched_add_before(node, new_node);
+       be_peephole_exchange(node, stack);
+       be_peephole_exchange(replaced, res);
+}
+
+static void peephole_sparc_RestoreZero(ir_node *node)
+{
+       /* restore gives us a free "add" instruction, let's try to use that to fold
+        * an instruction in. We can do the following:
+        *
+        * - Copy values                  (g0 + reg)
+        * - Produce constants            (g0 + immediate)
+        * - Perform an add               (reg + reg)
+        * - Perform a sub with immediate (reg + (-immediate))
+        *
+        * Note: In an ideal world, this would not be a peephole optimization but
+        * already performed during code selection. Since about all foldable ops are
+        * arguments of the return node. However we have a hard time doing this
+        * since we construct epilogue code only after register allocation
+        * (and therefore after code selection).
+        */
+       int n_tries = 10; /* limit our search */
+       ir_node *schedpoint = node;
+
+       while (sched_has_prev(schedpoint)) {
+               const arch_register_t *reg;
+               schedpoint = sched_prev(schedpoint);
+
+               if (--n_tries == 0)
+                       break;
+
+               if (arch_get_irn_n_outs(schedpoint) == 0)
+                       continue;
+
+               if (!mode_is_data(get_irn_mode(schedpoint)))
+                       return;
+
+               reg = arch_get_irn_register(schedpoint);
+               if (!is_restorezeroopt_reg(reg))
+                       continue;
+
+               if (be_is_Copy(schedpoint) && be_can_move_down(heights, schedpoint, node)) {
+                       ir_node *const op = be_get_Copy_op(schedpoint);
+                       replace_with_restore_imm(node, schedpoint, op, NULL, 0);
+               } else if (is_sparc_Or(schedpoint) &&
+                          arch_get_irn_flags(schedpoint) & ((arch_irn_flags_t)sparc_arch_irn_flag_immediate_form) &&
+                          arch_get_irn_register_in(schedpoint, 0) == &sparc_registers[REG_G0] &&
+                          be_can_move_down(heights, schedpoint, node)) {
+                       /* it's a constant */
+                       const sparc_attr_t *attr      = get_sparc_attr_const(schedpoint);
+                       ir_entity          *entity    = attr->immediate_value_entity;
+                       int32_t             immediate = attr->immediate_value;
+                       ir_node            *g0        = get_irn_n(schedpoint, 0);
+                       replace_with_restore_imm(node, schedpoint, g0, entity, immediate);
+               } else if (is_sparc_Add(schedpoint) &&
+                          be_can_move_down(heights, schedpoint, node)) {
+                       if (arch_get_irn_flags(schedpoint) & ((arch_irn_flags_t)sparc_arch_irn_flag_immediate_form)) {
+                               ir_node            *op     = get_irn_n(schedpoint, 0);
+                               const sparc_attr_t *attr   = get_sparc_attr_const(schedpoint);
+                               ir_entity          *entity = attr->immediate_value_entity;
+                               int32_t             imm    = attr->immediate_value;
+                               replace_with_restore_imm(node, schedpoint, op, entity, imm);
+                       } else {
+                               ir_node *op0 = get_irn_n(schedpoint, 0);
+                               ir_node *op1 = get_irn_n(schedpoint, 1);
+                               replace_with_restore_reg(node, schedpoint, op0, op1);
+                       }
+               } else if (is_sparc_Sub(schedpoint) &&
+                          arch_get_irn_flags(schedpoint) & ((arch_irn_flags_t)sparc_arch_irn_flag_immediate_form) &&
+                          arch_get_irn_register_in(schedpoint, 0) == &sparc_registers[REG_G0] &&
+                          be_can_move_down(heights, schedpoint, node)) {
+                       /* it's a constant */
+                       const sparc_attr_t *attr   = get_sparc_attr_const(schedpoint);
+                       ir_entity          *entity = attr->immediate_value_entity;
+                       int32_t             imm    = attr->immediate_value;
+                       if (entity == NULL && sparc_is_value_imm_encodeable(-imm)) {
+                               ir_node *g0 = get_irn_n(schedpoint, 0);
+                               replace_with_restore_imm(node, schedpoint, g0, NULL, -imm);
+                       } else {
+                               continue;
+                       }
+               }
+               /* when we're here then we performed a folding and are done */
+               return;
+       }
+}
+
 static void finish_sparc_Return(ir_node *node)
 {
        ir_node *schedpoint = node;
@@ -376,78 +614,6 @@ static void register_peephole_optimisation(ir_op *op, peephole_opt_func func)
        op->ops.generic = (op_func) func;
 }
 
-/**
- * transform reload node => load
- */
-static void transform_Reload(ir_node *node)
-{
-       ir_node   *block  = get_nodes_block(node);
-       dbg_info  *dbgi   = get_irn_dbg_info(node);
-       ir_node   *ptr    = get_irn_n(node, n_be_Spill_frame);
-       ir_node   *mem    = get_irn_n(node, n_be_Reload_mem);
-       ir_mode   *mode   = get_irn_mode(node);
-       ir_entity *entity = be_get_frame_entity(node);
-       const arch_register_t *reg;
-       ir_node   *proj;
-       ir_node   *load;
-
-       ir_node  *sched_point = sched_prev(node);
-
-       load = new_bd_sparc_Ld_imm(dbgi, block, ptr, mem, mode, entity, 0, true);
-       sched_add_after(sched_point, load);
-       sched_remove(node);
-
-       proj = new_rd_Proj(dbgi, load, mode, pn_sparc_Ld_res);
-
-       reg = arch_get_irn_register(node);
-       arch_set_irn_register(proj, reg);
-
-       exchange(node, proj);
-}
-
-/**
- * transform spill node => store
- */
-static void transform_Spill(ir_node *node)
-{
-       ir_node   *block  = get_nodes_block(node);
-       dbg_info  *dbgi   = get_irn_dbg_info(node);
-       ir_node   *ptr    = get_irn_n(node, n_be_Spill_frame);
-       ir_graph  *irg    = get_irn_irg(node);
-       ir_node   *mem    = get_irg_no_mem(irg);
-       ir_node   *val    = get_irn_n(node, n_be_Spill_val);
-       ir_mode   *mode   = get_irn_mode(val);
-       ir_entity *entity = be_get_frame_entity(node);
-       ir_node   *sched_point;
-       ir_node   *store;
-
-       sched_point = sched_prev(node);
-       store = new_bd_sparc_St_imm(dbgi, block, val, ptr, mem, mode, entity, 0, true);
-       sched_remove(node);
-       sched_add_after(sched_point, store);
-
-       exchange(node, store);
-}
-
-/**
- * walker to transform be_Spill and be_Reload nodes
- */
-static void sparc_after_ra_walker(ir_node *block, void *data)
-{
-       ir_node *node, *prev;
-       (void) data;
-
-       for (node = sched_last(block); !sched_is_begin(node); node = prev) {
-               prev = sched_prev(node);
-
-               if (be_is_Reload(node)) {
-                       transform_Reload(node);
-               } else if (be_is_Spill(node)) {
-                       transform_Spill(node);
-               }
-       }
-}
-
 static void sparc_collect_frame_entity_nodes(ir_node *node, void *data)
 {
        be_fec_env_t  *env = (be_fec_env_t*)data;
@@ -473,7 +639,7 @@ static void sparc_collect_frame_entity_nodes(ir_node *node, void *data)
                return;
        if (!attr->is_frame_entity)
                return;
-       if (arch_irn_get_flags(node) & sparc_arch_irn_flag_needs_64bit_spillslot)
+       if (arch_get_irn_flags(node) & sparc_arch_irn_flag_needs_64bit_spillslot)
                mode = mode_Lu;
        align  = get_mode_size_bytes(mode);
        be_node_needs_frame_entity(env, node, mode, align);
@@ -493,7 +659,7 @@ static void sparc_set_frame_entity(ir_node *node, ir_entity *entity)
        }
 }
 
-void sparc_finish(ir_graph *irg)
+void sparc_finish_graph(ir_graph *irg)
 {
        be_stack_layout_t *stack_layout = be_get_irg_stack_layout(irg);
        bool               at_begin     = stack_layout->sp_relative ? true : false;
@@ -502,32 +668,38 @@ void sparc_finish(ir_graph *irg)
        irg_walk_graph(irg, NULL, sparc_collect_frame_entity_nodes, fec_env);
        be_assign_entities(fec_env, sparc_set_frame_entity, at_begin);
        be_free_frame_entity_coalescer(fec_env);
-
-       irg_block_walk_graph(irg, NULL, sparc_after_ra_walker, NULL);
+       sparc_adjust_stack_entity_offsets(irg);
 
        sparc_introduce_prolog_epilog(irg);
 
        /* fix stack entity offsets */
        be_abi_fix_stack_nodes(irg);
-       be_abi_fix_stack_bias(irg);
+       sparc_fix_stack_bias(irg);
+
+       heights = heights_new(irg);
 
        /* perform peephole optimizations */
-       clear_irp_opcodes_generic_func();
+       ir_clear_opcodes_generic_func();
        register_peephole_optimisation(op_be_IncSP,        peephole_be_IncSP);
        register_peephole_optimisation(op_sparc_FrameAddr, peephole_sparc_FrameAddr);
+       register_peephole_optimisation(op_sparc_RestoreZero,
+                                      peephole_sparc_RestoreZero);
+       register_peephole_optimisation(op_sparc_Ldf, split_sparc_ldf);
        be_peephole_opt(irg);
 
        /* perform legalizations (mostly fix nodes with too big immediates) */
-       clear_irp_opcodes_generic_func();
+       ir_clear_opcodes_generic_func();
        register_peephole_optimisation(op_be_IncSP,        finish_be_IncSP);
        register_peephole_optimisation(op_sparc_FrameAddr, finish_sparc_FrameAddr);
-       register_peephole_optimisation(op_sparc_Ld,        finish_sparc_LdSt);
-       register_peephole_optimisation(op_sparc_Ldf,       finish_sparc_LdSt);
+       register_peephole_optimisation(op_sparc_Ld,        finish_sparc_Ld);
+       register_peephole_optimisation(op_sparc_Ldf,       finish_sparc_Ldf);
        register_peephole_optimisation(op_sparc_Return,    finish_sparc_Return);
        register_peephole_optimisation(op_sparc_Save,      finish_sparc_Save);
-       register_peephole_optimisation(op_sparc_St,        finish_sparc_LdSt);
-       register_peephole_optimisation(op_sparc_Stf,       finish_sparc_LdSt);
+       register_peephole_optimisation(op_sparc_St,        finish_sparc_St);
+       register_peephole_optimisation(op_sparc_Stf,       finish_sparc_Stf);
        be_peephole_opt(irg);
 
+       heights_free(heights);
+
        be_remove_dead_nodes_from_schedule(irg);
 }