Make new_type() static.
[libfirm] / ir / opt / opt_inline.c
index 92c2446..43dba4b 100644 (file)
@@ -1,5 +1,5 @@
 /*
- * Copyright (C) 1995-2008 University of Karlsruhe.  All right reserved.
+ * Copyright (C) 1995-2011 University of Karlsruhe.  All right reserved.
  *
  * This file is part of libFirm.
  *
@@ -21,7 +21,6 @@
  * @file
  * @brief    Dead node elimination and Procedure Inlining.
  * @author   Michael Beck, Goetz Lindenmaier
- * @version  $Id$
  */
 #include "config.h"
 
@@ -63,7 +62,7 @@
 #include "irtools.h"
 #include "iropt_dbg.h"
 #include "irpass_t.h"
-#include "irphase_t.h"
+#include "irnodemap.h"
 
 DEBUG_ONLY(static firm_dbg_module_t *dbg;)
 
@@ -89,357 +88,6 @@ static inline ir_node *get_new_node(ir_node *old_node)
        return (ir_node*) get_irn_link(old_node);
 }
 
-/** a pointer to the new phases */
-static ir_phase *new_phases[PHASE_LAST];
-
-/**
- * Copies the node to the new obstack. The Ins of the new node point to
- * the predecessors on the old obstack.  For block/phi nodes not all
- * predecessors might be copied.  n->link points to the new node.
- *
- * @param n    The node to be copied
- * @param irg  The irg onto which the node should be copied
- */
-static ir_node *copy_node(ir_node *n, ir_graph *irg)
-{
-       ir_op    *op    = get_irn_op(n);
-       int       arity = get_irn_arity(n);
-       dbg_info *dbgi  = get_irn_dbg_info(n);
-       ir_mode  *mode  = get_irn_mode(n);
-       ir_node  *nn;
-       ir_node  *block;
-       int       i;
-
-       if (op == op_Block) {
-               block = NULL;
-               n->attr.block.graph_arr = NULL;
-       } else {
-               block = get_nodes_block(n);
-       }
-       if (op->opar == oparity_dynamic) {
-               nn = new_ir_node(dbgi, irg, block, op, mode, -1, NULL);
-               for (i = 0; i < arity; ++i) {
-                       ir_node *in = get_irn_n(n, i);
-                       add_irn_n(nn, in);
-               }
-       } else {
-               ir_node **ins = get_irn_in(n)+1;
-               nn = new_ir_node(dbgi, irg, block, op, mode, arity, ins);
-       }
-
-       /* Copy the attributes.  These might point to additional data.  If this
-          was allocated on the old obstack the pointers now are dangling. */
-       copy_node_attr(irg, n, nn);
-       if (op == op_Block) {
-               /* we cannot allow blocks WITHOUT macroblock input */
-               set_Block_MacroBlock(nn, get_Block_MacroBlock(n));
-       }
-
-       /* copy phase information for this node */
-       for (i = PHASE_NOT_IRG_MANAGED+1; i < PHASE_LAST; i++) {
-               ir_graph *old_irg = get_irn_irg(n);
-               ir_phase *old_ph  = get_irg_phase(old_irg, i);
-               if (old_ph == NULL)
-                       continue;
-               if (!phase_get_irn_data(old_ph, n))
-                       continue;
-               phase_set_irn_data(new_phases[i], nn, phase_get_irn_data(old_ph, n));
-       }
-
-       set_new_node(n, nn);
-       hook_dead_node_elim_subst(current_ir_graph, n, nn);
-
-       return nn;
-}
-
-static void copy_node_with_number(ir_node *node, void *env)
-{
-       ir_graph *new_irg  = (ir_graph*) env;
-       ir_node  *new_node = copy_node(node, new_irg);
-
-       /* preserve the node numbers for easier debugging */
-       new_node->node_nr = node->node_nr;
-}
-
-/**
- * Reroute the inputs of a node from nodes in the old graph to copied nodes in
- * the new graph
- */
-static void set_preds(ir_node *node, void *env)
-{
-       ir_graph *new_irg = (ir_graph*) env;
-       ir_node  *new_node;
-       int       arity;
-       int       i;
-
-       new_node = get_new_node(node);
-
-       if (is_Block(node)) {
-               /* copy the macro block header */
-               ir_node *mbh = get_Block_MacroBlock(node);
-
-               /* get the macro block header */
-               ir_node *nmbh = get_new_node(mbh);
-               assert(nmbh != NULL);
-               set_Block_MacroBlock(new_node, nmbh);
-       } else {
-               ir_node *block     = get_nodes_block(node);
-               ir_node *new_block = get_new_node(block);
-               set_nodes_block(new_node, new_block);
-       }
-
-       arity = get_irn_arity(new_node);
-       for (i = 0; i < arity; ++i) {
-               ir_node *in     = get_irn_n(node, i);
-               ir_node *new_in = get_new_node(in);
-               set_irn_n(new_node, i, new_in);
-       }
-
-       /* Now the new node is complete. We can add it to the hash table for CSE. */
-       add_identities(new_irg->value_table, new_node);
-}
-
-/**
- * Copies the graph reachable from current_ir_graph->end to the obstack
- * in current_ir_graph and fixes the environment.
- * Then fixes the fields in current_ir_graph containing nodes of the
- * graph.
- *
- * @param copy_node_nr  If non-zero, the node number will be copied
- */
-static void copy_graph_env(ir_graph *irg)
-{
-       ir_node  *old_end;
-       ir_node  *new_anchor;
-       ir_phase *old_ph;
-       int i;
-
-       /* init the new_phases array */
-       for (i = PHASE_NOT_IRG_MANAGED+1; i < PHASE_LAST; i++) {
-               old_ph = get_irg_phase(irg, i);
-               if (old_ph == NULL) {
-                       new_phases[i] = NULL;
-               } else {
-                       new_phases[i] = xmalloc(sizeof(ir_phase));
-                       phase_init(new_phases[i], "", irg, old_ph->growth_factor,
-                                       old_ph->data_init, old_ph->priv);
-               }
-       }
-
-       /* remove end_except and end_reg nodes */
-       old_end = get_irg_end(irg);
-       set_irg_end_except(irg, old_end);
-       set_irg_end_reg   (irg, old_end);
-
-       /* copy nodes */
-       irg_walk(irg->anchor, copy_node_with_number, set_preds, irg);
-
-       /* fix the anchor */
-       new_anchor = get_new_node(irg->anchor);
-       assert(new_anchor != NULL);
-       irg->anchor = new_anchor;
-
-       /* copy the new phases into the irg */
-       for (i = PHASE_NOT_IRG_MANAGED+1; i < PHASE_LAST; i++) {
-               old_ph = get_irg_phase(irg, i);
-               if (old_ph == NULL)
-                       continue;
-
-               free_irg_phase(irg, i);
-               irg->phases[i] = new_phases[i];
-       }
-
-       free_End(old_end);
-}
-
-/**
- * Copies all reachable nodes to a new obstack.  Removes bad inputs
- * from block nodes and the corresponding inputs from Phi nodes.
- * Merges single exit blocks with single entry blocks and removes
- * 1-input Phis.
- * Adds all new nodes to a new hash table for CSE.  Does not
- * perform CSE, so the hash table might contain common subexpressions.
- */
-void dead_node_elimination(ir_graph *irg)
-{
-       ir_graph *rem;
-#ifdef INTERPROCEDURAL_VIEW
-       int rem_ipview = get_interprocedural_view();
-#endif
-       struct obstack *graveyard_obst = NULL;
-       struct obstack *rebirth_obst   = NULL;
-
-       edges_deactivate(irg);
-
-       /* inform statistics that we started a dead-node elimination run */
-       hook_dead_node_elim(irg, 1);
-
-       /* Remember external state of current_ir_graph. */
-       rem = current_ir_graph;
-       current_ir_graph = irg;
-#ifdef INTERPROCEDURAL_VIEW
-       set_interprocedural_view(0);
-#endif
-
-       assert(get_irg_phase_state(irg) != phase_building);
-
-       /* Handle graph state */
-       free_callee_info(irg);
-       free_irg_outs(irg);
-       free_trouts();
-       free_loop_information(irg);
-       set_irg_doms_inconsistent(irg);
-
-       /* A quiet place, where the old obstack can rest in peace,
-          until it will be cremated. */
-       graveyard_obst = irg->obst;
-
-       /* A new obstack, where the reachable nodes will be copied to. */
-       rebirth_obst = XMALLOC(struct obstack);
-       irg->obst = rebirth_obst;
-       obstack_init(irg->obst);
-       irg->last_node_idx = 0;
-
-       /* We also need a new value table for CSE */
-       del_identities(irg->value_table);
-       irg->value_table = new_identities();
-
-       /* Copy the graph from the old to the new obstack */
-       copy_graph_env(irg);
-
-       /* Free memory from old unoptimized obstack */
-       obstack_free(graveyard_obst, 0);  /* First empty the obstack ... */
-       xfree(graveyard_obst);            /* ... then free it.           */
-
-       /* inform statistics that the run is over */
-       hook_dead_node_elim(irg, 0);
-
-       current_ir_graph = rem;
-#ifdef INTERPROCEDURAL_VIEW
-       set_interprocedural_view(rem_ipview);
-#endif
-}
-
-ir_graph_pass_t *dead_node_elimination_pass(const char *name)
-{
-       return def_graph_pass(name ? name : "dce", dead_node_elimination);
-}
-
-/*
-   __                      _  __ __
-  (_     __    o     _    | \/  |_
-  __)|_| | \_/ | \_/(/_   |_/\__|__
-
-  The following stuff implements a facility that automatically patches
-  registered ir_node pointers to the new node when a dead node elimination occurs.
-*/
-
-struct _survive_dce_t {
-       struct obstack obst;
-       pmap *places;
-       pmap *new_places;
-       hook_entry_t dead_node_elim;
-       hook_entry_t dead_node_elim_subst;
-};
-
-typedef struct _survive_dce_list_t {
-       struct _survive_dce_list_t *next;
-       ir_node **place;
-} survive_dce_list_t;
-
-static void dead_node_hook(void *context, ir_graph *irg, int start)
-{
-       survive_dce_t *sd = context;
-       (void) irg;
-
-       /* Create a new map before the dead node elimination is performed. */
-       if (start) {
-               sd->new_places = pmap_create_ex(pmap_count(sd->places));
-       } else {
-               /* Patch back all nodes if dead node elimination is over and something is to be done. */
-               pmap_destroy(sd->places);
-               sd->places     = sd->new_places;
-               sd->new_places = NULL;
-       }
-}
-
-/**
- * Hook called when dead node elimination replaces old by nw.
- */
-static void dead_node_subst_hook(void *context, ir_graph *irg, ir_node *old, ir_node *nw)
-{
-       survive_dce_t *sd = context;
-       survive_dce_list_t *list = pmap_get(sd->places, old);
-       (void) irg;
-
-       /* If the node is to be patched back, write the new address to all registered locations. */
-       if (list) {
-               survive_dce_list_t *p;
-
-               for (p = list; p; p = p->next)
-                       *(p->place) = nw;
-
-               pmap_insert(sd->new_places, nw, list);
-       }
-}
-
-/**
- * Make a new Survive DCE environment.
- */
-survive_dce_t *new_survive_dce(void)
-{
-       survive_dce_t *res = XMALLOC(survive_dce_t);
-       obstack_init(&res->obst);
-       res->places     = pmap_create();
-       res->new_places = NULL;
-
-       res->dead_node_elim.hook._hook_dead_node_elim = dead_node_hook;
-       res->dead_node_elim.context                   = res;
-       res->dead_node_elim.next                      = NULL;
-
-       res->dead_node_elim_subst.hook._hook_dead_node_elim_subst = dead_node_subst_hook;
-       res->dead_node_elim_subst.context = res;
-       res->dead_node_elim_subst.next    = NULL;
-
-       register_hook(hook_dead_node_elim, &res->dead_node_elim);
-       register_hook(hook_dead_node_elim_subst, &res->dead_node_elim_subst);
-       return res;
-}
-
-/**
- * Free a Survive DCE environment.
- */
-void free_survive_dce(survive_dce_t *sd)
-{
-       obstack_free(&sd->obst, NULL);
-       pmap_destroy(sd->places);
-       unregister_hook(hook_dead_node_elim, &sd->dead_node_elim);
-       unregister_hook(hook_dead_node_elim_subst, &sd->dead_node_elim_subst);
-       xfree(sd);
-}
-
-/**
- * Register a node pointer to be patched upon DCE.
- * When DCE occurs, the node pointer specified by @p place will be
- * patched to the new address of the node it is pointing to.
- *
- * @param sd    The Survive DCE environment.
- * @param place The address of the node pointer.
- */
-void survive_dce_register_irn(survive_dce_t *sd, ir_node **place)
-{
-       if (*place != NULL) {
-               ir_node *irn      = *place;
-               survive_dce_list_t *curr = pmap_get(sd->places, irn);
-               survive_dce_list_t *nw   = OALLOC(&sd->obst, survive_dce_list_t);
-
-               nw->next  = curr;
-               nw->place = place;
-
-               pmap_insert(sd->places, irn, nw);
-       }
-}
-
 /*--------------------------------------------------------------------*/
 /*  Functionality for inlining                                         */
 /*--------------------------------------------------------------------*/
@@ -455,10 +103,10 @@ void survive_dce_register_irn(survive_dce_t *sd, ir_node **place)
  */
 static void copy_node_inline(ir_node *node, void *env)
 {
-       ir_graph *new_irg = (ir_graph*) env;
-       ir_node  *new_node;
+       ir_graph *new_irg  = (ir_graph*) env;
+       ir_node  *new_node = irn_copy_into_irg(node, new_irg);
 
-       new_node = copy_node(node, new_irg);
+       set_new_node(node, new_node);
        if (is_Sel(node)) {
                ir_graph  *old_irg        = get_irn_irg(node);
                ir_type   *old_frame_type = get_irg_frame_type(old_irg);
@@ -466,7 +114,7 @@ static void copy_node_inline(ir_node *node, void *env)
                assert(is_Sel(new_node));
                /* use copied entities from the new frame */
                if (get_entity_owner(old_entity) == old_frame_type) {
-                       ir_entity *new_entity = get_entity_link(old_entity);
+                       ir_entity *new_entity = (ir_entity*)get_entity_link(old_entity);
                        assert(new_entity != NULL);
                        set_Sel_entity(new_node, new_entity);
                }
@@ -479,7 +127,7 @@ static void set_preds_inline(ir_node *node, void *env)
 {
        ir_node *new_node;
 
-       set_preds(node, env);
+       irn_rewire_inputs(node);
 
        /* move constants into start block */
        new_node = get_new_node(node);
@@ -495,9 +143,15 @@ static void set_preds_inline(ir_node *node, void *env)
  */
 static void find_addr(ir_node *node, void *env)
 {
-       bool *allow_inline = env;
+       bool *allow_inline = (bool*)env;
 
-       if (is_Sel(node)) {
+       if (is_Block(node) && get_Block_entity(node)) {
+               /**
+                * Currently we can't handle blocks whose address was taken correctly
+                * when inlining
+                */
+               *allow_inline = false;
+       } else if (is_Sel(node)) {
                ir_graph *irg = current_ir_graph;
                if (get_Sel_ptr(node) == get_irg_frame(irg)) {
                        /* access to frame */
@@ -506,6 +160,9 @@ static void find_addr(ir_node *node, void *env)
                                /* access to value_type */
                                *allow_inline = false;
                        }
+                       if (is_parameter_entity(ent)) {
+                               *allow_inline = false;
+                       }
                }
        } else if (is_Alloc(node) && get_Alloc_where(node) == stack_alloc) {
                /* From GCC:
@@ -541,11 +198,11 @@ static bool can_inline(ir_node *call, ir_graph *called_graph)
        ir_entity          *called      = get_irg_entity(called_graph);
        ir_type            *called_type = get_entity_type(called);
        ir_type            *call_type   = get_Call_type(call);
-       int                 n_params    = get_method_n_params(called_type);
-       int                 n_arguments = get_method_n_params(call_type);
-       int                 n_res       = get_method_n_ress(called_type);
+       size_t              n_params    = get_method_n_params(called_type);
+       size_t              n_arguments = get_method_n_params(call_type);
+       size_t              n_res       = get_method_n_ress(called_type);
        irg_inline_property prop        = get_irg_inline_property(called_graph);
-       int                 i;
+       size_t              i;
        bool                res;
 
        if (prop == irg_inline_forbidden)
@@ -565,7 +222,7 @@ static bool can_inline(ir_node *call, ir_graph *called_graph)
         * It is implementation dependent what happens in that case.
         * We support inlining, if the bitsize of the types matches AND
         * the same arithmetic is used. */
-       for (i = n_params - 1; i >= 0; --i) {
+       for (i = 0; i < n_params; ++i) {
                ir_type *param_tp = get_method_param_type(called_type, i);
                ir_type *arg_tp   = get_method_param_type(call_type, i);
 
@@ -582,7 +239,7 @@ static bool can_inline(ir_node *call, ir_graph *called_graph)
                        /* otherwise we can simply "reinterpret" the bits */
                }
        }
-       for (i = n_res - 1; i >= 0; --i) {
+       for (i = 0; i < n_res; ++i) {
                ir_type *decl_res_tp = get_method_res_type(called_type, i);
                ir_type *used_res_tp = get_method_res_type(call_type, i);
 
@@ -634,14 +291,15 @@ static void copy_frame_entities(ir_graph *from, ir_graph *to)
 {
        ir_type *from_frame = get_irg_frame_type(from);
        ir_type *to_frame   = get_irg_frame_type(to);
-       int      n_members  = get_class_n_members(from_frame);
-       int      i;
+       size_t   n_members  = get_class_n_members(from_frame);
+       size_t   i;
        assert(from_frame != to_frame);
 
        for (i = 0; i < n_members; ++i) {
                ir_entity *old_ent = get_class_member(from_frame, i);
                ir_entity *new_ent = copy_entity_own(old_ent, to_frame);
                set_entity_link(old_ent, new_ent);
+               assert (!is_parameter_entity(old_ent));
        }
 }
 
@@ -650,7 +308,7 @@ int inline_method(ir_node *call, ir_graph *called_graph)
 {
        ir_node       *pre_call;
        ir_node       *post_call, *post_bl;
-       ir_node       *in[pn_Start_max];
+       ir_node       *in[pn_Start_max+1];
        ir_node       *end, *end_bl, *block;
        ir_node       **res_pred;
        ir_node       **cf_pred;
@@ -694,12 +352,11 @@ int inline_method(ir_node *call, ir_graph *called_graph)
        assert(get_irg_phase_state(irg) != phase_building);
        assert(get_irg_pinned(irg) == op_pin_state_pinned);
        assert(get_irg_pinned(called_graph) == op_pin_state_pinned);
-       set_irg_outs_inconsistent(irg);
-       set_irg_extblk_inconsistent(irg);
-       set_irg_doms_inconsistent(irg);
-       set_irg_loopinfo_inconsistent(irg);
+       clear_irg_state(irg, IR_GRAPH_STATE_CONSISTENT_DOMINANCE
+                          | IR_GRAPH_STATE_VALID_EXTENDED_BLOCKS
+                          | IR_GRAPH_STATE_CONSISTENT_ENTITY_USAGE);
        set_irg_callee_info_state(irg, irg_callee_info_inconsistent);
-       set_irg_entity_usage_state(irg, ir_entity_usage_not_computed);
+       clear_irg_state(irg, IR_GRAPH_STATE_CONSISTENT_ENTITY_USAGE);
        edges_deactivate(irg);
 
        /* here we know we WILL inline, so inform the statistics */
@@ -713,7 +370,8 @@ int inline_method(ir_node *call, ir_graph *called_graph)
        {
                ir_node *Xproj = NULL;
                ir_node *proj;
-               for (proj = get_irn_link(call); proj; proj = get_irn_link(proj)) {
+               for (proj = (ir_node*)get_irn_link(call); proj != NULL;
+                    proj = (ir_node*)get_irn_link(proj)) {
                        long proj_nr = get_Proj_proj(proj);
                        if (proj_nr == pn_Call_X_except) Xproj = proj;
                }
@@ -743,9 +401,8 @@ int inline_method(ir_node *call, ir_graph *called_graph)
        in[pn_Start_M]              = get_Call_mem(call);
        in[pn_Start_X_initial_exec] = new_r_Jmp(post_bl);
        in[pn_Start_P_frame_base]   = get_irg_frame(irg);
-       in[pn_Start_P_tls]          = get_irg_tls(irg);
        in[pn_Start_T_args]         = new_r_Tuple(post_bl, n_params, args_in);
-       pre_call = new_r_Tuple(post_bl, pn_Start_max, in);
+       pre_call = new_r_Tuple(post_bl, pn_Start_max+1, in);
        post_call = call;
 
        /* --
@@ -765,7 +422,6 @@ int inline_method(ir_node *call, ir_graph *called_graph)
        {
                ir_node *start_block;
                ir_node *start;
-               ir_node *bad;
                ir_node *nomem;
 
                start_block = get_irg_start_block(called_graph);
@@ -776,10 +432,6 @@ int inline_method(ir_node *call, ir_graph *called_graph)
                set_new_node(start, pre_call);
                mark_irn_visited(start);
 
-               bad = get_irg_bad(called_graph);
-               set_new_node(bad, get_irg_bad(irg));
-               mark_irn_visited(bad);
-
                nomem = get_irg_no_mem(called_graph);
                set_new_node(nomem, get_irg_no_mem(irg));
                mark_irn_visited(nomem);
@@ -787,7 +439,7 @@ int inline_method(ir_node *call, ir_graph *called_graph)
 
        /* entitiy link is used to link entities on old stackframe to the
         * new stackframe */
-       irp_reserve_resources(irp, IR_RESOURCE_ENTITY_LINK);
+       irp_reserve_resources(irp, IRP_RESOURCE_ENTITY_LINK);
 
        /* copy entities and nodes */
        assert(!irn_visited(get_irg_end(called_graph)));
@@ -795,7 +447,7 @@ int inline_method(ir_node *call, ir_graph *called_graph)
        irg_walk_core(get_irg_end(called_graph), copy_node_inline, set_preds_inline,
                      irg);
 
-       irp_free_resources(irp, IR_RESOURCE_ENTITY_LINK);
+       irp_free_resources(irp, IRP_RESOURCE_ENTITY_LINK);
 
        /* -- Merge the end of the inlined procedure with the call site -- */
        /* We will turn the old Call node into a Tuple with the following
@@ -843,7 +495,7 @@ int inline_method(ir_node *call, ir_graph *called_graph)
 
        /* build a Tuple for all results of the method.
         * add Phi node if there was more than one Return. */
-       turn_into_tuple(post_call, pn_Call_max);
+       turn_into_tuple(post_call, pn_Call_max+1);
        /* First the Memory-Phi */
        n_mem_phi = 0;
        for (i = 0; i < arity; i++) {
@@ -888,10 +540,9 @@ int inline_method(ir_node *call, ir_graph *called_graph)
                                }
                        }
                        if (n_ret > 0) {
-                               ir_mode *mode = get_irn_mode(cf_pred[0]);
-                               phi = new_r_Phi(post_bl, n_ret, cf_pred, mode);
+                               phi = new_r_Phi(post_bl, n_ret, cf_pred, res_mode);
                        } else {
-                               phi = new_r_Bad(irg);
+                               phi = new_r_Bad(irg, res_mode);
                        }
                        res_pred[j] = phi;
                        /* Conserve Phi-list for further inlinings -- but might be optimized */
@@ -903,14 +554,11 @@ int inline_method(ir_node *call, ir_graph *called_graph)
                result_tuple = new_r_Tuple(post_bl, n_res, res_pred);
                set_Tuple_pred(call, pn_Call_T_result, result_tuple);
        } else {
-               set_Tuple_pred(call, pn_Call_T_result, new_r_Bad(irg));
+               set_Tuple_pred(call, pn_Call_T_result, new_r_Bad(irg, mode_T));
        }
        /* handle the regular call */
        set_Tuple_pred(call, pn_Call_X_regular, new_r_Jmp(post_bl));
 
-       /* For now, we cannot inline calls with value_base */
-       set_Tuple_pred(call, pn_Call_P_value_res_base, new_r_Bad(irg));
-
        /* Finally the exception control flow.
           We have two possible situations:
           First if the Call branches to an exception handler:
@@ -942,7 +590,7 @@ int inline_method(ir_node *call, ir_graph *called_graph)
                                set_Tuple_pred(call, pn_Call_X_except, new_r_Jmp(block));
                        }
                } else {
-                       set_Tuple_pred(call, pn_Call_X_except, new_r_Bad(irg));
+                       set_Tuple_pred(call, pn_Call_X_except, new_r_Bad(irg, mode_X));
                }
        } else {
                ir_node *main_end_bl;
@@ -969,7 +617,7 @@ int inline_method(ir_node *call, ir_graph *called_graph)
                for (i = 0; i < n_exc; ++i)
                        end_preds[main_end_bl_arity + i] = cf_pred[i];
                set_irn_in(main_end_bl, n_exc + main_end_bl_arity, end_preds);
-               set_Tuple_pred(call, pn_Call_X_except, new_r_Bad(irg));
+               set_Tuple_pred(call, pn_Call_X_except, new_r_Bad(irg, mode_X));
                free(end_preds);
        }
        free(res_pred);
@@ -989,7 +637,7 @@ int inline_method(ir_node *call, ir_graph *called_graph)
 static struct obstack  temp_obst;
 
 /** Represents a possible inlinable call in a graph. */
-typedef struct _call_entry {
+typedef struct call_entry {
        ir_node    *call;       /**< The Call node. */
        ir_graph   *callee;     /**< The callee IR-graph. */
        list_head  list;        /**< List head for linking the next one. */
@@ -1002,7 +650,7 @@ typedef struct _call_entry {
 /**
  * environment for inlining small irgs
  */
-typedef struct _inline_env_t {
+typedef struct inline_env_t {
        struct obstack obst;  /**< An obstack where call_entries are allocated on. */
        list_head      calls; /**< The call entry list. */
 } inline_env_t;
@@ -1018,8 +666,8 @@ static ir_graph *get_call_called_irg(ir_node *call)
        ir_node *addr;
 
        addr = get_Call_ptr(call);
-       if (is_Global(addr)) {
-               ir_entity *ent = get_Global_entity(addr);
+       if (is_SymConst_addr_ent(addr)) {
+               ir_entity *ent = get_SymConst_entity(addr);
                /* we don't know which function gets finally bound to a weak symbol */
                if (get_entity_linkage(ent) & IR_LINKAGE_WEAK)
                        return NULL;
@@ -1041,7 +689,7 @@ static void collect_calls(ir_node *call, void *env)
 
                if (called_irg != NULL) {
                        /* The Call node calls a locally defined method.  Remember to inline. */
-                       inline_env_t *ienv  = env;
+                       inline_env_t *ienv  = (inline_env_t*)env;
                        call_entry   *entry = OALLOC(&ienv->obst, call_entry);
                        entry->call       = call;
                        entry->callee     = called_irg;
@@ -1107,17 +755,17 @@ void inline_small_irgs(ir_graph *irg, int size)
        current_ir_graph = rem;
 }
 
-struct inline_small_irgs_pass_t {
+typedef struct inline_small_irgs_pass_t {
        ir_graph_pass_t pass;
        int            size;
-};
+} inline_small_irgs_pass_t;
 
 /**
  * Wrapper to run inline_small_irgs() as a pass.
  */
 static int inline_small_irgs_wrapper(ir_graph *irg, void *context)
 {
-       struct inline_small_irgs_pass_t *pass = context;
+       inline_small_irgs_pass_t *pass = (inline_small_irgs_pass_t*)context;
 
        inline_small_irgs(irg, pass->size);
        return 0;
@@ -1126,8 +774,7 @@ static int inline_small_irgs_wrapper(ir_graph *irg, void *context)
 /* create a pass for inline_small_irgs() */
 ir_graph_pass_t *inline_small_irgs_pass(const char *name, int size)
 {
-       struct inline_small_irgs_pass_t *pass =
-               XMALLOCZ(struct inline_small_irgs_pass_t);
+       inline_small_irgs_pass_t *pass = XMALLOCZ(inline_small_irgs_pass_t);
 
        pass->size = size;
        return def_graph_pass_constructor(
@@ -1183,9 +830,9 @@ typedef struct walker_env {
  */
 static void collect_calls2(ir_node *call, void *ctx)
 {
-       wenv_t         *env = ctx;
+       wenv_t         *env = (wenv_t*)ctx;
        inline_irg_env *x = env->x;
-       ir_opcode      code = get_irn_opcode(call);
+       unsigned        code = get_irn_opcode(call);
        ir_graph       *callee;
        call_entry     *entry;
 
@@ -1205,8 +852,8 @@ static void collect_calls2(ir_node *call, void *ctx)
        if (env->ignore_runtime) {
                ir_node *symc = get_Call_ptr(call);
 
-               if (is_Global(symc)) {
-                       ir_entity *ent = get_Global_entity(symc);
+               if (is_SymConst_addr_ent(symc)) {
+                       ir_entity *ent = get_SymConst_entity(symc);
 
                        if (get_entity_additional_properties(ent) & mtp_property_runtime)
                                return;
@@ -1220,7 +867,7 @@ static void collect_calls2(ir_node *call, void *ctx)
        callee = get_call_called_irg(call);
        if (callee != NULL) {
                if (! env->ignore_callers) {
-                       inline_irg_env *callee_env = get_irg_link(callee);
+                       inline_irg_env *callee_env = (inline_irg_env*)get_irg_link(callee);
                        /* count all static callers */
                        ++callee_env->n_callers;
                        ++callee_env->n_callers_orig;
@@ -1243,11 +890,11 @@ static void collect_calls2(ir_node *call, void *ctx)
 
 /**
  * Returns TRUE if the number of callers is 0 in the irg's environment,
- * hence this irg is a leave.
+ * hence this irg is a leaf.
  */
-inline static int is_leave(ir_graph *irg)
+inline static int is_leaf(ir_graph *irg)
 {
-       inline_irg_env *env = get_irg_link(irg);
+       inline_irg_env *env = (inline_irg_env*)get_irg_link(irg);
        return env->n_call_nodes == 0;
 }
 
@@ -1257,7 +904,7 @@ inline static int is_leave(ir_graph *irg)
  */
 inline static int is_smaller(ir_graph *callee, unsigned size)
 {
-       inline_irg_env *env = get_irg_link(callee);
+       inline_irg_env *env = (inline_irg_env*)get_irg_link(callee);
        return env->n_nodes < size;
 }
 
@@ -1299,7 +946,7 @@ static void append_call_list(inline_irg_env *dst, inline_irg_env *src, int loop_
           we need Call nodes in our graph. Luckily the inliner leaves this information
           in the link field. */
        list_for_each_entry(call_entry, entry, &src->calls, list) {
-               nentry = duplicate_call_entry(entry, get_irn_link(entry->call), loop_depth);
+               nentry = duplicate_call_entry(entry, (ir_node*)get_irn_link(entry->call), loop_depth);
                list_add_tail(&nentry->list, &dst->calls);
        }
        dst->n_call_nodes += src->n_call_nodes;
@@ -1307,19 +954,19 @@ static void append_call_list(inline_irg_env *dst, inline_irg_env *src, int loop_
 }
 
 /*
- * Inlines small leave methods at call sites where the called address comes
+ * Inlines small leaf methods at call sites where the called address comes
  * from a Const node that references the entity representing the called
  * method.
  * The size argument is a rough measure for the code size of the method:
  * Methods where the obstack containing the firm graph is smaller than
  * size are inlined.
  */
-void inline_leave_functions(unsigned maxsize, unsigned leavesize,
+void inline_leaf_functions(unsigned maxsize, unsigned leafsize,
                             unsigned size, int ignore_runtime)
 {
        inline_irg_env   *env;
        ir_graph         *irg;
-       int              i, n_irgs;
+       size_t           i, n_irgs;
        ir_graph         *rem;
        int              did_inline;
        wenv_t           wenv;
@@ -1348,14 +995,14 @@ void inline_leave_functions(unsigned maxsize, unsigned leavesize,
                assert(get_irg_phase_state(irg) != phase_building);
                free_callee_info(irg);
 
-               assure_cf_loop(irg);
-               wenv.x = get_irg_link(irg);
+               assure_loopinfo(irg);
+               wenv.x = (inline_irg_env*)get_irg_link(irg);
                irg_walk_graph(irg, NULL, collect_calls2, &wenv);
        }
 
        /* -- and now inline. -- */
 
-       /* Inline leaves recursively -- we might construct new leaves. */
+       /* Inline leafs recursively -- we might construct new leafs. */
        do {
                did_inline = 0;
 
@@ -1364,7 +1011,7 @@ void inline_leave_functions(unsigned maxsize, unsigned leavesize,
                        int phiproj_computed = 0;
 
                        current_ir_graph = get_irp_irg(i);
-                       env              = get_irg_link(current_ir_graph);
+                       env              = (inline_irg_env*)get_irg_link(current_ir_graph);
 
                        ir_reserve_resources(current_ir_graph, IR_RESOURCE_IRN_LINK|IR_RESOURCE_PHI_LIST);
                        list_for_each_entry_safe(call_entry, entry, next, &env->calls, list) {
@@ -1382,8 +1029,8 @@ void inline_leave_functions(unsigned maxsize, unsigned leavesize,
                                        continue;
                                }
 
-                               if (is_leave(callee) && (
-                                   is_smaller(callee, leavesize) || prop >= irg_inline_forced)) {
+                               if (is_leaf(callee) && (
+                                   is_smaller(callee, leafsize) || prop >= irg_inline_forced)) {
                                        if (!phiproj_computed) {
                                                phiproj_computed = 1;
                                                collect_phiprojs(current_ir_graph);
@@ -1391,7 +1038,7 @@ void inline_leave_functions(unsigned maxsize, unsigned leavesize,
                                        did_inline = inline_method(call, callee);
 
                                        if (did_inline) {
-                                               inline_irg_env *callee_env = get_irg_link(callee);
+                                               inline_irg_env *callee_env = (inline_irg_env*)get_irg_link(callee);
 
                                                /* call was inlined, Phi/Projs for current graph must be recomputed */
                                                phiproj_computed = 0;
@@ -1418,7 +1065,7 @@ void inline_leave_functions(unsigned maxsize, unsigned leavesize,
                int phiproj_computed = 0;
 
                current_ir_graph = get_irp_irg(i);
-               env              = get_irg_link(current_ir_graph);
+               env              = (inline_irg_env*)get_irg_link(current_ir_graph);
 
                ir_reserve_resources(current_ir_graph, IR_RESOURCE_IRN_LINK|IR_RESOURCE_PHI_LIST);
 
@@ -1426,7 +1073,7 @@ void inline_leave_functions(unsigned maxsize, unsigned leavesize,
                list_for_each_entry_safe(call_entry, entry, next, &env->calls, list) {
                        irg_inline_property prop;
                        ir_graph            *callee;
-                       pmap_entry          *e;
+                       ir_graph            *calleee;
 
                        call   = entry->call;
                        callee = entry->callee;
@@ -1436,13 +1083,13 @@ void inline_leave_functions(unsigned maxsize, unsigned leavesize,
                                continue;
                        }
 
-                       e = pmap_find(copied_graphs, callee);
-                       if (e != NULL) {
+                       calleee = (ir_graph*)pmap_get(copied_graphs, callee);
+                       if (calleee != NULL) {
                                /*
                                 * Remap callee if we have a copy.
                                 * FIXME: Should we do this only for recursive Calls ?
                                 */
-                               callee = e->value;
+                               callee = calleee;
                        }
 
                        if (prop >= irg_inline_forced ||
@@ -1461,7 +1108,7 @@ void inline_leave_functions(unsigned maxsize, unsigned leavesize,
 
                                        /*
                                         * No copy yet, create one.
-                                        * Note that recursive methods are never leaves, so it is sufficient
+                                        * Note that recursive methods are never leafs, so it is sufficient
                                         * to test this condition here.
                                         */
                                        copy = create_irg_copy(callee);
@@ -1475,7 +1122,7 @@ void inline_leave_functions(unsigned maxsize, unsigned leavesize,
                                        callee_env = alloc_inline_irg_env();
                                        set_irg_link(copy, callee_env);
 
-                                       assure_cf_loop(copy);
+                                       assure_loopinfo(copy);
                                        wenv.x              = callee_env;
                                        wenv.ignore_callers = 1;
                                        irg_walk_graph(copy, NULL, collect_calls2, &wenv);
@@ -1505,7 +1152,7 @@ void inline_leave_functions(unsigned maxsize, unsigned leavesize,
                                        /* call was inlined, Phi/Projs for current graph must be recomputed */
                                        phiproj_computed = 0;
 
-                                       /* callee was inline. Append it's call list. */
+                                       /* callee was inline. Append its call list. */
                                        env->got_inline = 1;
                                        --env->n_call_nodes;
                                        append_call_list(env, callee_env, entry->loop_depth);
@@ -1514,7 +1161,7 @@ void inline_leave_functions(unsigned maxsize, unsigned leavesize,
                                        /* after we have inlined callee, all called methods inside callee
                                           are now called once more */
                                        list_for_each_entry(call_entry, centry, &callee_env->calls, list) {
-                                               inline_irg_env *penv = get_irg_link(centry->callee);
+                                               inline_irg_env *penv = (inline_irg_env*)get_irg_link(centry->callee);
                                                ++penv->n_callers;
                                        }
 
@@ -1529,7 +1176,7 @@ void inline_leave_functions(unsigned maxsize, unsigned leavesize,
 
        for (i = 0; i < n_irgs; ++i) {
                irg = get_irp_irg(i);
-               env = get_irg_link(irg);
+               env = (inline_irg_env*)get_irg_link(irg);
 
                if (env->got_inline) {
                        optimize_graph_df(irg);
@@ -1545,7 +1192,7 @@ void inline_leave_functions(unsigned maxsize, unsigned leavesize,
 
        /* kill the copied graphs: we don't need them anymore */
        foreach_pmap(copied_graphs, pm_entry) {
-               ir_graph *copy = pm_entry->value;
+               ir_graph *copy = (ir_graph*)pm_entry->value;
 
                /* reset the entity, otherwise it will be deleted in the next step ... */
                set_irg_entity(copy, NULL);
@@ -1557,45 +1204,44 @@ void inline_leave_functions(unsigned maxsize, unsigned leavesize,
        current_ir_graph = rem;
 }
 
-struct inline_leave_functions_pass_t {
+typedef struct inline_leaf_functions_pass_t {
        ir_prog_pass_t pass;
        unsigned       maxsize;
-       unsigned       leavesize;
+       unsigned       leafsize;
        unsigned       size;
        int            ignore_runtime;
-};
+} inline_leaf_functions_pass_t;
 
 /**
- * Wrapper to run inline_leave_functions() as a ir_prog pass.
+ * Wrapper to run inline_leaf_functions() as a ir_prog pass.
  */
-static int inline_leave_functions_wrapper(ir_prog *irp, void *context)
+static int inline_leaf_functions_wrapper(ir_prog *irp, void *context)
 {
-       struct inline_leave_functions_pass_t *pass = context;
+       inline_leaf_functions_pass_t *pass = (inline_leaf_functions_pass_t*)context;
 
        (void)irp;
-       inline_leave_functions(
-               pass->maxsize, pass->leavesize,
+       inline_leaf_functions(
+               pass->maxsize, pass->leafsize,
                pass->size, pass->ignore_runtime);
        return 0;
 }
 
-/* create a pass for inline_leave_functions() */
-ir_prog_pass_t *inline_leave_functions_pass(
-       const char *name, unsigned maxsize, unsigned leavesize,
+/* create a pass for inline_leaf_functions() */
+ir_prog_pass_t *inline_leaf_functions_pass(
+       const char *name, unsigned maxsize, unsigned leafsize,
        unsigned size, int ignore_runtime)
 {
-       struct inline_leave_functions_pass_t *pass =
-               XMALLOCZ(struct inline_leave_functions_pass_t);
+       inline_leaf_functions_pass_t *pass = XMALLOCZ(inline_leaf_functions_pass_t);
 
        pass->maxsize        = maxsize;
-       pass->leavesize      = leavesize;
+       pass->leafsize       = leafsize;
        pass->size           = size;
        pass->ignore_runtime = ignore_runtime;
 
        return def_prog_pass_constructor(
                &pass->pass,
-               name ? name : "inline_leave_functions",
-               inline_leave_functions_wrapper);
+               name ? name : "inline_leaf_functions",
+               inline_leaf_functions_wrapper);
 }
 
 /**
@@ -1670,7 +1316,9 @@ static void analyze_irg_local_weights(inline_irg_env *env, ir_graph *irg)
 {
        ir_entity *ent = get_irg_entity(irg);
        ir_type  *mtp;
-       int      nparams, i, proj_nr;
+       size_t   nparams;
+       int      i;
+       long     proj_nr;
        ir_node  *irg_args, *arg;
 
        mtp      = get_entity_type(ent);
@@ -1697,17 +1345,12 @@ static void analyze_irg_local_weights(inline_irg_env *env, ir_graph *irg)
  * After inlining, the local variable might be transformed into a
  * SSA variable by scalar_replacement().
  */
-static unsigned get_method_local_adress_weight(ir_graph *callee, int pos)
+static unsigned get_method_local_adress_weight(ir_graph *callee, size_t pos)
 {
-       inline_irg_env *env = get_irg_link(callee);
-
-       if (env->local_weights != NULL) {
-               if (pos < ARR_LEN(env->local_weights))
-                       return env->local_weights[pos];
-               return 0;
-       }
+       inline_irg_env *env = (inline_irg_env*)get_irg_link(callee);
 
-       analyze_irg_local_weights(env, callee);
+       if (env->local_weights == NULL)
+               analyze_irg_local_weights(env, callee);
 
        if (pos < ARR_LEN(env->local_weights))
                return env->local_weights[pos];
@@ -1724,10 +1367,12 @@ static int calc_inline_benefice(call_entry *entry, ir_graph *callee)
 {
        ir_node   *call = entry->call;
        ir_entity *ent  = get_irg_entity(callee);
+       ir_type   *callee_frame;
+       size_t    i, n_members, n_params;
        ir_node   *frame_ptr;
        ir_type   *mtp;
        int       weight = 0;
-       int       i, n_params, all_const;
+       int       all_const;
        unsigned  cc, v;
        irg_inline_property prop;
 
@@ -1740,6 +1385,18 @@ static int calc_inline_benefice(call_entry *entry, ir_graph *callee)
                return entry->benefice = INT_MIN;
        }
 
+       callee_frame = get_irg_frame_type(callee);
+       n_members = get_class_n_members(callee_frame);
+       for (i = 0; i < n_members; ++i) {
+               ir_entity *frame_ent = get_class_member(callee_frame, i);
+               if (is_parameter_entity(frame_ent)) {
+                       // TODO inliner should handle parameter entities by inserting Store operations
+                       DB((dbg, LEVEL_2, "In %+F Call to %+F: inlining forbidden due to parameter entity\n", call, callee));
+                       set_irg_inline_property(callee, irg_inline_forbidden);
+                       return entry->benefice = INT_MIN;
+               }
+       }
+
        if (get_irg_additional_properties(callee) & mtp_property_noreturn) {
                DB((dbg, LEVEL_2, "In %+F Call to %+F: not inlining noreturn or weak\n",
                    call, callee));
@@ -1752,7 +1409,7 @@ static int calc_inline_benefice(call_entry *entry, ir_graph *callee)
        cc       = get_method_calling_convention(mtp);
        if (cc & cc_reg_param) {
                /* register parameter, smaller costs for register parameters */
-               int max_regs = cc & ~cc_bits;
+               size_t max_regs = cc & ~cc_bits;
 
                if (max_regs < n_params)
                        weight += max_regs * 2 + (n_params - max_regs) * 5;
@@ -1790,7 +1447,7 @@ static int calc_inline_benefice(call_entry *entry, ir_graph *callee)
        }
        entry->all_const = all_const;
 
-       callee_env = get_irg_link(callee);
+       callee_env = (inline_irg_env*)get_irg_link(callee);
        if (callee_env->n_callers == 1 &&
            callee != current_ir_graph &&
            !entity_is_externally_visible(ent)) {
@@ -1805,7 +1462,7 @@ static int calc_inline_benefice(call_entry *entry, ir_graph *callee)
        if (callee_env->n_nodes < 30 && !callee_env->recursive)
                weight += 2000;
 
-       /* and finally for leaves: they do not increase the register pressure
+       /* and finally for leafs: they do not increase the register pressure
           because of callee safe registers */
        if (callee_env->n_call_nodes == 0)
                weight += 400;
@@ -1825,16 +1482,18 @@ static int calc_inline_benefice(call_entry *entry, ir_graph *callee)
        return entry->benefice = weight;
 }
 
-static ir_graph **irgs;
-static int      last_irg;
+typedef struct walk_env_t {
+       ir_graph **irgs;
+       size_t   last_irg;
+} walk_env_t;
 
 /**
  * Callgraph walker, collect all visited graphs.
  */
 static void callgraph_walker(ir_graph *irg, void *data)
 {
-       (void) data;
-       irgs[last_irg++] = irg;
+       walk_env_t *env = (walk_env_t *)data;
+       env->irgs[env->last_irg++] = irg;
 }
 
 /**
@@ -1844,22 +1503,24 @@ static void callgraph_walker(ir_graph *irg, void *data)
  */
 static ir_graph **create_irg_list(void)
 {
-       ir_entity **free_methods;
-       int       arr_len;
-       int       n_irgs = get_irp_n_irgs();
+       ir_entity  **free_methods;
+       size_t     n_irgs = get_irp_n_irgs();
+       walk_env_t env;
 
-       cgana(&arr_len, &free_methods);
+       cgana(&free_methods);
        xfree(free_methods);
 
        compute_callgraph();
 
-       last_irg = 0;
-       irgs     = XMALLOCNZ(ir_graph*, n_irgs);
+       env.irgs     = XMALLOCNZ(ir_graph*, n_irgs);
+       env.last_irg = 0;
 
-       callgraph_walk(NULL, callgraph_walker, NULL);
-       assert(n_irgs == last_irg);
+       callgraph_walk(NULL, callgraph_walker, &env);
+       assert(n_irgs == env.last_irg);
 
-       return irgs;
+       free_callgraph();
+
+       return env.irgs;
 }
 
 /**
@@ -1902,7 +1563,7 @@ static void inline_into(ir_graph *irg, unsigned maxsize,
                         int inline_threshold, pmap *copied_graphs)
 {
        int            phiproj_computed = 0;
-       inline_irg_env *env = get_irg_link(irg);
+       inline_irg_env *env = (inline_irg_env*)get_irg_link(irg);
        call_entry     *curr_call;
        wenv_t         wenv;
        pqueue_t       *pqueue;
@@ -1929,14 +1590,14 @@ static void inline_into(ir_graph *irg, unsigned maxsize,
        /* note that the list of possible calls is updated during the process */
        while (!pqueue_empty(pqueue)) {
                int                 did_inline;
-               call_entry          *curr_call  = pqueue_pop_front(pqueue);
+               call_entry          *curr_call  = (call_entry*)pqueue_pop_front(pqueue);
                ir_graph            *callee     = curr_call->callee;
                ir_node             *call_node  = curr_call->call;
-               inline_irg_env      *callee_env = get_irg_link(callee);
+               inline_irg_env      *callee_env = (inline_irg_env*)get_irg_link(callee);
                irg_inline_property prop        = get_irg_inline_property(callee);
+               ir_graph            *calleee;
                int                 loop_depth;
                const call_entry    *centry;
-               pmap_entry          *e;
 
                if ((prop < irg_inline_forced) && env->n_nodes + callee_env->n_nodes > maxsize) {
                        DB((dbg, LEVEL_2, "%+F: too big (%d) + %+F (%d)\n", irg,
@@ -1944,8 +1605,8 @@ static void inline_into(ir_graph *irg, unsigned maxsize,
                        continue;
                }
 
-               e = pmap_find(copied_graphs, callee);
-               if (e != NULL) {
+               calleee = (ir_graph*)pmap_get(copied_graphs, callee);
+               if (calleee != NULL) {
                        int benefice = curr_call->benefice;
                        /*
                         * Reduce the weight for recursive function IFF not all arguments are const.
@@ -1959,8 +1620,8 @@ static void inline_into(ir_graph *irg, unsigned maxsize,
                        /*
                         * Remap callee if we have a copy.
                         */
-                       callee     = e->value;
-                       callee_env = get_irg_link(callee);
+                       callee     = calleee;
+                       callee_env = (inline_irg_env*)get_irg_link(callee);
                }
 
                if (current_ir_graph == callee) {
@@ -1985,7 +1646,7 @@ static void inline_into(ir_graph *irg, unsigned maxsize,
 
                        /*
                         * No copy yet, create one.
-                        * Note that recursive methods are never leaves, so it is
+                        * Note that recursive methods are never leafs, so it is
                         * sufficient to test this condition here.
                         */
                        copy = create_irg_copy(callee);
@@ -1999,7 +1660,8 @@ static void inline_into(ir_graph *irg, unsigned maxsize,
                        callee_env = alloc_inline_irg_env();
                        set_irg_link(copy, callee_env);
 
-                       assure_cf_loop(copy);
+                       assure_loopinfo(copy);
+                       memset(&wenv, 0, sizeof(wenv));
                        wenv.x              = callee_env;
                        wenv.ignore_callers = 1;
                        irg_walk_graph(copy, NULL, collect_calls2, &wenv);
@@ -2032,14 +1694,14 @@ static void inline_into(ir_graph *irg, unsigned maxsize,
                /* remove it from the caller list */
                list_del(&curr_call->list);
 
-               /* callee was inline. Append it's call list. */
+               /* callee was inline. Append its call list. */
                env->got_inline = 1;
                --env->n_call_nodes;
 
                /* we just generate a bunch of new calls */
                loop_depth = curr_call->loop_depth;
                list_for_each_entry(call_entry, centry, &callee_env->calls, list) {
-                       inline_irg_env *penv = get_irg_link(centry->callee);
+                       inline_irg_env *penv = (inline_irg_env*)get_irg_link(centry->callee);
                        ir_node        *new_call;
                        call_entry     *new_entry;
 
@@ -2050,7 +1712,14 @@ static void inline_into(ir_graph *irg, unsigned maxsize,
                        /* Note that the src list points to Call nodes in the inlined graph,
                         * but we need Call nodes in our graph. Luckily the inliner leaves
                         * this information in the link field. */
-                       new_call = get_irn_link(centry->call);
+                       new_call = (ir_node*)get_irn_link(centry->call);
+                       if (get_irn_irg(new_call) != irg) {
+                               /* centry->call has not been copied, which means it is dead.
+                                * This might happen during inlining, if a const function,
+                                * which cannot be inlined is only used as an unused argument
+                                * of another function, which is inlined. */
+                               continue;
+                       }
                        assert(is_Call(new_call));
 
                        new_entry = duplicate_call_entry(centry, new_call, loop_depth);
@@ -2074,7 +1743,7 @@ void inline_functions(unsigned maxsize, int inline_threshold,
                       opt_ptr after_inline_opt)
 {
        inline_irg_env   *env;
-       int              i, n_irgs;
+       size_t           i, n_irgs;
        ir_graph         *rem;
        wenv_t           wenv;
        pmap             *copied_graphs;
@@ -2102,8 +1771,8 @@ void inline_functions(unsigned maxsize, int inline_threshold,
 
                free_callee_info(irg);
 
-               wenv.x = get_irg_link(irg);
-               assure_cf_loop(irg);
+               wenv.x = (inline_irg_env*)get_irg_link(irg);
+               assure_loopinfo(irg);
                irg_walk_graph(irg, NULL, collect_calls2, &wenv);
        }
 
@@ -2117,7 +1786,7 @@ void inline_functions(unsigned maxsize, int inline_threshold,
        for (i = 0; i < n_irgs; ++i) {
                ir_graph *irg = irgs[i];
 
-               env = get_irg_link(irg);
+               env = (inline_irg_env*)get_irg_link(irg);
                if (env->got_inline && after_inline_opt != NULL) {
                        /* this irg got calls inlined: optimize it */
                        after_inline_opt(irg);
@@ -2132,7 +1801,7 @@ void inline_functions(unsigned maxsize, int inline_threshold,
 
        /* kill the copied graphs: we don't need them anymore */
        foreach_pmap(copied_graphs, pm_entry) {
-               ir_graph *copy = pm_entry->value;
+               ir_graph *copy = (ir_graph*)pm_entry->value;
 
                /* reset the entity, otherwise it will be deleted in the next step ... */
                set_irg_entity(copy, NULL);
@@ -2146,19 +1815,19 @@ void inline_functions(unsigned maxsize, int inline_threshold,
        current_ir_graph = rem;
 }
 
-struct inline_functions_pass_t {
+typedef struct inline_functions_pass_t {
        ir_prog_pass_t pass;
        unsigned       maxsize;
        int            inline_threshold;
        opt_ptr        after_inline_opt;
-};
+} inline_functions_pass_t;
 
 /**
  * Wrapper to run inline_functions() as a ir_prog pass.
  */
 static int inline_functions_wrapper(ir_prog *irp, void *context)
 {
-       struct inline_functions_pass_t *pass = context;
+       inline_functions_pass_t *pass = (inline_functions_pass_t*)context;
 
        (void)irp;
        inline_functions(pass->maxsize, pass->inline_threshold,
@@ -2171,8 +1840,7 @@ ir_prog_pass_t *inline_functions_pass(
          const char *name, unsigned maxsize, int inline_threshold,
          opt_ptr after_inline_opt)
 {
-       struct inline_functions_pass_t *pass =
-               XMALLOCZ(struct inline_functions_pass_t);
+       inline_functions_pass_t *pass = XMALLOCZ(inline_functions_pass_t);
 
        pass->maxsize          = maxsize;
        pass->inline_threshold = inline_threshold;