+ * Copyright (C) 1995-2010 University of Karlsruhe. All right reserved.
+ *
+ * This file is part of libFirm.
+ *
+ * This file may be distributed and/or modified under the terms of the
+ * GNU General Public License version 2 as published by the Free Software
+ * Foundation and appearing in the file LICENSE.GPL included in the
+ * packaging of this file.
+ *
+ * Licensees holding valid libFirm Professional Edition licenses may use
+ * this file in accordance with the libFirm Commercial License.
+ * Agreement provided with the Software.
+ *
+ * This file is provided AS IS with NO WARRANTY OF ANY KIND, INCLUDING THE
+ * WARRANTY OF DESIGN, MERCHANTABILITY AND FITNESS FOR A PARTICULAR
+ * PURPOSE.
+ */
+
+/**
+ * @file
+ * @brief Statistics for Firm.
+ * @author Michael Beck
+ * @version $Id$
+ */
+#include "config.h"
+
+#ifdef FIRM_STATISTICS
+
+#include <stdio.h>
+#include <stdlib.h>
+#include <string.h>
+
+#include "irouts.h"
+#include "irdump.h"
+#include "hashptr.h"
+#include "firmstat_t.h"
+#include "irpass_t.h"
+#include "pattern.h"
+#include "dags.h"
+#include "stat_dmp.h"
+#include "xmalloc.h"
+#include "irhooks.h"
+#include "util.h"
+
+/*
+ * need this to be static:
+ * Special pseudo Opcodes that we need to count some interesting cases
+ */
+
+/**
+ * The Phi0, a node that is created during SSA construction
+ */
+static ir_op _op_Phi0;
+
+/** The PhiM, just to count memory Phi's. */
+static ir_op _op_PhiM;
+
+/** The Mul by Const node. */
+static ir_op _op_MulC;
+
+/** The Div by Const node. */
+static ir_op _op_DivC;
+
+/** The Div by Const node. */
+static ir_op _op_ModC;
+
+/** The memory Proj node. */
+static ir_op _op_ProjM;
+
+/** A Sel of a Sel */
+static ir_op _op_SelSel;
+
+/** A Sel of a Sel of a Sel */
+static ir_op _op_SelSelSel;
+
+/* ---------------------------------------------------------------------------------- */
+
+/** Marks the begin of a statistic (hook) function. */
+#define STAT_ENTER ++status->recursive
+
+/** Marks the end of a statistic (hook) functions. */
+#define STAT_LEAVE --status->recursive
+
+/** Allows to enter a statistic function only when we are not already in a hook. */
+#define STAT_ENTER_SINGLE do { if (status->recursive > 0) return; ++status->recursive; } while (0)
+
+/**
+ * global status
+ */
+static const unsigned status_disable = 0;
+static stat_info_t *status = (stat_info_t *)&status_disable;
+
+/**
+ * Compare two elements of the opcode hash.
+ */
+static int opcode_cmp(const void *elt, const void *key)
+{
+ const node_entry_t *e1 = (const node_entry_t*)elt;
+ const node_entry_t *e2 = (const node_entry_t*)key;
+
+ return e1->op->code - e2->op->code;
+} /* opcode_cmp */
+
+/**
+ * Compare two elements of the graph hash.
+ */
+static int graph_cmp(const void *elt, const void *key)
+{
+ const graph_entry_t *e1 = (const graph_entry_t*)elt;
+ const graph_entry_t *e2 = (const graph_entry_t*)key;
+
+ return e1->irg != e2->irg;
+} /* graph_cmp */
+
+/**
+ * Compare two elements of the optimization hash.
+ */
+static int opt_cmp(const void *elt, const void *key)
+{
+ const opt_entry_t *e1 = (const opt_entry_t*)elt;
+ const opt_entry_t *e2 = (const opt_entry_t*)key;
+
+ return e1->op->code != e2->op->code;
+} /* opt_cmp */
+
+/**
+ * Compare two elements of the block/extbb hash.
+ */
+static int block_cmp(const void *elt, const void *key)
+{
+ const block_entry_t *e1 = (const block_entry_t*)elt;
+ const block_entry_t *e2 = (const block_entry_t*)key;
+
+ /* it's enough to compare the block number */
+ return e1->block_nr != e2->block_nr;
+} /* block_cmp */
+
+/**
+ * Compare two elements of the be_block hash.
+ */
+static int be_block_cmp(const void *elt, const void *key)
+{
+ const be_block_entry_t *e1 = (const be_block_entry_t*)elt;
+ const be_block_entry_t *e2 = (const be_block_entry_t*)key;
+
+ return e1->block_nr != e2->block_nr;
+} /* be_block_cmp */
+
+/**
+ * Compare two elements of reg pressure hash.
+ */
+static int reg_pressure_cmp(const void *elt, const void *key)
+{
+ const reg_pressure_entry_t *e1 = (const reg_pressure_entry_t*)elt;
+ const reg_pressure_entry_t *e2 = (const reg_pressure_entry_t*)key;
+
+ return e1->class_name != e2->class_name;
+} /* reg_pressure_cmp */
+
+/**
+ * Compare two elements of the perm_stat hash.
+ */
+static int perm_stat_cmp(const void *elt, const void *key)
+{
+ const perm_stat_entry_t *e1 = (const perm_stat_entry_t*)elt;
+ const perm_stat_entry_t *e2 = (const perm_stat_entry_t*)key;
+
+ return e1->perm != e2->perm;
+} /* perm_stat_cmp */
+
+/**
+ * Compare two elements of the perm_class hash.
+ */
+static int perm_class_cmp(const void *elt, const void *key)
+{
+ const perm_class_entry_t *e1 = (const perm_class_entry_t*)elt;
+ const perm_class_entry_t *e2 = (const perm_class_entry_t*)key;
+
+ return e1->class_name != e2->class_name;
+} /* perm_class_cmp */
+
+/**
+ * Compare two elements of the ir_op hash.
+ */
+static int opcode_cmp_2(const void *elt, const void *key)
+{
+ const ir_op *e1 = (const ir_op*)elt;
+ const ir_op *e2 = (const ir_op*)key;
+
+ return e1->code != e2->code;
+} /* opcode_cmp_2 */
+
+/**
+ * Compare two elements of the address_mark set.
+ */
+static int address_mark_cmp(const void *elt, const void *key, size_t size)
+{
+ const address_mark_entry_t *e1 = (const address_mark_entry_t*)elt;
+ const address_mark_entry_t *e2 = (const address_mark_entry_t*)key;
+ (void) size;
+
+ /* compare only the nodes, the rest is used as data container */
+ return e1->node != e2->node;
+} /* address_mark_cmp */
+
+/**
+ * Clear all counter in a node_entry_t.
+ */
+static void opcode_clear_entry(node_entry_t *elem)
+{
+ cnt_clr(&elem->cnt_alive);
+ cnt_clr(&elem->new_node);
+ cnt_clr(&elem->into_Id);
+ cnt_clr(&elem->normalized);
+} /* opcode_clear_entry */
+
+/**
+ * Returns the associates node_entry_t for an ir_op (and allocates
+ * one if not yet available).
+ *
+ * @param op the IR operation
+ * @param hmap a hash map containing ir_op* -> node_entry_t*
+ */
+static node_entry_t *opcode_get_entry(const ir_op *op, hmap_node_entry_t *hmap)
+{
+ node_entry_t key;
+ node_entry_t *elem;
+
+ key.op = op;
+
+ elem = (node_entry_t*)pset_find(hmap, &key, op->code);
+ if (elem)
+ return elem;
+
+ elem = OALLOCZ(&status->cnts, node_entry_t);
+
+ /* clear counter */
+ opcode_clear_entry(elem);
+
+ elem->op = op;
+
+ return (node_entry_t*)pset_insert(hmap, elem, op->code);
+} /* opcode_get_entry */
+
+/**
+ * Returns the associates ir_op for an opcode
+ *
+ * @param code the IR opcode
+ * @param hmap the hash map containing opcode -> ir_op*
+ */
+static ir_op *opcode_find_entry(ir_opcode code, hmap_ir_op *hmap)
+{
+ ir_op key;
+
+ key.code = code;
+ return (ir_op*)pset_find(hmap, &key, code);
+} /* opcode_find_entry */
+
+/**
+ * Clears all counter in a graph_entry_t.
+ *
+ * @param elem the graph entry
+ * @param all if non-zero, clears all counters, else leave accumulated ones
+ */
+static void graph_clear_entry(graph_entry_t *elem, int all)
+{
+ int i;
+
+ /* clear accumulated / non-accumulated counter */
+ for (i = all ? 0 : _gcnt_non_acc; i < _gcnt_last; ++i) {
+ cnt_clr(&elem->cnt[i]);
+ } /* for */
+
+ if (elem->block_hash) {
+ del_pset(elem->block_hash);
+ elem->block_hash = NULL;
+ } /* if */
+
+ if (elem->extbb_hash) {
+ del_pset(elem->extbb_hash);
+ elem->extbb_hash = NULL;
+ } /* if */
+
+ obstack_free(&elem->recalc_cnts, NULL);
+ obstack_init(&elem->recalc_cnts);
+} /* graph_clear_entry */
+
+/**
+ * Returns the associated graph_entry_t for an IR graph.
+ *
+ * @param irg the IR graph, NULL for the global counter
+ * @param hmap the hash map containing ir_graph* -> graph_entry_t*
+ */
+static graph_entry_t *graph_get_entry(ir_graph *irg, hmap_graph_entry_t *hmap)
+{
+ graph_entry_t key;
+ graph_entry_t *elem;
+ size_t i;
+
+ key.irg = irg;
+
+ elem = (graph_entry_t*)pset_find(hmap, &key, HASH_PTR(irg));
+
+ if (elem) {
+ /* create hash map backend block information */
+ if (! elem->be_block_hash)
+ elem->be_block_hash = new_pset(be_block_cmp, 5);
+
+ return elem;
+ } /* if */
+
+ /* allocate a new one */
+ elem = OALLOCZ(&status->cnts, graph_entry_t);
+ obstack_init(&elem->recalc_cnts);
+
+ /* clear counter */
+ graph_clear_entry(elem, 1);
+
+ /* new hash table for opcodes here */
+ elem->opcode_hash = new_pset(opcode_cmp, 5);
+ elem->address_mark = new_set(address_mark_cmp, 5);
+ elem->irg = irg;
+
+ /* these hash tables are created on demand */
+ elem->block_hash = NULL;
+ elem->extbb_hash = NULL;
+
+ for (i = 0; i < sizeof(elem->opt_hash)/sizeof(elem->opt_hash[0]); ++i)
+ elem->opt_hash[i] = new_pset(opt_cmp, 4);
+
+ return (graph_entry_t*)pset_insert(hmap, elem, HASH_PTR(irg));
+} /* graph_get_entry */
+
+/**
+ * Clear all counter in an opt_entry_t.
+ */
+static void opt_clear_entry(opt_entry_t *elem)
+{
+ cnt_clr(&elem->count);
+} /* opt_clear_entry */
+
+/**
+ * Returns the associated opt_entry_t for an IR operation.
+ *
+ * @param op the IR operation
+ * @param hmap the hash map containing ir_op* -> opt_entry_t*
+ */
+static opt_entry_t *opt_get_entry(const ir_op *op, hmap_opt_entry_t *hmap)
+{
+ opt_entry_t key;
+ opt_entry_t *elem;
+
+ key.op = op;
+
+ elem = (opt_entry_t*)pset_find(hmap, &key, op->code);
+ if (elem)
+ return elem;
+
+ elem = OALLOCZ(&status->cnts, opt_entry_t);
+
+ /* clear new counter */
+ opt_clear_entry(elem);
+
+ elem->op = op;
+
+ return (opt_entry_t*)pset_insert(hmap, elem, op->code);
+} /* opt_get_entry */
+
+/**
+ * clears all counter in a block_entry_t
+ */
+static void block_clear_entry(block_entry_t *elem)
+{
+ int i;
+
+ for (i = 0; i < _bcnt_last; ++i)
+ cnt_clr(&elem->cnt[i]);
+} /* block_clear_entry */
+
+/**
+ * Returns the associated block_entry_t for an block.
+ *
+ * @param block_nr an IR block number
+ * @param hmap a hash map containing long -> block_entry_t
+ */
+static block_entry_t *block_get_entry(struct obstack *obst, long block_nr, hmap_block_entry_t *hmap)
+{
+ block_entry_t key;
+ block_entry_t *elem;
+
+ key.block_nr = block_nr;
+
+ elem = (block_entry_t*)pset_find(hmap, &key, block_nr);
+ if (elem)
+ return elem;
+
+ elem = OALLOCZ(obst, block_entry_t);
+
+ /* clear new counter */
+ block_clear_entry(elem);
+
+ elem->block_nr = block_nr;
+
+ return (block_entry_t*)pset_insert(hmap, elem, block_nr);
+} /* block_get_entry */
+
+/**
+ * Clear all sets in be_block_entry_t.
+ */
+static void be_block_clear_entry(be_block_entry_t *elem)
+{
+ if (elem->reg_pressure)
+ del_pset(elem->reg_pressure);
+
+ if (elem->sched_ready)
+ stat_delete_distrib_tbl(elem->sched_ready);
+
+ if (elem->perm_class_stat)
+ del_pset(elem->perm_class_stat);
+
+ elem->reg_pressure = new_pset(reg_pressure_cmp, 5);
+ elem->sched_ready = stat_new_int_distrib_tbl();
+ elem->perm_class_stat = new_pset(perm_class_cmp, 5);
+} /* be_block_clear_entry */
+
+/**
+ * Returns the associated be_block_entry_t for an block.
+ *
+ * @param block_nr an IR block number
+ * @param hmap a hash map containing long -> be_block_entry_t
+ */
+static be_block_entry_t *be_block_get_entry(struct obstack *obst, long block_nr, hmap_be_block_entry_t *hmap)
+{
+ be_block_entry_t key;
+ be_block_entry_t *elem;
+
+ key.block_nr = block_nr;
+
+ elem = (be_block_entry_t*)pset_find(hmap, &key, block_nr);
+ if (elem)
+ return elem;
+
+ elem = OALLOCZ(obst, be_block_entry_t);
+
+ /* clear new counter */
+ be_block_clear_entry(elem);
+
+ elem->block_nr = block_nr;
+
+ return (be_block_entry_t*)pset_insert(hmap, elem, block_nr);
+} /* be_block_get_entry */
+
+/**
+ * clears all sets in perm_class_entry_t
+ */
+static void perm_class_clear_entry(perm_class_entry_t *elem)
+{
+ if (elem->perm_stat)
+ del_pset(elem->perm_stat);
+
+ elem->perm_stat = new_pset(perm_stat_cmp, 5);
+} /* perm_class_clear_entry */
+
+/**
+ * Returns the associated perm_class entry for a register class.
+ *
+ * @param class_name the register class name
+ * @param hmap a hash map containing class_name -> perm_class_entry_t
+ */
+static perm_class_entry_t *perm_class_get_entry(struct obstack *obst, const char *class_name,
+ hmap_perm_class_entry_t *hmap)
+{
+ perm_class_entry_t key;
+ perm_class_entry_t *elem;
+
+ key.class_name = class_name;
+
+ elem = (perm_class_entry_t*)pset_find(hmap, &key, HASH_PTR(class_name));
+ if (elem)
+ return elem;
+
+ elem = OALLOCZ(obst, perm_class_entry_t);
+
+ /* clear new counter */
+ perm_class_clear_entry(elem);
+
+ elem->class_name = class_name;
+
+ return (perm_class_entry_t*)pset_insert(hmap, elem, HASH_PTR(class_name));
+} /* perm_class_get_entry */
+
+/**
+ * clears all sets in perm_stat_entry_t
+ */
+static void perm_stat_clear_entry(perm_stat_entry_t *elem)
+{
+ if (elem->chains)
+ stat_delete_distrib_tbl(elem->chains);
+
+ if (elem->cycles)
+ stat_delete_distrib_tbl(elem->cycles);
+
+ elem->chains = stat_new_int_distrib_tbl();
+ elem->cycles = stat_new_int_distrib_tbl();
+} /* perm_stat_clear_entry */
+
+/**
+ * Returns the associated perm_stat entry for a perm.
+ *
+ * @param perm the perm node
+ * @param hmap a hash map containing perm -> perm_stat_entry_t
+ */
+static perm_stat_entry_t *perm_stat_get_entry(struct obstack *obst, ir_node *perm, hmap_perm_stat_entry_t *hmap)
+{
+ perm_stat_entry_t key;
+ perm_stat_entry_t *elem;
+
+ key.perm = perm;
+
+ elem = (perm_stat_entry_t*)pset_find(hmap, &key, HASH_PTR(perm));
+ if (elem)
+ return elem;
+
+ elem = OALLOCZ(obst, perm_stat_entry_t);
+
+ /* clear new counter */
+ perm_stat_clear_entry(elem);
+
+ elem->perm = perm;
+
+ return (perm_stat_entry_t*)pset_insert(hmap, elem, HASH_PTR(perm));
+} /* perm_stat_get_entry */
+
+/**
+ * Clear optimizations counter,
+ */
+static void clear_optimization_counter(void)
+{
+ int i;
+ for (i = 0; i < FS_OPT_MAX; ++i)
+ cnt_clr(&status->num_opts[i]);
+}
+
+/**
+ * Returns the ir_op for an IR-node,
+ * handles special cases and return pseudo op codes.
+ *
+ * @param none an IR node
+ */
+static ir_op *stat_get_irn_op(ir_node *node)
+{
+ ir_op *op = get_irn_op(node);
+ unsigned opc = op->code;
+
+ switch (opc) {
+ case iro_Phi:
+ if (get_irn_arity(node) == 0) {
+ /* special case, a Phi0 node, count on extra counter */
+ op = status->op_Phi0 ? status->op_Phi0 : op;
+ } else if (get_irn_mode(node) == mode_M) {
+ /* special case, a Memory Phi node, count on extra counter */
+ op = status->op_PhiM ? status->op_PhiM : op;
+ } /* if */
+ break;
+ case iro_Proj:
+ if (get_irn_mode(node) == mode_M) {
+ /* special case, a Memory Proj node, count on extra counter */
+ op = status->op_ProjM ? status->op_ProjM : op;
+ } /* if */
+ break;
+ case iro_Mul:
+ if (is_Const(get_Mul_left(node)) || is_Const(get_Mul_right(node))) {
+ /* special case, a Multiply by a const, count on extra counter */
+ op = status->op_MulC ? status->op_MulC : op;
+ } /* if */
+ break;
+ case iro_Div:
+ if (is_Const(get_Div_right(node))) {
+ /* special case, a division by a const, count on extra counter */
+ op = status->op_DivC ? status->op_DivC : op;
+ } /* if */
+ break;
+ case iro_Mod:
+ if (is_Const(get_Mod_right(node))) {
+ /* special case, a module by a const, count on extra counter */
+ op = status->op_ModC ? status->op_ModC : op;
+ } /* if */
+ break;
+ case iro_Sel:
+ if (is_Sel(get_Sel_ptr(node))) {
+ /* special case, a Sel of a Sel, count on extra counter */
+ op = status->op_SelSel ? status->op_SelSel : op;
+ if (is_Sel(get_Sel_ptr(get_Sel_ptr(node)))) {
+ /* special case, a Sel of a Sel of a Sel, count on extra counter */
+ op = status->op_SelSelSel ? status->op_SelSelSel : op;
+ } /* if */
+ } /* if */
+ break;
+ default:
+ ;
+ } /* switch */
+
+ return op;
+} /* stat_get_irn_op */
+
+/**
+ * update the block counter
+ */
+static void undate_block_info(ir_node *node, graph_entry_t *graph)
+{
+ ir_op *op = get_irn_op(node);
+ ir_node *block;
+ block_entry_t *b_entry;
+ int i, arity;
+
+ /* check for block */
+ if (op == op_Block) {
+ arity = get_irn_arity(node);
+ b_entry = block_get_entry(&graph->recalc_cnts, get_irn_node_nr(node), graph->block_hash);
+ /* mark start end block to allow to filter them out */
+ if (node == get_irg_start_block(graph->irg))
+ b_entry->is_start = 1;
+ else if (node == get_irg_end_block(graph->irg))
+ b_entry->is_end = 1;
+
+ /* count all incoming edges */
+ for (i = 0; i < arity; ++i) {
+ ir_node *pred = get_irn_n(node, i);
+ ir_node *other_block = get_nodes_block(pred);
+ block_entry_t *b_entry_other = block_get_entry(&graph->recalc_cnts, get_irn_node_nr(other_block), graph->block_hash);
+
+ cnt_inc(&b_entry->cnt[bcnt_in_edges]); /* an edge coming from another block */
+ cnt_inc(&b_entry_other->cnt[bcnt_out_edges]);
+ } /* for */
+ return;
+ } /* if */
+
+ block = get_nodes_block(node);
+ b_entry = block_get_entry(&graph->recalc_cnts, get_irn_node_nr(block), graph->block_hash);
+
+ if (op == op_Phi && mode_is_datab(get_irn_mode(node))) {
+ /* count data Phi per block */
+ cnt_inc(&b_entry->cnt[bcnt_phi_data]);
+ } /* if */
+
+ /* we have a new node in our block */
+ cnt_inc(&b_entry->cnt[bcnt_nodes]);
+
+ /* don't count keep-alive edges */
+ if (is_End(node))
+ return;
+
+ arity = get_irn_arity(node);
+
+ for (i = 0; i < arity; ++i) {
+ ir_node *pred = get_irn_n(node, i);
+ ir_node *other_block;
+
+ other_block = get_nodes_block(pred);
+
+ if (other_block == block)
+ cnt_inc(&b_entry->cnt[bcnt_edges]); /* a in block edge */
+ else {
+ block_entry_t *b_entry_other = block_get_entry(&graph->recalc_cnts, get_irn_node_nr(other_block), graph->block_hash);
+
+ cnt_inc(&b_entry->cnt[bcnt_in_edges]); /* an edge coming from another block */
+ cnt_inc(&b_entry_other->cnt[bcnt_out_edges]);
+ } /* if */
+ } /* for */
+} /* undate_block_info */
+
+/**
+ * Update the extended block counter.
+ */
+static void update_extbb_info(ir_node *node, graph_entry_t *graph)
+{
+ ir_op *op = get_irn_op(node);
+ ir_extblk *extbb;
+ extbb_entry_t *eb_entry;
+ int i, arity;
+
+ /* check for block */
+ if (op == op_Block) {
+ extbb = get_nodes_extbb(node);
+ arity = get_irn_arity(node);
+ eb_entry = block_get_entry(&graph->recalc_cnts, get_extbb_node_nr(extbb), graph->extbb_hash);
+
+ /* count all incoming edges */
+ for (i = 0; i < arity; ++i) {
+ ir_node *pred = get_irn_n(node, i);
+ ir_extblk *other_extbb = get_nodes_extbb(pred);
+
+ if (extbb != other_extbb) {
+ extbb_entry_t *eb_entry_other = block_get_entry(&graph->recalc_cnts, get_extbb_node_nr(other_extbb), graph->extbb_hash);
+
+ cnt_inc(&eb_entry->cnt[bcnt_in_edges]); /* an edge coming from another extbb */
+ cnt_inc(&eb_entry_other->cnt[bcnt_out_edges]);
+ } /* if */
+ } /* for */
+ return;
+ } /* if */
+
+ extbb = get_nodes_extbb(node);
+ eb_entry = block_get_entry(&graph->recalc_cnts, get_extbb_node_nr(extbb), graph->extbb_hash);
+
+ if (op == op_Phi && mode_is_datab(get_irn_mode(node))) {
+ /* count data Phi per extbb */
+ cnt_inc(&eb_entry->cnt[bcnt_phi_data]);
+ } /* if */
+
+ /* we have a new node in our block */
+ cnt_inc(&eb_entry->cnt[bcnt_nodes]);
+
+ /* don't count keep-alive edges */
+ if (is_End(node))
+ return;
+
+ arity = get_irn_arity(node);
+
+ for (i = 0; i < arity; ++i) {
+ ir_node *pred = get_irn_n(node, i);
+ ir_extblk *other_extbb = get_nodes_extbb(pred);
+
+ if (other_extbb == extbb)
+ cnt_inc(&eb_entry->cnt[bcnt_edges]); /* a in extbb edge */
+ else {
+ extbb_entry_t *eb_entry_other = block_get_entry(&graph->recalc_cnts, get_extbb_node_nr(other_extbb), graph->extbb_hash);
+
+ cnt_inc(&eb_entry->cnt[bcnt_in_edges]); /* an edge coming from another extbb */
+ cnt_inc(&eb_entry_other->cnt[bcnt_out_edges]);
+ } /* if */
+ } /* for */
+} /* update_extbb_info */
+
+/**
+ * Calculates how many arguments of the call are const, updates
+ * param distribution.
+ */
+static void analyse_params_of_Call(graph_entry_t *graph, ir_node *call)
+{
+ int i, num_const_args = 0, num_local_adr = 0;
+ int n = get_Call_n_params(call);
+
+ for (i = 0; i < n; ++i) {
+ ir_node *param = get_Call_param(call, i);
+
+ if (is_irn_constlike(param))
+ ++num_const_args;
+ else if (is_Sel(param)) {
+ ir_node *base = param;
+
+ do {
+ base = get_Sel_ptr(base);
+ } while (is_Sel(base));
+
+ if (base == get_irg_frame(current_ir_graph))
+ ++num_local_adr;
+ }
+
+ } /* for */
+
+ if (num_const_args > 0)
+ cnt_inc(&graph->cnt[gcnt_call_with_cnst_arg]);
+ if (num_const_args == n)
+ cnt_inc(&graph->cnt[gcnt_call_with_all_cnst_arg]);
+ if (num_local_adr > 0)
+ cnt_inc(&graph->cnt[gcnt_call_with_local_adr]);
+
+ stat_inc_int_distrib_tbl(status->dist_param_cnt, n);
+} /* analyse_params_of_Call */
+
+/**
+ * Update info on calls.
+ *
+ * @param call The call
+ * @param graph The graph entry containing the call
+ */
+static void stat_update_call(ir_node *call, graph_entry_t *graph)
+{
+ ir_node *block = get_nodes_block(call);
+ ir_node *ptr = get_Call_ptr(call);
+ ir_entity *ent = NULL;
+ ir_graph *callee = NULL;
+
+ /*
+ * If the block is bad, the whole subgraph will collapse later
+ * so do not count this call.
+ * This happens in dead code.
+ */
+ if (is_Bad(block))
+ return;
+
+ cnt_inc(&graph->cnt[gcnt_all_calls]);
+
+ /* found a call, this function is not a leaf */
+ graph->is_leaf = 0;
+
+ if (is_SymConst(ptr)) {
+ if (get_SymConst_kind(ptr) == symconst_addr_ent) {
+ /* ok, we seems to know the entity */
+ ent = get_SymConst_entity(ptr);
+ callee = get_entity_irg(ent);
+
+ /* it is recursive, if it calls at least once */
+ if (callee == graph->irg)
+ graph->is_recursive = 1;
+ if (callee == NULL)
+ cnt_inc(&graph->cnt[gcnt_external_calls]);
+ } /* if */
+ } else {
+ /* indirect call, be could not predict */
+ cnt_inc(&graph->cnt[gcnt_indirect_calls]);
+
+ /* NOT a leaf call */
+ graph->is_leaf_call = LCS_NON_LEAF_CALL;
+ } /* if */
+
+ /* check, if it's a chain-call: Then, the call-block
+ * must dominate the end block. */
+ {
+ ir_node *curr = get_irg_end_block(graph->irg);
+ int depth = get_Block_dom_depth(block);
+
+ for (; curr != block && get_Block_dom_depth(curr) > depth;) {
+ curr = get_Block_idom(curr);
+
+ if (! curr || !is_Block(curr))
+ break;
+ } /* for */
+
+ if (curr != block)
+ graph->is_chain_call = 0;
+ }
+
+ /* check, if the callee is a leaf */
+ if (callee) {
+ graph_entry_t *called = graph_get_entry(callee, status->irg_hash);
+
+ if (called->is_analyzed) {
+ if (! called->is_leaf)
+ graph->is_leaf_call = LCS_NON_LEAF_CALL;
+ } /* if */
+ } /* if */
+
+ analyse_params_of_Call(graph, call);
+} /* stat_update_call */
+
+/**
+ * Update info on calls for graphs on the wait queue.
+ */
+static void stat_update_call_2(ir_node *call, graph_entry_t *graph)
+{
+ ir_node *block = get_nodes_block(call);
+ ir_node *ptr = get_Call_ptr(call);
+ ir_entity *ent = NULL;
+ ir_graph *callee = NULL;
+
+ /*
+ * If the block is bad, the whole subgraph will collapse later
+ * so do not count this call.
+ * This happens in dead code.
+ */
+ if (is_Bad(block))
+ return;
+
+ if (is_SymConst(ptr)) {
+ if (get_SymConst_kind(ptr) == symconst_addr_ent) {
+ /* ok, we seems to know the entity */
+ ent = get_SymConst_entity(ptr);
+ callee = get_entity_irg(ent);
+ } /* if */
+ } /* if */
+
+ /* check, if the callee is a leaf */
+ if (callee) {
+ graph_entry_t *called = graph_get_entry(callee, status->irg_hash);
+
+ assert(called->is_analyzed);
+
+ if (! called->is_leaf)
+ graph->is_leaf_call = LCS_NON_LEAF_CALL;
+ } else
+ graph->is_leaf_call = LCS_NON_LEAF_CALL;
+} /* stat_update_call_2 */
+
+/**
+ * Find the base address and entity of an Sel node.
+ *
+ * @param sel the node
+ *
+ * @return the base address.
+ */
+static ir_node *find_base_adr(ir_node *sel)
+{
+ ir_node *ptr = get_Sel_ptr(sel);
+
+ while (is_Sel(ptr)) {
+ sel = ptr;
+ ptr = get_Sel_ptr(sel);
+ }
+ return ptr;
+} /* find_base_adr */
+
+/**
+ * Update info on Load/Store address statistics.
+ */
+static void stat_update_address(ir_node *node, graph_entry_t *graph)
+{
+ unsigned opc = get_irn_opcode(node);
+ ir_node *base;
+ ir_graph *irg;
+
+ switch (opc) {
+ case iro_SymConst:
+ /* a global address */
+ cnt_inc(&graph->cnt[gcnt_global_adr]);
+ break;
+ case iro_Sel:
+ base = find_base_adr(node);
+ irg = current_ir_graph;
+ if (base == get_irg_frame(irg)) {
+ /* a local Variable. */
+ cnt_inc(&graph->cnt[gcnt_local_adr]);
+ } else {
+ /* Pointer access */
+ if (is_Proj(base) && skip_Proj(get_Proj_pred(base)) == get_irg_start(irg)) {
+ /* pointer access through parameter, check for THIS */
+ ir_entity *ent = get_irg_entity(irg);
+
+ if (ent != NULL) {
+ ir_type *ent_tp = get_entity_type(ent);
+
+ if (get_method_calling_convention(ent_tp) & cc_this_call) {
+ if (get_Proj_proj(base) == 0) {
+ /* THIS pointer */
+ cnt_inc(&graph->cnt[gcnt_this_adr]);
+ goto end_parameter;
+ } /* if */
+ } /* if */
+ } /* if */
+ /* other parameter */
+ cnt_inc(&graph->cnt[gcnt_param_adr]);
+end_parameter: ;
+ } else {
+ /* unknown Pointer access */
+ cnt_inc(&graph->cnt[gcnt_other_adr]);
+ } /* if */
+ } /* if */
+ default:
+ ;
+ } /* switch */
+} /* stat_update_address */
+
+/**
+ * Walker for reachable nodes count.
+ */
+static void update_node_stat(ir_node *node, void *env)
+{
+ graph_entry_t *graph = (graph_entry_t*)env;
+ node_entry_t *entry;
+
+ ir_op *op = stat_get_irn_op(node);
+ int i, arity = get_irn_arity(node);
+
+ entry = opcode_get_entry(op, graph->opcode_hash);
+
+ cnt_inc(&entry->cnt_alive);
+ cnt_add_i(&graph->cnt[gcnt_edges], arity);
+
+ /* count block edges */
+ undate_block_info(node, graph);
+
+ /* count extended block edges */
+ if (status->stat_options & FIRMSTAT_COUNT_EXTBB) {
+ if (graph->irg != get_const_code_irg())
+ update_extbb_info(node, graph);
+ } /* if */
+
+ /* handle statistics for special node types */
+
+ switch (op->code) {
+ case iro_Call:
+ /* check for properties that depends on calls like recursion/leaf/indirect call */
+ stat_update_call(node, graph);
+ break;
+ case iro_Load:
+ /* check address properties */
+ stat_update_address(get_Load_ptr(node), graph);
+ break;
+ case iro_Store:
+ /* check address properties */
+ stat_update_address(get_Store_ptr(node), graph);
+ break;
+ case iro_Phi:
+ /* check for non-strict Phi nodes */
+ for (i = arity - 1; i >= 0; --i) {
+ ir_node *pred = get_Phi_pred(node, i);
+ if (is_Unknown(pred)) {
+ /* found an Unknown predecessor, graph is not strict */
+ graph->is_strict = 0;
+ break;
+ }
+ }
+ default:
+ ;
+ } /* switch */
+
+ /* we want to count the constant IN nodes, not the CSE'ed constant's itself */
+ if (status->stat_options & FIRMSTAT_COUNT_CONSTS) {
+ int i;
+
+ for (i = get_irn_arity(node) - 1; i >= 0; --i) {
+ ir_node *pred = get_irn_n(node, i);
+
+ if (is_Const(pred)) {
+ /* check properties of constants */
+ stat_update_const(status, pred, graph);
+ } /* if */
+ } /* for */
+ } /* if */
+} /* update_node_stat */
+
+/**
+ * Walker for reachable nodes count for graphs on the wait_q.
+ */
+static void update_node_stat_2(ir_node *node, void *env)
+{
+ graph_entry_t *graph = (graph_entry_t*)env;
+
+ /* check for properties that depends on calls like recursion/leaf/indirect call */
+ if (is_Call(node))
+ stat_update_call_2(node, graph);
+} /* update_node_stat_2 */
+
+/**
+ * Get the current address mark.
+ */
+static unsigned get_adr_mark(graph_entry_t *graph, ir_node *node)
+{
+ address_mark_entry_t *value = (address_mark_entry_t*)set_find(graph->address_mark, &node, sizeof(*value), HASH_PTR(node));
+
+ return value ? value->mark : 0;
+} /* get_adr_mark */
+
+/**
+ * Set the current address mark.
+ */
+static void set_adr_mark(graph_entry_t *graph, ir_node *node, unsigned val)
+{
+ address_mark_entry_t *value = (address_mark_entry_t*)set_insert(graph->address_mark, &node, sizeof(*value), HASH_PTR(node));
+
+ value->mark = val;
+} /* set_adr_mark */
+
+#undef DUMP_ADR_MODE
+
+#ifdef DUMP_ADR_MODE
+/**
+ * a vcg attribute hook: Color a node with a different color if
+ * it's identified as a part of an address expression or at least referenced
+ * by an address expression.
+ */
+static int stat_adr_mark_hook(FILE *F, ir_node *node, ir_node *local)
+{
+ ir_node *n = local ? local : node;
+ ir_graph *irg = get_irn_irg(n);
+ graph_entry_t *graph = graph_get_entry(irg, status->irg_hash);
+ unsigned mark = get_adr_mark(graph, n);
+
+ if (mark & MARK_ADDRESS_CALC)
+ fprintf(F, "color: purple");
+ else if ((mark & (MARK_REF_ADR | MARK_REF_NON_ADR)) == MARK_REF_ADR)
+ fprintf(F, "color: pink");
+ else if ((mark & (MARK_REF_ADR | MARK_REF_NON_ADR)) == (MARK_REF_ADR|MARK_REF_NON_ADR))
+ fprintf(F, "color: lightblue");
+ else
+ return 0;
+
+ /* I know the color! */
+ return 1;
+} /* stat_adr_mark_hook */
+#endif /* DUMP_ADR_MODE */
+
+/**
+ * Return the "operational" mode of a Firm node.
+ */
+static ir_mode *get_irn_op_mode(ir_node *node)
+{
+ switch (get_irn_opcode(node)) {
+ case iro_Load:
+ return get_Load_mode(node);
+ case iro_Store:
+ return get_irn_mode(get_Store_value(node));
+ case iro_Div:
+ return get_irn_mode(get_Div_left(node));
+ case iro_Mod:
+ return get_irn_mode(get_Mod_left(node));
+ case iro_Cmp:
+ /* Cmp is no address calculation, or is it? */
+ default:
+ return get_irn_mode(node);
+ } /* switch */
+} /* get_irn_op_mode */
+
+/**
+ * Post-walker that marks every node that is an address calculation.
+ *
+ * Users of a node must be visited first. We ensure this by
+ * calling it in the post of an outs walk. This should work even in cycles,
+ * while the normal pre-walk will not.
+ */
+static void mark_address_calc(ir_node *node, void *env)
+{
+ graph_entry_t *graph = (graph_entry_t*)env;
+ ir_mode *mode = get_irn_op_mode(node);
+ int i, n;
+ unsigned mark_preds = MARK_REF_NON_ADR;
+
+ if (! mode_is_data(mode))
+ return;
+
+ if (mode_is_reference(mode)) {
+ /* a reference is calculated here, we are sure */
+ set_adr_mark(graph, node, MARK_ADDRESS_CALC);
+
+ mark_preds = MARK_REF_ADR;
+ } else {
+ unsigned mark = get_adr_mark(graph, node);
+
+ if ((mark & (MARK_REF_ADR | MARK_REF_NON_ADR)) == MARK_REF_ADR) {
+ /*
+ * this node has no reference mode, but is only
+ * referenced by address calculations
+ */
+ mark_preds = MARK_REF_ADR;
+ } /* if */
+ } /* if */
+
+ /* mark all predecessors */
+ for (i = 0, n = get_irn_arity(node); i < n; ++i) {
+ ir_node *pred = get_irn_n(node, i);
+
+ mode = get_irn_op_mode(pred);
+ if (! mode_is_data(mode))
+ continue;
+
+ set_adr_mark(graph, pred, get_adr_mark(graph, pred) | mark_preds);
+ } /* for */
+} /* mark_address_calc */
+
+/**
+ * Post-walker that marks every node that is an address calculation.
+ *
+ * Users of a node must be visited first. We ensure this by
+ * calling it in the post of an outs walk. This should work even in cycles,
+ * while the normal pre-walk will not.
+ */
+static void count_adr_ops(ir_node *node, void *env)
+{
+ graph_entry_t *graph = (graph_entry_t*)env;
+ unsigned mark = get_adr_mark(graph, node);
+
+ if (mark & MARK_ADDRESS_CALC)
+ cnt_inc(&graph->cnt[gcnt_pure_adr_ops]);
+ else if ((mark & (MARK_REF_ADR | MARK_REF_NON_ADR)) == MARK_REF_ADR)
+ cnt_inc(&graph->cnt[gcnt_pure_adr_ops]);
+ else if ((mark & (MARK_REF_ADR | MARK_REF_NON_ADR)) == (MARK_REF_ADR|MARK_REF_NON_ADR))
+ cnt_inc(&graph->cnt[gcnt_all_adr_ops]);
+} /* count_adr_ops */
+
+/**
+ * Called for every graph when the graph is either deleted or stat_dump_snapshot()
+ * is called, must recalculate all statistic info.
+ *
+ * @param global The global entry
+ * @param graph The current entry
+ */
+static void update_graph_stat(graph_entry_t *global, graph_entry_t *graph)
+{
+ node_entry_t *entry;
+ int i;
+
+ /* clear first the alive counter in the graph */
+ foreach_pset(graph->opcode_hash, node_entry_t*, entry) {
+ cnt_clr(&entry->cnt_alive);
+ } /* foreach_pset */
+
+ /* set pessimistic values */
+ graph->is_leaf = 1;
+ graph->is_leaf_call = LCS_UNKNOWN;
+ graph->is_recursive = 0;
+ graph->is_chain_call = 1;
+ graph->is_strict = 1;
+
+ /* create new block counter */
+ graph->block_hash = new_pset(block_cmp, 5);
+
+ /* we need dominator info */
+ if (graph->irg != get_const_code_irg()) {
+ assure_doms(graph->irg);
+
+ if (status->stat_options & FIRMSTAT_COUNT_EXTBB) {
+ /* we need extended basic blocks */
+ compute_extbb(graph->irg);
+
+ /* create new extbb counter */
+ graph->extbb_hash = new_pset(block_cmp, 5);
+ } /* if */
+ } /* if */
+
+ /* count the nodes in the graph */
+ irg_walk_graph(graph->irg, update_node_stat, NULL, graph);
+
+#if 0
+ /* Uncomment this code if chain-call means call exact one. */
+ entry = opcode_get_entry(op_Call, graph->opcode_hash);
+
+ /* check if we have more than 1 call */
+ if (cnt_gt(entry->cnt_alive, 1))
+ graph->is_chain_call = 0;