X-Git-Url: http://nsz.repo.hu/git/?a=blobdiff_plain;ds=sidebyside;f=ir%2Fana%2Firlivechk.c;h=c7dcf5172b89826df671a22baec4c1710f00bf1e;hb=63c3d0ec75a683d72bd00f830ea00f3cb0e76443;hp=973d9ac7e41e49162a48c9382c7d8c70a91c6a60;hpb=c2b87d6336e0fc8a5bad0dc43cf6fca6c8288421;p=libfirm diff --git a/ir/ana/irlivechk.c b/ir/ana/irlivechk.c index 973d9ac7e..c7dcf5172 100644 --- a/ir/ana/irlivechk.c +++ b/ir/ana/irlivechk.c @@ -22,7 +22,7 @@ * @date 21.04.2007 * @author Sebastian Hack * @version $Id$ - * @summary + * @brief * * Liveness checks as developed by Benoit Boissinot, Fabrice Rastello and myself. * @@ -37,9 +37,7 @@ * * The precomputation remains valid as long as the CFG is not altered. */ -#ifdef HAVE_CONFIG_H #include -#endif #include @@ -60,7 +58,7 @@ #include "statev.h" -typedef struct _bl_info_t { +typedef struct bl_info_t { const ir_node *block; /**< The block. */ int be_tgt_calc : 1; @@ -70,34 +68,33 @@ typedef struct _bl_info_t { bitset_t *red_reachable; /**< Holds all id's if blocks reachable in the CFG modulo back edges. */ - bitset_t *be_tgt_reach; /**< target blocks of back edges whose + bitset_t *be_tgt_reach; /**< target blocks of back edges whose sources are reachable from this block in the reduced graph. */ } bl_info_t; #define get_block_info(lv, bl) ((bl_info_t *) phase_get_irn_data(&(lv)->ph, bl)) -struct _lv_chk_t { - ir_phase ph; +struct lv_chk_t { + ir_phase ph; const dfs_t *dfs; + int n_blocks; + bitset_t *back_edge_src; + bitset_t *back_edge_tgt; + bl_info_t **map; DEBUG_ONLY(firm_dbg_module_t *dbg;) - int n_blocks; - bitset_t *back_edge_src; - bitset_t *back_edge_tgt; - bl_info_t **map; }; -static void *init_block_data(ir_phase *ph, const ir_node *irn, void *old) +static void *init_block_data(ir_phase *ph, const ir_node *irn) { - lv_chk_t *lv = container_of(ph, lv_chk_t, ph); - bl_info_t *bi = phase_alloc(ph, sizeof(bi[0])); + lv_chk_t *lv = firm_container_of(ph, lv_chk_t, ph); + bl_info_t *bi = (bl_info_t*) phase_alloc(ph, sizeof(bi[0])); bi->id = get_Block_dom_tree_pre_num(irn); bi->block = irn; bi->red_reachable = bitset_obstack_alloc(phase_obst(ph), lv->n_blocks); bi->be_tgt_reach = bitset_obstack_alloc(phase_obst(ph), lv->n_blocks); bi->be_tgt_calc = 0; - (void) old; return bi; } @@ -106,15 +103,16 @@ static void *init_block_data(ir_phase *ph, const ir_node *irn, void *old) * @param irn A node. * @return 1 if the node shall be considered in liveness, 0 if not. */ -static INLINE int is_liveness_node(const ir_node *irn) +static inline int is_liveness_node(const ir_node *irn) { - switch(get_irn_opcode(irn)) { + switch (get_irn_opcode(irn)) { case iro_Block: case iro_Bad: case iro_End: case iro_Anchor: return 0; - default:; + default: + break; } return 1; @@ -134,7 +132,7 @@ static void red_trans_closure(lv_chk_t *lv) int i, n; for (i = 0, n = dfs_get_n_nodes(lv->dfs); i < n; ++i) { - const ir_node *bl = dfs_get_post_num_node(lv->dfs, i); + const ir_node *bl = (const ir_node*) dfs_get_post_num_node(lv->dfs, i); bl_info_t *bi = get_block_info(lv, bl); const ir_edge_t *edge; @@ -170,7 +168,7 @@ static void compute_back_edge_chain(lv_chk_t *lv, const ir_node *bl) bitset_t *tmp = bitset_alloca(lv->n_blocks); bl_info_t *bi = get_block_info(lv, bl); - bitset_pos_t elm; + size_t elm; DBG((lv->dbg, LEVEL_2, "computing T_%d\n", bi->id)); @@ -205,9 +203,9 @@ static void compute_back_edge_chain(lv_chk_t *lv, const ir_node *bl) } -static INLINE void compute_back_edge_chains(lv_chk_t *lv) +static inline void compute_back_edge_chains(lv_chk_t *lv) { - bitset_pos_t elm; + size_t elm; int i, n; DBG((lv->dbg, LEVEL_2, "back edge sources: %B\n", lv->back_edge_src)); @@ -216,7 +214,7 @@ static INLINE void compute_back_edge_chains(lv_chk_t *lv) } for (i = 0, n = dfs_get_n_nodes(lv->dfs); i < n; ++i) { - const ir_node *bl = dfs_get_post_num_node(lv->dfs, i); + const ir_node *bl = (const ir_node*) dfs_get_post_num_node(lv->dfs, i); bl_info_t *bi = get_block_info(lv, bl); const ir_edge_t *edge; @@ -236,7 +234,7 @@ static INLINE void compute_back_edge_chains(lv_chk_t *lv) } for (i = 0, n = dfs_get_n_nodes(lv->dfs); i < n; ++i) { - const ir_node *bl = dfs_get_post_num_node(lv->dfs, i); + const ir_node *bl = (const ir_node*) dfs_get_post_num_node(lv->dfs, i); bl_info_t *bi = get_block_info(lv, bl); bitset_set(bi->be_tgt_reach, bi->id); } @@ -244,7 +242,7 @@ static INLINE void compute_back_edge_chains(lv_chk_t *lv) lv_chk_t *lv_chk_new(ir_graph *irg, const dfs_t *dfs) { - lv_chk_t *res = xmalloc(sizeof(res[0])); + lv_chk_t *res = XMALLOC(lv_chk_t); struct obstack *obst; int i; @@ -253,7 +251,7 @@ lv_chk_t *lv_chk_new(ir_graph *irg, const dfs_t *dfs) compute_doms(irg); stat_ev_tim_push(); - phase_init(&res->ph, "liveness check", irg, PHASE_DEFAULT_GROWTH, init_block_data, NULL); + phase_init(&res->ph, irg, init_block_data); obst = phase_obst(&res->ph); FIRM_DBG_REGISTER(res->dbg, "ir.ana.lvchk"); @@ -262,8 +260,7 @@ lv_chk_t *lv_chk_new(ir_graph *irg, const dfs_t *dfs) res->n_blocks = dfs_get_n_nodes(res->dfs); res->back_edge_src = bitset_obstack_alloc(obst, res->n_blocks); res->back_edge_tgt = bitset_obstack_alloc(obst, res->n_blocks); - res->map = obstack_alloc(obst, res->n_blocks * sizeof(res->map[0])); - memset(res->map, 0, res->n_blocks * sizeof(res->map[0])); + res->map = OALLOCNZ(obst, bl_info_t*, res->n_blocks); #if 0 { @@ -281,7 +278,7 @@ lv_chk_t *lv_chk_new(ir_graph *irg, const dfs_t *dfs) /* fill the map which maps pre_num to block infos */ for (i = res->n_blocks - 1; i >= 0; --i) { ir_node *irn = (ir_node *) dfs_get_pre_num_node(res->dfs, i); - bl_info_t *bi = phase_get_or_set_irn_data(&res->ph, irn); + bl_info_t *bi = (bl_info_t*) phase_get_or_set_irn_data(&res->ph, irn); assert(bi->id < res->n_blocks); assert(res->map[bi->id] == NULL); res->map[bi->id] = bi; @@ -296,7 +293,7 @@ lv_chk_t *lv_chk_new(ir_graph *irg, const dfs_t *dfs) #ifndef NDEBUG DBG((res->dbg, LEVEL_1, "liveness chk in %+F\n", irg)); for (i = res->n_blocks - 1; i >= 0; --i) { - const ir_node *irn = dfs_get_pre_num_node(res->dfs, i); + const ir_node *irn = (const ir_node*) dfs_get_pre_num_node(res->dfs, i); bl_info_t *bi = get_block_info(res, irn); DBG((res->dbg, LEVEL_1, "lv_chk for %d -> %+F\n", i, irn)); DBG((res->dbg, LEVEL_1, "\tred reach: %B\n", bi->red_reachable)); @@ -313,10 +310,11 @@ lv_chk_t *lv_chk_new(ir_graph *irg, const dfs_t *dfs) void lv_chk_free(lv_chk_t *lv) { - obstack_free(phase_obst(&lv->ph), NULL); + phase_deinit(&lv->ph); xfree(lv); } +#if 0 /** * Check if a node is live at the end of a block. * This function is for internal use as its code is shared between @@ -336,11 +334,11 @@ void lv_chk_free(lv_chk_t *lv) */ unsigned lv_chk_bl_in_mask(const lv_chk_t *lv, const ir_node *bl, const ir_node *var) { - stat_ev_cnt_decl(uses); - ir_node *def_bl; const ir_edge_t *edge; + stat_ev_cnt_decl(uses); + int res = 0; assert(is_Block(bl) && "can only check for liveness in a block"); @@ -416,11 +414,11 @@ end: unsigned lv_chk_bl_end_mask(const lv_chk_t *lv, const ir_node *bl, const ir_node *var) { - stat_ev_cnt_decl(uses); - ir_node *def_bl; const ir_edge_t *edge; + stat_ev_cnt_decl(uses); + int res = 0; assert(is_Block(bl) && "can only check for liveness in a block"); @@ -431,9 +429,7 @@ unsigned lv_chk_bl_end_mask(const lv_chk_t *lv, const ir_node *bl, const ir_node def_bl = get_nodes_block(var); if (!block_dominates(def_bl, bl)) { goto end; - } - - else { + } else { bitset_t *uses = bitset_alloca(lv->n_blocks); bitset_t *tmp = bitset_alloca(lv->n_blocks); int min_dom = get_Block_dom_tree_pre_num(def_bl) + 1; @@ -488,6 +484,7 @@ unsigned lv_chk_bl_end_mask(const lv_chk_t *lv, const ir_node *bl, const ir_node end: return res; } +#endif /** * Check a nodes liveness situation of a block. @@ -500,11 +497,10 @@ end: */ unsigned lv_chk_bl_xxx(const lv_chk_t *lv, const ir_node *bl, const ir_node *var) { - stat_ev_cnt_decl(uses); - stat_ev_cnt_decl(iter); - int res = 0; ir_node *def_bl; + stat_ev_cnt_decl(uses); + stat_ev_cnt_decl(iter); assert(is_Block(bl) && "can only check for liveness in a block"); @@ -564,7 +560,7 @@ unsigned lv_chk_bl_xxx(const lv_chk_t *lv, const ir_node *bl, const ir_node *var * We try to gather as much information as possible during looking * at the uses. * - * Note that we know for shure that bl != def_bl. That is sometimes + * Note that we know for sure that bl != def_bl. That is sometimes * silently exploited below. */ else { @@ -573,7 +569,8 @@ unsigned lv_chk_bl_xxx(const lv_chk_t *lv, const ir_node *bl, const ir_node *var bitset_t *uses = bitset_alloca(lv->n_blocks); bitset_t *Tq; - unsigned i, min_dom, max_dom; + size_t i; + unsigned min_dom, max_dom; const ir_edge_t *edge; /* if the block has no DFS info, it cannot be reached. @@ -647,18 +644,18 @@ unsigned lv_chk_bl_xxx(const lv_chk_t *lv, const ir_node *bl, const ir_node *var * will be left. */ DBG((lv->dbg, LEVEL_2, "\tbe tgt reach: %B, dom span: [%d, %d]\n", Tq, min_dom, max_dom)); i = bitset_next_set(Tq, min_dom); - while(i <= max_dom) { + while (i <= max_dom) { bl_info_t *ti = lv->map[i]; int use_in_current_block = bitset_is_set(uses, ti->id); stat_ev_cnt_inc(iter); /* - * This is somehat tricky. Since this routine handles both, live in + * This is somewhat tricky. Since this routine handles both, live in * and end/out we have to handle all the border cases correctly. * Each node is in its own red_reachable set (see calculation * function above). That means, that in the case where bl == t, the - * intersection check of uses and rechability below will always + * intersection check of uses and reachability below will always * find an intersection, namely t. * * However, if a block contains a use and the variable is dead