X-Git-Url: http://nsz.repo.hu/git/?a=blobdiff_plain;f=ir%2Fir%2Firvrfy.c;h=0a4e8d4a02fb1c481c8f411e45c12b60f081623d;hb=af0115376ccdcc72e0643caf221407218f4504f0;hp=034ff42631e5bff3b614183a9cf18380ca25536a;hpb=9c6ef39a9a1f9401eb46b5b577f109f704fa92b9;p=libfirm diff --git a/ir/ir/irvrfy.c b/ir/ir/irvrfy.c index 034ff4263..0a4e8d4a0 100644 --- a/ir/ir/irvrfy.c +++ b/ir/ir/irvrfy.c @@ -23,9 +23,7 @@ * @author Christian Schaefer, Goetz Lindenmaier, Till Riedel, Michael Beck * @version $Id$ */ -#ifdef HAVE_CONFIG_H -# include "config.h" -#endif +#include "config.h" #include "irprog.h" #include "irop_t.h" @@ -60,11 +58,13 @@ static const char *get_mode_name_ex(ir_mode *mode) { return get_mode_name(mode); } -/** the last IRG, on which a verify error was found */ +/** the last IRG, on which a verification error was found */ static ir_graph *last_irg_error = NULL; /** * print the name of the entity of an verification failure + * + * @param node the node caused the failure */ static void show_entity_failure(ir_node *node) { ir_graph *irg = get_irn_irg(node); @@ -283,7 +283,7 @@ static void show_phi_inputs(ir_node *phi, ir_node *block) { * @param ptr the node representing the address */ static ir_entity *get_ptr_entity(ir_node *ptr) { - if (get_irn_op(ptr) == op_Sel) { + if (is_Sel(ptr)) { return get_Sel_entity(ptr); } else if (is_SymConst_addr_ent(ptr)) { return get_SymConst_entity(ptr); @@ -305,9 +305,7 @@ static int verify_node_Proj_Start(ir_node *n, ir_node *p) { (proj == pn_Start_M && mode == mode_M) || (proj == pn_Start_P_frame_base && mode_is_reference(mode)) || (proj == pn_Start_P_tls && mode_is_reference(mode)) || - (proj == pn_Start_T_args && mode == mode_T) || - (proj == pn_Start_P_value_arg_base && mode_is_reference(mode)) || - (proj == pn_Start_P_value_arg_base && mode == mode_T) /* FIXME: only one of those */ + (proj == pn_Start_T_args && mode == mode_T) ), "wrong Proj from Start", 0, show_proj_failure(p); @@ -394,15 +392,15 @@ static int verify_node_Proj_Call(ir_node *n, ir_node *p) { /* if we have exception flow, we must have a real Memory input */ if (proj == pn_Call_X_regular) ASSERT_AND_RET( - get_irn_op(get_Call_mem(n)) != op_NoMem, + !is_NoMem(get_Call_mem(n)), "Regular Proj from FunctionCall", 0); else if (proj == pn_Call_X_except) ASSERT_AND_RET( - get_irn_op(get_Call_mem(n)) != op_NoMem, + !is_NoMem(get_Call_mem(n)), "Exception Proj from FunctionCall", 0); else if (proj == pn_Call_M_regular || proj == pn_Call_M_except) ASSERT_AND_RET( - (get_irn_op(get_Call_mem(n)) != op_NoMem || 1), + (!is_NoMem(get_Call_mem(n)) || 1), "Memory Proj from FunctionCall", 0); return 1; } @@ -690,15 +688,6 @@ static int verify_node_Proj_Proj(ir_node *pred, ir_node *p) { show_proj_mode_failure(p, get_method_param_type(mt, proj)); ); } - } else if (nr == pn_Start_P_value_arg_base) { - ASSERT_AND_RET( - (proj >= 0 && mode_is_reference(mode)), - "wrong Proj from Proj from Start", 0 - ); - ASSERT_AND_RET( - (proj < get_method_n_params(mt)), - "More Projs for args than args in type", 0 - ); } break; @@ -868,18 +857,23 @@ static int verify_node_Block(ir_node *n, ir_graph *irg) { /* Blocks with more than one predecessor must be header blocks */ ASSERT_AND_RET(get_Block_n_cfgpreds(n) == 1, "partBlock with more than one predecessor", 0); - pred = get_Block_cfgpred(n, 0); - if (is_Proj(pred)) { - /* the predecessor MUST be a regular Proj */ - ir_node *frag_op = get_Proj_pred(pred); - ASSERT_AND_RET(is_fragile_op(frag_op) && get_Proj_proj(pred) == pn_Generic_X_regular, - "partBlock with non-regular predecessor", 0); + if (get_irg_phase_state(irg) != phase_backend) { + pred = get_Block_cfgpred(n, 0); + if (is_Proj(pred)) { + /* the predecessor MUST be a regular Proj */ + ir_node *frag_op = get_Proj_pred(pred); + ASSERT_AND_RET( + is_fragile_op(frag_op) && get_Proj_proj(pred) == pn_Generic_X_regular, + "partBlock with non-regular predecessor", 0); + } else { + /* We allow Jmps to be predecessors of partBlocks. This can happen due to optimization + of fragile nodes during construction. It does not violate our assumption of dominance + so let it. */ + ASSERT_AND_RET(is_Jmp(pred) || is_Bad(pred), + "partBlock with non-regular predecessor", 0); + } } else { - /* We allow Jmps to be predecessors of partBlocks. This can happen due to optimization - of fragile nodes during construction. It does not violate our assumption of dominance - so let it. */ - ASSERT_AND_RET(is_Jmp(pred) || is_Bad(pred), - "partBlock with non-regular predecessor", 0); + /* relax in backend: Bound nodes are probably lowered into conditional jumps */ } } @@ -1274,15 +1268,12 @@ static int verify_node_Sub(ir_node *n, ir_graph *irg) { (mymode ==op1mode && mymode == op2mode && mode_is_data(op1mode)) || /* Pointer Sub: BB x ref x int --> ref */ (op1mode == mymode && mode_is_int(op2mode) && mode_is_reference(mymode)) || - /* Pointer Sub: BB x int x ref --> ref */ - (mode_is_int(op1mode) && op2mode == mymode && mode_is_reference(mymode)) || /* Pointer Sub: BB x ref x ref --> int */ (op1mode == op2mode && mode_is_reference(op2mode) && mode_is_int(mymode)) ), "Sub node", 0, show_binop_failure(n, "/* common Sub: BB x numP x numP --> numP */ |\n" "/* Pointer Sub: BB x ref x int --> ref */ |\n" - "/* Pointer Sub: BB x int x ref --> ref */ |\n" "/* Pointer Sub: BB x ref x ref --> int */" ); ); return 1; @@ -1538,21 +1529,21 @@ static int verify_node_Shift(ir_node *n, ir_graph *irg) { #define verify_node_Shrs verify_node_Shift /** - * verify a Rot node + * verify a Rotl node */ -static int verify_node_Rot(ir_node *n, ir_graph *irg) { +static int verify_node_Rotl(ir_node *n, ir_graph *irg) { ir_mode *mymode = get_irn_mode(n); - ir_mode *op1mode = get_irn_mode(get_Rot_left(n)); - ir_mode *op2mode = get_irn_mode(get_Rot_right(n)); + ir_mode *op1mode = get_irn_mode(get_Rotl_left(n)); + ir_mode *op2mode = get_irn_mode(get_Rotl_right(n)); (void) irg; ASSERT_AND_RET_DBG( - /* Rot: BB x int x int --> int */ + /* Rotl: BB x int x int --> int */ mode_is_int(op1mode) && mode_is_int(op2mode) && mymode == op1mode, - "Rot node", 0, - show_binop_failure(n, "/* Rot: BB x int x int --> int */"); + "Rotl node", 0, + show_binop_failure(n, "/* Rotl: BB x int x int --> int */"); ); return 1; } @@ -1610,9 +1601,9 @@ static int verify_node_Phi(ir_node *n, ir_graph *irg) { } /* Phi: BB x dataM^n --> dataM */ - for (i = get_irn_arity(n) - 1; i >= 0; --i) { - ir_node *pred = get_irn_n(n, i); - if (!is_Bad(pred) && (get_irn_op(pred) != op_Unknown)) { + for (i = get_Phi_n_preds(n) - 1; i >= 0; --i) { + ir_node *pred = get_Phi_pred(n, i); + if (!is_Bad(pred)) { ASSERT_AND_RET_DBG( get_irn_mode(pred) == mymode, "Phi node", 0, @@ -1675,9 +1666,6 @@ static int verify_node_Load(ir_node *n, ir_graph *irg) { ASSERT_AND_RET(op1mode == mode_M, "Load node", 0); if (get_irg_phase_state(irg) != phase_backend) { ASSERT_AND_RET(mode_is_reference(op2mode), "Load node", 0 ); - } else { - ASSERT_AND_RET(mode_is_reference(op2mode) || - (mode_is_int(op2mode) && get_mode_size_bits(op2mode) == get_mode_size_bits(mode_P)), "Load node", 0 ); } ASSERT_AND_RET( mymode == mode_T, "Load node", 0 ); @@ -1713,9 +1701,6 @@ static int verify_node_Store(ir_node *n, ir_graph *irg) { ASSERT_AND_RET(op1mode == mode_M && mode_is_datab(op3mode), "Store node", 0 ); if (get_irg_phase_state(irg) != phase_backend) { ASSERT_AND_RET(mode_is_reference(op2mode), "Store node", 0 ); - } else { - ASSERT_AND_RET(mode_is_reference(op2mode) || - (mode_is_int(op2mode) && get_mode_size_bits(op2mode) == get_mode_size_bits(mode_P)), "Store node", 0 ); } ASSERT_AND_RET(mymode == mode_T, "Store node", 0); @@ -1749,7 +1734,7 @@ static int verify_node_Alloc(ir_node *n, ir_graph *irg) { !mode_is_signed(op2mode) && mymode == mode_T, "Alloc node", 0, - show_binop_failure(n, "/* Alloc: BB x M x int_u --> M x X x ref */"); + show_node_failure(n); ); return 1; } @@ -1822,11 +1807,11 @@ static int verify_node_Mux(ir_node *n, ir_graph *irg) { (void) irg; ASSERT_AND_RET( - /* Mux: BB x b x numP x numP --> numP */ + /* Mux: BB x b x datab x datab --> datab */ op1mode == mode_b && op2mode == mymode && op3mode == mymode && - mode_is_data(mymode), + mode_is_datab(mymode), "Mux node", 0 ); return 1; @@ -1847,11 +1832,6 @@ static int verify_node_CopyB(ir_node *n, ir_graph *irg) { if (get_irg_phase_state(irg) != phase_backend) { ASSERT_AND_RET(mode_is_reference(op2mode) && mode_is_reference(op3mode), "CopyB node", 0 ); - } else { - ASSERT_AND_RET(mode_is_reference(op2mode) || - (mode_is_int(op2mode) && get_mode_size_bits(op2mode) == get_mode_size_bits(mode_P)), "CopyB node", 0 ); - ASSERT_AND_RET(mode_is_reference(op3mode) || - (mode_is_int(op3mode) && get_mode_size_bits(op3mode) == get_mode_size_bits(mode_P)), "CopyB node", 0 ); } ASSERT_AND_RET( @@ -1961,7 +1941,8 @@ int irn_vrfy_irg(ir_node *n, ir_graph *irg) { { unsigned idx = get_irn_idx(n); ir_node *node_from_map = get_idx_irn(irg, idx); - ASSERT_AND_RET_DBG(node_from_map == n, "Node index and index map entry differ", 0, ir_printf("node %+F node in map %+F(%p)", n, node_from_map, node_from_map)); + ASSERT_AND_RET_DBG(node_from_map == n, "Node index and index map entry differ", 0, + ir_printf("node %+F node in map %+F(%p)\n", n, node_from_map, node_from_map)); } } @@ -2143,7 +2124,7 @@ static void check_bads(ir_node *node, void *env) { } if ((venv->flags & TUPLE) == 0) { - if (get_irn_op(node) == op_Tuple) { + if (is_Tuple(node)) { venv->res |= TUPLE; if (get_node_verification_mode() == FIRM_VERIFICATION_REPORT) { @@ -2250,7 +2231,7 @@ void firm_set_default_verifyer(ir_opcode code, ir_op_ops *ops) { CASE(Shl); CASE(Shr); CASE(Shrs); - CASE(Rot); + CASE(Rotl); CASE(Conv); CASE(Cast); CASE(Phi);