X-Git-Url: http://nsz.repo.hu/git/?a=blobdiff_plain;f=ir%2Fir%2Firvrfy.c;h=20b8890e74575c030befb71cbf2481e7533d3b2d;hb=bb9f2e36362333c6635b89f5258171b06c786608;hp=1236a12b3482454b823927e8578d7b8feee0c168;hpb=9baf0cd6e299828073f897e4bd0a284379ad9cdb;p=libfirm diff --git a/ir/ir/irvrfy.c b/ir/ir/irvrfy.c index 1236a12b3..20b8890e7 100644 --- a/ir/ir/irvrfy.c +++ b/ir/ir/irvrfy.c @@ -1,5 +1,5 @@ /* - * Copyright (C) 1995-2007 University of Karlsruhe. All right reserved. + * Copyright (C) 1995-2008 University of Karlsruhe. All right reserved. * * This file is part of libFirm. * @@ -23,9 +23,7 @@ * @author Christian Schaefer, Goetz Lindenmaier, Till Riedel, Michael Beck * @version $Id$ */ -#ifdef HAVE_CONFIG_H -# include "config.h" -#endif +#include "config.h" #include "irprog.h" #include "irop_t.h" @@ -49,6 +47,8 @@ void vrfy_enable_entity_tests(int enable) { vrfy_entities = enable; } +#ifndef NDEBUG + /** * little helper for NULL modes */ @@ -58,11 +58,13 @@ static const char *get_mode_name_ex(ir_mode *mode) { return get_mode_name(mode); } -/** the last IRG, on which a verify error was found */ +/** the last IRG, on which a verification error was found */ static ir_graph *last_irg_error = NULL; /** * print the name of the entity of an verification failure + * + * @param node the node caused the failure */ static void show_entity_failure(ir_node *node) { ir_graph *irg = get_irn_irg(node); @@ -273,11 +275,17 @@ static void show_phi_inputs(ir_node *phi, ir_node *block) { get_irn_node_nr(block), get_irn_arity(block)); } -/** If the address is Sel or SymConst, return the entity. */ +#endif /* #ifndef NDEBUG */ + +/** + * If the address is Sel or SymConst, return the entity. + * + * @param ptr the node representing the address + */ static ir_entity *get_ptr_entity(ir_node *ptr) { - if (get_irn_op(ptr) == op_Sel) { + if (is_Sel(ptr)) { return get_Sel_entity(ptr); - } else if ((get_irn_op(ptr) == op_SymConst) && (get_SymConst_kind(ptr) == symconst_addr_ent)) { + } else if (is_SymConst_addr_ent(ptr)) { return get_SymConst_entity(ptr); } return NULL; @@ -296,7 +304,6 @@ static int verify_node_Proj_Start(ir_node *n, ir_node *p) { (proj == pn_Start_X_initial_exec && mode == mode_X) || (proj == pn_Start_M && mode == mode_M) || (proj == pn_Start_P_frame_base && mode_is_reference(mode)) || - (proj == pn_Start_P_globals && mode_is_reference(mode)) || (proj == pn_Start_P_tls && mode_is_reference(mode)) || (proj == pn_Start_T_args && mode == mode_T) || (proj == pn_Start_P_value_arg_base && mode_is_reference(mode)) || @@ -384,17 +391,18 @@ static int verify_node_Proj_Call(ir_node *n, ir_node *p) { "wrong Proj from Call", 0, show_proj_failure(p); ); + /* if we have exception flow, we must have a real Memory input */ if (proj == pn_Call_X_regular) ASSERT_AND_RET( - get_irn_op(get_Call_mem(n)) != op_NoMem, + !is_NoMem(get_Call_mem(n)), "Regular Proj from FunctionCall", 0); else if (proj == pn_Call_X_except) ASSERT_AND_RET( - get_irn_op(get_Call_mem(n)) != op_NoMem, + !is_NoMem(get_Call_mem(n)), "Exception Proj from FunctionCall", 0); else if (proj == pn_Call_M_regular || proj == pn_Call_M_except) ASSERT_AND_RET( - (get_irn_op(get_Call_mem(n)) != op_NoMem || 1), + (!is_NoMem(get_Call_mem(n)) || 1), "Memory Proj from FunctionCall", 0); return 1; } @@ -541,6 +549,11 @@ static int verify_node_Proj_Cmp(ir_node *n, ir_node *p) { "wrong Proj from Cmp", 0, show_proj_failure(p); ); + ASSERT_AND_RET_DBG( + (mode_is_float(get_irn_mode(get_Cmp_left(n))) || !(proj & pn_Cmp_Uo)), + "unordered Proj for non-float Cmp", 0, + show_proj_failure(p); + ); return 1; } @@ -670,11 +683,13 @@ static int verify_node_Proj_Proj(ir_node *pred, ir_node *p) { if ((mode_is_reference(mode)) && is_compound_type(get_method_param_type(mt, proj))) /* value argument */ break; + if (get_irg_phase_state(get_irn_irg(pred)) != phase_backend) { ASSERT_AND_RET_DBG( - (mode == get_type_mode(get_method_param_type(mt, proj))), - "Mode of Proj from Start doesn't match mode of param type.", 0, - show_proj_mode_failure(p, get_method_param_type(mt, proj)); - ); + (mode == get_type_mode(get_method_param_type(mt, proj))), + "Mode of Proj from Start doesn't match mode of param type.", 0, + show_proj_mode_failure(p, get_method_param_type(mt, proj)); + ); + } } else if (nr == pn_Start_P_value_arg_base) { ASSERT_AND_RET( (proj >= 0 && mode_is_reference(mode)), @@ -745,9 +760,11 @@ static int verify_node_Proj_CallBegin(ir_node *n, ir_node *p) { static int verify_node_Proj_EndReg(ir_node *n, ir_node *p) { (void) n; (void) p; +#ifdef INTERPROCEDURAL_VIEW ASSERT_AND_RET( (get_irp_ip_view_state() != ip_view_no), "EndReg may only appear if ip view is constructed.", 0); +#endif return 1; } @@ -757,9 +774,11 @@ static int verify_node_Proj_EndReg(ir_node *n, ir_node *p) { static int verify_node_Proj_EndExcept(ir_node *n, ir_node *p) { (void) n; (void) p; +#ifdef INTERPROCEDURAL_VIEW ASSERT_AND_RET( (get_irp_ip_view_state() != ip_view_no), "EndExcept may only appear if ip view is constructed.", 0); +#endif return 1; } @@ -840,29 +859,53 @@ verify_node_Proj(ir_node *p, ir_graph *irg) { */ static int verify_node_Block(ir_node *n, ir_graph *irg) { int i; + ir_node *mb = get_Block_MacroBlock(n); + + ASSERT_AND_RET(is_Block(mb) || is_Bad(mb), "Block node with wrong MacroBlock", 0); + + if (is_Block(mb) && mb != n) { + ir_node *pred; + + /* Blocks with more than one predecessor must be header blocks */ + ASSERT_AND_RET(get_Block_n_cfgpreds(n) == 1, "partBlock with more than one predecessor", 0); + if (get_irg_phase_state(irg) != phase_backend) { + pred = get_Block_cfgpred(n, 0); + if (is_Proj(pred)) { + /* the predecessor MUST be a regular Proj */ + ir_node *frag_op = get_Proj_pred(pred); + ASSERT_AND_RET( + is_fragile_op(frag_op) && get_Proj_proj(pred) == pn_Generic_X_regular, + "partBlock with non-regular predecessor", 0); + } else { + /* We allow Jmps to be predecessors of partBlocks. This can happen due to optimization + of fragile nodes during construction. It does not violate our assumption of dominance + so let it. */ + ASSERT_AND_RET(is_Jmp(pred) || is_Bad(pred), + "partBlock with non-regular predecessor", 0); + } + } else { + /* relax in backend: Bound nodes are probably lowered into conditional jumps */ + } + } for (i = get_Block_n_cfgpreds(n) - 1; i >= 0; --i) { ir_node *pred = get_Block_cfgpred(n, i); ASSERT_AND_RET( - ( - is_Bad(pred) || - is_Unknown(pred) || - (get_irn_mode(pred) == mode_X) - ), - "Block node", 0); + is_Bad(pred) || (get_irn_mode(pred) == mode_X), + "Block node must have a mode_X predecessor", 0); } - /* End block may only have Return, Raise or fragile ops as preds. */ if (n == get_irg_end_block(irg) && get_irg_phase_state(irg) != phase_backend) + /* End block may only have Return, Raise or fragile ops as preds. */ for (i = get_Block_n_cfgpreds(n) - 1; i >= 0; --i) { ir_node *pred = skip_Proj(get_Block_cfgpred(n, i)); - if (is_Proj(pred) || get_irn_op(pred) == op_Tuple) + if (is_Proj(pred) || is_Tuple(pred)) break; /* We can not test properly. How many tuples are there? */ ASSERT_AND_RET( ( - is_Return(pred) || - is_Bad(pred) || - (get_irn_op(pred) == op_Raise) || + is_Return(pred) || + is_Bad(pred) || + is_Raise(pred) || is_fragile_op(pred) ), "End Block node", 0); @@ -924,8 +967,10 @@ static int verify_node_Break(ir_node *n, ir_graph *irg) { ir_mode *mymode = get_irn_mode(n); (void) irg; +#ifdef INTERPROCEDURAL_VIEW ASSERT_AND_RET((get_irp_ip_view_state() != ip_view_no), "Break may only appear if ip view is constructed.", 0); +#endif ASSERT_AND_RET( /* Jmp: BB --> X */ mymode == mode_X, "Break node", 0 @@ -977,18 +1022,20 @@ static int verify_node_Return(ir_node *n, ir_graph *irg) { for (i = get_Return_n_ress(n) - 1; i >= 0; --i) { ir_type *res_type = get_method_res_type(mt, i); - if (is_atomic_type(res_type)) { - ASSERT_AND_RET_DBG( - get_irn_mode(get_Return_res(n, i)) == get_type_mode(res_type), - "Mode of result for Return doesn't match mode of result type.", 0, - show_return_modes(irg, n, mt, i); - ); - } else { - ASSERT_AND_RET_DBG( - mode_is_reference(get_irn_mode(get_Return_res(n, i))), - "Mode of result for Return doesn't match mode of result type.", 0, - show_return_modes(irg, n, mt, i); - ); + if (get_irg_phase_state(irg) != phase_backend) { + if (is_atomic_type(res_type)) { + ASSERT_AND_RET_DBG( + get_irn_mode(get_Return_res(n, i)) == get_type_mode(res_type), + "Mode of result for Return doesn't match mode of result type.", 0, + show_return_modes(irg, n, mt, i); + ); + } else { + ASSERT_AND_RET_DBG( + mode_is_reference(get_irn_mode(get_Return_res(n, i))), + "Mode of result for Return doesn't match mode of result type.", 0, + show_return_modes(irg, n, mt, i); + ); + } } } return 1; @@ -1096,6 +1143,22 @@ static int verify_node_InstOf(ir_node *n, ir_graph *irg) { return 1; } +/** + * Check if the pinned state is right. + */ +static int verify_right_pinned(ir_node *n) { + ir_node *mem; + + if (get_irn_pinned(n) == op_pin_state_pinned) + return 1; + mem = get_Call_mem(n); + + /* if it's not pinned, its memory predecessor must be NoMem or Pin */ + if (is_NoMem(mem) || is_Pin(mem)) + return 1; + return 0; +} + /** * verify a Call node */ @@ -1112,10 +1175,7 @@ static int verify_node_Call(ir_node *n, ir_graph *irg) { ASSERT_AND_RET( op1mode == mode_M && mode_is_reference(op2mode), "Call node", 0 ); /* operand M x ref */ /* NoMem nodes are only allowed as memory input if the Call is NOT pinned */ - ASSERT_AND_RET( - (get_irn_op(get_Call_mem(n)) == op_NoMem) || - (get_irn_op(get_Call_mem(n)) != op_NoMem && get_irn_pinned(n) == op_pin_state_pinned), - "Call node with wrong memory input", 0 ); + ASSERT_AND_RET(verify_right_pinned(n),"Call node with wrong memory input", 0 ); mt = get_Call_type(n); if(get_unknown_type() == mt) { @@ -1134,32 +1194,37 @@ static int verify_node_Call(ir_node *n, ir_graph *irg) { get_Call_n_params(n) >= get_method_n_params(mt), "Number of args for Call doesn't match number of args in variadic type.", 0, - fprintf(stderr, "Call has %d params, method %s type %d\n", - get_Call_n_params(n), get_type_name(mt), get_method_n_params(mt)); + ir_fprintf(stderr, "Call %+F has %d params, method %s type %d\n", + n, get_Call_n_params(n), get_type_name(mt), get_method_n_params(mt)); ); } else { - ASSERT_AND_RET( + ASSERT_AND_RET_DBG( get_Call_n_params(n) == get_method_n_params(mt), "Number of args for Call doesn't match number of args in non variadic type.", - 0); + 0, + ir_fprintf(stderr, "Call %+F has %d params, method %s type %d\n", + n, get_Call_n_params(n), get_type_name(mt), get_method_n_params(mt)); + ); } for (i = 0; i < get_method_n_params(mt); i++) { ir_type *t = get_method_param_type(mt, i); - if (is_atomic_type(t)) { - ASSERT_AND_RET_DBG( - get_irn_mode(get_Call_param(n, i)) == get_type_mode(t), - "Mode of arg for Call doesn't match mode of arg type.", 0, - show_call_param(n, mt); - ); - } else { - /* call with a compound type, mode must be reference */ - ASSERT_AND_RET_DBG( - mode_is_reference(get_irn_mode(get_Call_param(n, i))), - "Mode of arg for Call doesn't match mode of arg type.", 0, - show_call_param(n, mt); - ); + if (get_irg_phase_state(irg) != phase_backend) { + if (is_atomic_type(t)) { + ASSERT_AND_RET_DBG( + get_irn_mode(get_Call_param(n, i)) == get_type_mode(t), + "Mode of arg for Call doesn't match mode of arg type.", 0, + show_call_param(n, mt); + ); + } else { + /* call with a compound type, mode must be reference */ + ASSERT_AND_RET_DBG( + mode_is_reference(get_irn_mode(get_Call_param(n, i))), + "Mode of arg for Call doesn't match mode of arg type.", 0, + show_call_param(n, mt); + ); + } } } @@ -1214,15 +1279,12 @@ static int verify_node_Sub(ir_node *n, ir_graph *irg) { (mymode ==op1mode && mymode == op2mode && mode_is_data(op1mode)) || /* Pointer Sub: BB x ref x int --> ref */ (op1mode == mymode && mode_is_int(op2mode) && mode_is_reference(mymode)) || - /* Pointer Sub: BB x int x ref --> ref */ - (mode_is_int(op1mode) && op2mode == mymode && mode_is_reference(mymode)) || /* Pointer Sub: BB x ref x ref --> int */ (op1mode == op2mode && mode_is_reference(op2mode) && mode_is_int(mymode)) ), "Sub node", 0, show_binop_failure(n, "/* common Sub: BB x numP x numP --> numP */ |\n" "/* Pointer Sub: BB x ref x int --> ref */ |\n" - "/* Pointer Sub: BB x int x ref --> ref */ |\n" "/* Pointer Sub: BB x ref x ref --> int */" ); ); return 1; @@ -1255,18 +1317,39 @@ static int verify_node_Mul(ir_node *n, ir_graph *irg) { ASSERT_AND_RET_DBG( ( - /* Mul: BB x int1 x int1 --> int2 */ - (mode_is_int(op1mode) && op2mode == op1mode && mode_is_int(mymode)) || + /* Mul: BB x int_n x int_n --> int_n|int_2n */ + (mode_is_int(op1mode) && op2mode == op1mode && mode_is_int(mymode) && + (op1mode == mymode || get_mode_size_bits(op1mode) * 2 == get_mode_size_bits(mymode))) || /* Mul: BB x float x float --> float */ (mode_is_float(op1mode) && op2mode == op1mode && mymode == op1mode) ), "Mul node",0, - show_binop_failure(n, "/* Mul: BB x int1 x int1 --> int2 */ |\n" + show_binop_failure(n, "/* Mul: BB x int_n x int_n --> int_n|int_2n */ |\n" "/* Mul: BB x float x float --> float */"); ); return 1; } +/** + * verify a Mulh node + */ +static int verify_node_Mulh(ir_node *n, ir_graph *irg) { + ir_mode *mymode = get_irn_mode(n); + ir_mode *op1mode = get_irn_mode(get_Mulh_left(n)); + ir_mode *op2mode = get_irn_mode(get_Mulh_right(n)); + (void) irg; + + ASSERT_AND_RET_DBG( + ( + /* Mulh: BB x int x int --> int */ + (mode_is_int(op1mode) && op2mode == op1mode && op1mode == mymode) + ), + "Mulh node",0, + show_binop_failure(n, "/* Mulh: BB x int x int --> int */"); + ); + return 1; +} + /** * verify a Quot node */ @@ -1422,7 +1505,7 @@ static int verify_node_Cmp(ir_node *n, ir_graph *irg) { ASSERT_AND_RET_DBG( /* Cmp: BB x datab x datab --> b16 */ - mode_is_data (op1mode) && + mode_is_datab(op1mode) && op2mode == op1mode && mymode == mode_T, "Cmp node", 0, @@ -1457,21 +1540,21 @@ static int verify_node_Shift(ir_node *n, ir_graph *irg) { #define verify_node_Shrs verify_node_Shift /** - * verify a Rot node + * verify a Rotl node */ -static int verify_node_Rot(ir_node *n, ir_graph *irg) { +static int verify_node_Rotl(ir_node *n, ir_graph *irg) { ir_mode *mymode = get_irn_mode(n); - ir_mode *op1mode = get_irn_mode(get_Rot_left(n)); - ir_mode *op2mode = get_irn_mode(get_Rot_right(n)); + ir_mode *op1mode = get_irn_mode(get_Rotl_left(n)); + ir_mode *op2mode = get_irn_mode(get_Rotl_right(n)); (void) irg; ASSERT_AND_RET_DBG( - /* Rot: BB x int x int --> int */ + /* Rotl: BB x int x int --> int */ mode_is_int(op1mode) && mode_is_int(op2mode) && mymode == op1mode, - "Rot node", 0, - show_binop_failure(n, "/* Rot: BB x int x int --> int */"); + "Rotl node", 0, + show_binop_failure(n, "/* Rotl: BB x int x int --> int */"); ); return 1; } @@ -1485,10 +1568,10 @@ static int verify_node_Conv(ir_node *n, ir_graph *irg) { (void) irg; ASSERT_AND_RET_DBG( - /* Conv: BB x datab1 --> datab2 */ - mode_is_datab(op1mode) && mode_is_datab(mymode), + get_irg_phase_state(irg) == phase_backend || + (mode_is_datab(op1mode) && mode_is_data(mymode)), "Conv node", 0, - show_unop_failure(n, "/* Conv: BB x datab1 --> datab2 */"); + show_unop_failure(n, "/* Conv: BB x datab --> data */"); ); return 1; } @@ -1519,7 +1602,7 @@ static int verify_node_Phi(ir_node *n, ir_graph *irg) { int i; (void) irg; - if (! is_Bad(block) && get_irg_phase_state(get_irn_irg(n)) != phase_building) { + if (! is_Bad(block) && get_irg_phase_state(get_irn_irg(n)) != phase_building && get_irn_arity(n) > 0) { /* a Phi node MUST have the same number of inputs as its block */ ASSERT_AND_RET_DBG( get_irn_arity(n) == get_irn_arity(block), @@ -1529,9 +1612,9 @@ static int verify_node_Phi(ir_node *n, ir_graph *irg) { } /* Phi: BB x dataM^n --> dataM */ - for (i = get_irn_arity(n) - 1; i >= 0; --i) { - ir_node *pred = get_irn_n(n, i); - if (!is_Bad(pred) && (get_irn_op(pred) != op_Unknown)) { + for (i = get_Phi_n_preds(n) - 1; i >= 0; --i) { + ir_node *pred = get_Phi_pred(n, i); + if (!is_Bad(pred)) { ASSERT_AND_RET_DBG( get_irn_mode(pred) == mymode, "Phi node", 0, @@ -1575,8 +1658,10 @@ static int verify_node_Phi(ir_node *n, ir_graph *irg) { static int verify_node_Filter(ir_node *n, ir_graph *irg) { (void) n; (void) irg; +#ifdef INTERPROCEDURAL_VIEW ASSERT_AND_RET((get_irp_ip_view_state() != ip_view_no), "Filter may only appear if ip view is constructed.", 0); +#endif /* We should further do tests as for Proj and Phi. */ return 1; } @@ -1627,7 +1712,7 @@ static int verify_node_Store(ir_node *n, ir_graph *irg) { ir_mode *op2mode = get_irn_mode(get_Store_ptr(n)); ir_mode *op3mode = get_irn_mode(get_Store_value(n)); - ASSERT_AND_RET(op1mode == mode_M && mode_is_data(op3mode), "Store node", 0 ); + ASSERT_AND_RET(op1mode == mode_M && mode_is_datab(op3mode), "Store node", 0 ); if (get_irg_phase_state(irg) != phase_backend) { ASSERT_AND_RET(mode_is_reference(op2mode), "Store node", 0 ); } else { @@ -1666,7 +1751,7 @@ static int verify_node_Alloc(ir_node *n, ir_graph *irg) { !mode_is_signed(op2mode) && mymode == mode_T, "Alloc node", 0, - show_binop_failure(n, "/* Alloc: BB x M x int_u --> M x X x ref */"); + show_node_failure(n); ); return 1; } @@ -1739,11 +1824,11 @@ static int verify_node_Mux(ir_node *n, ir_graph *irg) { (void) irg; ASSERT_AND_RET( - /* Mux: BB x b x numP x numP --> numP */ + /* Mux: BB x b x datab x datab --> datab */ op1mode == mode_b && op2mode == mymode && op3mode == mymode && - mode_is_data(mymode), + mode_is_datab(mymode), "Mux node", 0 ); return 1; @@ -1777,10 +1862,7 @@ static int verify_node_CopyB(ir_node *n, ir_graph *irg) { /* NoMem nodes are only allowed as memory input if the CopyB is NOT pinned. This should happen RARELY, as CopyB COPIES MEMORY */ - ASSERT_AND_RET( - (get_irn_op(get_CopyB_mem(n)) == op_NoMem) || - (get_irn_op(get_CopyB_mem(n)) != op_NoMem && get_irn_pinned(n) == op_pin_state_pinned), - "CopyB node with wrong memory input", 0 ); + ASSERT_AND_RET(verify_right_pinned(n), "CopyB node with wrong memory input", 0 ); return 1; } @@ -1815,8 +1897,18 @@ static int verify_node_Bound(ir_node *n, ir_graph *irg) { * @return non-zero on success, 0 on dominance error */ static int check_dominance_for_node(ir_node *use) { + if (is_Block(use)) { + ir_node *mbh = get_Block_MacroBlock(use); + + if (mbh != use) { + /* must be a partBlock */ + if (is_Block(mbh)) { + ASSERT_AND_RET(block_dominates(mbh, use), "MacroBlock header must dominate a partBlock", 0); + } + } + } /* This won't work for blocks and the end node */ - if (!is_Block(use) && use != get_irg_end(current_ir_graph) && use != current_ir_graph->anchor) { + else if (use != get_irg_end(current_ir_graph) && use != current_ir_graph->anchor) { int i; ir_node *bl = get_nodes_block(use); @@ -1903,11 +1995,12 @@ int irn_vrfy_irg(ir_node *n, ir_graph *irg) { } int irn_vrfy(ir_node *n) { - int res = 1; #ifdef DEBUG_libfirm - res = irn_vrfy_irg(n, current_ir_graph); + return irn_vrfy_irg(n, current_ir_graph); +#else + (void)n; + return 1; #endif - return res; } /*-----------------------------------------------------------------*/ @@ -1950,7 +2043,10 @@ int irg_verify(ir_graph *irg, unsigned flags) { rem = current_ir_graph; current_ir_graph = irg; - last_irg_error = NULL; + +#ifndef NDEBUG + last_irg_error = NULL; +#endif /* NDEBUG */ assert(get_irg_pinned(irg) == op_pin_state_pinned && "Verification need pinned graph"); @@ -1974,6 +2070,9 @@ int irg_verify(ir_graph *irg, unsigned flags) { else fprintf(stderr, "irg_verify: Verifying graph %p failed\n", (void *)current_ir_graph); } +#else + (void)irg; + (void)flags; #endif /* DEBUG_libfirm */ return res; @@ -2022,7 +2121,7 @@ static void check_bads(ir_node *node, void *env) { fprintf(stderr, "irg_vrfy_bads: Block %ld has Bad predecessor\n", get_irn_node_nr(node)); } if (get_node_verification_mode() == FIRM_VERIFICATION_ON) { - dump_ir_block_graph(current_ir_graph, "-assert"); + dump_ir_block_graph_sched(current_ir_graph, "-assert"); assert(0 && "Bad CF detected"); } } @@ -2039,21 +2138,21 @@ static void check_bads(ir_node *node, void *env) { fprintf(stderr, "irg_vrfy_bads: node %ld has Bad Block\n", get_irn_node_nr(node)); } if (get_node_verification_mode() == FIRM_VERIFICATION_ON) { - dump_ir_block_graph(current_ir_graph, "-assert"); + dump_ir_block_graph_sched(current_ir_graph, "-assert"); assert(0 && "Bad CF detected"); } } } if ((venv->flags & TUPLE) == 0) { - if (get_irn_op(node) == op_Tuple) { + if (is_Tuple(node)) { venv->res |= TUPLE; if (get_node_verification_mode() == FIRM_VERIFICATION_REPORT) { fprintf(stderr, "irg_vrfy_bads: node %ld is a Tuple\n", get_irn_node_nr(node)); } if (get_node_verification_mode() == FIRM_VERIFICATION_ON) { - dump_ir_block_graph(current_ir_graph, "-assert"); + dump_ir_block_graph_sched(current_ir_graph, "-assert"); assert(0 && "Tuple detected"); } } @@ -2074,7 +2173,7 @@ static void check_bads(ir_node *node, void *env) { fprintf(stderr, "irg_vrfy_bads: Phi %ld has Bad Input\n", get_irn_node_nr(node)); } if (get_node_verification_mode() == FIRM_VERIFICATION_ON) { - dump_ir_block_graph(current_ir_graph, "-assert"); + dump_ir_block_graph_sched(current_ir_graph, "-assert"); assert(0 && "Bad CF detected"); } } @@ -2088,7 +2187,7 @@ static void check_bads(ir_node *node, void *env) { fprintf(stderr, "irg_vrfy_bads: node %ld has Bad Input\n", get_irn_node_nr(node)); } if (get_node_verification_mode() == FIRM_VERIFICATION_ON) { - dump_ir_block_graph(current_ir_graph, "-assert"); + dump_ir_block_graph_sched(current_ir_graph, "-assert"); assert(0 && "Bad NON-CF detected"); } } @@ -2139,6 +2238,7 @@ void firm_set_default_verifyer(ir_opcode code, ir_op_ops *ops) { CASE(Sub); CASE(Minus); CASE(Mul); + CASE(Mulh); CASE(Quot); CASE(DivMod); CASE(Div); @@ -2152,7 +2252,7 @@ void firm_set_default_verifyer(ir_opcode code, ir_op_ops *ops) { CASE(Shl); CASE(Shr); CASE(Shrs); - CASE(Rot); + CASE(Rotl); CASE(Conv); CASE(Cast); CASE(Phi);