X-Git-Url: http://nsz.repo.hu/git/?a=blobdiff_plain;f=ir%2Fbe%2Fia32%2Fia32_intrinsics.c;h=12728b9e2f1faaf4c65022bae158b3544194ef66;hb=40db13dafa724caf3f38f76539cf107cd4c1a376;hp=8dd0b28b27851efd70a83406c5585de49a21b169;hpb=9e56dbf066d3698df877e2a9f78caeca29ea5a03;p=libfirm diff --git a/ir/be/ia32/ia32_intrinsics.c b/ir/be/ia32/ia32_intrinsics.c index 8dd0b28b2..12728b9e2 100644 --- a/ir/be/ia32/ia32_intrinsics.c +++ b/ir/be/ia32/ia32_intrinsics.c @@ -1,5 +1,5 @@ /* - * Copyright (C) 1995-2007 University of Karlsruhe. All right reserved. + * Copyright (C) 1995-2008 University of Karlsruhe. All right reserved. * * This file is part of libFirm. * @@ -22,38 +22,41 @@ * @brief This file implements the mapping of 64Bit intrinsic * functions to code or library calls. * @author Michael Beck - * @version $Id$ */ -#ifdef HAVE_CONFIG_H #include "config.h" -#endif +#include "iredges.h" #include "irgmod.h" #include "irop.h" #include "irnode_t.h" #include "ircons.h" #include "irprog_t.h" -#include "lowering.h" +#include "iroptimize.h" +#include "lower_dw.h" #include "array.h" +#include "error.h" #include "ia32_new_nodes.h" #include "bearch_ia32_t.h" #include "gen_ia32_regalloc_if.h" +#include "begnuas.h" /** The array of all intrinsics that must be mapped. */ static i_record *intrinsics; -/** An array to cache all entities */ -static ir_entity *i_ents[iro_MaxOpcode]; +/** An array to cache all entities. */ +static ir_entity *i_ents[iro_Last + 1]; /* * Maps all intrinsic calls that the backend support * and map all instructions the backend did not support * to runtime calls. */ -void ia32_handle_intrinsics(void) { - if (intrinsics && ARR_LEN(intrinsics) > 0) - lower_intrinsics(intrinsics, ARR_LEN(intrinsics)); +void ia32_handle_intrinsics(void) +{ + if (intrinsics && ARR_LEN(intrinsics) > 0) { + lower_intrinsics(intrinsics, ARR_LEN(intrinsics), /*part_block_used=*/1); + } } #define BINOP_Left_Low 0 @@ -61,30 +64,117 @@ void ia32_handle_intrinsics(void) { #define BINOP_Right_Low 2 #define BINOP_Right_High 3 +/** + * Reroute edges from the pn_Call_T_result proj of a call. + * + * @param resproj the pn_Call_T_result Proj + * @param l_res the lower 32 bit result + * @param h_res the upper 32 bit result or NULL + */ +static void reroute_result(ir_node *resproj, ir_node *l_res, ir_node *h_res) +{ + foreach_out_edge_safe(resproj, edge) { + ir_node *proj = get_edge_src_irn(edge); + long pn = get_Proj_proj(proj); + + if (pn == 0) { + edges_reroute(proj, l_res); + } else if (pn == 1 && h_res != NULL) { + edges_reroute(proj, h_res); + } else { + panic("Unsupported Result-Proj from Call found"); + } + } +} + /** * Replace a call be a tuple of l_res, h_res. + * + * @param call the call node to replace + * @param l_res the lower 32 bit result + * @param h_res the upper 32 bit result or NULL + * @param irg the graph to replace on + * @param block the block to replace on (always the call block) */ -static void resolve_call(ir_node *call, ir_node *l_res, ir_node *h_res, ir_graph *irg, ir_node *block) { - ir_node *res, *in[2]; - - in[0] = l_res; - in[1] = h_res; - res = new_r_Tuple(irg, block, h_res == NULL ? 1 : 2, in); - - turn_into_tuple(call, pn_Call_max); - set_Tuple_pred(call, pn_Call_M_regular, get_irg_no_mem(irg)); - set_Tuple_pred(call, pn_Call_X_regular, new_r_Jmp(irg, block)); - set_Tuple_pred(call, pn_Call_X_except, get_irg_bad(irg)); - set_Tuple_pred(call, pn_Call_T_result, res); - set_Tuple_pred(call, pn_Call_M_except, get_irg_no_mem(irg)); - set_Tuple_pred(call, pn_Call_P_value_res_base, get_irg_bad(irg)); +static void resolve_call(ir_node *call, ir_node *l_res, ir_node *h_res, ir_graph *irg, ir_node *block) +{ + ir_node *jmp, *res, *in[2]; + ir_node *nomem = get_irg_no_mem(irg); + int old_cse; + + if (edges_activated(irg)) { + /* use rerouting to prevent some warning in the backend */ + foreach_out_edge_safe(call, edge) { + ir_node *proj = get_edge_src_irn(edge); + pn_Call pn = (pn_Call)get_Proj_proj(proj); + + switch (pn) { + case pn_Call_X_regular: + /* Beware: + * We do not check here if this call really has exception and regular Proj's. + * new_r_Jmp might than be CSEd with the real exit jmp and then bad things happen + * (in movgen.c from 186.crafty for example). + * So be sure the newly created Jmp cannot CSE. + */ + old_cse = get_opt_cse(); + set_opt_cse(0); + jmp = new_r_Jmp(block); + set_opt_cse(old_cse); + edges_reroute(proj, jmp); + break; + + case pn_Call_X_except: + /* should not happen here */ + edges_reroute(proj, new_r_Bad(irg, mode_X)); + break; + case pn_Call_M: + /* should not happen here */ + edges_reroute(proj, nomem); + break; + case pn_Call_T_result: + reroute_result(proj, l_res, h_res); + break; + default: + panic("Wrong Proj from Call"); + } + kill_node(proj); + } + kill_node(call); + } else { + /* no edges, build Tuple */ + if (h_res == NULL) + res = l_res; + else { + in[0] = l_res; + in[1] = h_res; + res = new_r_Tuple(block, 2, in); + } + + /* + * Beware: + * We do not check here if this call really has exception and regular Proj's. + * new_r_Jmp might than be CSEd with the real exit jmp and then bad things happen + * (in movgen.c from 186.crafty for example). + * So be sure the newly created Jmp cannot CSE. + */ + old_cse = get_opt_cse(); + set_opt_cse(0); + jmp = new_r_Jmp(block); + set_opt_cse(old_cse); + + turn_into_tuple(call, pn_Call_max+1); + set_Tuple_pred(call, pn_Call_M, nomem); + set_Tuple_pred(call, pn_Call_X_regular, jmp); + set_Tuple_pred(call, pn_Call_X_except, new_r_Bad(irg, mode_X)); + set_Tuple_pred(call, pn_Call_T_result, res); + } } /** * Map an Add (a_l, a_h, b_l, b_h) */ -static int map_Add(ir_node *call, void *ctx) { - ir_graph *irg = current_ir_graph; +static int map_Add(ir_node *call, void *ctx) +{ dbg_info *dbg = get_irn_dbg_info(call); ir_node *block = get_nodes_block(call); ir_node **params = get_Call_param_arr(call); @@ -93,27 +183,32 @@ static int map_Add(ir_node *call, void *ctx) { ir_node *a_h = params[BINOP_Left_High]; ir_node *b_l = params[BINOP_Right_Low]; ir_node *b_h = params[BINOP_Right_High]; - ir_mode *l_res_mode = get_type_mode(get_method_res_type(method, 0)); - ir_mode *h_res_mode = get_type_mode(get_method_res_type(method, 1)); - ir_node *l_res, *h_res, *add; + ir_mode *l_mode = get_type_mode(get_method_res_type(method, 0)); + ir_mode *h_mode = get_type_mode(get_method_res_type(method, 1)); + ir_mode *mode_flags = ia32_reg_classes[CLASS_ia32_flags].mode; + ir_node *add_low, *add_high, *flags; + ir_node *l_res, *h_res; (void) ctx; /* l_res = a_l + b_l */ /* h_res = a_h + b_h + carry */ - add = new_rd_ia32_Add64Bit(dbg, irg, block, a_l, a_h, b_l, b_h); - l_res = new_r_Proj(irg, block, add, l_res_mode, pn_ia32_Add64Bit_low_res); - h_res = new_r_Proj(irg, block, add, h_res_mode, pn_ia32_Add64Bit_high_res); + add_low = new_bd_ia32_l_Add(dbg, block, a_l, b_l, mode_T); + flags = new_r_Proj(add_low, mode_flags, pn_ia32_flags); + add_high = new_bd_ia32_l_Adc(dbg, block, a_h, b_h, flags, h_mode); + + l_res = new_r_Proj(add_low, l_mode, pn_ia32_res); + h_res = add_high; - resolve_call(call, l_res, h_res, irg, block); + resolve_call(call, l_res, h_res, current_ir_graph, block); return 1; } /** * Map a Sub (a_l, a_h, b_l, b_h) */ -static int map_Sub(ir_node *call, void *ctx) { - ir_graph *irg = current_ir_graph; +static int map_Sub(ir_node *call, void *ctx) +{ dbg_info *dbg = get_irn_dbg_info(call); ir_node *block = get_nodes_block(call); ir_node **params = get_Call_param_arr(call); @@ -122,124 +217,78 @@ static int map_Sub(ir_node *call, void *ctx) { ir_node *a_h = params[BINOP_Left_High]; ir_node *b_l = params[BINOP_Right_Low]; ir_node *b_h = params[BINOP_Right_High]; - ir_mode *l_res_mode = get_type_mode(get_method_res_type(method, 0)); - ir_mode *h_res_mode = get_type_mode(get_method_res_type(method, 1)); - ir_node *l_res, *h_res, *res; + ir_mode *l_mode = get_type_mode(get_method_res_type(method, 0)); + ir_mode *h_mode = get_type_mode(get_method_res_type(method, 1)); + ir_mode *mode_flags = ia32_reg_classes[CLASS_ia32_flags].mode; + ir_node *sub_low, *sub_high, *flags; + ir_node *l_res, *h_res; (void) ctx; /* l_res = a_l - b_l */ /* h_res = a_h - b_h - carry */ - res = new_rd_ia32_Sub64Bit(dbg, irg, block, a_l, a_h, b_l, b_h); - l_res = new_r_Proj(irg, block, res, l_res_mode, pn_ia32_Sub64Bit_low_res); - h_res = new_r_Proj(irg, block, res, h_res_mode, pn_ia32_Sub64Bit_high_res); + sub_low = new_bd_ia32_l_Sub(dbg, block, a_l, b_l, mode_T); + flags = new_r_Proj(sub_low, mode_flags, pn_ia32_flags); + sub_high = new_bd_ia32_l_Sbb(dbg, block, a_h, b_h, flags, h_mode); - resolve_call(call, l_res, h_res, irg, block); - return 1; -} + l_res = new_r_Proj(sub_low, l_mode, pn_ia32_res); + h_res = sub_high; -/** - * Map a Shl (a_l, a_h, count) - */ -static int map_Shl(ir_node *call, void *ctx) { - ir_graph *irg = current_ir_graph; - dbg_info *dbg = get_irn_dbg_info(call); - ir_node *block = get_nodes_block(call); - ir_node **params = get_Call_param_arr(call); - ir_type *method = get_Call_type(call); - ir_node *a_l = params[BINOP_Left_Low]; - ir_node *a_h = params[BINOP_Left_High]; - ir_node *cnt = params[BINOP_Right_Low]; - ir_mode *l_res_mode = get_type_mode(get_method_res_type(method, 0)); - ir_mode *h_res_mode = get_type_mode(get_method_res_type(method, 1)); - ir_node *l_res, *h_res; - (void) ctx; - - /* h_res = SHLD a_h, a_l, cnt */ - h_res = new_rd_ia32_l_ShlD(dbg, irg, block, a_h, a_l, cnt, l_res_mode); - - /* l_res = SHL a_l, cnt */ - l_res = new_rd_ia32_l_Shl(dbg, irg, block, a_l, cnt, h_res_mode); - - //add_irn_dep(l_res, h_res); - - resolve_call(call, l_res, h_res, irg, block); + resolve_call(call, l_res, h_res, current_ir_graph, block); return 1; } /** - * Map a Shr (a_l, a_h, count) + * Checks where node high is a sign extension of low. */ -static int map_Shr(ir_node *call, void *ctx) { - ir_graph *irg = current_ir_graph; - dbg_info *dbg = get_irn_dbg_info(call); - ir_node *block = get_nodes_block(call); - ir_node **params = get_Call_param_arr(call); - ir_type *method = get_Call_type(call); - ir_node *a_l = params[BINOP_Left_Low]; - ir_node *a_h = params[BINOP_Left_High]; - ir_node *cnt = params[BINOP_Right_Low]; - ir_mode *l_res_mode = get_type_mode(get_method_res_type(method, 0)); - ir_mode *h_res_mode = get_type_mode(get_method_res_type(method, 1)); - ir_node *l_res, *h_res; - (void) ctx; - - /* l_res = SHRD a_l, a_h, cnt */ - l_res = new_rd_ia32_l_ShrD(dbg, irg, block, a_l, a_h, cnt, l_res_mode); - - /* h_res = SHR a_h, cnt */ - h_res = new_rd_ia32_l_Shr(dbg, irg, block, a_h, cnt, h_res_mode); +static int is_sign_extend(ir_node *low, ir_node *high) +{ + if (is_Shrs(high)) { + ir_node *high_l; + ir_node *high_r; + ir_tarval *shift_count; - //add_irn_dep(h_res, l_res); + high_r = get_Shrs_right(high); + if (!is_Const(high_r)) return 0; - resolve_call(call, l_res, h_res, irg, block); - return 1; -} + shift_count = get_Const_tarval(high_r); + if (!tarval_is_long(shift_count)) return 0; + if (get_tarval_long(shift_count) != 31) return 0; -/** - * Map a Shrs (a_l, a_h, count) - */ -static int map_Shrs(ir_node *call, void *ctx) { - ir_graph *irg = current_ir_graph; - dbg_info *dbg = get_irn_dbg_info(call); - ir_node *block = get_nodes_block(call); - ir_node **params = get_Call_param_arr(call); - ir_type *method = get_Call_type(call); - ir_node *a_l = params[BINOP_Left_Low]; - ir_node *a_h = params[BINOP_Left_High]; - ir_node *cnt = params[BINOP_Right_Low]; - ir_mode *l_res_mode = get_type_mode(get_method_res_type(method, 0)); - ir_mode *h_res_mode = get_type_mode(get_method_res_type(method, 1)); - ir_node *l_res, *h_res; - (void) ctx; + high_l = get_Shrs_left(high); - /* l_res = SHRD a_l, a_h, cnt */ - l_res = new_rd_ia32_l_ShrD(dbg, irg, block, a_l, a_h, cnt, l_res_mode); + if (is_Conv(low) && get_Conv_op(low) == high_l) return 1; + if (is_Conv(high_l) && get_Conv_op(high_l) == low) return 1; + } else if (is_Const(low) && is_Const(high)) { + ir_tarval *tl = get_Const_tarval(low); + ir_tarval *th = get_Const_tarval(high); - /* h_res = SAR a_h, cnt */ - h_res = new_rd_ia32_l_Sar(dbg, irg, block, a_h, cnt, h_res_mode); + if (tarval_is_long(th) && tarval_is_long(tl)) { + long l = get_tarval_long(tl); + long h = get_tarval_long(th); - //add_irn_dep(h_res, l_res); + return (h == 0 && l >= 0) || (h == -1 && l < 0); + } + } - resolve_call(call, l_res, h_res, irg, block); - return 1; + return 0; } /** * Map a Mul (a_l, a_h, b_l, b_h) */ -static int map_Mul(ir_node *call, void *ctx) { - ir_graph *irg = current_ir_graph; - dbg_info *dbg = get_irn_dbg_info(call); - ir_node *block = get_nodes_block(call); - ir_node **params = get_Call_param_arr(call); - ir_type *method = get_Call_type(call); - ir_node *a_l = params[BINOP_Left_Low]; - ir_node *a_h = params[BINOP_Left_High]; - ir_node *b_l = params[BINOP_Right_Low]; - ir_node *b_h = params[BINOP_Right_High]; - ir_mode *l_res_mode = get_type_mode(get_method_res_type(method, 0)); - ir_mode *h_res_mode = get_type_mode(get_method_res_type(method, 1)); +static int map_Mul(ir_node *call, void *ctx) +{ + dbg_info *dbg = get_irn_dbg_info(call); + ir_node *block = get_nodes_block(call); + ir_node **params = get_Call_param_arr(call); + ir_type *method = get_Call_type(call); + ir_node *a_l = params[BINOP_Left_Low]; + ir_node *a_h = params[BINOP_Left_High]; + ir_node *b_l = params[BINOP_Right_Low]; + ir_node *b_h = params[BINOP_Right_High]; + ir_mode *l_mode = get_type_mode(get_method_res_type(method, 0)); + ir_mode *h_mode = get_type_mode(get_method_res_type(method, 1)); ir_node *l_res, *h_res, *mul, *pEDX, *add; (void) ctx; @@ -252,16 +301,26 @@ static int map_Mul(ir_node *call, void *ctx) { t3 = a_l * b_h h_res = t2 + t3 */ - mul = new_rd_ia32_l_Mul(dbg, irg, block, a_l, b_l); - pEDX = new_rd_Proj(dbg, irg, block, mul, l_res_mode, pn_ia32_l_Mul_EDX); - l_res = new_rd_Proj(dbg, irg, block, mul, l_res_mode, pn_ia32_l_Mul_EAX); - mul = new_rd_ia32_l_Mul(dbg, irg, block, a_h, b_l); - add = new_rd_ia32_l_Add(dbg, irg, block, mul, pEDX, h_res_mode); - mul = new_rd_ia32_l_Mul(dbg, irg, block, a_l, b_h); - h_res = new_rd_ia32_l_Add(dbg, irg, block, add, mul, h_res_mode); - - resolve_call(call, l_res, h_res, irg, block); + /* handle the often used case of 32x32=64 mul */ + if (is_sign_extend(a_l, a_h) && is_sign_extend(b_l, b_h)) { + mul = new_bd_ia32_l_IMul(dbg, block, a_l, b_l); + h_res = new_rd_Proj(dbg, mul, h_mode, pn_ia32_l_IMul_res_high); + l_res = new_rd_Proj(dbg, mul, l_mode, pn_ia32_l_IMul_res_low); + } else { + /* note that zero extension is handled hare efficiently */ + mul = new_bd_ia32_l_Mul(dbg, block, a_l, b_l); + pEDX = new_rd_Proj(dbg, mul, h_mode, pn_ia32_l_Mul_res_high); + l_res = new_rd_Proj(dbg, mul, l_mode, pn_ia32_l_Mul_res_low); + + b_l = new_rd_Conv(dbg, block, b_l, h_mode); + mul = new_rd_Mul( dbg, block, a_h, b_l, h_mode); + add = new_rd_Add( dbg, block, mul, pEDX, h_mode); + a_l = new_rd_Conv(dbg, block, a_l, h_mode); + mul = new_rd_Mul( dbg, block, a_l, b_h, h_mode); + h_res = new_rd_Add( dbg, block, add, mul, h_mode); + } + resolve_call(call, l_res, h_res, current_ir_graph, block); return 1; } @@ -269,240 +328,125 @@ static int map_Mul(ir_node *call, void *ctx) { /** * Map a Minus (a_l, a_h) */ -static int map_Minus(ir_node *call, void *ctx) { - ir_graph *irg = current_ir_graph; - dbg_info *dbg = get_irn_dbg_info(call); - ir_node *block = get_nodes_block(call); - ir_node **params = get_Call_param_arr(call); - ir_type *method = get_Call_type(call); - ir_node *a_l = params[BINOP_Left_Low]; - ir_node *a_h = params[BINOP_Left_High]; - ir_mode *l_res_mode = get_type_mode(get_method_res_type(method, 0)); - ir_mode *h_res_mode = get_type_mode(get_method_res_type(method, 1)); - ir_node *l_res, *h_res, *cnst, *res; +static int map_Minus(ir_node *call, void *ctx) +{ + dbg_info *dbg = get_irn_dbg_info(call); + ir_node *block = get_nodes_block(call); + ir_node **params = get_Call_param_arr(call); + ir_type *method = get_Call_type(call); + ir_node *a_l = params[BINOP_Left_Low]; + ir_node *a_h = params[BINOP_Left_High]; + ir_mode *l_mode = get_type_mode(get_method_res_type(method, 0)); + ir_mode *h_mode = get_type_mode(get_method_res_type(method, 1)); + ir_node *l_res, *h_res, *res; (void) ctx; - /* too bad: we need 0 in a register here */ - cnst = new_Const_long(h_res_mode, 0); - - /* l_res = 0 - a_l */ - /* h_res = 0 - a_h - carry */ + res = new_bd_ia32_Minus64Bit(dbg, block, a_l, a_h); + l_res = new_r_Proj(res, l_mode, pn_ia32_Minus64Bit_low_res); + h_res = new_r_Proj(res, h_mode, pn_ia32_Minus64Bit_high_res); - res = new_rd_ia32_Minus64Bit(dbg, irg, block, cnst, a_l, a_h); - l_res = new_r_Proj(irg, block, res, l_res_mode, pn_ia32_Minus64Bit_low_res); - h_res = new_r_Proj(irg, block, res, h_res_mode, pn_ia32_Minus64Bit_high_res); - - resolve_call(call, l_res, h_res, irg, block); + resolve_call(call, l_res, h_res, current_ir_graph, block); return 1; } +#define ID(x) new_id_from_chars(x, sizeof(x)-1) + /** - * Map a Abs (a_l, a_h) + * Maps a Div. Change into a library call. */ -static int map_Abs(ir_node *call, void *ctx) { - ir_graph *irg = current_ir_graph; - dbg_info *dbg = get_irn_dbg_info(call); - ir_node *block = get_nodes_block(call); - ir_node **params = get_Call_param_arr(call); - ir_type *method = get_Call_type(call); - ir_node *a_l = params[BINOP_Left_Low]; - ir_node *a_h = params[BINOP_Left_High]; - ir_mode *l_res_mode = get_type_mode(get_method_res_type(method, 0)); - ir_mode *h_res_mode = get_type_mode(get_method_res_type(method, 1)); - ir_node *l_res, *h_res, *sign, *sub_l, *sub_h, *res; - (void) ctx; - - /* - Code inspired by gcc output :) (although gcc doubles the - operation for t1 as t2 and uses t1 for operations with low part - and t2 for operations with high part which is actually unnecessary - because t1 and t2 represent the same value) - - t1 = SHRS a_h, 31 - t2 = a_l ^ t1 - t3 = a_h ^ t1 - l_res = t2 - t1 - h_res = t3 - t1 - carry - - */ - - sign = new_rd_ia32_l_Sar(dbg, irg, block, a_h, new_Const_long(h_res_mode, 31), h_res_mode); - sub_l = new_rd_ia32_l_Xor(dbg, irg, block, a_l, sign, l_res_mode); - sub_h = new_rd_ia32_l_Xor(dbg, irg, block, a_h, sign, h_res_mode); - res = new_rd_ia32_Sub64Bit(dbg, irg, block, sub_l, sub_h, sign, sign); - l_res = new_r_Proj(irg, block, res, l_res_mode, pn_ia32_Sub64Bit_low_res); - h_res = new_r_Proj(irg, block, res, h_res_mode, pn_ia32_Sub64Bit_high_res); +static int map_Div(ir_node *call, void *ctx) +{ + ia32_intrinsic_env_t *env = (ia32_intrinsic_env_t*)ctx; + ir_type *method = get_Call_type(call); + ir_mode *h_mode = get_type_mode(get_method_res_type(method, 1)); + ir_node *ptr; + ir_entity *ent; + ir_graph *irg = get_irn_irg(call); + symconst_symbol sym; + + if (mode_is_signed(h_mode)) { + /* 64bit signed Division */ + ent = env->divdi3; + if (ent == NULL) { + /* create library entity */ + ident *id = ID("__divdi3"); + ent = env->divdi3 = create_compilerlib_entity(id, method); + } + } else { + /* 64bit unsigned Division */ + ent = env->udivdi3; + if (ent == NULL) { + /* create library entity */ + ident *id = ID("__udivdi3"); + ent = env->udivdi3 = create_compilerlib_entity(id, method); + } + } - resolve_call(call, l_res, h_res, irg, block); + ptr = get_Call_ptr(call); + sym.entity_p = ent; + ptr = new_r_SymConst(irg, get_irn_mode(ptr), sym, symconst_addr_ent); + set_Call_ptr(call, ptr); return 1; } -typedef enum { - IA32_INTRINSIC_DIV, - IA32_INTRINSIC_MOD, -} ia32_intrinsic_divmod_t; - /** - * Maps a Div/Mod (a_l, a_h, b_l, b_h) + * Maps a Mod. Change into a library call */ -static int DivMod_mapper(ir_node *call, void *ctx, ia32_intrinsic_divmod_t dmtp) { - ia32_intrinsic_env_t *env = ctx; - ir_graph *irg = current_ir_graph; - dbg_info *dbg = get_irn_dbg_info(call); - ir_node *block = get_nodes_block(call); - ir_node **params = get_Call_param_arr(call); - ir_type *method = get_Call_type(call); - ir_node *a_l = params[BINOP_Left_Low]; - ir_node *a_h = params[BINOP_Left_High]; - ir_node *b_l = params[BINOP_Right_Low]; - ir_node *b_h = params[BINOP_Right_High]; - ir_mode *l_res_mode = get_type_mode(get_method_res_type(method, 0)); - ir_mode *h_res_mode = get_type_mode(get_method_res_type(method, 1)); - int mode_bytes = get_mode_size_bytes(ia32_reg_classes[CLASS_ia32_gp].mode); - ir_entity *ent_a = env->irg == irg ? env->ll_div_op1 : NULL; - ir_entity *ent_b = env->irg == irg ? env->ll_div_op2 : NULL; - ir_node *l_res, *h_res, *frame; - ir_node *store_l, *store_h; - ir_node *op_mem[2], *mem, *fa_mem, *fb_mem; - ir_node *fa, *fb, *fres; - - /* allocate memory on frame to store args */ - if (! ent_a) { - ent_a = env->ll_div_op1 = - frame_alloc_area(get_irg_frame_type(irg), 2 * mode_bytes, 16, 0); - env->irg = irg; - } - - if (! ent_b) { - ent_b = env->ll_div_op2 = - frame_alloc_area(get_irg_frame_type(irg), 2 * mode_bytes, 16, 0); - env->irg = irg; - } - - frame = get_irg_frame(irg); - - /* store first arg */ - store_l = new_rd_ia32_l_Store(dbg, irg, block, frame, a_l, get_irg_no_mem(irg)); - set_ia32_frame_ent(store_l, ent_a); - set_ia32_use_frame(store_l); - set_ia32_ls_mode(store_l, get_irn_mode(a_l)); - op_mem[0] = store_l; - - store_h = new_rd_ia32_l_Store(dbg, irg, block, frame, a_h, get_irg_no_mem(irg)); - set_ia32_frame_ent(store_h, ent_a); - add_ia32_am_offs_int(store_h, mode_bytes); - set_ia32_use_frame(store_h); - set_ia32_ls_mode(store_h, get_irn_mode(a_h)); - op_mem[1] = store_h; - - mem = new_r_Sync(irg, block, 2, op_mem); - - /* load first arg into FPU */ - fa = new_rd_ia32_l_vfild(dbg, irg, block, frame, mem); - set_ia32_frame_ent(fa, ent_a); - set_ia32_use_frame(fa); - set_ia32_ls_mode(fa, mode_D); - fa_mem = new_r_Proj(irg, block, fa, mode_M, pn_ia32_l_vfild_M); - fa = new_r_Proj(irg, block, fa, mode_E, pn_ia32_l_vfild_res); - - /* store second arg */ - store_l = new_rd_ia32_l_Store(dbg, irg, block, frame, b_l, get_irg_no_mem(irg)); - set_ia32_frame_ent(store_l, ent_b); - set_ia32_use_frame(store_l); - set_ia32_ls_mode(store_l, get_irn_mode(b_l)); - op_mem[0] = store_l; - - store_h = new_rd_ia32_l_Store(dbg, irg, block, frame, b_h, get_irg_no_mem(irg)); - set_ia32_frame_ent(store_h, ent_b); - add_ia32_am_offs_int(store_h, mode_bytes); - set_ia32_use_frame(store_h); - set_ia32_ls_mode(store_h, get_irn_mode(b_h)); - op_mem[1] = store_h; - - mem = new_r_Sync(irg, block, 2, op_mem); - - /* load second arg into FPU */ - fb = new_rd_ia32_l_vfild(dbg, irg, block, frame, mem); - set_ia32_frame_ent(fb, ent_b); - set_ia32_use_frame(fb); - set_ia32_ls_mode(fb, mode_D); - fb_mem = new_r_Proj(irg, block, fb, mode_M, pn_ia32_l_vfild_M); - fb = new_r_Proj(irg, block, fb, mode_E, pn_ia32_l_vfild_res); - - op_mem[0] = fa_mem; - op_mem[1] = fb_mem; - - mem = new_r_Sync(irg, block, 2, op_mem); - - /* perform division */ - switch (dmtp) { - case IA32_INTRINSIC_DIV: - fres = new_rd_ia32_l_vfdiv(dbg, irg, block, fa, fb); - fres = new_rd_Proj(dbg, irg, block, fres, mode_E, pn_ia32_l_vfdiv_res); - break; - case IA32_INTRINSIC_MOD: - fres = new_rd_ia32_l_vfprem(dbg, irg, block, fa, fb, mode_E); - break; - default: - assert(0); +static int map_Mod(ir_node *call, void *ctx) +{ + ia32_intrinsic_env_t *env = (ia32_intrinsic_env_t*)ctx; + ir_type *method = get_Call_type(call); + ir_mode *h_mode = get_type_mode(get_method_res_type(method, 1)); + ir_node *ptr; + ir_entity *ent; + ir_graph *irg = get_irn_irg(call); + symconst_symbol sym; + + if (mode_is_signed(h_mode)) { + /* 64bit signed Modulo */ + ent = env->moddi3; + if (ent == NULL) { + /* create library entity */ + ident *id = ID("__moddi3"); + ent = env->moddi3 = create_compilerlib_entity(id, method); + } + } else { + /* 64bit signed Modulo */ + ent = env->umoddi3; + if (ent == NULL) { + /* create library entity */ + ident *id = ID("__umoddi3"); + ent = env->umoddi3 = create_compilerlib_entity(id, method); + } } - /* store back result, we use ent_a here */ - fres = new_rd_ia32_l_vfist(dbg, irg, block, frame, fres, mem); - set_ia32_frame_ent(fres, ent_a); - set_ia32_use_frame(fres); - set_ia32_ls_mode(fres, mode_D); - mem = fres; - - /* load low part of the result */ - l_res = new_rd_ia32_l_Load(dbg, irg, block, frame, mem); - set_ia32_frame_ent(l_res, ent_a); - set_ia32_use_frame(l_res); - set_ia32_ls_mode(l_res, l_res_mode); - l_res = new_r_Proj(irg, block, l_res, l_res_mode, pn_ia32_l_Load_res); - - /* load hight part of the result */ - h_res = new_rd_ia32_l_Load(dbg, irg, block, frame, mem); - set_ia32_frame_ent(h_res, ent_a); - add_ia32_am_offs_int(h_res, mode_bytes); - set_ia32_use_frame(h_res); - set_ia32_ls_mode(h_res, h_res_mode); - h_res = new_r_Proj(irg, block, h_res, h_res_mode, pn_ia32_l_Load_res); - - /* lower the call */ - resolve_call(call, l_res, h_res, irg, block); + ptr = get_Call_ptr(call); + sym.entity_p = ent; + ptr = new_r_SymConst(irg, get_irn_mode(ptr), sym, symconst_addr_ent); + set_Call_ptr(call, ptr); return 1; } -static int map_Div(ir_node *call, void *ctx) { - return DivMod_mapper(call, ctx, IA32_INTRINSIC_DIV); -} - -static int map_Mod(ir_node *call, void *ctx) { - return DivMod_mapper(call, ctx, IA32_INTRINSIC_MOD); -} - /** - * Maps a Conv (a_l, a_h) + * Maps a Conv. */ -static int map_Conv(ir_node *call, void *ctx) { - ia32_intrinsic_env_t *env = ctx; - ir_graph *irg = current_ir_graph; - dbg_info *dbg = get_irn_dbg_info(call); - ir_node *block = get_nodes_block(call); - ir_node **params = get_Call_param_arr(call); - ir_type *method = get_Call_type(call); - int n = get_Call_n_params(call); - int gp_bytes = get_mode_size_bytes(ia32_reg_classes[CLASS_ia32_gp].mode); - ir_entity *ent; - ir_node *l_res, *h_res, *frame, *fres; - ir_node *store_l, *store_h; - ir_node *op_mem[2], *mem; +static int map_Conv(ir_node *call, void *ctx) +{ + ir_graph *irg = current_ir_graph; + dbg_info *dbg = get_irn_dbg_info(call); + ir_node *block = get_nodes_block(call); + ir_node **params = get_Call_param_arr(call); + ir_type *method = get_Call_type(call); + int n = get_Call_n_params(call); + ir_node *l_res, *h_res; + (void) ctx; if (n == 1) { + ir_node *float_to_ll; + /* We have a Conv float -> long long here */ ir_node *a_f = params[0]; ir_mode *l_res_mode = get_type_mode(get_method_res_type(method, 0)); @@ -510,113 +454,90 @@ static int map_Conv(ir_node *call, void *ctx) { assert(mode_is_float(get_irn_mode(a_f)) && "unexpected Conv call"); - /* allocate memory on frame to store args */ - ent = env->irg == irg ? env->d_ll_conv : NULL; - if (! ent) { - ent = env->d_ll_conv = frame_alloc_area(get_irg_frame_type(irg), 2 * gp_bytes, 16, 0); - env->irg = irg; + if (mode_is_signed(h_res_mode)) { + /* convert from float to signed 64bit */ + float_to_ll = new_bd_ia32_l_FloattoLL(dbg, block, a_f); + + l_res = new_r_Proj(float_to_ll, l_res_mode, + pn_ia32_l_FloattoLL_res_low); + h_res = new_r_Proj(float_to_ll, h_res_mode, + pn_ia32_l_FloattoLL_res_high); + } else { + /* Convert from float to unsigned 64bit. */ + ir_tarval *flt_tv = new_tarval_from_str("9223372036854775808", 19, ia32_mode_E); + ir_node *flt_corr = new_r_Const(irg, flt_tv); + ir_node *lower_blk = block; + ir_node *upper_blk; + ir_node *cmp, *proj, *cond, *blk, *int_phi, *flt_phi; + ir_node *in[2]; + + part_block(call); + upper_blk = get_nodes_block(call); + + a_f = new_rd_Conv(dbg, upper_blk, a_f, ia32_mode_E); + cmp = new_rd_Cmp(dbg, upper_blk, a_f, flt_corr, ir_relation_less); + cond = new_rd_Cond(dbg, upper_blk, cmp); + in[0] = new_r_Proj(cond, mode_X, pn_Cond_true); + in[1] = new_r_Proj(cond, mode_X, pn_Cond_false); + blk = new_r_Block(irg, 1, &in[1]); + in[1] = new_r_Jmp(blk); + + set_irn_in(lower_blk, 2, in); + + /* create to Phis */ + in[0] = new_r_Const(irg, get_mode_null(h_res_mode)); + in[1] = new_r_Const_long(irg, h_res_mode, 0x80000000); + + int_phi = new_r_Phi(lower_blk, 2, in, h_res_mode); + + in[0] = a_f; + in[1] = new_rd_Sub(dbg, upper_blk, a_f, flt_corr, ia32_mode_E); + + flt_phi = new_r_Phi(lower_blk, 2, in, ia32_mode_E); + + /* fix Phi links for next part_block() */ + if (is_Phi(int_phi)) + add_Block_phi(lower_blk, int_phi); + if (is_Phi(flt_phi)) + add_Block_phi(lower_blk, flt_phi); + + float_to_ll = new_bd_ia32_l_FloattoLL(dbg, lower_blk, flt_phi); + + l_res = new_r_Proj(float_to_ll, l_res_mode, + pn_ia32_l_FloattoLL_res_low); + h_res = new_r_Proj(float_to_ll, h_res_mode, + pn_ia32_l_FloattoLL_res_high); + + h_res = new_rd_Add(dbg, lower_blk, h_res, int_phi, h_res_mode); + + /* move the call and its Proj's to the lower block */ + set_nodes_block(call, lower_blk); + + for (proj = (ir_node*)get_irn_link(call); proj != NULL; + proj = (ir_node*)get_irn_link(proj)) { + set_nodes_block(proj, lower_blk); + } + block = lower_blk; } - - /* Store arg */ - frame = get_irg_frame(irg); - - /* - Now we create a node to move the value from a XMM register into - x87 FPU because it is unknown here, which FPU is used. - This node is killed in transformation phase when not needed. - Otherwise it is split up into a movsd + fld - */ - a_f = new_rd_ia32_l_SSEtoX87(dbg, irg, block, frame, a_f, get_irg_no_mem(irg), mode_D); - set_ia32_frame_ent(a_f, ent); - set_ia32_use_frame(a_f); - set_ia32_ls_mode(a_f, mode_D); - - /* store from FPU as Int */ - a_f = new_rd_ia32_l_vfist(dbg, irg, block, frame, a_f, get_irg_no_mem(irg)); - set_ia32_frame_ent(a_f, ent); - set_ia32_use_frame(a_f); - set_ia32_ls_mode(a_f, mode_D); - mem = a_f; - - /* load low part of the result */ - l_res = new_rd_ia32_l_Load(dbg, irg, block, frame, mem); - set_ia32_frame_ent(l_res, ent); - set_ia32_use_frame(l_res); - set_ia32_ls_mode(l_res, l_res_mode); - l_res = new_r_Proj(irg, block, l_res, l_res_mode, pn_ia32_l_Load_res); - - /* load hight part of the result */ - h_res = new_rd_ia32_l_Load(dbg, irg, block, frame, mem); - set_ia32_frame_ent(h_res, ent); - add_ia32_am_offs_int(h_res, gp_bytes); - set_ia32_use_frame(h_res); - set_ia32_ls_mode(h_res, h_res_mode); - h_res = new_r_Proj(irg, block, h_res, h_res_mode, pn_ia32_l_Load_res); - /* lower the call */ resolve_call(call, l_res, h_res, irg, block); - } - else if (n == 2) { + } else if (n == 2) { + ir_node *ll_to_float; + /* We have a Conv long long -> float here */ ir_node *a_l = params[BINOP_Left_Low]; ir_node *a_h = params[BINOP_Left_High]; - ir_mode *mode_a_l = get_irn_mode(a_l); - ir_mode *mode_a_h = get_irn_mode(a_h); ir_mode *fres_mode = get_type_mode(get_method_res_type(method, 0)); - assert(! mode_is_float(mode_a_l) && ! mode_is_float(mode_a_h) && "unexpected Conv call"); - - /* allocate memory on frame to store args */ - ent = env->irg == irg ? env->ll_d_conv : NULL; - if (! ent) { - ent = env->ll_d_conv = frame_alloc_area(get_irg_frame_type(irg), 2 * gp_bytes, 16, 0); - env->irg = irg; - } + assert(! mode_is_float(get_irn_mode(a_l)) + && ! mode_is_float(get_irn_mode(a_h))); - /* Store arg */ - frame = get_irg_frame(irg); - - /* store first arg (low part) */ - store_l = new_rd_ia32_l_Store(dbg, irg, block, frame, a_l, get_irg_no_mem(irg)); - set_ia32_frame_ent(store_l, ent); - set_ia32_use_frame(store_l); - set_ia32_ls_mode(store_l, get_irn_mode(a_l)); - op_mem[0] = store_l; - - /* store second arg (high part) */ - store_h = new_rd_ia32_l_Store(dbg, irg, block, frame, a_h, get_irg_no_mem(irg)); - set_ia32_frame_ent(store_h, ent); - add_ia32_am_offs_int(store_h, gp_bytes); - set_ia32_use_frame(store_h); - set_ia32_ls_mode(store_h, get_irn_mode(a_h)); - op_mem[1] = store_h; - - mem = new_r_Sync(irg, block, 2, op_mem); - - /* Load arg into x87 FPU (implicit convert) */ - fres = new_rd_ia32_l_vfild(dbg, irg, block, frame, mem); - set_ia32_frame_ent(fres, ent); - set_ia32_use_frame(fres); - set_ia32_ls_mode(fres, mode_D); - mem = new_r_Proj(irg, block, fres, mode_M, pn_ia32_l_vfild_M); - fres = new_r_Proj(irg, block, fres, fres_mode, pn_ia32_l_vfild_res); - - /* - Now we create a node to move the loaded value into a XMM - register because it is unknown here, which FPU is used. - This node is killed in transformation phase when not needed. - Otherwise it is split up into a fst + movsd - */ - fres = new_rd_ia32_l_X87toSSE(dbg, irg, block, frame, fres, mem, fres_mode); - set_ia32_frame_ent(fres, ent); - set_ia32_use_frame(fres); - set_ia32_ls_mode(fres, fres_mode); + ll_to_float = new_bd_ia32_l_LLtoFloat(dbg, block, a_h, a_l, fres_mode); /* lower the call */ - resolve_call(call, fres, NULL, irg, block); - } - else { - assert(0 && "unexpected Conv call"); + resolve_call(call, ll_to_float, NULL, irg, block); + } else { + panic("unexpected Conv call %+F", call); } return 1; @@ -643,18 +564,6 @@ ir_entity *ia32_create_intrinsic_fkt(ir_type *method, const ir_op *op, ent = &i_ents[iro_Sub]; mapper = map_Sub; break; - case iro_Shl: - ent = &i_ents[iro_Shl]; - mapper = map_Shl; - break; - case iro_Shr: - ent = &i_ents[iro_Shr]; - mapper = map_Shr; - break; - case iro_Shrs: - ent = &i_ents[iro_Shrs]; - mapper = map_Shrs; - break; case iro_Mul: ent = &i_ents[iro_Mul]; mapper = map_Mul; @@ -663,10 +572,6 @@ ir_entity *ia32_create_intrinsic_fkt(ir_type *method, const ir_op *op, ent = &i_ents[iro_Minus]; mapper = map_Minus; break; - case iro_Abs: - ent = &i_ents[iro_Abs]; - mapper = map_Abs; - break; case iro_Div: ent = &i_ents[iro_Div]; mapper = map_Div; @@ -685,10 +590,9 @@ ir_entity *ia32_create_intrinsic_fkt(ir_type *method, const ir_op *op, } if (ent && ! *ent) { -#define IDENT(s) new_id_from_chars(s, sizeof(s)-1) - - ident *id = mangle(IDENT("L"), get_op_ident(op)); + ident *id = id_mangle(ID("L"), get_op_ident(op)); *ent = new_entity(get_glob_type(), id, method); + set_entity_visibility(*ent, ir_visibility_private); } elt.i_call.kind = INTRINSIC_CALL;