2 * Copyright (C) 1995-2011 University of Karlsruhe. All right reserved.
4 * This file is part of libFirm.
6 * This file may be distributed and/or modified under the terms of the
7 * GNU General Public License version 2 as published by the Free Software
8 * Foundation and appearing in the file LICENSE.GPL included in the
9 * packaging of this file.
11 * Licensees holding valid libFirm Professional Edition licenses may use
12 * this file in accordance with the libFirm Commercial License.
13 * Agreement provided with the Software.
15 * This file is provided AS IS with NO WARRANTY OF ANY KIND, INCLUDING THE
16 * WARRANTY OF DESIGN, MERCHANTABILITY AND FITNESS FOR A PARTICULAR
22 * @brief Lower double word operations, i.e. 64bit -> 32bit, 32bit -> 16bit etc.
24 * @author Michael Beck
37 #include "irnodeset.h"
38 #include "irgraph_t.h"
43 #include "dbginfo_t.h"
44 #include "iropt_dbg.h"
50 #include "iroptimize.h"
61 /** A map from (op, imode, omode) to Intrinsic functions entities. */
62 static set *intrinsic_fkt;
64 /** A map from (imode, omode) to conv function types. */
65 static set *conv_types;
67 /** A map from a method type to its lowered type. */
68 static pmap *lowered_type;
70 /** A map from a builtin type to its lower and higher type. */
71 static pmap *lowered_builtin_type_high;
72 static pmap *lowered_builtin_type_low;
74 /** The types for the binop and unop intrinsics. */
75 static ir_type *binop_tp_u, *binop_tp_s, *unop_tp_u, *unop_tp_s, *tp_s, *tp_u;
77 static ir_nodeset_t created_mux_nodes;
79 /** the debug handle */
80 DEBUG_ONLY(static firm_dbg_module_t *dbg = NULL;)
83 * An entry in the (op, imode, omode) -> entity map.
85 typedef struct op_mode_entry {
86 const ir_op *op; /**< the op */
87 const ir_mode *imode; /**< the input mode */
88 const ir_mode *omode; /**< the output mode */
89 ir_entity *ent; /**< the associated entity of this (op, imode, omode) triple */
93 * An entry in the (imode, omode) -> tp map.
95 typedef struct conv_tp_entry {
96 const ir_mode *imode; /**< the input mode */
97 const ir_mode *omode; /**< the output mode */
98 ir_type *mtd; /**< the associated method type of this (imode, omode) pair */
102 MUST_BE_LOWERED = 1, /**< graph must be lowered */
103 CF_CHANGED = 2, /**< control flow was changed */
107 * The lower environment.
109 typedef struct lower_dw_env_t {
110 lower64_entry_t **entries; /**< entries per node */
112 struct obstack obst; /**< an obstack holding the temporary data */
113 ir_tarval *tv_mode_bytes; /**< a tarval containing the number of bytes in the lowered modes */
114 ir_tarval *tv_mode_bits; /**< a tarval containing the number of bits in the lowered modes */
115 pdeq *waitq; /**< a wait queue of all nodes that must be handled later */
116 ir_node **lowered_phis; /**< list of lowered phis */
117 ir_mode *high_signed; /**< doubleword signed type */
118 ir_mode *high_unsigned; /**< doubleword unsigned type */
119 ir_mode *low_signed; /**< word signed type */
120 ir_mode *low_unsigned; /**< word unsigned type */
121 ident *first_id; /**< .l for little and .h for big endian */
122 ident *next_id; /**< .h for little and .l for big endian */
123 const lwrdw_param_t *params; /**< transformation parameter */
124 unsigned flags; /**< some flags */
125 unsigned n_entries; /**< number of entries */
128 static lower_dw_env_t *env;
130 static void lower_node(ir_node *node);
133 * Create a method type for a Conv emulation from imode to omode.
135 static ir_type *get_conv_type(ir_mode *imode, ir_mode *omode)
137 conv_tp_entry_t key, *entry;
144 entry = set_insert(conv_tp_entry_t, conv_types, &key, sizeof(key), hash_ptr(imode) ^ hash_ptr(omode));
146 int n_param = 1, n_res = 1;
148 if (imode == env->high_signed || imode == env->high_unsigned)
150 if (omode == env->high_signed || omode == env->high_unsigned)
153 /* create a new one */
154 mtd = new_type_method(n_param, n_res);
156 /* set param types and result types */
158 if (imode == env->high_signed) {
159 if (env->params->little_endian) {
160 set_method_param_type(mtd, n_param++, tp_u);
161 set_method_param_type(mtd, n_param++, tp_s);
163 set_method_param_type(mtd, n_param++, tp_s);
164 set_method_param_type(mtd, n_param++, tp_u);
166 } else if (imode == env->high_unsigned) {
167 set_method_param_type(mtd, n_param++, tp_u);
168 set_method_param_type(mtd, n_param++, tp_u);
170 ir_type *tp = get_type_for_mode(imode);
171 set_method_param_type(mtd, n_param++, tp);
175 if (omode == env->high_signed) {
176 if (env->params->little_endian) {
177 set_method_res_type(mtd, n_res++, tp_u);
178 set_method_res_type(mtd, n_res++, tp_s);
180 set_method_res_type(mtd, n_res++, tp_s);
181 set_method_res_type(mtd, n_res++, tp_u);
183 } else if (omode == env->high_unsigned) {
184 set_method_res_type(mtd, n_res++, tp_u);
185 set_method_res_type(mtd, n_res++, tp_u);
187 ir_type *tp = get_type_for_mode(omode);
188 set_method_res_type(mtd, n_res++, tp);
198 * Add an additional control flow input to a block.
199 * Patch all Phi nodes. The new Phi inputs are copied from
200 * old input number nr.
202 static void add_block_cf_input_nr(ir_node *block, int nr, ir_node *cf)
204 int i, arity = get_irn_arity(block);
209 NEW_ARR_A(ir_node *, in, arity + 1);
210 for (i = 0; i < arity; ++i)
211 in[i] = get_irn_n(block, i);
214 set_irn_in(block, i + 1, in);
216 foreach_out_edge(block, edge) {
217 ir_node *phi = get_edge_src_irn(edge);
221 for (i = 0; i < arity; ++i)
222 in[i] = get_irn_n(phi, i);
224 set_irn_in(phi, i + 1, in);
229 * Add an additional control flow input to a block.
230 * Patch all Phi nodes. The new Phi inputs are copied from
231 * old input from cf tmpl.
233 static void add_block_cf_input(ir_node *block, ir_node *tmpl, ir_node *cf)
235 int i, arity = get_irn_arity(block);
238 for (i = 0; i < arity; ++i) {
239 if (get_irn_n(block, i) == tmpl) {
245 add_block_cf_input_nr(block, nr, cf);
249 * Return the "operational" mode of a Firm node.
251 static ir_mode *get_irn_op_mode(ir_node *node)
253 switch (get_irn_opcode(node)) {
255 return get_Load_mode(node);
257 return get_irn_mode(get_Store_value(node));
259 return get_irn_mode(get_Div_left(node));
261 return get_irn_mode(get_Mod_left(node));
263 return get_irn_mode(get_Cmp_left(node));
265 return get_irn_mode(node);
270 * Walker, prepare the node links and determine which nodes need to be lowered
273 static void prepare_links(ir_node *node)
275 ir_mode *mode = get_irn_op_mode(node);
276 lower64_entry_t *link;
278 if (mode == env->high_signed || mode == env->high_unsigned) {
279 unsigned idx = get_irn_idx(node);
280 /* ok, found a node that will be lowered */
281 link = OALLOCZ(&env->obst, lower64_entry_t);
283 if (idx >= env->n_entries) {
284 /* enlarge: this happens only for Rotl nodes which is RARELY */
285 unsigned old = env->n_entries;
286 unsigned n_idx = idx + (idx >> 3);
288 ARR_RESIZE(lower64_entry_t *, env->entries, n_idx);
289 memset(&env->entries[old], 0, (n_idx - old) * sizeof(env->entries[0]));
290 env->n_entries = n_idx;
292 env->entries[idx] = link;
293 env->flags |= MUST_BE_LOWERED;
294 } else if (is_Conv(node)) {
295 /* Conv nodes have two modes */
296 ir_node *pred = get_Conv_op(node);
297 mode = get_irn_mode(pred);
299 if (mode == env->high_signed || mode == env->high_unsigned) {
300 /* must lower this node either but don't need a link */
301 env->flags |= MUST_BE_LOWERED;
304 } else if (is_Call(node)) {
305 /* Special case: If the result of the Call is never used, we won't
306 * find a Proj with a mode that potentially triggers MUST_BE_LOWERED
307 * to be set. Thus, if we see a call, we check its result types and
308 * decide whether MUST_BE_LOWERED has to be set.
310 ir_type *tp = get_Call_type(node);
313 n_res = get_method_n_ress(tp);
314 for (i = 0; i < n_res; ++i) {
315 ir_type *rtp = get_method_res_type(tp, i);
317 if (is_Primitive_type(rtp)) {
318 ir_mode *rmode = get_type_mode(rtp);
320 if (rmode == env->high_signed || rmode == env->high_unsigned) {
321 env->flags |= MUST_BE_LOWERED;
328 lower64_entry_t *get_node_entry(ir_node *node)
330 unsigned idx = get_irn_idx(node);
331 assert(idx < env->n_entries);
332 return env->entries[idx];
335 void ir_set_dw_lowered(ir_node *old, ir_node *new_low, ir_node *new_high)
337 lower64_entry_t *entry = get_node_entry(old);
338 entry->low_word = new_low;
339 entry->high_word = new_high;
342 ir_mode *ir_get_low_unsigned_mode(void)
344 return env->low_unsigned;
348 * Translate a Constant: create two.
350 static void lower_Const(ir_node *node, ir_mode *mode)
352 ir_graph *irg = get_irn_irg(node);
353 dbg_info *dbg = get_irn_dbg_info(node);
354 ir_mode *low_mode = env->low_unsigned;
355 ir_tarval *tv = get_Const_tarval(node);
356 ir_tarval *tv_l = tarval_convert_to(tv, low_mode);
357 ir_node *res_low = new_rd_Const(dbg, irg, tv_l);
358 ir_tarval *tv_shrs = tarval_shrs(tv, env->tv_mode_bits);
359 ir_tarval *tv_h = tarval_convert_to(tv_shrs, mode);
360 ir_node *res_high = new_rd_Const(dbg, irg, tv_h);
362 ir_set_dw_lowered(node, res_low, res_high);
366 * Translate a Load: create two.
368 static void lower_Load(ir_node *node, ir_mode *mode)
370 ir_mode *low_mode = env->low_unsigned;
371 ir_graph *irg = get_irn_irg(node);
372 ir_node *adr = get_Load_ptr(node);
373 ir_node *mem = get_Load_mem(node);
378 ir_node *block = get_nodes_block(node);
379 ir_cons_flags volatility = get_Load_volatility(node) == volatility_is_volatile
380 ? cons_volatile : cons_none;
382 if (env->params->little_endian) {
384 high = new_r_Add(block, adr, new_r_Const(irg, env->tv_mode_bytes), get_irn_mode(adr));
386 low = new_r_Add(block, adr, new_r_Const(irg, env->tv_mode_bytes), get_irn_mode(adr));
390 /* create two loads */
391 dbg = get_irn_dbg_info(node);
392 low = new_rd_Load(dbg, block, mem, low, low_mode, volatility);
393 proj_m = new_r_Proj(low, mode_M, pn_Load_M);
394 high = new_rd_Load(dbg, block, proj_m, high, mode, volatility);
396 foreach_out_edge_safe(node, edge) {
397 ir_node *proj = get_edge_src_irn(edge);
401 switch (get_Proj_proj(proj)) {
402 case pn_Load_M: /* Memory result. */
403 /* put it to the second one */
404 set_Proj_pred(proj, high);
406 case pn_Load_X_except: /* Execution result if exception occurred. */
407 /* put it to the first one */
408 set_Proj_pred(proj, low);
410 case pn_Load_res: { /* Result of load operation. */
411 ir_node *res_low = new_r_Proj(low, low_mode, pn_Load_res);
412 ir_node *res_high = new_r_Proj(high, mode, pn_Load_res);
413 ir_set_dw_lowered(proj, res_low, res_high);
417 assert(0 && "unexpected Proj number");
419 /* mark this proj: we have handled it already, otherwise we might fall
420 * into out new nodes. */
421 mark_irn_visited(proj);
426 * Translate a Store: create two.
428 static void lower_Store(ir_node *node, ir_mode *mode)
431 ir_node *block, *adr, *mem;
432 ir_node *low, *high, *proj_m;
434 ir_node *value = get_Store_value(node);
435 const lower64_entry_t *entry = get_node_entry(value);
436 ir_cons_flags volatility = get_Store_volatility(node) == volatility_is_volatile
437 ? cons_volatile : cons_none;
442 if (! entry->low_word) {
443 /* not ready yet, wait */
444 pdeq_putr(env->waitq, node);
448 irg = get_irn_irg(node);
449 adr = get_Store_ptr(node);
450 mem = get_Store_mem(node);
451 block = get_nodes_block(node);
453 if (env->params->little_endian) {
455 high = new_r_Add(block, adr, new_r_Const(irg, env->tv_mode_bytes), get_irn_mode(adr));
457 low = new_r_Add(block, adr, new_r_Const(irg, env->tv_mode_bytes), get_irn_mode(adr));
461 /* create two Stores */
462 dbg = get_irn_dbg_info(node);
463 low = new_rd_Store(dbg, block, mem, low, entry->low_word, volatility);
464 proj_m = new_r_Proj(low, mode_M, pn_Store_M);
465 high = new_rd_Store(dbg, block, proj_m, high, entry->high_word, volatility);
467 foreach_out_edge_safe(node, edge) {
468 ir_node *proj = get_edge_src_irn(edge);
472 switch (get_Proj_proj(proj)) {
473 case pn_Store_M: /* Memory result. */
474 /* put it to the second one */
475 set_Proj_pred(proj, high);
477 case pn_Store_X_except: /* Execution result if exception occurred. */
478 /* put it to the first one */
479 set_Proj_pred(proj, low);
482 assert(0 && "unexpected Proj number");
484 /* mark this proj: we have handled it already, otherwise we might fall into
486 mark_irn_visited(proj);
491 * Return a node containing the address of the intrinsic emulation function.
493 * @param method the method type of the emulation function
494 * @param op the emulated ir_op
495 * @param imode the input mode of the emulated opcode
496 * @param omode the output mode of the emulated opcode
497 * @param env the lower environment
499 static ir_node *get_intrinsic_address(ir_type *method, ir_op *op,
500 ir_mode *imode, ir_mode *omode)
504 op_mode_entry_t key, *entry;
511 entry = set_insert(op_mode_entry_t, intrinsic_fkt, &key, sizeof(key),
512 hash_ptr(op) ^ hash_ptr(imode) ^ (hash_ptr(omode) << 8));
514 /* create a new one */
515 ent = env->params->create_intrinsic(method, op, imode, omode, env->params->ctx);
517 assert(ent && "Intrinsic creator must return an entity");
523 return new_r_SymConst(env->irg, mode_P_code, sym, symconst_addr_ent);
529 * Create an intrinsic Call.
531 static void lower_Div(ir_node *node, ir_mode *mode)
533 ir_node *left = get_Div_left(node);
534 ir_node *right = get_Div_right(node);
535 ir_node *block = get_nodes_block(node);
536 dbg_info *dbgi = get_irn_dbg_info(node);
537 ir_type *mtp = mode_is_signed(mode) ? binop_tp_s : binop_tp_u;
538 ir_mode *opmode = get_irn_op_mode(node);
539 ir_node *addr = get_intrinsic_address(mtp, get_irn_op(node), opmode, opmode);
544 if (env->params->little_endian) {
545 in[0] = get_lowered_low(left);
546 in[1] = get_lowered_high(left);
547 in[2] = get_lowered_low(right);
548 in[3] = get_lowered_high(right);
550 in[0] = get_lowered_high(left);
551 in[1] = get_lowered_low(left);
552 in[2] = get_lowered_high(right);
553 in[3] = get_lowered_low(right);
555 call = new_rd_Call(dbgi, block, get_Div_mem(node), addr, 4, in, mtp);
556 resproj = new_r_Proj(call, mode_T, pn_Call_T_result);
557 set_irn_pinned(call, get_irn_pinned(node));
559 foreach_out_edge_safe(node, edge) {
560 ir_node *proj = get_edge_src_irn(edge);
564 switch (get_Proj_proj(proj)) {
565 case pn_Div_M: /* Memory result. */
566 /* reroute to the call */
567 set_Proj_pred(proj, call);
568 set_Proj_proj(proj, pn_Call_M);
570 case pn_Div_X_regular:
571 set_Proj_pred(proj, call);
572 set_Proj_proj(proj, pn_Call_X_regular);
574 case pn_Div_X_except:
575 set_Proj_pred(proj, call);
576 set_Proj_proj(proj, pn_Call_X_except);
579 if (env->params->little_endian) {
580 ir_node *res_low = new_r_Proj(resproj, env->low_unsigned, 0);
581 ir_node *res_high = new_r_Proj(resproj, mode, 1);
582 ir_set_dw_lowered(proj, res_low, res_high);
584 ir_node *res_low = new_r_Proj(resproj, env->low_unsigned, 1);
585 ir_node *res_high = new_r_Proj(resproj, mode, 0);
586 ir_set_dw_lowered(proj, res_low, res_high);
590 assert(0 && "unexpected Proj number");
592 /* mark this proj: we have handled it already, otherwise we might fall into
594 mark_irn_visited(proj);
601 * Create an intrinsic Call.
603 static void lower_Mod(ir_node *node, ir_mode *mode)
605 ir_node *left = get_Mod_left(node);
606 ir_node *right = get_Mod_right(node);
607 dbg_info *dbgi = get_irn_dbg_info(node);
608 ir_node *block = get_nodes_block(node);
609 ir_type *mtp = mode_is_signed(mode) ? binop_tp_s : binop_tp_u;
610 ir_mode *opmode = get_irn_op_mode(node);
611 ir_node *addr = get_intrinsic_address(mtp, get_irn_op(node), opmode, opmode);
616 if (env->params->little_endian) {
617 in[0] = get_lowered_low(left);
618 in[1] = get_lowered_high(left);
619 in[2] = get_lowered_low(right);
620 in[3] = get_lowered_high(right);
622 in[0] = get_lowered_high(left);
623 in[1] = get_lowered_low(left);
624 in[2] = get_lowered_high(right);
625 in[3] = get_lowered_low(right);
627 call = new_rd_Call(dbgi, block, get_Mod_mem(node), addr, 4, in, mtp);
628 resproj = new_r_Proj(call, mode_T, pn_Call_T_result);
629 set_irn_pinned(call, get_irn_pinned(node));
631 foreach_out_edge_safe(node, edge) {
632 ir_node *proj = get_edge_src_irn(edge);
636 switch (get_Proj_proj(proj)) {
637 case pn_Mod_M: /* Memory result. */
638 /* reroute to the call */
639 set_Proj_pred(proj, call);
640 set_Proj_proj(proj, pn_Call_M);
642 case pn_Div_X_regular:
643 set_Proj_pred(proj, call);
644 set_Proj_proj(proj, pn_Call_X_regular);
646 case pn_Mod_X_except:
647 set_Proj_pred(proj, call);
648 set_Proj_proj(proj, pn_Call_X_except);
651 if (env->params->little_endian) {
652 ir_node *res_low = new_r_Proj(resproj, env->low_unsigned, 0);
653 ir_node *res_high = new_r_Proj(resproj, mode, 1);
654 ir_set_dw_lowered(proj, res_low, res_high);
656 ir_node *res_low = new_r_Proj(resproj, env->low_unsigned, 1);
657 ir_node *res_high = new_r_Proj(resproj, mode, 0);
658 ir_set_dw_lowered(proj, res_low, res_high);
662 assert(0 && "unexpected Proj number");
664 /* mark this proj: we have handled it already, otherwise we might fall
665 * into out new nodes. */
666 mark_irn_visited(proj);
673 * Create an intrinsic Call.
675 static void lower_binop(ir_node *node, ir_mode *mode)
677 ir_node *left = get_binop_left(node);
678 ir_node *right = get_binop_right(node);
679 dbg_info *dbgi = get_irn_dbg_info(node);
680 ir_node *block = get_nodes_block(node);
681 ir_graph *irg = get_irn_irg(block);
682 ir_type *mtp = mode_is_signed(mode) ? binop_tp_s : binop_tp_u;
683 ir_node *addr = get_intrinsic_address(mtp, get_irn_op(node), mode, mode);
688 if (env->params->little_endian) {
689 in[0] = get_lowered_low(left);
690 in[1] = get_lowered_high(left);
691 in[2] = get_lowered_low(right);
692 in[3] = get_lowered_high(right);
694 in[0] = get_lowered_high(left);
695 in[1] = get_lowered_low(left);
696 in[2] = get_lowered_high(right);
697 in[3] = get_lowered_low(right);
699 call = new_rd_Call(dbgi, block, get_irg_no_mem(irg), addr, 4, in, mtp);
700 resproj = new_r_Proj(call, mode_T, pn_Call_T_result);
701 set_irn_pinned(call, get_irn_pinned(node));
703 if (env->params->little_endian) {
704 ir_node *res_low = new_r_Proj(resproj, env->low_unsigned, 0);
705 ir_node *res_high = new_r_Proj(resproj, mode, 1);
706 ir_set_dw_lowered(node, res_low, res_high);
708 ir_node *res_low = new_r_Proj(resproj, env->low_unsigned, 1);
709 ir_node *res_high = new_r_Proj(resproj, mode, 0);
710 ir_set_dw_lowered(node, res_low, res_high);
714 static ir_node *create_conv(ir_node *block, ir_node *node, ir_mode *dest_mode)
716 if (get_irn_mode(node) == dest_mode)
718 return new_r_Conv(block, node, dest_mode);
722 * Moves node and all predecessors of node from from_bl to to_bl.
723 * Does not move predecessors of Phi nodes (or block nodes).
725 static void move(ir_node *node, ir_node *from_bl, ir_node *to_bl)
730 set_nodes_block(node, to_bl);
733 if (get_irn_mode(node) == mode_T) {
734 foreach_out_edge(node, edge) {
735 ir_node *proj = get_edge_src_irn(edge);
738 move(proj, from_bl, to_bl);
742 /* We must not move predecessors of Phi nodes, even if they are in
743 * from_bl. (because these are values from an earlier loop iteration
744 * which are not predecessors of node here)
750 arity = get_irn_arity(node);
751 for (i = 0; i < arity; i++) {
752 ir_node *pred = get_irn_n(node, i);
753 ir_mode *pred_mode = get_irn_mode(pred);
754 if (get_nodes_block(pred) == from_bl)
755 move(pred, from_bl, to_bl);
756 if (pred_mode == env->high_signed || pred_mode == env->high_unsigned) {
757 ir_node *pred_low = get_lowered_low(pred);
758 ir_node *pred_high = get_lowered_high(pred);
759 if (get_nodes_block(pred_low) == from_bl)
760 move(pred_low, from_bl, to_bl);
761 if (pred_high != NULL && get_nodes_block(pred_high) == from_bl)
762 move(pred_high, from_bl, to_bl);
768 * We need a custom version of part_block_edges because during transformation
769 * not all data-dependencies are explicit yet if a lowered nodes users are not
771 * We can fix this by modifying move to look for such implicit dependencies.
772 * Additionally we have to keep the proj_2_block map updated
774 static ir_node *part_block_dw(ir_node *node)
776 ir_graph *irg = get_irn_irg(node);
777 ir_node *old_block = get_nodes_block(node);
778 int n_cfgpreds = get_Block_n_cfgpreds(old_block);
779 ir_node **cfgpreds = get_Block_cfgpred_arr(old_block);
780 ir_node *new_block = new_r_Block(irg, n_cfgpreds, cfgpreds);
782 /* old_block has no predecessors anymore for now */
783 set_irn_in(old_block, 0, NULL);
785 /* move node and its predecessors to new_block */
786 move(node, old_block, new_block);
788 /* move Phi nodes to new_block */
789 foreach_out_edge_safe(old_block, edge) {
790 ir_node *phi = get_edge_src_irn(edge);
793 set_nodes_block(phi, new_block);
798 typedef ir_node* (*new_rd_shr_func)(dbg_info *dbgi, ir_node *block,
799 ir_node *left, ir_node *right,
802 static void lower_shr_helper(ir_node *node, ir_mode *mode,
803 new_rd_shr_func new_rd_shrs)
805 ir_node *right = get_binop_right(node);
806 ir_node *left = get_binop_left(node);
807 ir_mode *shr_mode = get_irn_mode(node);
808 unsigned modulo_shift = get_mode_modulo_shift(shr_mode);
809 ir_mode *low_unsigned = env->low_unsigned;
810 unsigned modulo_shift2 = get_mode_modulo_shift(mode);
811 ir_graph *irg = get_irn_irg(node);
812 ir_node *left_low = get_lowered_low(left);
813 ir_node *left_high = get_lowered_high(left);
814 dbg_info *dbgi = get_irn_dbg_info(node);
815 ir_node *lower_block;
825 ir_node *lower_in[2];
826 ir_node *phi_low_in[2];
827 ir_node *phi_high_in[2];
829 /* this version is optimized for modulo shift architectures
830 * (and can't handle anything else) */
831 if (modulo_shift != get_mode_size_bits(shr_mode)
832 || modulo_shift2<<1 != modulo_shift) {
833 panic("Shr lowering only implemented for modulo shift shr operations");
835 if (!is_po2(modulo_shift) || !is_po2(modulo_shift2)) {
836 panic("Shr lowering only implemented for power-of-2 modes");
838 /* without 2-complement the -x instead of (bit_width-x) trick won't work */
839 if (get_mode_arithmetic(shr_mode) != irma_twos_complement) {
840 panic("Shr lowering only implemented for two-complement modes");
843 block = get_nodes_block(node);
845 /* if the right operand is a 64bit value, we're only interested in the
847 if (get_irn_mode(right) == env->high_unsigned) {
848 right = get_lowered_low(right);
850 /* shift should never have signed mode on the right */
851 assert(get_irn_mode(right) != env->high_signed);
852 right = create_conv(block, right, low_unsigned);
855 lower_block = part_block_dw(node);
856 env->flags |= CF_CHANGED;
857 block = get_nodes_block(node);
859 /* add a Cmp to test if highest bit is set <=> whether we shift more
860 * than half the word width */
861 cnst = new_r_Const_long(irg, low_unsigned, modulo_shift2);
862 andn = new_r_And(block, right, cnst, low_unsigned);
863 cnst = new_r_Const(irg, get_mode_null(low_unsigned));
864 cmp = new_rd_Cmp(dbgi, block, andn, cnst, ir_relation_equal);
865 cond = new_rd_Cond(dbgi, block, cmp);
866 proj_true = new_r_Proj(cond, mode_X, pn_Cond_true);
867 proj_false = new_r_Proj(cond, mode_X, pn_Cond_false);
869 /* the true block => shift_width < 1word */
871 /* In theory the low value (for 64bit shifts) is:
872 * Or(High << (32-x)), Low >> x)
873 * In practice High << 32-x will fail when x is zero (since we have
874 * modulo shift and 32 will be 0). So instead we use:
875 * Or(High<<1<<~x, Low >> x)
877 ir_node *in[1] = { proj_true };
878 ir_node *block_true = new_r_Block(irg, ARRAY_SIZE(in), in);
879 ir_node *res_high = new_rd_shrs(dbgi, block_true, left_high,
881 ir_node *shift_low = new_rd_Shr(dbgi, block_true, left_low, right,
883 ir_node *not_shiftval = new_rd_Not(dbgi, block_true, right,
885 ir_node *conv = create_conv(block_true, left_high,
887 ir_node *one = new_r_Const(irg, get_mode_one(low_unsigned));
888 ir_node *carry0 = new_rd_Shl(dbgi, block_true, conv, one,
890 ir_node *carry1 = new_rd_Shl(dbgi, block_true, carry0,
891 not_shiftval, low_unsigned);
892 ir_node *res_low = new_rd_Or(dbgi, block_true, shift_low, carry1,
894 lower_in[0] = new_r_Jmp(block_true);
895 phi_low_in[0] = res_low;
896 phi_high_in[0] = res_high;
899 /* false block => shift_width > 1word */
901 ir_node *in[1] = { proj_false };
902 ir_node *block_false = new_r_Block(irg, ARRAY_SIZE(in), in);
903 ir_node *conv = create_conv(block_false, left_high, low_unsigned);
904 ir_node *res_low = new_rd_shrs(dbgi, block_false, conv, right,
906 int cnsti = modulo_shift2-1;
907 ir_node *cnst2 = new_r_Const_long(irg, low_unsigned, cnsti);
909 if (new_rd_shrs == new_rd_Shrs) {
910 res_high = new_rd_shrs(dbgi, block_false, left_high, cnst2, mode);
912 res_high = new_r_Const(irg, get_mode_null(mode));
914 lower_in[1] = new_r_Jmp(block_false);
915 phi_low_in[1] = res_low;
916 phi_high_in[1] = res_high;
919 /* patch lower block */
920 set_irn_in(lower_block, ARRAY_SIZE(lower_in), lower_in);
921 phi_low = new_r_Phi(lower_block, ARRAY_SIZE(phi_low_in), phi_low_in,
923 phi_high = new_r_Phi(lower_block, ARRAY_SIZE(phi_high_in), phi_high_in,
925 ir_set_dw_lowered(node, phi_low, phi_high);
928 static void lower_Shr(ir_node *node, ir_mode *mode)
930 lower_shr_helper(node, mode, new_rd_Shr);
933 static void lower_Shrs(ir_node *node, ir_mode *mode)
935 lower_shr_helper(node, mode, new_rd_Shrs);
938 static void lower_Shl(ir_node *node, ir_mode *mode)
940 ir_node *right = get_binop_right(node);
941 ir_node *left = get_binop_left(node);
942 ir_mode *shr_mode = get_irn_mode(node);
943 unsigned modulo_shift = get_mode_modulo_shift(shr_mode);
944 ir_mode *low_unsigned = env->low_unsigned;
945 unsigned modulo_shift2 = get_mode_modulo_shift(mode);
946 ir_graph *irg = get_irn_irg(node);
947 ir_node *left_low = get_lowered_low(left);
948 ir_node *left_high = get_lowered_high(left);
949 dbg_info *dbgi = get_irn_dbg_info(node);
950 ir_node *lower_block = get_nodes_block(node);
960 ir_node *lower_in[2];
961 ir_node *phi_low_in[2];
962 ir_node *phi_high_in[2];
964 /* this version is optimized for modulo shift architectures
965 * (and can't handle anything else) */
966 if (modulo_shift != get_mode_size_bits(shr_mode)
967 || modulo_shift2<<1 != modulo_shift) {
968 panic("Shl lowering only implemented for modulo shift shl operations");
970 if (!is_po2(modulo_shift) || !is_po2(modulo_shift2)) {
971 panic("Shl lowering only implemented for power-of-2 modes");
973 /* without 2-complement the -x instead of (bit_width-x) trick won't work */
974 if (get_mode_arithmetic(shr_mode) != irma_twos_complement) {
975 panic("Shl lowering only implemented for two-complement modes");
978 /* if the right operand is a 64bit value, we're only interested in the
980 if (get_irn_mode(right) == env->high_unsigned) {
981 right = get_lowered_low(right);
983 /* shift should never have signed mode on the right */
984 assert(get_irn_mode(right) != env->high_signed);
985 right = create_conv(lower_block, right, low_unsigned);
989 env->flags |= CF_CHANGED;
990 block = get_nodes_block(node);
992 /* add a Cmp to test if highest bit is set <=> whether we shift more
993 * than half the word width */
994 cnst = new_r_Const_long(irg, low_unsigned, modulo_shift2);
995 andn = new_r_And(block, right, cnst, low_unsigned);
996 cnst = new_r_Const(irg, get_mode_null(low_unsigned));
997 cmp = new_rd_Cmp(dbgi, block, andn, cnst, ir_relation_equal);
998 cond = new_rd_Cond(dbgi, block, cmp);
999 proj_true = new_r_Proj(cond, mode_X, pn_Cond_true);
1000 proj_false = new_r_Proj(cond, mode_X, pn_Cond_false);
1002 /* the true block => shift_width < 1word */
1004 ir_node *in[1] = { proj_true };
1005 ir_node *block_true = new_r_Block(irg, ARRAY_SIZE(in), in);
1007 ir_node *res_low = new_rd_Shl(dbgi, block_true, left_low,
1008 right, low_unsigned);
1009 ir_node *shift_high = new_rd_Shl(dbgi, block_true, left_high, right,
1011 ir_node *not_shiftval = new_rd_Not(dbgi, block_true, right,
1013 ir_node *conv = create_conv(block_true, left_low, mode);
1014 ir_node *one = new_r_Const(irg, get_mode_one(low_unsigned));
1015 ir_node *carry0 = new_rd_Shr(dbgi, block_true, conv, one, mode);
1016 ir_node *carry1 = new_rd_Shr(dbgi, block_true, carry0,
1017 not_shiftval, mode);
1018 ir_node *res_high = new_rd_Or(dbgi, block_true, shift_high, carry1,
1020 lower_in[0] = new_r_Jmp(block_true);
1021 phi_low_in[0] = res_low;
1022 phi_high_in[0] = res_high;
1025 /* false block => shift_width > 1word */
1027 ir_node *in[1] = { proj_false };
1028 ir_node *block_false = new_r_Block(irg, ARRAY_SIZE(in), in);
1029 ir_node *res_low = new_r_Const(irg, get_mode_null(low_unsigned));
1030 ir_node *conv = create_conv(block_false, left_low, mode);
1031 ir_node *res_high = new_rd_Shl(dbgi, block_false, conv, right, mode);
1032 lower_in[1] = new_r_Jmp(block_false);
1033 phi_low_in[1] = res_low;
1034 phi_high_in[1] = res_high;
1037 /* patch lower block */
1038 set_irn_in(lower_block, ARRAY_SIZE(lower_in), lower_in);
1039 phi_low = new_r_Phi(lower_block, ARRAY_SIZE(phi_low_in), phi_low_in,
1041 phi_high = new_r_Phi(lower_block, ARRAY_SIZE(phi_high_in), phi_high_in,
1043 ir_set_dw_lowered(node, phi_low, phi_high);
1047 * Rebuild Rotl nodes into Or(Shl, Shr) and prepare all nodes.
1049 static void prepare_links_and_handle_rotl(ir_node *node, void *data)
1052 if (is_Rotl(node)) {
1053 ir_mode *mode = get_irn_op_mode(node);
1055 ir_node *left, *shl, *shr, *ornode, *block, *sub, *c;
1056 ir_mode *omode, *rmode;
1059 optimization_state_t state;
1061 if (mode != env->high_signed && mode != env->high_unsigned) {
1062 prepare_links(node);
1066 /* replace the Rotl(x,y) by an Or(Shl(x,y), Shr(x,64-y)) */
1067 right = get_Rotl_right(node);
1068 irg = get_irn_irg(node);
1069 dbg = get_irn_dbg_info(node);
1070 omode = get_irn_mode(node);
1071 left = get_Rotl_left(node);
1072 block = get_nodes_block(node);
1073 shl = new_rd_Shl(dbg, block, left, right, omode);
1074 rmode = get_irn_mode(right);
1075 c = new_r_Const_long(irg, rmode, get_mode_size_bits(omode));
1076 sub = new_rd_Sub(dbg, block, c, right, rmode);
1077 shr = new_rd_Shr(dbg, block, left, sub, omode);
1079 /* switch optimization off here, or we will get the Rotl back */
1080 save_optimization_state(&state);
1081 set_opt_algebraic_simplification(0);
1082 ornode = new_rd_Or(dbg, block, shl, shr, omode);
1083 restore_optimization_state(&state);
1085 exchange(node, ornode);
1087 /* do lowering on the new nodes */
1092 prepare_links(ornode);
1096 prepare_links(node);
1100 * Translate an Unop.
1102 * Create an intrinsic Call.
1104 static void lower_unop(ir_node *node, ir_mode *mode)
1106 ir_node *op = get_unop_op(node);
1107 dbg_info *dbgi = get_irn_dbg_info(node);
1108 ir_node *block = get_nodes_block(node);
1109 ir_graph *irg = get_irn_irg(block);
1110 ir_type *mtp = mode_is_signed(mode) ? unop_tp_s : unop_tp_u;
1111 ir_op *irop = get_irn_op(node);
1112 ir_node *addr = get_intrinsic_address(mtp, irop, mode, mode);
1113 ir_node *nomem = get_irg_no_mem(irg);
1118 if (env->params->little_endian) {
1119 in[0] = get_lowered_low(op);
1120 in[1] = get_lowered_high(op);
1122 in[0] = get_lowered_high(op);
1123 in[1] = get_lowered_low(op);
1125 call = new_rd_Call(dbgi, block, nomem, addr, 2, in, mtp);
1126 resproj = new_r_Proj(call, mode_T, pn_Call_T_result);
1127 set_irn_pinned(call, get_irn_pinned(node));
1129 if (env->params->little_endian) {
1130 ir_node *res_low = new_r_Proj(resproj, env->low_unsigned, 0);
1131 ir_node *res_high = new_r_Proj(resproj, mode, 1);
1132 ir_set_dw_lowered(node, res_low, res_high);
1134 ir_node *res_low = new_r_Proj(resproj, env->low_unsigned, 1);
1135 ir_node *res_high = new_r_Proj(resproj, mode, 0);
1136 ir_set_dw_lowered(node, res_low, res_high);
1141 * Translate a logical binop.
1143 * Create two logical binops.
1145 static void lower_binop_logical(ir_node *node, ir_mode *mode,
1146 ir_node *(*constr_rd)(dbg_info *db, ir_node *block, ir_node *op1, ir_node *op2, ir_mode *mode) )
1148 ir_node *left = get_binop_left(node);
1149 ir_node *right = get_binop_right(node);
1150 const lower64_entry_t *left_entry = get_node_entry(left);
1151 const lower64_entry_t *right_entry = get_node_entry(right);
1152 dbg_info *dbgi = get_irn_dbg_info(node);
1153 ir_node *block = get_nodes_block(node);
1155 = constr_rd(dbgi, block, left_entry->low_word, right_entry->low_word,
1158 = constr_rd(dbgi, block, left_entry->high_word, right_entry->high_word,
1160 ir_set_dw_lowered(node, res_low, res_high);
1163 static void lower_And(ir_node *node, ir_mode *mode)
1165 lower_binop_logical(node, mode, new_rd_And);
1168 static void lower_Or(ir_node *node, ir_mode *mode)
1170 lower_binop_logical(node, mode, new_rd_Or);
1173 static void lower_Eor(ir_node *node, ir_mode *mode)
1175 lower_binop_logical(node, mode, new_rd_Eor);
1181 * Create two logical Nots.
1183 static void lower_Not(ir_node *node, ir_mode *mode)
1185 ir_node *op = get_Not_op(node);
1186 const lower64_entry_t *op_entry = get_node_entry(op);
1187 dbg_info *dbgi = get_irn_dbg_info(node);
1188 ir_node *block = get_nodes_block(node);
1190 = new_rd_Not(dbgi, block, op_entry->low_word, env->low_unsigned);
1192 = new_rd_Not(dbgi, block, op_entry->high_word, mode);
1193 ir_set_dw_lowered(node, res_low, res_high);
1196 static void lower_Proj(ir_node *node, ir_mode *op_mode)
1198 ir_mode *mode = get_irn_mode(node);
1201 if (mode != env->high_signed && mode != env->high_unsigned)
1204 pred = get_Proj_pred(node);
1205 if (is_Tuple(pred)) {
1206 long pn = get_Proj_proj(node);
1207 ir_node *op = get_irn_n(pred, pn);
1208 const lower64_entry_t *entry = get_node_entry(op);
1209 ir_set_dw_lowered(node, entry->low_word, entry->high_word);
1213 static bool is_equality_cmp(const ir_node *node)
1215 ir_relation relation = get_Cmp_relation(node);
1216 ir_node *left = get_Cmp_left(node);
1217 ir_node *right = get_Cmp_right(node);
1218 ir_mode *mode = get_irn_mode(left);
1220 /* this probably makes no sense if unordered is involved */
1221 assert(!mode_is_float(mode));
1223 if (relation == ir_relation_equal || relation == ir_relation_less_greater)
1226 if (!is_Const(right) || !is_Const_null(right))
1228 if (mode_is_signed(mode)) {
1229 return relation == ir_relation_less_greater;
1231 return relation == ir_relation_greater;
1235 static ir_node *get_cfop_destination(const ir_node *cfop)
1237 const ir_edge_t *first = get_irn_out_edge_first(cfop);
1238 /* we should only have 1 destination */
1239 assert(get_irn_n_edges(cfop) == 1);
1240 return get_edge_src_irn(first);
1243 static void lower_Switch(ir_node *node, ir_mode *high_mode)
1245 ir_node *selector = get_Switch_selector(node);
1246 ir_mode *mode = get_irn_mode(selector);
1248 if (mode == env->high_signed || mode == env->high_unsigned) {
1249 /* we can't really handle Switch with 64bit offsets */
1250 panic("Switch with 64bit jumptable not supported");
1252 lower_node(selector);
1258 static void lower_Cond(ir_node *node, ir_mode *high_mode)
1260 ir_node *left, *right, *block;
1261 ir_node *sel = get_Cond_selector(node);
1263 const lower64_entry_t *lentry, *rentry;
1264 ir_node *projT = NULL, *projF = NULL;
1265 ir_node *new_bl, *irn;
1266 ir_node *projHF, *projHT;
1268 ir_relation relation;
1279 left = get_Cmp_left(sel);
1280 cmp_mode = get_irn_mode(left);
1281 if (cmp_mode != env->high_signed && cmp_mode != env->high_unsigned) {
1286 right = get_Cmp_right(sel);
1289 lentry = get_node_entry(left);
1290 rentry = get_node_entry(right);
1292 /* all right, build the code */
1293 foreach_out_edge_safe(node, edge) {
1294 ir_node *proj = get_edge_src_irn(edge);
1298 proj_nr = get_Proj_proj(proj);
1300 if (proj_nr == pn_Cond_true) {
1301 assert(projT == NULL && "more than one Proj(true)");
1304 assert(proj_nr == pn_Cond_false);
1305 assert(projF == NULL && "more than one Proj(false)");
1308 mark_irn_visited(proj);
1310 assert(projT && projF);
1312 /* create a new high compare */
1313 block = get_nodes_block(node);
1314 irg = get_Block_irg(block);
1315 dbg = get_irn_dbg_info(sel);
1316 relation = get_Cmp_relation(sel);
1318 if (is_equality_cmp(sel)) {
1319 /* x ==/!= y ==> or(x_low^y_low,x_high^y_high) ==/!= 0 */
1320 ir_mode *mode = env->low_unsigned;
1321 ir_node *low_left = new_rd_Conv(dbg, block, lentry->low_word, mode);
1322 ir_node *high_left = new_rd_Conv(dbg, block, lentry->high_word, mode);
1323 ir_node *low_right = new_rd_Conv(dbg, block, rentry->low_word, mode);
1324 ir_node *high_right = new_rd_Conv(dbg, block, rentry->high_word, mode);
1325 ir_node *xor_low = new_rd_Eor(dbg, block, low_left, low_right, mode);
1326 ir_node *xor_high = new_rd_Eor(dbg, block, high_left, high_right, mode);
1327 ir_node *ornode = new_rd_Or(dbg, block, xor_low, xor_high, mode);
1328 ir_node *cmp = new_rd_Cmp(dbg, block, ornode, new_r_Const(irg, get_mode_null(mode)), relation);
1329 set_Cond_selector(node, cmp);
1333 if (relation == ir_relation_equal) {
1335 /* simple case:a == b <==> a_h == b_h && a_l == b_l */
1336 dst_blk = get_cfop_destination(projF);
1338 irn = new_rd_Cmp(dbg, block, lentry->high_word, rentry->high_word,
1340 dbg = get_irn_dbg_info(node);
1341 irn = new_rd_Cond(dbg, block, irn);
1343 projHF = new_r_Proj(irn, mode_X, pn_Cond_false);
1344 mark_irn_visited(projHF);
1345 exchange(projF, projHF);
1347 projHT = new_r_Proj(irn, mode_X, pn_Cond_true);
1348 mark_irn_visited(projHT);
1350 new_bl = new_r_Block(irg, 1, &projHT);
1352 dbg = get_irn_dbg_info(sel);
1353 irn = new_rd_Cmp(dbg, new_bl, lentry->low_word, rentry->low_word,
1355 dbg = get_irn_dbg_info(node);
1356 irn = new_rd_Cond(dbg, new_bl, irn);
1358 proj = new_r_Proj(irn, mode_X, pn_Cond_false);
1359 mark_irn_visited(proj);
1360 add_block_cf_input(dst_blk, projHF, proj);
1362 proj = new_r_Proj(irn, mode_X, pn_Cond_true);
1363 mark_irn_visited(proj);
1364 exchange(projT, proj);
1365 } else if (relation == ir_relation_less_greater) {
1367 /* simple case:a != b <==> a_h != b_h || a_l != b_l */
1368 dst_blk = get_cfop_destination(projT);
1370 irn = new_rd_Cmp(dbg, block, lentry->high_word, rentry->high_word,
1371 ir_relation_less_greater);
1372 dbg = get_irn_dbg_info(node);
1373 irn = new_rd_Cond(dbg, block, irn);
1375 projHT = new_r_Proj(irn, mode_X, pn_Cond_true);
1376 mark_irn_visited(projHT);
1377 exchange(projT, projHT);
1379 projHF = new_r_Proj(irn, mode_X, pn_Cond_false);
1380 mark_irn_visited(projHF);
1382 new_bl = new_r_Block(irg, 1, &projHF);
1384 dbg = get_irn_dbg_info(sel);
1385 irn = new_rd_Cmp(dbg, new_bl, lentry->low_word, rentry->low_word,
1386 ir_relation_less_greater);
1387 dbg = get_irn_dbg_info(node);
1388 irn = new_rd_Cond(dbg, new_bl, irn);
1390 proj = new_r_Proj(irn, mode_X, pn_Cond_true);
1391 mark_irn_visited(proj);
1392 add_block_cf_input(dst_blk, projHT, proj);
1394 proj = new_r_Proj(irn, mode_X, pn_Cond_false);
1395 mark_irn_visited(proj);
1396 exchange(projF, proj);
1399 /* a rel b <==> a_h REL b_h || (a_h == b_h && a_l rel b_l) */
1400 ir_node *dstT, *dstF, *newbl_eq, *newbl_l;
1403 dstT = get_cfop_destination(projT);
1404 dstF = get_cfop_destination(projF);
1406 irn = new_rd_Cmp(dbg, block, lentry->high_word, rentry->high_word,
1407 relation & ~ir_relation_equal);
1408 dbg = get_irn_dbg_info(node);
1409 irn = new_rd_Cond(dbg, block, irn);
1411 projHT = new_r_Proj(irn, mode_X, pn_Cond_true);
1412 mark_irn_visited(projHT);
1414 projHF = new_r_Proj(irn, mode_X, pn_Cond_false);
1415 mark_irn_visited(projHF);
1417 newbl_eq = new_r_Block(irg, 1, &projHF);
1419 irn = new_rd_Cmp(dbg, block, lentry->high_word, rentry->high_word,
1421 irn = new_rd_Cond(dbg, newbl_eq, irn);
1423 projEqF = new_r_Proj(irn, mode_X, pn_Cond_false);
1424 mark_irn_visited(projEqF);
1426 proj = new_r_Proj(irn, mode_X, pn_Cond_true);
1427 mark_irn_visited(proj);
1429 newbl_l = new_r_Block(irg, 1, &proj);
1431 dbg = get_irn_dbg_info(sel);
1432 irn = new_rd_Cmp(dbg, newbl_l, lentry->low_word, rentry->low_word,
1434 dbg = get_irn_dbg_info(node);
1435 irn = new_rd_Cond(dbg, newbl_l, irn);
1437 proj = new_r_Proj(irn, mode_X, pn_Cond_true);
1438 mark_irn_visited(proj);
1439 add_block_cf_input(dstT, projT, proj);
1441 proj = new_r_Proj(irn, mode_X, pn_Cond_false);
1442 mark_irn_visited(proj);
1443 add_block_cf_input(dstF, projF, proj);
1445 exchange(projT, projHT);
1446 exchange(projF, projEqF);
1449 /* we have changed the control flow */
1450 env->flags |= CF_CHANGED;
1454 * Translate a Conv to higher_signed
1456 static void lower_Conv_to_Ll(ir_node *node)
1458 ir_mode *omode = get_irn_mode(node);
1459 ir_node *op = get_Conv_op(node);
1460 ir_mode *imode = get_irn_mode(op);
1461 ir_graph *irg = get_irn_irg(node);
1462 ir_node *block = get_nodes_block(node);
1463 dbg_info *dbg = get_irn_dbg_info(node);
1467 ir_mode *low_unsigned = env->low_unsigned;
1469 = mode_is_signed(omode) ? env->low_signed : low_unsigned;
1471 if (mode_is_int(imode) || mode_is_reference(imode)) {
1472 if (imode == env->high_signed || imode == env->high_unsigned) {
1473 /* a Conv from Lu to Ls or Ls to Lu */
1474 const lower64_entry_t *op_entry = get_node_entry(op);
1475 res_low = op_entry->low_word;
1476 res_high = new_rd_Conv(dbg, block, op_entry->high_word, low_signed);
1478 /* simple case: create a high word */
1479 if (imode != low_unsigned)
1480 op = new_rd_Conv(dbg, block, op, low_unsigned);
1484 if (mode_is_signed(imode)) {
1485 int c = get_mode_size_bits(low_signed) - 1;
1486 ir_node *cnst = new_r_Const_long(irg, low_unsigned, c);
1487 if (get_irn_mode(op) != low_signed)
1488 op = new_rd_Conv(dbg, block, op, low_signed);
1489 res_high = new_rd_Shrs(dbg, block, op, cnst, low_signed);
1491 res_high = new_r_Const(irg, get_mode_null(low_signed));
1494 } else if (imode == mode_b) {
1495 res_low = new_rd_Conv(dbg, block, op, low_unsigned);
1496 res_high = new_r_Const(irg, get_mode_null(low_signed));
1498 ir_node *irn, *call;
1499 ir_type *mtp = get_conv_type(imode, omode);
1501 irn = get_intrinsic_address(mtp, get_irn_op(node), imode, omode);
1502 call = new_rd_Call(dbg, block, get_irg_no_mem(irg), irn, 1, &op, mtp);
1503 set_irn_pinned(call, get_irn_pinned(node));
1504 irn = new_r_Proj(call, mode_T, pn_Call_T_result);
1506 if (env->params->little_endian) {
1507 res_low = new_r_Proj(irn, low_unsigned, 0);
1508 res_high = new_r_Proj(irn, low_signed, 1);
1510 res_low = new_r_Proj(irn, low_unsigned, 1);
1511 res_high = new_r_Proj(irn, low_signed, 0);
1514 ir_set_dw_lowered(node, res_low, res_high);
1518 * Translate a Conv from higher_unsigned
1520 static void lower_Conv_from_Ll(ir_node *node)
1522 ir_node *op = get_Conv_op(node);
1523 ir_mode *omode = get_irn_mode(node);
1524 ir_node *block = get_nodes_block(node);
1525 dbg_info *dbg = get_irn_dbg_info(node);
1526 ir_graph *irg = get_irn_irg(node);
1527 const lower64_entry_t *entry = get_node_entry(op);
1529 if (mode_is_int(omode) || mode_is_reference(omode)) {
1530 op = entry->low_word;
1532 /* simple case: create a high word */
1533 if (omode != env->low_unsigned)
1534 op = new_rd_Conv(dbg, block, op, omode);
1536 set_Conv_op(node, op);
1537 } else if (omode == mode_b) {
1538 /* llu ? true : false <=> (low|high) ? true : false */
1539 ir_mode *mode = env->low_unsigned;
1540 ir_node *ornode = new_rd_Or(dbg, block, entry->low_word,
1541 entry->high_word, mode);
1542 set_Conv_op(node, ornode);
1544 ir_node *irn, *call, *in[2];
1545 ir_mode *imode = get_irn_mode(op);
1546 ir_type *mtp = get_conv_type(imode, omode);
1549 irn = get_intrinsic_address(mtp, get_irn_op(node), imode, omode);
1550 if (env->params->little_endian) {
1551 in[0] = entry->low_word;
1552 in[1] = entry->high_word;
1554 in[0] = entry->high_word;
1555 in[1] = entry->low_word;
1558 call = new_rd_Call(dbg, block, get_irg_no_mem(irg), irn, 2, in, mtp);
1559 set_irn_pinned(call, get_irn_pinned(node));
1560 irn = new_r_Proj(call, mode_T, pn_Call_T_result);
1561 res = new_r_Proj(irn, omode, 0);
1563 exchange(node, res);
1570 static void lower_Cmp(ir_node *cmp, ir_mode *m)
1572 ir_node *l = get_Cmp_left(cmp);
1573 ir_mode *cmp_mode = get_irn_mode(l);
1574 ir_node *r, *low, *high, *t, *res;
1575 ir_relation relation;
1578 const lower64_entry_t *lentry;
1579 const lower64_entry_t *rentry;
1582 if (cmp_mode != env->high_signed && cmp_mode != env->high_unsigned)
1585 r = get_Cmp_right(cmp);
1586 lentry = get_node_entry(l);
1587 rentry = get_node_entry(r);
1588 relation = get_Cmp_relation(cmp);
1589 block = get_nodes_block(cmp);
1590 dbg = get_irn_dbg_info(cmp);
1592 /* easy case for x ==/!= 0 (see lower_Cond for details) */
1593 if (is_equality_cmp(cmp)) {
1594 ir_graph *irg = get_irn_irg(cmp);
1595 ir_mode *mode = env->low_unsigned;
1596 ir_node *low_left = new_rd_Conv(dbg, block, lentry->low_word, mode);
1597 ir_node *high_left = new_rd_Conv(dbg, block, lentry->high_word, mode);
1598 ir_node *low_right = new_rd_Conv(dbg, block, rentry->low_word, mode);
1599 ir_node *high_right = new_rd_Conv(dbg, block, rentry->high_word, mode);
1600 ir_node *xor_low = new_rd_Eor(dbg, block, low_left, low_right, mode);
1601 ir_node *xor_high = new_rd_Eor(dbg, block, high_left, high_right, mode);
1602 ir_node *ornode = new_rd_Or(dbg, block, xor_low, xor_high, mode);
1603 ir_node *new_cmp = new_rd_Cmp(dbg, block, ornode, new_r_Const(irg, get_mode_null(mode)), relation);
1604 exchange(cmp, new_cmp);
1608 if (relation == ir_relation_equal) {
1609 /* simple case:a == b <==> a_h == b_h && a_l == b_l */
1610 low = new_rd_Cmp(dbg, block, lentry->low_word, rentry->low_word,
1612 high = new_rd_Cmp(dbg, block, lentry->high_word, rentry->high_word,
1614 res = new_rd_And(dbg, block, low, high, mode_b);
1615 } else if (relation == ir_relation_less_greater) {
1616 /* simple case:a != b <==> a_h != b_h || a_l != b_l */
1617 low = new_rd_Cmp(dbg, block, lentry->low_word, rentry->low_word,
1619 high = new_rd_Cmp(dbg, block, lentry->high_word, rentry->high_word,
1621 res = new_rd_Or(dbg, block, low, high, mode_b);
1623 /* a rel b <==> a_h REL b_h || (a_h == b_h && a_l rel b_l) */
1624 ir_node *high1 = new_rd_Cmp(dbg, block, lentry->high_word,
1625 rentry->high_word, relation & ~ir_relation_equal);
1626 low = new_rd_Cmp(dbg, block, lentry->low_word, rentry->low_word,
1628 high = new_rd_Cmp(dbg, block, lentry->high_word, rentry->high_word,
1630 t = new_rd_And(dbg, block, low, high, mode_b);
1631 res = new_rd_Or(dbg, block, high1, t, mode_b);
1639 static void lower_Conv(ir_node *node, ir_mode *mode)
1641 mode = get_irn_mode(node);
1643 if (mode == env->high_signed || mode == env->high_unsigned) {
1644 lower_Conv_to_Ll(node);
1646 ir_mode *op_mode = get_irn_mode(get_Conv_op(node));
1648 if (op_mode == env->high_signed || op_mode == env->high_unsigned) {
1649 lower_Conv_from_Ll(node);
1654 static void fix_parameter_entities(ir_graph *irg, ir_type *orig_mtp)
1656 size_t orig_n_params = get_method_n_params(orig_mtp);
1657 ir_entity **parameter_entities;
1659 parameter_entities = ALLOCANZ(ir_entity*, orig_n_params);
1661 ir_type *frame_type = get_irg_frame_type(irg);
1662 size_t n = get_compound_n_members(frame_type);
1666 /* collect parameter entities */
1667 for (i = 0; i < n; ++i) {
1668 ir_entity *entity = get_compound_member(frame_type, i);
1670 if (!is_parameter_entity(entity))
1672 p = get_entity_parameter_number(entity);
1673 if (p == IR_VA_START_PARAMETER_NUMBER)
1675 assert(p < orig_n_params);
1676 assert(parameter_entities[p] == NULL);
1677 parameter_entities[p] = entity;
1680 /* adjust indices */
1682 for (i = 0; i < orig_n_params; ++i, ++n_param) {
1683 ir_entity *entity = parameter_entities[i];
1687 set_entity_parameter_number(entity, n_param);
1689 tp = get_method_param_type(orig_mtp, i);
1690 if (is_Primitive_type(tp)) {
1691 ir_mode *mode = get_type_mode(tp);
1692 if (mode == env->high_signed || mode == env->high_unsigned) {
1694 /* note that we do not change the type of the parameter
1695 * entities, as calling convention fixup later still needs to
1696 * know which is/was a lowered doubleword.
1697 * So we just mark/remember it for later */
1698 if (entity != NULL) {
1699 assert(entity->attr.parameter.doubleword_low_mode == NULL);
1700 entity->attr.parameter.doubleword_low_mode
1701 = env->low_unsigned;
1709 * Lower the method type.
1711 * @param env the lower environment
1712 * @param mtp the method type to lower
1714 * @return the lowered type
1716 static ir_type *lower_mtp(ir_type *mtp)
1720 size_t orig_n_params;
1724 bool must_be_lowered;
1726 res = pmap_get(ir_type, lowered_type, mtp);
1729 if (type_visited(mtp))
1731 mark_type_visited(mtp);
1733 orig_n_params = get_method_n_params(mtp);
1734 orig_n_res = get_method_n_ress(mtp);
1735 n_param = orig_n_params;
1737 must_be_lowered = false;
1739 /* count new number of params */
1740 for (i = orig_n_params; i > 0;) {
1741 ir_type *tp = get_method_param_type(mtp, --i);
1743 if (is_Primitive_type(tp)) {
1744 ir_mode *mode = get_type_mode(tp);
1746 if (mode == env->high_signed || mode == env->high_unsigned) {
1748 must_be_lowered = true;
1753 /* count new number of results */
1754 for (i = orig_n_res; i > 0;) {
1755 ir_type *tp = get_method_res_type(mtp, --i);
1757 if (is_Primitive_type(tp)) {
1758 ir_mode *mode = get_type_mode(tp);
1760 if (mode == env->high_signed || mode == env->high_unsigned) {
1762 must_be_lowered = true;
1766 if (!must_be_lowered) {
1767 set_type_link(mtp, NULL);
1771 res = new_d_type_method(n_param, n_res, get_type_dbg_info(mtp));
1773 /* set param types and result types */
1774 for (i = n_param = 0; i < orig_n_params; ++i) {
1775 ir_type *tp = get_method_param_type(mtp, i);
1777 if (is_Primitive_type(tp)) {
1778 ir_mode *mode = get_type_mode(tp);
1780 if (mode == env->high_signed) {
1781 if (env->params->little_endian) {
1782 set_method_param_type(res, n_param++, tp_u);
1783 set_method_param_type(res, n_param++, tp_s);
1785 set_method_param_type(res, n_param++, tp_s);
1786 set_method_param_type(res, n_param++, tp_u);
1788 } else if (mode == env->high_unsigned) {
1789 set_method_param_type(res, n_param++, tp_u);
1790 set_method_param_type(res, n_param++, tp_u);
1792 set_method_param_type(res, n_param, tp);
1796 set_method_param_type(res, n_param, tp);
1800 for (i = n_res = 0; i < orig_n_res; ++i) {
1801 ir_type *tp = get_method_res_type(mtp, i);
1803 if (is_Primitive_type(tp)) {
1804 ir_mode *mode = get_type_mode(tp);
1806 if (mode == env->high_signed) {
1807 if (env->params->little_endian) {
1808 set_method_res_type(res, n_res++, tp_u);
1809 set_method_res_type(res, n_res++, tp_s);
1811 set_method_res_type(res, n_res++, tp_s);
1812 set_method_res_type(res, n_res++, tp_u);
1814 } else if (mode == env->high_unsigned) {
1815 set_method_res_type(res, n_res++, tp_u);
1816 set_method_res_type(res, n_res++, tp_u);
1818 set_method_res_type(res, n_res++, tp);
1821 set_method_res_type(res, n_res++, tp);
1825 set_method_variadicity(res, get_method_variadicity(mtp));
1826 set_method_calling_convention(res, get_method_calling_convention(mtp));
1827 set_method_additional_properties(res, get_method_additional_properties(mtp));
1829 set_higher_type(res, mtp);
1830 set_type_link(res, mtp);
1832 mark_type_visited(res);
1833 pmap_insert(lowered_type, mtp, res);
1838 * Translate a Return.
1840 static void lower_Return(ir_node *node, ir_mode *mode)
1842 ir_graph *irg = get_irn_irg(node);
1843 ir_entity *ent = get_irg_entity(irg);
1844 ir_type *mtp = get_entity_type(ent);
1850 /* check if this return must be lowered */
1851 for (i = 0, n = get_Return_n_ress(node); i < n; ++i) {
1852 ir_node *pred = get_Return_res(node, i);
1853 ir_mode *rmode = get_irn_op_mode(pred);
1855 if (rmode == env->high_signed || rmode == env->high_unsigned)
1861 ent = get_irg_entity(irg);
1862 mtp = get_entity_type(ent);
1864 /* create a new in array */
1865 NEW_ARR_A(ir_node *, in, get_method_n_ress(mtp) + 1);
1867 in[j++] = get_Return_mem(node);
1869 for (i = 0, n = get_Return_n_ress(node); i < n; ++i) {
1870 ir_node *pred = get_Return_res(node, i);
1871 ir_mode *pred_mode = get_irn_mode(pred);
1873 if (pred_mode == env->high_signed || pred_mode == env->high_unsigned) {
1874 const lower64_entry_t *entry = get_node_entry(pred);
1875 if (env->params->little_endian) {
1876 in[j++] = entry->low_word;
1877 in[j++] = entry->high_word;
1879 in[j++] = entry->high_word;
1880 in[j++] = entry->low_word;
1886 assert(j == get_method_n_ress(mtp)+1);
1888 set_irn_in(node, j, in);
1892 * Translate the parameters.
1894 static void lower_Start(ir_node *node, ir_mode *high_mode)
1896 ir_graph *irg = get_irn_irg(node);
1897 ir_entity *ent = get_irg_entity(irg);
1898 ir_type *mtp = get_entity_type(ent);
1899 ir_type *orig_mtp = (ir_type*)get_type_link(mtp);
1902 size_t i, j, n_params;
1905 /* if type link is NULL then the type was not lowered, hence no changes
1906 * at Start necessary */
1907 if (orig_mtp == NULL)
1910 n_params = get_method_n_params(orig_mtp);
1912 NEW_ARR_A(long, new_projs, n_params);
1914 /* Calculate mapping of proj numbers in new_projs */
1915 for (i = j = 0; i < n_params; ++i, ++j) {
1916 ir_type *ptp = get_method_param_type(orig_mtp, i);
1919 if (is_Primitive_type(ptp)) {
1920 ir_mode *amode = get_type_mode(ptp);
1921 if (amode == env->high_signed || amode == env->high_unsigned)
1926 /* find args Proj */
1928 foreach_out_edge(node, edge) {
1929 ir_node *proj = get_edge_src_irn(edge);
1932 if (get_Proj_proj(proj) == pn_Start_T_args) {
1940 /* fix all Proj's and create new ones */
1941 foreach_out_edge_safe(args, edge) {
1942 ir_node *proj = get_edge_src_irn(edge);
1943 ir_mode *mode = get_irn_mode(proj);
1944 ir_mode *mode_l = env->low_unsigned;
1955 pred = get_Proj_pred(proj);
1956 proj_nr = get_Proj_proj(proj);
1958 if (mode == env->high_signed) {
1959 mode_h = env->low_signed;
1960 } else if (mode == env->high_unsigned) {
1961 mode_h = env->low_unsigned;
1963 long new_pn = new_projs[proj_nr];
1964 set_Proj_proj(proj, new_pn);
1968 /* Switch off CSE or we might get an already existing Proj. */
1969 old_cse = get_opt_cse();
1971 dbg = get_irn_dbg_info(proj);
1972 if (env->params->little_endian) {
1973 res_low = new_rd_Proj(dbg, pred, mode_l, new_projs[proj_nr]);
1974 res_high = new_rd_Proj(dbg, pred, mode_h, new_projs[proj_nr] + 1);
1976 res_high = new_rd_Proj(dbg, pred, mode_h, new_projs[proj_nr]);
1977 res_low = new_rd_Proj(dbg, pred, mode_l, new_projs[proj_nr] + 1);
1979 set_opt_cse(old_cse);
1980 ir_set_dw_lowered(proj, res_low, res_high);
1987 static void lower_Call(ir_node *node, ir_mode *mode)
1989 ir_type *tp = get_Call_type(node);
1991 size_t n_params, n_res;
1992 bool need_lower = false;
1995 long *res_numbers = NULL;
1999 n_params = get_method_n_params(tp);
2000 for (p = 0; p < n_params; ++p) {
2001 ir_type *ptp = get_method_param_type(tp, p);
2003 if (is_Primitive_type(ptp)) {
2004 ir_mode *pmode = get_type_mode(ptp);
2005 if (pmode == env->high_signed || pmode == env->high_unsigned) {
2011 n_res = get_method_n_ress(tp);
2013 NEW_ARR_A(long, res_numbers, n_res);
2015 for (i = j = 0; i < n_res; ++i, ++j) {
2016 ir_type *ptp = get_method_res_type(tp, i);
2019 if (is_Primitive_type(ptp)) {
2020 ir_mode *rmode = get_type_mode(ptp);
2021 if (rmode == env->high_signed || rmode == env->high_unsigned) {
2032 /* let's lower it */
2034 set_Call_type(node, tp);
2036 NEW_ARR_A(ir_node *, in, get_method_n_params(tp) + 2);
2038 in[0] = get_Call_mem(node);
2039 in[1] = get_Call_ptr(node);
2041 for (j = 2, i = 0; i < n_params; ++i) {
2042 ir_node *pred = get_Call_param(node, i);
2043 ir_mode *pred_mode = get_irn_mode(pred);
2045 if (pred_mode == env->high_signed || pred_mode == env->high_unsigned) {
2046 const lower64_entry_t *pred_entry = get_node_entry(pred);
2047 if (env->params->little_endian) {
2048 in[j++] = pred_entry->low_word;
2049 in[j++] = pred_entry->high_word;
2051 in[j++] = pred_entry->high_word;
2052 in[j++] = pred_entry->low_word;
2059 set_irn_in(node, j, in);
2061 /* find results T */
2063 foreach_out_edge(node, edge) {
2064 ir_node *proj = get_edge_src_irn(edge);
2067 if (get_Proj_proj(proj) == pn_Call_T_result) {
2072 if (resproj == NULL)
2075 /* fix the results */
2076 foreach_out_edge_safe(resproj, edge) {
2077 ir_node *proj = get_edge_src_irn(edge);
2078 ir_mode *proj_mode = get_irn_mode(proj);
2079 ir_mode *mode_l = env->low_unsigned;
2089 pred = get_Proj_pred(proj);
2090 proj_nr = get_Proj_proj(proj);
2092 if (proj_mode == env->high_signed) {
2093 mode_h = env->low_signed;
2094 } else if (proj_mode == env->high_unsigned) {
2095 mode_h = env->low_unsigned;
2097 long new_nr = res_numbers[proj_nr];
2098 set_Proj_proj(proj, new_nr);
2102 dbg = get_irn_dbg_info(proj);
2103 if (env->params->little_endian) {
2104 res_low = new_rd_Proj(dbg, pred, mode_l, res_numbers[proj_nr]);
2105 res_high = new_rd_Proj(dbg, pred, mode_h, res_numbers[proj_nr] + 1);
2107 res_high = new_rd_Proj(dbg, pred, mode_h, res_numbers[proj_nr]);
2108 res_low = new_rd_Proj(dbg, pred, mode_l, res_numbers[proj_nr] + 1);
2110 ir_set_dw_lowered(proj, res_low, res_high);
2115 * Translate an Unknown into two.
2117 static void lower_Unknown(ir_node *node, ir_mode *mode)
2119 ir_mode *low_mode = env->low_unsigned;
2120 ir_graph *irg = get_irn_irg(node);
2121 ir_node *res_low = new_r_Unknown(irg, low_mode);
2122 ir_node *res_high = new_r_Unknown(irg, mode);
2123 ir_set_dw_lowered(node, res_low, res_high);
2127 * Translate a Bad into two.
2129 static void lower_Bad(ir_node *node, ir_mode *mode)
2131 ir_mode *low_mode = env->low_unsigned;
2132 ir_graph *irg = get_irn_irg(node);
2133 ir_node *res_low = new_r_Bad(irg, low_mode);
2134 ir_node *res_high = new_r_Bad(irg, mode);
2135 ir_set_dw_lowered(node, res_low, res_high);
2141 * First step: just create two templates
2143 static void lower_Phi(ir_node *phi)
2145 ir_mode *mode = get_irn_mode(phi);
2160 /* enqueue predecessors */
2161 arity = get_Phi_n_preds(phi);
2162 for (i = 0; i < arity; ++i) {
2163 ir_node *pred = get_Phi_pred(phi, i);
2164 pdeq_putr(env->waitq, pred);
2167 if (mode != env->high_signed && mode != env->high_unsigned)
2170 /* first create a new in array */
2171 NEW_ARR_A(ir_node *, in_l, arity);
2172 NEW_ARR_A(ir_node *, in_h, arity);
2173 irg = get_irn_irg(phi);
2174 mode_l = env->low_unsigned;
2175 mode_h = mode == env->high_signed ? env->low_signed : env->low_unsigned;
2176 unk_l = new_r_Dummy(irg, mode_l);
2177 unk_h = new_r_Dummy(irg, mode_h);
2178 for (i = 0; i < arity; ++i) {
2183 dbg = get_irn_dbg_info(phi);
2184 block = get_nodes_block(phi);
2185 phi_l = new_rd_Phi(dbg, block, arity, in_l, mode_l);
2186 phi_h = new_rd_Phi(dbg, block, arity, in_h, mode_h);
2188 ir_set_dw_lowered(phi, phi_l, phi_h);
2190 /* remember that we need to fixup the predecessors later */
2191 ARR_APP1(ir_node*, env->lowered_phis, phi);
2194 static void fixup_phi(ir_node *phi)
2196 const lower64_entry_t *entry = get_node_entry(phi);
2197 ir_node *phi_l = entry->low_word;
2198 ir_node *phi_h = entry->high_word;
2199 int arity = get_Phi_n_preds(phi);
2202 /* exchange phi predecessors which are lowered by now */
2203 for (i = 0; i < arity; ++i) {
2204 ir_node *pred = get_Phi_pred(phi, i);
2205 const lower64_entry_t *pred_entry = get_node_entry(pred);
2207 set_Phi_pred(phi_l, i, pred_entry->low_word);
2208 set_Phi_pred(phi_h, i, pred_entry->high_word);
2215 static void lower_Mux(ir_node *mux, ir_mode *mode)
2217 ir_node *truen = get_Mux_true(mux);
2218 ir_node *falsen = get_Mux_false(mux);
2219 ir_node *sel = get_Mux_sel(mux);
2220 const lower64_entry_t *true_entry = get_node_entry(truen);
2221 const lower64_entry_t *false_entry = get_node_entry(falsen);
2222 ir_node *true_l = true_entry->low_word;
2223 ir_node *true_h = true_entry->high_word;
2224 ir_node *false_l = false_entry->low_word;
2225 ir_node *false_h = false_entry->high_word;
2226 dbg_info *dbgi = get_irn_dbg_info(mux);
2227 ir_node *block = get_nodes_block(mux);
2229 = new_rd_Mux(dbgi, block, sel, false_l, true_l, env->low_unsigned);
2231 = new_rd_Mux(dbgi, block, sel, false_h, true_h, mode);
2232 ir_set_dw_lowered(mux, res_low, res_high);
2236 * Translate an ASM node.
2238 static void lower_ASM(ir_node *asmn, ir_mode *mode)
2240 ir_mode *high_signed = env->high_signed;
2241 ir_mode *high_unsigned = env->high_unsigned;
2242 int n_outs = get_ASM_n_output_constraints(asmn);
2243 ir_asm_constraint *output_constraints = get_ASM_output_constraints(asmn);
2244 ir_asm_constraint *input_constraints = get_ASM_input_constraints(asmn);
2245 unsigned n_64bit_outs = 0;
2249 for (int i = get_irn_arity(asmn) - 1; i >= 0; --i) {
2250 ir_node *op = get_irn_n(asmn, i);
2251 ir_mode *op_mode = get_irn_mode(op);
2252 if (op_mode == high_signed || op_mode == high_unsigned) {
2253 panic("lowering ASM 64bit input unimplemented");
2257 for (int o = 0; o < n_outs; ++o) {
2258 const ir_asm_constraint *constraint = &output_constraints[o];
2259 if (constraint->mode == high_signed || constraint->mode == high_unsigned) {
2260 const char *constr = get_id_str(constraint->constraint);
2262 /* TODO: How to do this architecture neutral? This is very
2263 * i386 specific... */
2264 if (constr[0] != '=' || constr[1] != 'A') {
2265 panic("lowering ASM 64bit output only supports '=A' currently");
2270 if (n_64bit_outs == 0)
2273 dbg_info *dbgi = get_irn_dbg_info(asmn);
2274 ir_node *block = get_nodes_block(asmn);
2275 ir_node *mem = get_ASM_mem(asmn);
2277 int n_clobber = get_ASM_n_clobbers(asmn);
2278 long *proj_map = ALLOCAN(long, n_outs);
2279 ident **clobbers = get_ASM_clobbers(asmn);
2280 ident *asm_text = get_ASM_text(asmn);
2281 ir_asm_constraint *new_outputs
2282 = ALLOCAN(ir_asm_constraint, n_outs+n_64bit_outs);
2285 for (int o = 0; o < n_outs; ++o) {
2286 const ir_asm_constraint *constraint = &output_constraints[o];
2287 if (constraint->mode == high_signed || constraint->mode == high_unsigned) {
2288 new_outputs[new_n_outs].pos = constraint->pos;
2289 new_outputs[new_n_outs].constraint = new_id_from_str("=a");
2290 new_outputs[new_n_outs].mode = env->low_unsigned;
2291 proj_map[o] = new_n_outs;
2293 new_outputs[new_n_outs].pos = constraint->pos;
2294 new_outputs[new_n_outs].constraint = new_id_from_str("=d");
2295 if (constraint->mode == high_signed)
2296 new_outputs[new_n_outs].mode = env->low_signed;
2298 new_outputs[new_n_outs].mode = env->low_unsigned;
2301 new_outputs[new_n_outs] = *constraint;
2302 proj_map[o] = new_n_outs;
2306 assert(new_n_outs == n_outs+(int)n_64bit_outs);
2308 int n_inputs = get_ASM_n_inputs(asmn);
2309 ir_node **new_ins = ALLOCAN(ir_node*, n_inputs);
2310 for (int i = 0; i < n_inputs; ++i)
2311 new_ins[i] = get_ASM_input(asmn, i);
2313 new_asm = new_rd_ASM(dbgi, block, mem, n_inputs, new_ins, input_constraints,
2314 new_n_outs, new_outputs, n_clobber, clobbers,
2317 foreach_out_edge_safe(asmn, edge) {
2318 ir_node *proj = get_edge_src_irn(edge);
2319 ir_mode *proj_mode = get_irn_mode(proj);
2324 pn = get_Proj_proj(proj);
2329 pn = new_n_outs + pn - n_outs;
2331 if (proj_mode == high_signed || proj_mode == high_unsigned) {
2333 = proj_mode == high_signed ? env->low_signed : env->low_unsigned;
2334 ir_node *np_low = new_r_Proj(new_asm, env->low_unsigned, pn);
2335 ir_node *np_high = new_r_Proj(new_asm, high_mode, pn+1);
2336 ir_set_dw_lowered(proj, np_low, np_high);
2338 ir_node *np = new_r_Proj(new_asm, proj_mode, pn);
2345 * Lower the builtin type to its higher part.
2347 * @param mtp the builtin type to lower
2349 * @return the lowered type
2351 static ir_type *lower_Builtin_type_high(ir_type *mtp)
2357 bool must_be_lowered;
2359 res = pmap_get(ir_type, lowered_builtin_type_high, mtp);
2363 n_params = get_method_n_params(mtp);
2364 n_results = get_method_n_ress(mtp);
2365 must_be_lowered = false;
2367 /* check for double word parameter */
2368 for (i = n_params; i > 0;) {
2369 ir_type *tp = get_method_param_type(mtp, --i);
2371 if (is_Primitive_type(tp)) {
2372 ir_mode *mode = get_type_mode(tp);
2374 if (mode == env->high_signed || mode == env->high_unsigned) {
2375 must_be_lowered = true;
2381 if (!must_be_lowered) {
2382 set_type_link(mtp, NULL);
2386 res = new_d_type_method(n_params, n_results, get_type_dbg_info(mtp));
2388 /* set param types and result types */
2389 for (i = 0; i < n_params; ++i) {
2390 ir_type *tp = get_method_param_type(mtp, i);
2392 if (is_Primitive_type(tp)) {
2393 ir_mode *mode = get_type_mode(tp);
2395 if (mode == env->high_signed) {
2396 if (env->params->little_endian) {
2397 set_method_param_type(res, i, tp_u);
2399 set_method_param_type(res, i, tp_s);
2401 } else if (mode == env->high_unsigned) {
2402 set_method_param_type(res, i, tp_u);
2404 set_method_param_type(res, i, tp);
2407 set_method_param_type(res, i, tp);
2410 for (i = n_results = 0; i < n_results; ++i) {
2411 ir_type *tp = get_method_res_type(mtp, i);
2413 if (is_Primitive_type(tp)) {
2414 ir_mode *mode = get_type_mode(tp);
2416 if (mode == env->high_signed) {
2417 if (env->params->little_endian) {
2418 set_method_res_type(res, i, tp_u);
2420 set_method_res_type(res, i, tp_s);
2422 } else if (mode == env->high_unsigned) {
2423 set_method_res_type(res, i, tp_u);
2425 set_method_res_type(res, i, tp);
2428 set_method_res_type(res, i, tp);
2432 set_method_variadicity(res, get_method_variadicity(mtp));
2433 set_method_calling_convention(res, get_method_calling_convention(mtp));
2434 set_method_additional_properties(res, get_method_additional_properties(mtp));
2436 pmap_insert(lowered_builtin_type_high, mtp, res);
2441 * Lower the builtin type to its lower part.
2443 * @param mtp the builtin type to lower
2445 * @return the lowered type
2447 static ir_type *lower_Builtin_type_low(ir_type *mtp)
2453 bool must_be_lowered;
2455 res = pmap_get(ir_type, lowered_builtin_type_low, mtp);
2459 n_params = get_method_n_params(mtp);
2460 n_results = get_method_n_ress(mtp);
2461 must_be_lowered = false;
2463 /* check for double word parameter */
2464 for (i = n_params; i > 0;) {
2465 ir_type *tp = get_method_param_type(mtp, --i);
2467 if (is_Primitive_type(tp)) {
2468 ir_mode *mode = get_type_mode(tp);
2470 if (mode == env->high_signed || mode == env->high_unsigned) {
2471 must_be_lowered = true;
2477 if (!must_be_lowered) {
2478 set_type_link(mtp, NULL);
2482 res = new_d_type_method(n_params, n_results, get_type_dbg_info(mtp));
2484 /* set param types and result types */
2485 for (i = 0; i < n_params; ++i) {
2486 ir_type *tp = get_method_param_type(mtp, i);
2488 if (is_Primitive_type(tp)) {
2489 ir_mode *mode = get_type_mode(tp);
2491 if (mode == env->high_signed) {
2492 if (env->params->little_endian) {
2493 set_method_param_type(res, i, tp_s);
2495 set_method_param_type(res, i, tp_u);
2497 } else if (mode == env->high_unsigned) {
2498 set_method_param_type(res, i, tp_u);
2500 set_method_param_type(res, i, tp);
2503 set_method_param_type(res, i, tp);
2506 for (i = 0; i < n_results; ++i) {
2507 ir_type *tp = get_method_res_type(mtp, i);
2509 if (is_Primitive_type(tp)) {
2510 ir_mode *mode = get_type_mode(tp);
2512 if (mode == env->high_signed) {
2513 if (env->params->little_endian) {
2514 set_method_res_type(res, i, tp_s);
2516 set_method_res_type(res, i, tp_u);
2518 } else if (mode == env->high_unsigned) {
2519 set_method_res_type(res, i, tp_u);
2521 set_method_res_type(res, i, tp);
2524 set_method_res_type(res, i, tp);
2528 set_method_variadicity(res, get_method_variadicity(mtp));
2529 set_method_calling_convention(res, get_method_calling_convention(mtp));
2530 set_method_additional_properties(res, get_method_additional_properties(mtp));
2532 pmap_insert(lowered_builtin_type_low, mtp, res);
2537 * lowers a builtin which reduces a 64bit value to a simple summary value
2538 * (popcount, ffs, ...)
2540 static void lower_reduce_builtin(ir_node *builtin, ir_mode *mode)
2542 ir_builtin_kind kind = get_Builtin_kind(builtin);
2543 ir_node *operand = get_Builtin_param(builtin, 0);
2544 ir_mode *operand_mode = get_irn_mode(operand);
2545 if (operand_mode != env->high_signed && operand_mode != env->high_unsigned)
2549 arch_allow_ifconv_func allow_ifconv = be_get_backend_param()->allow_ifconv;
2550 int arity = get_irn_arity(builtin);
2551 dbg_info *dbgi = get_irn_dbg_info(builtin);
2552 ir_graph *irg = get_irn_irg(builtin);
2553 ir_type *type = get_Builtin_type(builtin);
2554 ir_type *lowered_type_high = lower_Builtin_type_high(type);
2555 ir_type *lowered_type_low = lower_Builtin_type_low(type);
2556 ir_type *result_type = get_method_res_type(lowered_type_low, 0);
2557 ir_mode *result_mode = get_type_mode(result_type);
2558 ir_node *block = get_nodes_block(builtin);
2559 ir_node *mem = get_Builtin_mem(builtin);
2560 const lower64_entry_t *entry = get_node_entry(operand);
2561 ir_mode *high_mode = get_irn_mode(entry->high_word);
2562 ir_node *in_high[1] = {entry->high_word};
2563 ir_node *in_low[1] = {entry->low_word};
2566 assert(is_NoMem(mem));
2571 ir_node *number_of_bits = new_r_Const_long(irg, result_mode, get_mode_size_bits(env->low_unsigned));
2572 ir_node *zero_high = new_rd_Const(dbgi, irg, get_mode_null(high_mode));
2573 ir_node *zero_unsigned = new_rd_Const(dbgi, irg, get_mode_null(env->low_unsigned));
2574 ir_node *zero_result = new_rd_Const(dbgi, irg, get_mode_null(result_mode));
2575 ir_node *cmp_low = new_rd_Cmp(dbgi, block, entry->low_word, zero_unsigned, ir_relation_equal);
2576 ir_node *cmp_high = new_rd_Cmp(dbgi, block, entry->high_word, zero_high, ir_relation_equal);
2577 ir_node *ffs_high = new_rd_Builtin(dbgi, block, mem, 1, in_high, kind, lowered_type_high);
2578 ir_node *high_proj = new_r_Proj(ffs_high, result_mode, pn_Builtin_max+1);
2579 ir_node *high = new_rd_Add(dbgi, block, high_proj, number_of_bits, result_mode);
2580 ir_node *ffs_low = new_rd_Builtin(dbgi, block, mem, 1, in_low, kind, lowered_type_low);
2581 ir_node *low = new_r_Proj(ffs_low, result_mode, pn_Builtin_max+1);
2582 ir_node *mux_high = new_rd_Mux(dbgi, block, cmp_high, high, zero_result, result_mode);
2584 if (! allow_ifconv(cmp_high, high, zero_result))
2585 ir_nodeset_insert(&created_mux_nodes, mux_high);
2587 res = new_rd_Mux(dbgi, block, cmp_low, low, mux_high, result_mode);
2589 if (! allow_ifconv(cmp_low, low, mux_high))
2590 ir_nodeset_insert(&created_mux_nodes, res);
2594 ir_node *zero = new_rd_Const(dbgi, irg, get_mode_null(high_mode));
2595 ir_node *cmp_high = new_rd_Cmp(dbgi, block, entry->high_word, zero, ir_relation_equal);
2596 ir_node *clz_high = new_rd_Builtin(dbgi, block, mem, 1, in_high, kind, lowered_type_high);
2597 ir_node *high = new_r_Proj(clz_high, result_mode, pn_Builtin_max+1);
2598 ir_node *clz_low = new_rd_Builtin(dbgi, block, mem, 1, in_low, kind, lowered_type_low);
2599 ir_node *low_proj = new_r_Proj(clz_low, result_mode, pn_Builtin_max+1);
2600 ir_node *number_of_bits = new_r_Const_long(irg, result_mode, get_mode_size_bits(mode));
2601 ir_node *low = new_rd_Add(dbgi, block, low_proj, number_of_bits, result_mode);
2603 res = new_rd_Mux(dbgi, block, cmp_high, high, low, result_mode);
2605 if (! allow_ifconv(cmp_high, high, low))
2606 ir_nodeset_insert(&created_mux_nodes, res);
2610 ir_node *zero_unsigned = new_rd_Const(dbgi, irg, get_mode_null(env->low_unsigned));
2611 ir_node *cmp_low = new_rd_Cmp(dbgi, block, entry->low_word, zero_unsigned, ir_relation_equal);
2612 ir_node *ffs_high = new_rd_Builtin(dbgi, block, mem, 1, in_high, kind, lowered_type_high);
2613 ir_node *high_proj = new_r_Proj(ffs_high, result_mode, pn_Builtin_max+1);
2614 ir_node *number_of_bits = new_r_Const_long(irg, result_mode, get_mode_size_bits(env->low_unsigned));
2615 ir_node *high = new_rd_Add(dbgi, block, high_proj, number_of_bits, result_mode);
2616 ir_node *ffs_low = new_rd_Builtin(dbgi, block, mem, 1, in_low, kind, lowered_type_low);
2617 ir_node *low = new_r_Proj(ffs_low, result_mode, pn_Builtin_max+1);
2619 res = new_rd_Mux(dbgi, block, cmp_low, low, high, result_mode);
2621 if (! allow_ifconv(cmp_low, low, high))
2622 ir_nodeset_insert(&created_mux_nodes, res);
2625 case ir_bk_popcount: {
2626 ir_node *popcount_high = new_rd_Builtin(dbgi, block, mem, 1, in_high, kind, lowered_type_high);
2627 ir_node *popcount_low = new_rd_Builtin(dbgi, block, mem, 1, in_low, kind, lowered_type_low);
2628 ir_node *high = new_r_Proj(popcount_high, result_mode, pn_Builtin_max+1);
2629 ir_node *low = new_r_Proj(popcount_low, result_mode, pn_Builtin_max+1);
2631 res = new_rd_Add(dbgi, block, high, low, result_mode);
2634 case ir_bk_parity: {
2635 ir_node *parity_high;
2636 ir_node *parity_low;
2642 parity_high = new_rd_Builtin(dbgi, block, mem, 1, in_high, kind, lowered_type_high);
2643 high = new_r_Proj(parity_high, result_mode, pn_Builtin_max+1);
2644 parity_low = new_rd_Builtin(dbgi, block, mem, 1, in_low, kind, lowered_type_low);
2645 low = new_r_Proj(parity_low, result_mode, pn_Builtin_max+1);
2646 res = new_rd_Eor(dbgi, block, high, low, result_mode);
2650 panic("unexpected builtin");
2653 turn_into_tuple(builtin, 2);
2654 set_irn_n(builtin, pn_Builtin_M, mem);
2655 set_irn_n(builtin, pn_Builtin_max+1, res);
2660 * lowers builtins performing arithmetic (bswap)
2662 static void lower_arithmetic_builtin(ir_node *builtin, ir_mode *mode)
2664 ir_builtin_kind kind = get_Builtin_kind(builtin);
2665 ir_node *operand = get_Builtin_param(builtin, 0);
2666 ir_mode *operand_mode = get_irn_mode(operand);
2668 if (operand_mode != env->high_signed && operand_mode != env->high_unsigned)
2672 dbg_info *dbgi = get_irn_dbg_info(builtin);
2673 ir_type *type = get_Builtin_type(builtin);
2674 ir_type *lowered_type_high = lower_Builtin_type_high(type);
2675 ir_type *lowered_type_low = lower_Builtin_type_low(type);
2676 ir_node *block = get_nodes_block(builtin);
2677 ir_node *mem = get_Builtin_mem(builtin);
2678 const lower64_entry_t *entry = get_node_entry(operand);
2679 ir_mode *mode_high = get_irn_mode(entry->high_word);
2685 ir_node *in_high[1] = { entry->high_word };
2686 ir_node *in_low[1] = { entry->low_word };
2687 ir_node *swap_high = new_rd_Builtin(dbgi, block, mem, 1, in_high, kind, lowered_type_high);
2688 ir_node *swap_low = new_rd_Builtin(dbgi, block, mem, 1, in_low, kind, lowered_type_low);
2689 ir_node *high = new_r_Proj(swap_high, mode_high, pn_Builtin_max+1);
2690 ir_node *low = new_r_Proj(swap_low, env->low_unsigned, pn_Builtin_max+1);
2691 if (mode_high == env->low_signed) {
2692 res_high = new_rd_Conv(dbgi, block, low, env->low_signed);
2693 res_low = new_rd_Conv(dbgi, block, high, env->low_unsigned);
2701 panic("unexpected builtin");
2704 /* search result Proj */
2705 foreach_out_edge_safe(builtin, edge) {
2706 ir_node *proj = get_edge_src_irn(edge);
2710 if (get_Proj_proj(proj) == pn_Builtin_max+1) {
2711 ir_set_dw_lowered(proj, res_low, res_high);
2718 * Lower double word builtins.
2720 static void lower_Builtin(ir_node *builtin, ir_mode *mode)
2722 ir_builtin_kind kind = get_Builtin_kind(builtin);
2726 case ir_bk_debugbreak:
2727 case ir_bk_return_address:
2728 case ir_bk_frame_address:
2729 case ir_bk_prefetch:
2732 case ir_bk_inner_trampoline:
2733 /* Nothing to do. */
2736 lower_arithmetic_builtin(builtin, mode);
2741 case ir_bk_popcount:
2743 lower_reduce_builtin(builtin, mode);
2746 panic("unknown builtin");
2750 * check for opcodes that must always be lowered.
2752 static bool always_lower(unsigned code)
2771 * Compare two op_mode_entry_t's.
2773 static int cmp_op_mode(const void *elt, const void *key, size_t size)
2775 const op_mode_entry_t *e1 = (const op_mode_entry_t*)elt;
2776 const op_mode_entry_t *e2 = (const op_mode_entry_t*)key;
2779 return (e1->op != e2->op) | (e1->imode != e2->imode) | (e1->omode != e2->omode);
2783 * Compare two conv_tp_entry_t's.
2785 static int cmp_conv_tp(const void *elt, const void *key, size_t size)
2787 const conv_tp_entry_t *e1 = (const conv_tp_entry_t*)elt;
2788 const conv_tp_entry_t *e2 = (const conv_tp_entry_t*)key;
2791 return (e1->imode != e2->imode) | (e1->omode != e2->omode);
2795 * Enter a lowering function into an ir_op.
2797 void ir_register_dw_lower_function(ir_op *op, lower_dw_func func)
2799 op->ops.generic = (op_func)func;
2802 /* Determine which modes need to be lowered */
2803 static void setup_modes(void)
2805 unsigned size_bits = env->params->doubleword_size;
2806 ir_mode *doubleword_signed = NULL;
2807 ir_mode *doubleword_unsigned = NULL;
2808 size_t n_modes = ir_get_n_modes();
2809 ir_mode_arithmetic arithmetic;
2810 unsigned modulo_shift;
2813 /* search for doubleword modes... */
2814 for (i = 0; i < n_modes; ++i) {
2815 ir_mode *mode = ir_get_mode(i);
2816 if (!mode_is_int(mode))
2818 if (get_mode_size_bits(mode) != size_bits)
2820 if (mode_is_signed(mode)) {
2821 if (doubleword_signed != NULL) {
2822 /* sigh - the lowerer should really just lower all mode with
2823 * size_bits it finds. Unfortunately this required a bigger
2825 panic("multiple double word signed modes found");
2827 doubleword_signed = mode;
2829 if (doubleword_unsigned != NULL) {
2830 /* sigh - the lowerer should really just lower all mode with
2831 * size_bits it finds. Unfortunately this required a bigger
2833 panic("multiple double word unsigned modes found");
2835 doubleword_unsigned = mode;
2838 if (doubleword_signed == NULL || doubleword_unsigned == NULL) {
2839 panic("Couldn't find doubleword modes");
2842 arithmetic = get_mode_arithmetic(doubleword_signed);
2843 modulo_shift = get_mode_modulo_shift(doubleword_signed);
2845 assert(get_mode_size_bits(doubleword_unsigned) == size_bits);
2846 assert(size_bits % 2 == 0);
2847 assert(get_mode_sign(doubleword_signed) == 1);
2848 assert(get_mode_sign(doubleword_unsigned) == 0);
2849 assert(get_mode_sort(doubleword_signed) == irms_int_number);
2850 assert(get_mode_sort(doubleword_unsigned) == irms_int_number);
2851 assert(get_mode_arithmetic(doubleword_unsigned) == arithmetic);
2852 assert(get_mode_modulo_shift(doubleword_unsigned) == modulo_shift);
2854 /* try to guess a sensible modulo shift for the new mode.
2855 * (This is IMO another indication that this should really be a node
2856 * attribute instead of a mode thing) */
2857 if (modulo_shift == size_bits) {
2858 modulo_shift = modulo_shift / 2;
2859 } else if (modulo_shift == 0) {
2862 panic("Don't know what new modulo shift to use for lowered doubleword mode");
2866 /* produce lowered modes */
2867 env->high_signed = doubleword_signed;
2868 env->high_unsigned = doubleword_unsigned;
2869 env->low_signed = new_int_mode("WS", arithmetic, size_bits, 1,
2871 env->low_unsigned = new_int_mode("WU", arithmetic, size_bits, 0,
2875 static void enqueue_preds(ir_node *node)
2877 int arity = get_irn_arity(node);
2880 for (i = 0; i < arity; ++i) {
2881 ir_node *pred = get_irn_n(node, i);
2882 pdeq_putr(env->waitq, pred);
2886 static void lower_node(ir_node *node)
2894 lower64_entry_t *entry;
2896 if (irn_visited_else_mark(node))
2899 /* cycles are always broken at Phi and Block nodes. So we don't need special
2900 * magic in all the other lower functions */
2901 if (is_Block(node)) {
2902 enqueue_preds(node);
2904 } else if (is_Phi(node)) {
2909 /* depth-first: descend into operands */
2910 if (!is_Block(node)) {
2911 ir_node *block = get_nodes_block(node);
2915 if (!is_Cond(node)) {
2916 arity = get_irn_arity(node);
2917 for (i = 0; i < arity; ++i) {
2918 ir_node *pred = get_irn_n(node, i);
2923 op = get_irn_op(node);
2924 func = (lower_dw_func) op->ops.generic;
2928 idx = get_irn_idx(node);
2929 entry = idx < env->n_entries ? env->entries[idx] : NULL;
2930 if (entry != NULL || always_lower(get_irn_opcode(node))) {
2931 mode = get_irn_op_mode(node);
2932 if (mode == env->high_signed) {
2933 mode = env->low_signed;
2935 mode = env->low_unsigned;
2937 DB((dbg, LEVEL_1, " %+F\n", node));
2942 static void clear_node_and_phi_links(ir_node *node, void *data)
2945 if (get_irn_mode(node) == mode_T) {
2946 set_irn_link(node, node);
2948 set_irn_link(node, NULL);
2951 set_Block_phis(node, NULL);
2952 else if (is_Phi(node))
2953 set_Phi_next(node, NULL);
2956 static void lower_irg(ir_graph *irg)
2960 ir_type *lowered_mtp;
2963 obstack_init(&env->obst);
2965 /* just here for debugging */
2966 current_ir_graph = irg;
2969 n_idx = get_irg_last_idx(irg);
2970 n_idx = n_idx + (n_idx >> 2); /* add 25% */
2971 env->n_entries = n_idx;
2972 env->entries = NEW_ARR_F(lower64_entry_t*, n_idx);
2973 memset(env->entries, 0, sizeof(env->entries[0]) * n_idx);
2978 ent = get_irg_entity(irg);
2979 mtp = get_entity_type(ent);
2980 lowered_mtp = lower_mtp(mtp);
2982 if (lowered_mtp != mtp) {
2983 set_entity_type(ent, lowered_mtp);
2984 env->flags |= MUST_BE_LOWERED;
2986 fix_parameter_entities(irg, mtp);
2989 /* first step: link all nodes and allocate data */
2990 ir_reserve_resources(irg, IR_RESOURCE_PHI_LIST | IR_RESOURCE_IRN_LINK);
2991 visit_all_identities(irg, clear_node_and_phi_links, NULL);
2992 irg_walk_graph(irg, NULL, prepare_links_and_handle_rotl, env);
2994 if (env->flags & MUST_BE_LOWERED) {
2996 ir_reserve_resources(irg, IR_RESOURCE_IRN_VISITED);
2997 inc_irg_visited(irg);
2999 assert(pdeq_empty(env->waitq));
3000 pdeq_putr(env->waitq, get_irg_end(irg));
3002 env->lowered_phis = NEW_ARR_F(ir_node*, 0);
3003 while (!pdeq_empty(env->waitq)) {
3004 ir_node *node = (ir_node*)pdeq_getl(env->waitq);
3008 /* we need to fixup phis */
3009 for (i = 0; i < ARR_LEN(env->lowered_phis); ++i) {
3010 ir_node *phi = env->lowered_phis[i];
3013 DEL_ARR_F(env->lowered_phis);
3016 ir_free_resources(irg, IR_RESOURCE_IRN_VISITED);
3018 if (env->flags & CF_CHANGED) {
3019 /* control flow changed, dominance info is invalid */
3020 clear_irg_properties(irg, IR_GRAPH_PROPERTY_CONSISTENT_DOMINANCE);
3022 edges_deactivate(irg);
3025 ir_free_resources(irg, IR_RESOURCE_PHI_LIST | IR_RESOURCE_IRN_LINK);
3027 DEL_ARR_F(env->entries);
3028 obstack_free(&env->obst, NULL);
3031 static const lwrdw_param_t *param;
3033 void ir_prepare_dw_lowering(const lwrdw_param_t *new_param)
3035 assert(new_param != NULL);
3036 FIRM_DBG_REGISTER(dbg, "firm.lower.dw");
3040 ir_clear_opcodes_generic_func();
3041 ir_register_dw_lower_function(op_ASM, lower_ASM);
3042 ir_register_dw_lower_function(op_Add, lower_binop);
3043 ir_register_dw_lower_function(op_And, lower_And);
3044 ir_register_dw_lower_function(op_Bad, lower_Bad);
3045 ir_register_dw_lower_function(op_Builtin, lower_Builtin);
3046 ir_register_dw_lower_function(op_Call, lower_Call);
3047 ir_register_dw_lower_function(op_Cmp, lower_Cmp);
3048 ir_register_dw_lower_function(op_Cond, lower_Cond);
3049 ir_register_dw_lower_function(op_Const, lower_Const);
3050 ir_register_dw_lower_function(op_Conv, lower_Conv);
3051 ir_register_dw_lower_function(op_Div, lower_Div);
3052 ir_register_dw_lower_function(op_Eor, lower_Eor);
3053 ir_register_dw_lower_function(op_Load, lower_Load);
3054 ir_register_dw_lower_function(op_Minus, lower_unop);
3055 ir_register_dw_lower_function(op_Mod, lower_Mod);
3056 ir_register_dw_lower_function(op_Mul, lower_binop);
3057 ir_register_dw_lower_function(op_Mux, lower_Mux);
3058 ir_register_dw_lower_function(op_Not, lower_Not);
3059 ir_register_dw_lower_function(op_Or, lower_Or);
3060 ir_register_dw_lower_function(op_Proj, lower_Proj);
3061 ir_register_dw_lower_function(op_Return, lower_Return);
3062 ir_register_dw_lower_function(op_Shl, lower_Shl);
3063 ir_register_dw_lower_function(op_Shr, lower_Shr);
3064 ir_register_dw_lower_function(op_Shrs, lower_Shrs);
3065 ir_register_dw_lower_function(op_Start, lower_Start);
3066 ir_register_dw_lower_function(op_Store, lower_Store);
3067 ir_register_dw_lower_function(op_Sub, lower_binop);
3068 ir_register_dw_lower_function(op_Switch, lower_Switch);
3069 ir_register_dw_lower_function(op_Unknown, lower_Unknown);
3073 * Callback to lower only the Mux nodes we created.
3075 static int lower_mux_cb(ir_node *mux)
3077 return ir_nodeset_contains(&created_mux_nodes, mux);
3083 void ir_lower_dw_ops(void)
3085 lower_dw_env_t lenv;
3088 memset(&lenv, 0, sizeof(lenv));
3089 lenv.params = param;
3094 /* create the necessary maps */
3095 if (! intrinsic_fkt)
3096 intrinsic_fkt = new_set(cmp_op_mode, iro_Last + 1);
3098 conv_types = new_set(cmp_conv_tp, 16);
3100 lowered_type = pmap_create();
3101 if (! lowered_builtin_type_low)
3102 lowered_builtin_type_low = pmap_create();
3103 if (! lowered_builtin_type_high)
3104 lowered_builtin_type_high = pmap_create();
3106 /* create a primitive unsigned and signed type */
3108 tp_u = get_type_for_mode(lenv.low_unsigned);
3110 tp_s = get_type_for_mode(lenv.low_signed);
3112 /* create method types for the created binop calls */
3114 binop_tp_u = new_type_method(4, 2);
3115 set_method_param_type(binop_tp_u, 0, tp_u);
3116 set_method_param_type(binop_tp_u, 1, tp_u);
3117 set_method_param_type(binop_tp_u, 2, tp_u);
3118 set_method_param_type(binop_tp_u, 3, tp_u);
3119 set_method_res_type(binop_tp_u, 0, tp_u);
3120 set_method_res_type(binop_tp_u, 1, tp_u);
3123 binop_tp_s = new_type_method(4, 2);
3124 if (env->params->little_endian) {
3125 set_method_param_type(binop_tp_s, 0, tp_u);
3126 set_method_param_type(binop_tp_s, 1, tp_s);
3127 set_method_param_type(binop_tp_s, 2, tp_u);
3128 set_method_param_type(binop_tp_s, 3, tp_s);
3129 set_method_res_type(binop_tp_s, 0, tp_u);
3130 set_method_res_type(binop_tp_s, 1, tp_s);
3132 set_method_param_type(binop_tp_s, 0, tp_s);
3133 set_method_param_type(binop_tp_s, 1, tp_u);
3134 set_method_param_type(binop_tp_s, 2, tp_s);
3135 set_method_param_type(binop_tp_s, 3, tp_u);
3136 set_method_res_type(binop_tp_s, 0, tp_s);
3137 set_method_res_type(binop_tp_s, 1, tp_u);
3141 unop_tp_u = new_type_method(2, 2);
3142 set_method_param_type(unop_tp_u, 0, tp_u);
3143 set_method_param_type(unop_tp_u, 1, tp_u);
3144 set_method_res_type(unop_tp_u, 0, tp_u);
3145 set_method_res_type(unop_tp_u, 1, tp_u);
3148 unop_tp_s = new_type_method(2, 2);
3149 if (env->params->little_endian) {
3150 set_method_param_type(unop_tp_s, 0, tp_u);
3151 set_method_param_type(unop_tp_s, 1, tp_s);
3152 set_method_res_type(unop_tp_s, 0, tp_u);
3153 set_method_res_type(unop_tp_s, 1, tp_s);
3155 set_method_param_type(unop_tp_s, 0, tp_s);
3156 set_method_param_type(unop_tp_s, 1, tp_u);
3157 set_method_res_type(unop_tp_s, 0, tp_s);
3158 set_method_res_type(unop_tp_s, 1, tp_u);
3162 lenv.tv_mode_bytes = new_tarval_from_long(param->doubleword_size/(2*8), lenv.low_unsigned);
3163 lenv.tv_mode_bits = new_tarval_from_long(param->doubleword_size/2, lenv.low_unsigned);
3164 lenv.waitq = new_pdeq();
3165 lenv.first_id = new_id_from_chars(param->little_endian ? ".l" : ".h", 2);
3166 lenv.next_id = new_id_from_chars(param->little_endian ? ".h" : ".l", 2);
3168 irp_reserve_resources(irp, IRP_RESOURCE_TYPE_LINK | IRP_RESOURCE_TYPE_VISITED);
3169 inc_master_type_visited();
3170 /* transform all graphs */
3171 for (i = 0, n = get_irp_n_irgs(); i < n; ++i) {
3172 ir_graph *irg = get_irp_irg(i);
3174 ir_nodeset_init(&created_mux_nodes);
3178 if (ir_nodeset_size(&created_mux_nodes) > 0)
3179 lower_mux(irg, lower_mux_cb);
3181 ir_nodeset_destroy(&created_mux_nodes);
3183 irp_free_resources(irp, IRP_RESOURCE_TYPE_LINK | IRP_RESOURCE_TYPE_VISITED);
3184 del_pdeq(lenv.waitq);
3189 /* Default implementation. */
3190 ir_entity *def_create_intrinsic_fkt(ir_type *method, const ir_op *op,
3191 const ir_mode *imode, const ir_mode *omode,
3199 if (imode == omode) {
3200 snprintf(buf, sizeof(buf), "__l%s%s", get_op_name(op), get_mode_name(imode));
3202 snprintf(buf, sizeof(buf), "__l%s%s%s", get_op_name(op),
3203 get_mode_name(imode), get_mode_name(omode));
3205 id = new_id_from_str(buf);
3207 ent = new_entity(get_glob_type(), id, method);
3208 set_entity_ld_ident(ent, get_entity_ident(ent));
3209 set_entity_visibility(ent, ir_visibility_external);