6d98e4f33b1d9ab530f90de32c6410b325367cdf
[libfirm] / ir / be / beflags.c
1 /*
2  * Copyright (C) 1995-2007 University of Karlsruhe.  All right reserved.
3  *
4  * This file is part of libFirm.
5  *
6  * This file may be distributed and/or modified under the terms of the
7  * GNU General Public License version 2 as published by the Free Software
8  * Foundation and appearing in the file LICENSE.GPL included in the
9  * packaging of this file.
10  *
11  * Licensees holding valid libFirm Professional Edition licenses may use
12  * this file in accordance with the libFirm Commercial License.
13  * Agreement provided with the Software.
14  *
15  * This file is provided AS IS with NO WARRANTY OF ANY KIND, INCLUDING THE
16  * WARRANTY OF DESIGN, MERCHANTABILITY AND FITNESS FOR A PARTICULAR
17  * PURPOSE.
18  */
19
20 /**
21  * @file
22  * @brief       modifies schedule so flags dependencies are respected.
23  * @author      Matthias Braun, Christoph Mallon
24  * @version     $Id: besched.h 14693 2007-06-21 15:35:49Z beck $
25  *
26  * Fixup schedule to respect flag constraints by moving and rematerialisation of
27  * nodes.
28  *
29  * Flags are modeled as register classes with ignore registers. However to avoid
30  * bloating the graph, only flag-consumer -> producer dependencies are
31  * explicitely modeled in the graph. Nodes that just change the flags are only
32  * marked with the arch_irn_flags_modify_flags flag.
33  *
34  * Flags are usually a limited resource that can't (or at least shouldn't) be
35  * spilled. So in some situations (for example 2 adc-nodes that use the flags of
36  * a single add node on x86) operations have to be repeated to work correctly.
37  */
38 #ifdef HAVE_CONFIG_H
39 #include "config.h"
40 #endif
41
42 #include "irgwalk.h"
43 #include "irnode_t.h"
44 #include "irtools.h"
45 #include "ircons.h"
46 #include "iredges_t.h"
47 #include "irprintf.h"
48 #include "error.h"
49
50 #include "beflags.h"
51 #include "bearch_t.h"
52 #include "beirg_t.h"
53 #include "besched_t.h"
54 #include "benode_t.h"
55
56 static const arch_env_t            *arch_env   = NULL;
57 static const arch_register_class_t *flag_class = NULL;
58 static const arch_register_t       *flags_reg  = NULL;
59 static func_rematerialize           remat      = NULL;
60 static int                          changed;
61
62 static ir_node *default_remat(ir_node *node, ir_node *after)
63 {
64         ir_node *block, *copy;
65         if(is_Block(after))
66                 block = after;
67         else
68                 block = get_nodes_block(after);
69
70         copy = exact_copy(node);
71         set_nodes_block(copy, block);
72         sched_add_after(after, copy);
73
74         return copy;
75 }
76
77 /**
78  * tests wether we can legally move node node after node after
79  * (only works for nodes in same block)
80  */
81 static int can_move(ir_node *node, ir_node *after)
82 {
83         const ir_edge_t *edge;
84         assert(get_nodes_block(node) == get_nodes_block(after));
85
86         /* TODO respect dep edges */
87         assert(get_irn_n_edges_kind(node, EDGE_KIND_DEP) == 0);
88
89         /** all users have to be after the after node */
90         foreach_out_edge(node, edge) {
91                 ir_node *out = get_edge_src_irn(edge);
92                 if(is_Proj(out)) {
93                         const ir_edge_t *edge2;
94                         assert(get_irn_n_edges_kind(out, EDGE_KIND_DEP) == 0);
95                         foreach_out_edge(out, edge2) {
96                                 ir_node *out2 = get_edge_src_irn(edge2);
97                                 /* phi represents a usage at block end */
98                                 if(is_Phi(out2))
99                                         continue;
100                                 if(is_Sync(out2)) {
101                                         const ir_edge_t *edge3;
102                                         foreach_out_edge(out2, edge3) {
103                                                 ir_node *out3 = get_edge_src_irn(edge3);
104                                                 /* phi represents a usage at block end */
105                                                 if(is_Phi(out3))
106                                                         continue;
107                                                 assert(!is_Sync(out3));
108                                                 if(sched_get_time_step(out3) <= sched_get_time_step(after)) {
109                                                         return 0;
110                                                 }
111                                         }
112                                 } else if(sched_get_time_step(out2) <= sched_get_time_step(after)) {
113                                         return 0;
114                                 }
115                         }
116                 } else {
117                         /* phi represents a usage at block end */
118                         if(is_Phi(out))
119                                 continue;
120                         if(sched_get_time_step(out) <= sched_get_time_step(after)) {
121                                 return 0;
122                         }
123                 }
124         }
125
126         return 1;
127 }
128
129 static void rematerialize_or_move(ir_node *flags_needed, ir_node *node,
130                                   ir_node *flag_consumers, int pn)
131 {
132         ir_node *n;
133         ir_node *copy;
134         ir_node *value;
135
136         if(!is_Block(node) &&
137                         get_nodes_block(flags_needed) == get_nodes_block(node) &&
138                         can_move(flags_needed, node)) {
139                 /* move it */
140                 sched_remove(flags_needed);
141                 sched_add_after(node, flags_needed);
142                 return;
143         }
144
145         changed = 1;
146         copy    = remat(flags_needed, node);
147
148         if(get_irn_mode(copy) == mode_T) {
149                 ir_node *block = get_nodes_block(copy);
150                 ir_mode *mode  = flag_class->mode;
151                 value = new_rd_Proj(NULL, current_ir_graph, block,
152                                                         copy, mode, pn);
153         } else {
154                 value = copy;
155         }
156
157         n = flag_consumers;
158         do {
159                 int i;
160                 int arity = get_irn_arity(n);
161                 for(i = 0; i < arity; ++i) {
162                         ir_node *in = get_irn_n(n, i);
163                         in = skip_Proj(in);
164                         if(in == flags_needed) {
165                                 set_irn_n(n, i, value);
166                                 break;
167                         }
168                 }
169                 n = get_irn_link(n);
170         } while(n != NULL);
171 }
172
173 static int is_modify_flags(ir_node *node) {
174         int i, arity;
175
176         if(arch_irn_is(arch_env, node, modify_flags))
177                 return 1;
178         if(!be_is_Keep(node))
179                 return 0;
180
181         arity = get_irn_arity(node);
182         for(i = 0; i < arity; ++i) {
183                 ir_node *in = get_irn_n(node, i);
184                 in = skip_Proj(in);
185                 if(arch_irn_is(arch_env, in, modify_flags))
186                         return 1;
187         }
188
189         return 0;
190 }
191
192 /**
193  * walks up the schedule and makes sure there are no flag-destroying nodes
194  * between a flag-consumer -> flag-producer chain. Fixes problematic situations
195  * by moving and/or rematerialisation of the flag-producers.
196  * (This can be extended in the future to do some register allocation on targets
197  *  like ppc32 where we conceptually have 8 flag registers)
198  */
199 static void fix_flags_walker(ir_node *block, void *env)
200 {
201         ir_node *node;
202         ir_node *flags_needed   = NULL;
203         ir_node *flag_consumers = NULL;
204         int      pn = -1;
205         (void) env;
206
207         sched_foreach_reverse(block, node) {
208                 int i, arity;
209                 ir_node *new_flags_needed = NULL;
210
211                 if(is_Phi(node))
212                         break;
213
214                 if(node == flags_needed) {
215                         /* all ok */
216                         flags_needed   = NULL;
217                         flag_consumers = NULL;
218                 }
219
220                 /* test whether node destroys the flags */
221                 if(flags_needed != NULL && is_modify_flags(node)) {
222                         /* rematerialize */
223                         rematerialize_or_move(flags_needed, node, flag_consumers, pn);
224                         flags_needed   = NULL;
225                         flag_consumers = NULL;
226                 }
227
228                 /* test whether the current node needs flags */
229                 arity = get_irn_arity(node);
230                 for(i = 0; i < arity; ++i) {
231                         const arch_register_class_t *cls
232                                 = arch_get_irn_reg_class(arch_env, node, i);
233                         if(cls == flag_class) {
234                                 assert(new_flags_needed == NULL);
235                                 new_flags_needed = get_irn_n(node, i);
236                         }
237                 }
238
239                 if(new_flags_needed == NULL)
240                         continue;
241
242                 /* spiller can't (correctly) remat flag consumers at the moment */
243                 assert(!arch_irn_is(arch_env, node, rematerializable));
244
245                 if(skip_Proj(new_flags_needed) != flags_needed) {
246                         if(flags_needed != NULL) {
247                                 /* rematerialize node */
248                                 rematerialize_or_move(flags_needed, node, flag_consumers, pn);
249                                 flags_needed   = NULL;
250                                 flag_consumers = NULL;
251                         }
252
253                         flags_needed = new_flags_needed;
254                         arch_set_irn_register(arch_env, flags_needed, flags_reg);
255                         if(is_Proj(flags_needed)) {
256                                 pn           = get_Proj_proj(flags_needed);
257                                 flags_needed = get_Proj_pred(flags_needed);
258                         }
259                         flag_consumers = node;
260                         set_irn_link(flag_consumers, NULL);
261                         assert(arch_irn_is(arch_env, flags_needed, rematerializable));
262                 } else {
263                         /* link all consumers in a list */
264                         set_irn_link(node, flag_consumers);
265                         flag_consumers = node;
266                 }
267         }
268
269         if(flags_needed != NULL) {
270                 assert(get_nodes_block(flags_needed) != block);
271                 rematerialize_or_move(flags_needed, node, flag_consumers, pn);
272                 flags_needed   = NULL;
273                 flag_consumers = NULL;
274         }
275
276         assert(flags_needed   == NULL);
277         assert(flag_consumers == NULL);
278 }
279
280 void be_sched_fix_flags(be_irg_t *birg, const arch_register_class_t *flag_cls,
281                         func_rematerialize remat_func)
282 {
283         ir_graph *irg = be_get_birg_irg(birg);
284
285         arch_env   = be_get_birg_arch_env(birg);
286         flag_class = flag_cls;
287         flags_reg  = & flag_class->regs[0];
288         remat      = remat_func;
289         changed    = 0;
290         if(remat == NULL)
291                 remat = &default_remat;
292
293         set_using_irn_link(irg);
294         irg_block_walk_graph(irg, fix_flags_walker, NULL, NULL);
295         clear_using_irn_link(irg);
296
297         if(changed) {
298                 be_remove_dead_nodes_from_schedule(birg);
299         }
300 }