1 /* IRA allocation based on graph coloring.
2 Copyright (C) 2006-2024 Free Software Foundation, Inc.
3 Contributed by Vladimir Makarov <vmakarov@redhat.com>.
5 This file is part of GCC.
7 GCC is free software; you can redistribute it and/or modify it under
8 the terms of the GNU General Public License as published by the Free
9 Software Foundation; either version 3, or (at your option) any later
12 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
13 WARRANTY; without even the implied warranty of MERCHANTABILITY or
14 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
17 You should have received a copy of the GNU General Public License
18 along with GCC; see the file COPYING3. If not see
19 <http://www.gnu.org/licenses/>. */
23 #include "coretypes.h"
32 #include "insn-config.h"
39 /* To prevent soft conflict detection becoming quadratic in the
40 loop depth. Only for very pathological cases, so it hardly
41 seems worth a --param. */
42 const int max_soft_conflict_loop_depth
= 64;
44 typedef struct allocno_hard_regs
*allocno_hard_regs_t
;
46 /* The structure contains information about hard registers can be
47 assigned to allocnos. Usually it is allocno profitable hard
48 registers but in some cases this set can be a bit different. Major
49 reason of the difference is a requirement to use hard register sets
50 that form a tree or a forest (set of trees), i.e. hard register set
51 of a node should contain hard register sets of its subnodes. */
52 struct allocno_hard_regs
54 /* Hard registers can be assigned to an allocno. */
56 /* Overall (spilling) cost of all allocnos with given register
61 typedef struct allocno_hard_regs_node
*allocno_hard_regs_node_t
;
63 /* A node representing allocno hard registers. Such nodes form a
64 forest (set of trees). Each subnode of given node in the forest
65 refers for hard register set (usually allocno profitable hard
66 register set) which is a subset of one referred from given
68 struct allocno_hard_regs_node
70 /* Set up number of the node in preorder traversing of the forest. */
72 /* Used for different calculation like finding conflict size of an
75 /* Used for calculation of conflict size of an allocno. The
76 conflict size of the allocno is maximal number of given allocno
77 hard registers needed for allocation of the conflicting allocnos.
78 Given allocno is trivially colored if this number plus the number
79 of hard registers needed for given allocno is not greater than
80 the number of given allocno hard register set. */
82 /* The number of hard registers given by member hard_regs. */
84 /* The following member is used to form the final forest. */
86 /* Pointer to the corresponding profitable hard registers. */
87 allocno_hard_regs_t hard_regs
;
88 /* Parent, first subnode, previous and next node with the same
89 parent in the forest. */
90 allocno_hard_regs_node_t parent
, first
, prev
, next
;
93 /* Info about changing hard reg costs of an allocno. */
94 struct update_cost_record
96 /* Hard regno for which we changed the cost. */
98 /* Divisor used when we changed the cost of HARD_REGNO. */
100 /* Next record for given allocno. */
101 struct update_cost_record
*next
;
104 /* To decrease footprint of ira_allocno structure we store all data
105 needed only for coloring in the following structure. */
106 struct allocno_color_data
108 /* TRUE value means that the allocno was not removed yet from the
109 conflicting graph during coloring. */
110 unsigned int in_graph_p
: 1;
111 /* TRUE if it is put on the stack to make other allocnos
113 unsigned int may_be_spilled_p
: 1;
114 /* TRUE if the allocno is trivially colorable. */
115 unsigned int colorable_p
: 1;
116 /* Number of hard registers of the allocno class really
117 available for the allocno allocation. It is number of the
118 profitable hard regs. */
119 int available_regs_num
;
120 /* Sum of frequencies of hard register preferences of all
121 conflicting allocnos which are not the coloring stack yet. */
122 int conflict_allocno_hard_prefs
;
123 /* Allocnos in a bucket (used in coloring) chained by the following
125 ira_allocno_t next_bucket_allocno
;
126 ira_allocno_t prev_bucket_allocno
;
127 /* Used for temporary purposes. */
129 /* Used to exclude repeated processing. */
131 /* Profitable hard regs available for this pseudo allocation. It
132 means that the set excludes unavailable hard regs and hard regs
133 conflicting with given pseudo. They should be of the allocno
135 HARD_REG_SET profitable_hard_regs
;
136 /* The allocno hard registers node. */
137 allocno_hard_regs_node_t hard_regs_node
;
138 /* Array of structures allocno_hard_regs_subnode representing
139 given allocno hard registers node (the 1st element in the array)
140 and all its subnodes in the tree (forest) of allocno hard
141 register nodes (see comments above). */
142 int hard_regs_subnodes_start
;
143 /* The length of the previous array. */
144 int hard_regs_subnodes_num
;
145 /* Records about updating allocno hard reg costs from copies. If
146 the allocno did not get expected hard register, these records are
147 used to restore original hard reg costs of allocnos connected to
148 this allocno by copies. */
149 struct update_cost_record
*update_cost_records
;
150 /* Threads. We collect allocnos connected by copies into threads
151 and try to assign hard regs to allocnos by threads. */
152 /* Allocno representing all thread. */
153 ira_allocno_t first_thread_allocno
;
154 /* Allocnos in thread forms a cycle list through the following
156 ira_allocno_t next_thread_allocno
;
157 /* All thread frequency. Defined only for first thread allocno. */
159 /* Sum of frequencies of hard register preferences of the allocno. */
164 typedef struct allocno_color_data
*allocno_color_data_t
;
166 /* Container for storing allocno data concerning coloring. */
167 static allocno_color_data_t allocno_color_data
;
169 /* Macro to access the data concerning coloring. */
170 #define ALLOCNO_COLOR_DATA(a) ((allocno_color_data_t) ALLOCNO_ADD_DATA (a))
172 /* Used for finding allocno colorability to exclude repeated allocno
173 processing and for updating preferencing to exclude repeated
174 allocno processing during assignment. */
175 static int curr_allocno_process
;
177 /* This file contains code for regional graph coloring, spill/restore
178 code placement optimization, and code helping the reload pass to do
181 /* Bitmap of allocnos which should be colored. */
182 static bitmap coloring_allocno_bitmap
;
184 /* Bitmap of allocnos which should be taken into account during
185 coloring. In general case it contains allocnos from
186 coloring_allocno_bitmap plus other already colored conflicting
188 static bitmap consideration_allocno_bitmap
;
190 /* All allocnos sorted according their priorities. */
191 static ira_allocno_t
*sorted_allocnos
;
193 /* Vec representing the stack of allocnos used during coloring. */
194 static vec
<ira_allocno_t
> allocno_stack_vec
;
196 /* Helper for qsort comparison callbacks - return a positive integer if
197 X > Y, or a negative value otherwise. Use a conditional expression
198 instead of a difference computation to insulate from possible overflow
199 issues, e.g. X - Y < 0 for some X > 0 and Y < 0. */
200 #define SORTGT(x,y) (((x) > (y)) ? 1 : -1)
204 /* Definition of vector of allocno hard registers. */
206 /* Vector of unique allocno hard registers. */
207 static vec
<allocno_hard_regs_t
> allocno_hard_regs_vec
;
209 struct allocno_hard_regs_hasher
: nofree_ptr_hash
<allocno_hard_regs
>
211 static inline hashval_t
hash (const allocno_hard_regs
*);
212 static inline bool equal (const allocno_hard_regs
*,
213 const allocno_hard_regs
*);
216 /* Returns hash value for allocno hard registers V. */
218 allocno_hard_regs_hasher::hash (const allocno_hard_regs
*hv
)
220 return iterative_hash (&hv
->set
, sizeof (HARD_REG_SET
), 0);
223 /* Compares allocno hard registers V1 and V2. */
225 allocno_hard_regs_hasher::equal (const allocno_hard_regs
*hv1
,
226 const allocno_hard_regs
*hv2
)
228 return hv1
->set
== hv2
->set
;
231 /* Hash table of unique allocno hard registers. */
232 static hash_table
<allocno_hard_regs_hasher
> *allocno_hard_regs_htab
;
234 /* Return allocno hard registers in the hash table equal to HV. */
235 static allocno_hard_regs_t
236 find_hard_regs (allocno_hard_regs_t hv
)
238 return allocno_hard_regs_htab
->find (hv
);
241 /* Insert allocno hard registers HV in the hash table (if it is not
242 there yet) and return the value which in the table. */
243 static allocno_hard_regs_t
244 insert_hard_regs (allocno_hard_regs_t hv
)
246 allocno_hard_regs
**slot
= allocno_hard_regs_htab
->find_slot (hv
, INSERT
);
253 /* Initialize data concerning allocno hard registers. */
255 init_allocno_hard_regs (void)
257 allocno_hard_regs_vec
.create (200);
258 allocno_hard_regs_htab
259 = new hash_table
<allocno_hard_regs_hasher
> (200);
262 /* Add (or update info about) allocno hard registers with SET and
264 static allocno_hard_regs_t
265 add_allocno_hard_regs (HARD_REG_SET set
, int64_t cost
)
267 struct allocno_hard_regs temp
;
268 allocno_hard_regs_t hv
;
270 gcc_assert (! hard_reg_set_empty_p (set
));
272 if ((hv
= find_hard_regs (&temp
)) != NULL
)
276 hv
= ((struct allocno_hard_regs
*)
277 ira_allocate (sizeof (struct allocno_hard_regs
)));
280 allocno_hard_regs_vec
.safe_push (hv
);
281 insert_hard_regs (hv
);
286 /* Finalize data concerning allocno hard registers. */
288 finish_allocno_hard_regs (void)
291 allocno_hard_regs_t hv
;
294 allocno_hard_regs_vec
.iterate (i
, &hv
);
297 delete allocno_hard_regs_htab
;
298 allocno_hard_regs_htab
= NULL
;
299 allocno_hard_regs_vec
.release ();
302 /* Sort hard regs according to their frequency of usage. */
304 allocno_hard_regs_compare (const void *v1p
, const void *v2p
)
306 allocno_hard_regs_t hv1
= *(const allocno_hard_regs_t
*) v1p
;
307 allocno_hard_regs_t hv2
= *(const allocno_hard_regs_t
*) v2p
;
309 if (hv2
->cost
> hv1
->cost
)
311 else if (hv2
->cost
< hv1
->cost
)
313 return SORTGT (allocno_hard_regs_hasher::hash(hv2
), allocno_hard_regs_hasher::hash(hv1
));
318 /* Used for finding a common ancestor of two allocno hard registers
319 nodes in the forest. We use the current value of
320 'node_check_tick' to mark all nodes from one node to the top and
321 then walking up from another node until we find a marked node.
323 It is also used to figure out allocno colorability as a mark that
324 we already reset value of member 'conflict_size' for the forest
325 node corresponding to the processed allocno. */
326 static int node_check_tick
;
328 /* Roots of the forest containing hard register sets can be assigned
330 static allocno_hard_regs_node_t hard_regs_roots
;
332 /* Definition of vector of allocno hard register nodes. */
334 /* Vector used to create the forest. */
335 static vec
<allocno_hard_regs_node_t
> hard_regs_node_vec
;
337 /* Create and return allocno hard registers node containing allocno
338 hard registers HV. */
339 static allocno_hard_regs_node_t
340 create_new_allocno_hard_regs_node (allocno_hard_regs_t hv
)
342 allocno_hard_regs_node_t new_node
;
344 new_node
= ((struct allocno_hard_regs_node
*)
345 ira_allocate (sizeof (struct allocno_hard_regs_node
)));
347 new_node
->hard_regs
= hv
;
348 new_node
->hard_regs_num
= hard_reg_set_size (hv
->set
);
349 new_node
->first
= NULL
;
350 new_node
->used_p
= false;
354 /* Add allocno hard registers node NEW_NODE to the forest on its level
357 add_new_allocno_hard_regs_node_to_forest (allocno_hard_regs_node_t
*roots
,
358 allocno_hard_regs_node_t new_node
)
360 new_node
->next
= *roots
;
361 if (new_node
->next
!= NULL
)
362 new_node
->next
->prev
= new_node
;
363 new_node
->prev
= NULL
;
367 /* Add allocno hard registers HV (or its best approximation if it is
368 not possible) to the forest on its level given by ROOTS. */
370 add_allocno_hard_regs_to_forest (allocno_hard_regs_node_t
*roots
,
371 allocno_hard_regs_t hv
)
373 unsigned int i
, start
;
374 allocno_hard_regs_node_t node
, prev
, new_node
;
375 HARD_REG_SET temp_set
;
376 allocno_hard_regs_t hv2
;
378 start
= hard_regs_node_vec
.length ();
379 for (node
= *roots
; node
!= NULL
; node
= node
->next
)
381 if (hv
->set
== node
->hard_regs
->set
)
383 if (hard_reg_set_subset_p (hv
->set
, node
->hard_regs
->set
))
385 add_allocno_hard_regs_to_forest (&node
->first
, hv
);
388 if (hard_reg_set_subset_p (node
->hard_regs
->set
, hv
->set
))
389 hard_regs_node_vec
.safe_push (node
);
390 else if (hard_reg_set_intersect_p (hv
->set
, node
->hard_regs
->set
))
392 temp_set
= hv
->set
& node
->hard_regs
->set
;
393 hv2
= add_allocno_hard_regs (temp_set
, hv
->cost
);
394 add_allocno_hard_regs_to_forest (&node
->first
, hv2
);
397 if (hard_regs_node_vec
.length ()
400 /* Create a new node which contains nodes in hard_regs_node_vec. */
401 CLEAR_HARD_REG_SET (temp_set
);
403 i
< hard_regs_node_vec
.length ();
406 node
= hard_regs_node_vec
[i
];
407 temp_set
|= node
->hard_regs
->set
;
409 hv
= add_allocno_hard_regs (temp_set
, hv
->cost
);
410 new_node
= create_new_allocno_hard_regs_node (hv
);
413 i
< hard_regs_node_vec
.length ();
416 node
= hard_regs_node_vec
[i
];
417 if (node
->prev
== NULL
)
420 node
->prev
->next
= node
->next
;
421 if (node
->next
!= NULL
)
422 node
->next
->prev
= node
->prev
;
424 new_node
->first
= node
;
431 add_new_allocno_hard_regs_node_to_forest (roots
, new_node
);
433 hard_regs_node_vec
.truncate (start
);
436 /* Add allocno hard registers nodes starting with the forest level
437 given by FIRST which contains biggest set inside SET. */
439 collect_allocno_hard_regs_cover (allocno_hard_regs_node_t first
,
442 allocno_hard_regs_node_t node
;
444 ira_assert (first
!= NULL
);
445 for (node
= first
; node
!= NULL
; node
= node
->next
)
446 if (hard_reg_set_subset_p (node
->hard_regs
->set
, set
))
447 hard_regs_node_vec
.safe_push (node
);
448 else if (hard_reg_set_intersect_p (set
, node
->hard_regs
->set
))
449 collect_allocno_hard_regs_cover (node
->first
, set
);
452 /* Set up field parent as PARENT in all allocno hard registers nodes
453 in forest given by FIRST. */
455 setup_allocno_hard_regs_nodes_parent (allocno_hard_regs_node_t first
,
456 allocno_hard_regs_node_t parent
)
458 allocno_hard_regs_node_t node
;
460 for (node
= first
; node
!= NULL
; node
= node
->next
)
462 node
->parent
= parent
;
463 setup_allocno_hard_regs_nodes_parent (node
->first
, node
);
467 /* Return allocno hard registers node which is a first common ancestor
468 node of FIRST and SECOND in the forest. */
469 static allocno_hard_regs_node_t
470 first_common_ancestor_node (allocno_hard_regs_node_t first
,
471 allocno_hard_regs_node_t second
)
473 allocno_hard_regs_node_t node
;
476 for (node
= first
; node
!= NULL
; node
= node
->parent
)
477 node
->check
= node_check_tick
;
478 for (node
= second
; node
!= NULL
; node
= node
->parent
)
479 if (node
->check
== node_check_tick
)
481 return first_common_ancestor_node (second
, first
);
484 /* Print hard reg set SET to F. */
486 print_hard_reg_set (FILE *f
, HARD_REG_SET set
, bool new_line_p
)
490 for (start
= end
= -1, i
= 0; i
< FIRST_PSEUDO_REGISTER
; i
++)
492 bool reg_included
= TEST_HARD_REG_BIT (set
, i
);
500 if (start
>= 0 && (!reg_included
|| i
== FIRST_PSEUDO_REGISTER
- 1))
503 fprintf (f
, " %d", start
);
504 else if (start
== end
+ 1)
505 fprintf (f
, " %d %d", start
, end
);
507 fprintf (f
, " %d-%d", start
, end
);
515 /* Dump a hard reg set SET to stderr. */
517 debug_hard_reg_set (HARD_REG_SET set
)
519 print_hard_reg_set (stderr
, set
, true);
522 /* Print allocno hard register subforest given by ROOTS and its LEVEL
525 print_hard_regs_subforest (FILE *f
, allocno_hard_regs_node_t roots
,
529 allocno_hard_regs_node_t node
;
531 for (node
= roots
; node
!= NULL
; node
= node
->next
)
534 for (i
= 0; i
< level
* 2; i
++)
536 fprintf (f
, "%d:(", node
->preorder_num
);
537 print_hard_reg_set (f
, node
->hard_regs
->set
, false);
538 fprintf (f
, ")@%" PRId64
"\n", node
->hard_regs
->cost
);
539 print_hard_regs_subforest (f
, node
->first
, level
+ 1);
543 /* Print the allocno hard register forest to F. */
545 print_hard_regs_forest (FILE *f
)
547 fprintf (f
, " Hard reg set forest:\n");
548 print_hard_regs_subforest (f
, hard_regs_roots
, 1);
551 /* Print the allocno hard register forest to stderr. */
553 ira_debug_hard_regs_forest (void)
555 print_hard_regs_forest (stderr
);
558 /* Remove unused allocno hard registers nodes from forest given by its
561 remove_unused_allocno_hard_regs_nodes (allocno_hard_regs_node_t
*roots
)
563 allocno_hard_regs_node_t node
, prev
, next
, last
;
565 for (prev
= NULL
, node
= *roots
; node
!= NULL
; node
= next
)
570 remove_unused_allocno_hard_regs_nodes (&node
->first
);
575 for (last
= node
->first
;
576 last
!= NULL
&& last
->next
!= NULL
;
582 *roots
= node
->first
;
584 prev
->next
= node
->first
;
604 /* Set up fields preorder_num starting with START_NUM in all allocno
605 hard registers nodes in forest given by FIRST. Return biggest set
606 PREORDER_NUM increased by 1. */
608 enumerate_allocno_hard_regs_nodes (allocno_hard_regs_node_t first
,
609 allocno_hard_regs_node_t parent
,
612 allocno_hard_regs_node_t node
;
614 for (node
= first
; node
!= NULL
; node
= node
->next
)
616 node
->preorder_num
= start_num
++;
617 node
->parent
= parent
;
618 start_num
= enumerate_allocno_hard_regs_nodes (node
->first
, node
,
624 /* Number of allocno hard registers nodes in the forest. */
625 static int allocno_hard_regs_nodes_num
;
627 /* Table preorder number of allocno hard registers node in the forest
628 -> the allocno hard registers node. */
629 static allocno_hard_regs_node_t
*allocno_hard_regs_nodes
;
632 typedef struct allocno_hard_regs_subnode
*allocno_hard_regs_subnode_t
;
634 /* The structure is used to describes all subnodes (not only immediate
635 ones) in the mentioned above tree for given allocno hard register
636 node. The usage of such data accelerates calculation of
637 colorability of given allocno. */
638 struct allocno_hard_regs_subnode
640 /* The conflict size of conflicting allocnos whose hard register
641 sets are equal sets (plus supersets if given node is given
642 allocno hard registers node) of one in the given node. */
643 int left_conflict_size
;
644 /* The summary conflict size of conflicting allocnos whose hard
645 register sets are strict subsets of one in the given node.
646 Overall conflict size is
647 left_conflict_subnodes_size
648 + MIN (max_node_impact - left_conflict_subnodes_size,
651 short left_conflict_subnodes_size
;
652 short max_node_impact
;
655 /* Container for hard regs subnodes of all allocnos. */
656 static allocno_hard_regs_subnode_t allocno_hard_regs_subnodes
;
658 /* Table (preorder number of allocno hard registers node in the
659 forest, preorder number of allocno hard registers subnode) -> index
660 of the subnode relative to the node. -1 if it is not a
662 static int *allocno_hard_regs_subnode_index
;
664 /* Setup arrays ALLOCNO_HARD_REGS_NODES and
665 ALLOCNO_HARD_REGS_SUBNODE_INDEX. */
667 setup_allocno_hard_regs_subnode_index (allocno_hard_regs_node_t first
)
669 allocno_hard_regs_node_t node
, parent
;
672 for (node
= first
; node
!= NULL
; node
= node
->next
)
674 allocno_hard_regs_nodes
[node
->preorder_num
] = node
;
675 for (parent
= node
; parent
!= NULL
; parent
= parent
->parent
)
677 index
= parent
->preorder_num
* allocno_hard_regs_nodes_num
;
678 allocno_hard_regs_subnode_index
[index
+ node
->preorder_num
]
679 = node
->preorder_num
- parent
->preorder_num
;
681 setup_allocno_hard_regs_subnode_index (node
->first
);
685 /* Count all allocno hard registers nodes in tree ROOT. */
687 get_allocno_hard_regs_subnodes_num (allocno_hard_regs_node_t root
)
691 for (root
= root
->first
; root
!= NULL
; root
= root
->next
)
692 len
+= get_allocno_hard_regs_subnodes_num (root
);
696 /* Build the forest of allocno hard registers nodes and assign each
697 allocno a node from the forest. */
699 form_allocno_hard_regs_nodes_forest (void)
701 unsigned int i
, j
, size
, len
;
704 allocno_hard_regs_t hv
;
707 allocno_hard_regs_node_t node
, allocno_hard_regs_node
;
708 allocno_color_data_t allocno_data
;
711 init_allocno_hard_regs ();
712 hard_regs_roots
= NULL
;
713 hard_regs_node_vec
.create (100);
714 for (i
= 0; i
< FIRST_PSEUDO_REGISTER
; i
++)
715 if (! TEST_HARD_REG_BIT (ira_no_alloc_regs
, i
))
717 CLEAR_HARD_REG_SET (temp
);
718 SET_HARD_REG_BIT (temp
, i
);
719 hv
= add_allocno_hard_regs (temp
, 0);
720 node
= create_new_allocno_hard_regs_node (hv
);
721 add_new_allocno_hard_regs_node_to_forest (&hard_regs_roots
, node
);
723 start
= allocno_hard_regs_vec
.length ();
724 EXECUTE_IF_SET_IN_BITMAP (coloring_allocno_bitmap
, 0, i
, bi
)
727 allocno_data
= ALLOCNO_COLOR_DATA (a
);
729 if (hard_reg_set_empty_p (allocno_data
->profitable_hard_regs
))
731 hv
= (add_allocno_hard_regs
732 (allocno_data
->profitable_hard_regs
,
733 ALLOCNO_MEMORY_COST (a
) - ALLOCNO_CLASS_COST (a
)));
735 temp
= ~ira_no_alloc_regs
;
736 add_allocno_hard_regs (temp
, 0);
737 qsort (allocno_hard_regs_vec
.address () + start
,
738 allocno_hard_regs_vec
.length () - start
,
739 sizeof (allocno_hard_regs_t
), allocno_hard_regs_compare
);
741 allocno_hard_regs_vec
.iterate (i
, &hv
);
744 add_allocno_hard_regs_to_forest (&hard_regs_roots
, hv
);
745 ira_assert (hard_regs_node_vec
.length () == 0);
747 /* We need to set up parent fields for right work of
748 first_common_ancestor_node. */
749 setup_allocno_hard_regs_nodes_parent (hard_regs_roots
, NULL
);
750 EXECUTE_IF_SET_IN_BITMAP (coloring_allocno_bitmap
, 0, i
, bi
)
753 allocno_data
= ALLOCNO_COLOR_DATA (a
);
754 if (hard_reg_set_empty_p (allocno_data
->profitable_hard_regs
))
756 hard_regs_node_vec
.truncate (0);
757 collect_allocno_hard_regs_cover (hard_regs_roots
,
758 allocno_data
->profitable_hard_regs
);
759 allocno_hard_regs_node
= NULL
;
760 for (j
= 0; hard_regs_node_vec
.iterate (j
, &node
); j
++)
761 allocno_hard_regs_node
764 : first_common_ancestor_node (node
, allocno_hard_regs_node
));
765 /* That is a temporary storage. */
766 allocno_hard_regs_node
->used_p
= true;
767 allocno_data
->hard_regs_node
= allocno_hard_regs_node
;
769 ira_assert (hard_regs_roots
->next
== NULL
);
770 hard_regs_roots
->used_p
= true;
771 remove_unused_allocno_hard_regs_nodes (&hard_regs_roots
);
772 allocno_hard_regs_nodes_num
773 = enumerate_allocno_hard_regs_nodes (hard_regs_roots
, NULL
, 0);
774 allocno_hard_regs_nodes
775 = ((allocno_hard_regs_node_t
*)
776 ira_allocate (allocno_hard_regs_nodes_num
777 * sizeof (allocno_hard_regs_node_t
)));
778 size
= allocno_hard_regs_nodes_num
* allocno_hard_regs_nodes_num
;
779 allocno_hard_regs_subnode_index
780 = (int *) ira_allocate (size
* sizeof (int));
781 for (i
= 0; i
< size
; i
++)
782 allocno_hard_regs_subnode_index
[i
] = -1;
783 setup_allocno_hard_regs_subnode_index (hard_regs_roots
);
785 EXECUTE_IF_SET_IN_BITMAP (coloring_allocno_bitmap
, 0, i
, bi
)
788 allocno_data
= ALLOCNO_COLOR_DATA (a
);
789 if (hard_reg_set_empty_p (allocno_data
->profitable_hard_regs
))
791 len
= get_allocno_hard_regs_subnodes_num (allocno_data
->hard_regs_node
);
792 allocno_data
->hard_regs_subnodes_start
= start
;
793 allocno_data
->hard_regs_subnodes_num
= len
;
796 allocno_hard_regs_subnodes
797 = ((allocno_hard_regs_subnode_t
)
798 ira_allocate (sizeof (struct allocno_hard_regs_subnode
) * start
));
799 hard_regs_node_vec
.release ();
802 /* Free tree of allocno hard registers nodes given by its ROOT. */
804 finish_allocno_hard_regs_nodes_tree (allocno_hard_regs_node_t root
)
806 allocno_hard_regs_node_t child
, next
;
808 for (child
= root
->first
; child
!= NULL
; child
= next
)
811 finish_allocno_hard_regs_nodes_tree (child
);
816 /* Finish work with the forest of allocno hard registers nodes. */
818 finish_allocno_hard_regs_nodes_forest (void)
820 allocno_hard_regs_node_t node
, next
;
822 ira_free (allocno_hard_regs_subnodes
);
823 for (node
= hard_regs_roots
; node
!= NULL
; node
= next
)
826 finish_allocno_hard_regs_nodes_tree (node
);
828 ira_free (allocno_hard_regs_nodes
);
829 ira_free (allocno_hard_regs_subnode_index
);
830 finish_allocno_hard_regs ();
833 /* Set up left conflict sizes and left conflict subnodes sizes of hard
834 registers subnodes of allocno A. Return TRUE if allocno A is
835 trivially colorable. */
837 setup_left_conflict_sizes_p (ira_allocno_t a
)
839 int i
, k
, nobj
, start
;
840 int conflict_size
, left_conflict_subnodes_size
, node_preorder_num
;
841 allocno_color_data_t data
;
842 HARD_REG_SET profitable_hard_regs
;
843 allocno_hard_regs_subnode_t subnodes
;
844 allocno_hard_regs_node_t node
;
845 HARD_REG_SET node_set
;
847 nobj
= ALLOCNO_NUM_OBJECTS (a
);
848 data
= ALLOCNO_COLOR_DATA (a
);
849 subnodes
= allocno_hard_regs_subnodes
+ data
->hard_regs_subnodes_start
;
850 profitable_hard_regs
= data
->profitable_hard_regs
;
851 node
= data
->hard_regs_node
;
852 node_preorder_num
= node
->preorder_num
;
853 node_set
= node
->hard_regs
->set
;
855 for (k
= 0; k
< nobj
; k
++)
857 ira_object_t obj
= ALLOCNO_OBJECT (a
, k
);
858 ira_object_t conflict_obj
;
859 ira_object_conflict_iterator oci
;
861 FOR_EACH_OBJECT_CONFLICT (obj
, conflict_obj
, oci
)
864 ira_allocno_t conflict_a
= OBJECT_ALLOCNO (conflict_obj
);
865 allocno_hard_regs_node_t conflict_node
, temp_node
;
866 HARD_REG_SET conflict_node_set
;
867 allocno_color_data_t conflict_data
;
869 conflict_data
= ALLOCNO_COLOR_DATA (conflict_a
);
870 if (! ALLOCNO_COLOR_DATA (conflict_a
)->in_graph_p
871 || ! hard_reg_set_intersect_p (profitable_hard_regs
,
873 ->profitable_hard_regs
))
875 conflict_node
= conflict_data
->hard_regs_node
;
876 conflict_node_set
= conflict_node
->hard_regs
->set
;
877 if (hard_reg_set_subset_p (node_set
, conflict_node_set
))
881 ira_assert (hard_reg_set_subset_p (conflict_node_set
, node_set
));
882 temp_node
= conflict_node
;
884 if (temp_node
->check
!= node_check_tick
)
886 temp_node
->check
= node_check_tick
;
887 temp_node
->conflict_size
= 0;
889 size
= (ira_reg_class_max_nregs
890 [ALLOCNO_CLASS (conflict_a
)][ALLOCNO_MODE (conflict_a
)]);
891 if (ALLOCNO_NUM_OBJECTS (conflict_a
) > 1)
892 /* We will deal with the subwords individually. */
894 temp_node
->conflict_size
+= size
;
897 for (i
= 0; i
< data
->hard_regs_subnodes_num
; i
++)
899 allocno_hard_regs_node_t temp_node
;
901 temp_node
= allocno_hard_regs_nodes
[i
+ node_preorder_num
];
902 ira_assert (temp_node
->preorder_num
== i
+ node_preorder_num
);
903 subnodes
[i
].left_conflict_size
= (temp_node
->check
!= node_check_tick
904 ? 0 : temp_node
->conflict_size
);
905 if (hard_reg_set_subset_p (temp_node
->hard_regs
->set
,
906 profitable_hard_regs
))
907 subnodes
[i
].max_node_impact
= temp_node
->hard_regs_num
;
910 HARD_REG_SET temp_set
;
911 int j
, n
, hard_regno
;
912 enum reg_class aclass
;
914 temp_set
= temp_node
->hard_regs
->set
& profitable_hard_regs
;
915 aclass
= ALLOCNO_CLASS (a
);
916 for (n
= 0, j
= ira_class_hard_regs_num
[aclass
] - 1; j
>= 0; j
--)
918 hard_regno
= ira_class_hard_regs
[aclass
][j
];
919 if (TEST_HARD_REG_BIT (temp_set
, hard_regno
))
922 subnodes
[i
].max_node_impact
= n
;
924 subnodes
[i
].left_conflict_subnodes_size
= 0;
926 start
= node_preorder_num
* allocno_hard_regs_nodes_num
;
927 for (i
= data
->hard_regs_subnodes_num
- 1; i
> 0; i
--)
930 allocno_hard_regs_node_t parent
;
932 size
= (subnodes
[i
].left_conflict_subnodes_size
933 + MIN (subnodes
[i
].max_node_impact
934 - subnodes
[i
].left_conflict_subnodes_size
,
935 subnodes
[i
].left_conflict_size
));
936 parent
= allocno_hard_regs_nodes
[i
+ node_preorder_num
]->parent
;
937 gcc_checking_assert(parent
);
939 = allocno_hard_regs_subnode_index
[start
+ parent
->preorder_num
];
940 gcc_checking_assert(parent_i
>= 0);
941 subnodes
[parent_i
].left_conflict_subnodes_size
+= size
;
943 left_conflict_subnodes_size
= subnodes
[0].left_conflict_subnodes_size
;
945 = (left_conflict_subnodes_size
946 + MIN (subnodes
[0].max_node_impact
- left_conflict_subnodes_size
,
947 subnodes
[0].left_conflict_size
));
948 conflict_size
+= ira_reg_class_max_nregs
[ALLOCNO_CLASS (a
)][ALLOCNO_MODE (a
)];
949 data
->colorable_p
= conflict_size
<= data
->available_regs_num
;
950 return data
->colorable_p
;
953 /* Update left conflict sizes of hard registers subnodes of allocno A
954 after removing allocno REMOVED_A with SIZE from the conflict graph.
955 Return TRUE if A is trivially colorable. */
957 update_left_conflict_sizes_p (ira_allocno_t a
,
958 ira_allocno_t removed_a
, int size
)
960 int i
, conflict_size
, before_conflict_size
, diff
, start
;
961 int node_preorder_num
, parent_i
;
962 allocno_hard_regs_node_t node
, removed_node
, parent
;
963 allocno_hard_regs_subnode_t subnodes
;
964 allocno_color_data_t data
= ALLOCNO_COLOR_DATA (a
);
966 ira_assert (! data
->colorable_p
);
967 node
= data
->hard_regs_node
;
968 node_preorder_num
= node
->preorder_num
;
969 removed_node
= ALLOCNO_COLOR_DATA (removed_a
)->hard_regs_node
;
970 ira_assert (hard_reg_set_subset_p (removed_node
->hard_regs
->set
,
971 node
->hard_regs
->set
)
972 || hard_reg_set_subset_p (node
->hard_regs
->set
,
973 removed_node
->hard_regs
->set
));
974 start
= node_preorder_num
* allocno_hard_regs_nodes_num
;
975 i
= allocno_hard_regs_subnode_index
[start
+ removed_node
->preorder_num
];
978 subnodes
= allocno_hard_regs_subnodes
+ data
->hard_regs_subnodes_start
;
980 = (subnodes
[i
].left_conflict_subnodes_size
981 + MIN (subnodes
[i
].max_node_impact
982 - subnodes
[i
].left_conflict_subnodes_size
,
983 subnodes
[i
].left_conflict_size
));
984 subnodes
[i
].left_conflict_size
-= size
;
988 = (subnodes
[i
].left_conflict_subnodes_size
989 + MIN (subnodes
[i
].max_node_impact
990 - subnodes
[i
].left_conflict_subnodes_size
,
991 subnodes
[i
].left_conflict_size
));
992 if ((diff
= before_conflict_size
- conflict_size
) == 0)
994 ira_assert (conflict_size
< before_conflict_size
);
995 parent
= allocno_hard_regs_nodes
[i
+ node_preorder_num
]->parent
;
999 = allocno_hard_regs_subnode_index
[start
+ parent
->preorder_num
];
1003 before_conflict_size
1004 = (subnodes
[i
].left_conflict_subnodes_size
1005 + MIN (subnodes
[i
].max_node_impact
1006 - subnodes
[i
].left_conflict_subnodes_size
,
1007 subnodes
[i
].left_conflict_size
));
1008 subnodes
[i
].left_conflict_subnodes_size
-= diff
;
1012 + ira_reg_class_max_nregs
[ALLOCNO_CLASS (a
)][ALLOCNO_MODE (a
)]
1013 > data
->available_regs_num
))
1015 data
->colorable_p
= true;
1019 /* Return true if allocno A has empty profitable hard regs. */
1021 empty_profitable_hard_regs (ira_allocno_t a
)
1023 allocno_color_data_t data
= ALLOCNO_COLOR_DATA (a
);
1025 return hard_reg_set_empty_p (data
->profitable_hard_regs
);
1028 /* Set up profitable hard registers for each allocno being
1031 setup_profitable_hard_regs (void)
1034 int j
, k
, nobj
, hard_regno
, nregs
, class_size
;
1037 enum reg_class aclass
;
1039 allocno_color_data_t data
;
1041 /* Initial set up from allocno classes and explicitly conflicting
1043 EXECUTE_IF_SET_IN_BITMAP (coloring_allocno_bitmap
, 0, i
, bi
)
1045 a
= ira_allocnos
[i
];
1046 if ((aclass
= ALLOCNO_CLASS (a
)) == NO_REGS
)
1048 data
= ALLOCNO_COLOR_DATA (a
);
1049 if (ALLOCNO_UPDATED_HARD_REG_COSTS (a
) == NULL
1050 && ALLOCNO_CLASS_COST (a
) > ALLOCNO_MEMORY_COST (a
)
1051 /* Do not empty profitable regs for static chain pointer
1052 pseudo when non-local goto is used. */
1053 && ! non_spilled_static_chain_regno_p (ALLOCNO_REGNO (a
)))
1054 CLEAR_HARD_REG_SET (data
->profitable_hard_regs
);
1057 mode
= ALLOCNO_MODE (a
);
1058 data
->profitable_hard_regs
1059 = ira_useful_class_mode_regs
[aclass
][mode
];
1060 nobj
= ALLOCNO_NUM_OBJECTS (a
);
1061 for (k
= 0; k
< nobj
; k
++)
1063 ira_object_t obj
= ALLOCNO_OBJECT (a
, k
);
1065 data
->profitable_hard_regs
1066 &= ~OBJECT_TOTAL_CONFLICT_HARD_REGS (obj
);
1070 /* Exclude hard regs already assigned for conflicting objects. */
1071 EXECUTE_IF_SET_IN_BITMAP (consideration_allocno_bitmap
, 0, i
, bi
)
1073 a
= ira_allocnos
[i
];
1074 if ((aclass
= ALLOCNO_CLASS (a
)) == NO_REGS
1075 || ! ALLOCNO_ASSIGNED_P (a
)
1076 || (hard_regno
= ALLOCNO_HARD_REGNO (a
)) < 0)
1078 mode
= ALLOCNO_MODE (a
);
1079 nregs
= hard_regno_nregs (hard_regno
, mode
);
1080 nobj
= ALLOCNO_NUM_OBJECTS (a
);
1081 for (k
= 0; k
< nobj
; k
++)
1083 ira_object_t obj
= ALLOCNO_OBJECT (a
, k
);
1084 ira_object_t conflict_obj
;
1085 ira_object_conflict_iterator oci
;
1087 FOR_EACH_OBJECT_CONFLICT (obj
, conflict_obj
, oci
)
1089 ira_allocno_t conflict_a
= OBJECT_ALLOCNO (conflict_obj
);
1091 /* We can process the conflict allocno repeatedly with
1093 if (nregs
== nobj
&& nregs
> 1)
1095 int num
= OBJECT_SUBWORD (conflict_obj
);
1097 if (REG_WORDS_BIG_ENDIAN
)
1099 (ALLOCNO_COLOR_DATA (conflict_a
)->profitable_hard_regs
,
1100 hard_regno
+ nobj
- num
- 1);
1103 (ALLOCNO_COLOR_DATA (conflict_a
)->profitable_hard_regs
,
1107 ALLOCNO_COLOR_DATA (conflict_a
)->profitable_hard_regs
1108 &= ~ira_reg_mode_hard_regset
[hard_regno
][mode
];
1112 /* Exclude too costly hard regs. */
1113 EXECUTE_IF_SET_IN_BITMAP (coloring_allocno_bitmap
, 0, i
, bi
)
1115 int min_cost
= INT_MAX
;
1118 a
= ira_allocnos
[i
];
1119 if ((aclass
= ALLOCNO_CLASS (a
)) == NO_REGS
1120 || empty_profitable_hard_regs (a
))
1122 data
= ALLOCNO_COLOR_DATA (a
);
1123 if ((costs
= ALLOCNO_UPDATED_HARD_REG_COSTS (a
)) != NULL
1124 || (costs
= ALLOCNO_HARD_REG_COSTS (a
)) != NULL
)
1126 class_size
= ira_class_hard_regs_num
[aclass
];
1127 for (j
= 0; j
< class_size
; j
++)
1129 hard_regno
= ira_class_hard_regs
[aclass
][j
];
1130 if (! TEST_HARD_REG_BIT (data
->profitable_hard_regs
,
1133 if (ALLOCNO_UPDATED_MEMORY_COST (a
) < costs
[j
]
1134 /* Do not remove HARD_REGNO for static chain pointer
1135 pseudo when non-local goto is used. */
1136 && ! non_spilled_static_chain_regno_p (ALLOCNO_REGNO (a
)))
1137 CLEAR_HARD_REG_BIT (data
->profitable_hard_regs
,
1139 else if (min_cost
> costs
[j
])
1140 min_cost
= costs
[j
];
1143 else if (ALLOCNO_UPDATED_MEMORY_COST (a
)
1144 < ALLOCNO_UPDATED_CLASS_COST (a
)
1145 /* Do not empty profitable regs for static chain
1146 pointer pseudo when non-local goto is used. */
1147 && ! non_spilled_static_chain_regno_p (ALLOCNO_REGNO (a
)))
1148 CLEAR_HARD_REG_SET (data
->profitable_hard_regs
);
1149 if (ALLOCNO_UPDATED_CLASS_COST (a
) > min_cost
)
1150 ALLOCNO_UPDATED_CLASS_COST (a
) = min_cost
;
1156 /* This page contains functions used to choose hard registers for
1159 /* Pool for update cost records. */
1160 static object_allocator
<update_cost_record
> update_cost_record_pool
1161 ("update cost records");
1163 /* Return new update cost record with given params. */
1164 static struct update_cost_record
*
1165 get_update_cost_record (int hard_regno
, int divisor
,
1166 struct update_cost_record
*next
)
1168 struct update_cost_record
*record
;
1170 record
= update_cost_record_pool
.allocate ();
1171 record
->hard_regno
= hard_regno
;
1172 record
->divisor
= divisor
;
1173 record
->next
= next
;
1177 /* Free memory for all records in LIST. */
1179 free_update_cost_record_list (struct update_cost_record
*list
)
1181 struct update_cost_record
*next
;
1183 while (list
!= NULL
)
1186 update_cost_record_pool
.remove (list
);
1191 /* Free memory allocated for all update cost records. */
1193 finish_update_cost_records (void)
1195 update_cost_record_pool
.release ();
1198 /* Array whose element value is TRUE if the corresponding hard
1199 register was already allocated for an allocno. */
1200 static bool allocated_hardreg_p
[FIRST_PSEUDO_REGISTER
];
1202 /* Describes one element in a queue of allocnos whose costs need to be
1203 updated. Each allocno in the queue is known to have an allocno
1205 struct update_cost_queue_elem
1207 /* This element is in the queue iff CHECK == update_cost_check. */
1210 /* COST_HOP_DIVISOR**N, where N is the length of the shortest path
1211 connecting this allocno to the one being allocated. */
1214 /* Allocno from which we started chaining costs of connected
1216 ira_allocno_t start
;
1218 /* Allocno from which we are chaining costs of connected allocnos.
1219 It is used not go back in graph of allocnos connected by
1223 /* The next allocno in the queue, or null if this is the last element. */
1227 /* The first element in a queue of allocnos whose copy costs need to be
1228 updated. Null if the queue is empty. */
1229 static ira_allocno_t update_cost_queue
;
1231 /* The last element in the queue described by update_cost_queue.
1232 Not valid if update_cost_queue is null. */
1233 static struct update_cost_queue_elem
*update_cost_queue_tail
;
1235 /* A pool of elements in the queue described by update_cost_queue.
1236 Elements are indexed by ALLOCNO_NUM. */
1237 static struct update_cost_queue_elem
*update_cost_queue_elems
;
1239 /* The current value of update_costs_from_copies call count. */
1240 static int update_cost_check
;
1242 /* Allocate and initialize data necessary for function
1243 update_costs_from_copies. */
1245 initiate_cost_update (void)
1249 size
= ira_allocnos_num
* sizeof (struct update_cost_queue_elem
);
1250 update_cost_queue_elems
1251 = (struct update_cost_queue_elem
*) ira_allocate (size
);
1252 memset (update_cost_queue_elems
, 0, size
);
1253 update_cost_check
= 0;
1256 /* Deallocate data used by function update_costs_from_copies. */
1258 finish_cost_update (void)
1260 ira_free (update_cost_queue_elems
);
1261 finish_update_cost_records ();
1264 /* When we traverse allocnos to update hard register costs, the cost
1265 divisor will be multiplied by the following macro value for each
1266 hop from given allocno to directly connected allocnos. */
1267 #define COST_HOP_DIVISOR 4
1269 /* Start a new cost-updating pass. */
1271 start_update_cost (void)
1273 update_cost_check
++;
1274 update_cost_queue
= NULL
;
1277 /* Add (ALLOCNO, START, FROM, DIVISOR) to the end of update_cost_queue, unless
1278 ALLOCNO is already in the queue, or has NO_REGS class. */
1280 queue_update_cost (ira_allocno_t allocno
, ira_allocno_t start
,
1281 ira_allocno_t from
, int divisor
)
1283 struct update_cost_queue_elem
*elem
;
1285 elem
= &update_cost_queue_elems
[ALLOCNO_NUM (allocno
)];
1286 if (elem
->check
!= update_cost_check
1287 && ALLOCNO_CLASS (allocno
) != NO_REGS
)
1289 elem
->check
= update_cost_check
;
1290 elem
->start
= start
;
1292 elem
->divisor
= divisor
;
1294 if (update_cost_queue
== NULL
)
1295 update_cost_queue
= allocno
;
1297 update_cost_queue_tail
->next
= allocno
;
1298 update_cost_queue_tail
= elem
;
1302 /* Try to remove the first element from update_cost_queue. Return
1303 false if the queue was empty, otherwise make (*ALLOCNO, *START,
1304 *FROM, *DIVISOR) describe the removed element. */
1306 get_next_update_cost (ira_allocno_t
*allocno
, ira_allocno_t
*start
,
1307 ira_allocno_t
*from
, int *divisor
)
1309 struct update_cost_queue_elem
*elem
;
1311 if (update_cost_queue
== NULL
)
1314 *allocno
= update_cost_queue
;
1315 elem
= &update_cost_queue_elems
[ALLOCNO_NUM (*allocno
)];
1316 *start
= elem
->start
;
1318 *divisor
= elem
->divisor
;
1319 update_cost_queue
= elem
->next
;
1323 /* Increase costs of HARD_REGNO by UPDATE_COST and conflict cost by
1324 UPDATE_CONFLICT_COST for ALLOCNO. Return true if we really
1325 modified the cost. */
1327 update_allocno_cost (ira_allocno_t allocno
, int hard_regno
,
1328 int update_cost
, int update_conflict_cost
)
1331 enum reg_class aclass
= ALLOCNO_CLASS (allocno
);
1333 i
= ira_class_hard_reg_index
[aclass
][hard_regno
];
1336 ira_allocate_and_set_or_copy_costs
1337 (&ALLOCNO_UPDATED_HARD_REG_COSTS (allocno
), aclass
,
1338 ALLOCNO_UPDATED_CLASS_COST (allocno
),
1339 ALLOCNO_HARD_REG_COSTS (allocno
));
1340 ira_allocate_and_set_or_copy_costs
1341 (&ALLOCNO_UPDATED_CONFLICT_HARD_REG_COSTS (allocno
),
1342 aclass
, 0, ALLOCNO_CONFLICT_HARD_REG_COSTS (allocno
));
1343 ALLOCNO_UPDATED_HARD_REG_COSTS (allocno
)[i
] += update_cost
;
1344 ALLOCNO_UPDATED_CONFLICT_HARD_REG_COSTS (allocno
)[i
] += update_conflict_cost
;
1348 /* Return TRUE if the object OBJ conflicts with the allocno A. */
1350 object_conflicts_with_allocno_p (ira_object_t obj
, ira_allocno_t a
)
1352 if (!OBJECT_CONFLICT_VEC_P (obj
))
1353 for (int word
= 0; word
< ALLOCNO_NUM_OBJECTS (a
); word
++)
1355 ira_object_t another_obj
= ALLOCNO_OBJECT (a
, word
);
1356 if (OBJECT_CONFLICT_ID (another_obj
) >= OBJECT_MIN (obj
)
1357 && OBJECT_CONFLICT_ID (another_obj
) <= OBJECT_MAX (obj
)
1358 && TEST_MINMAX_SET_BIT (OBJECT_CONFLICT_BITVEC (obj
),
1359 OBJECT_CONFLICT_ID (another_obj
),
1360 OBJECT_MIN (obj
), OBJECT_MAX (obj
)))
1365 /* If this linear walk ever becomes a bottleneck we could add a
1366 conflict_vec_sorted_p flag and if not set, sort the conflicts after
1367 their ID so we can use a binary search. That would also require
1368 tracking the actual number of conflicts in the vector to not rely
1369 on the NULL termination. */
1370 ira_object_conflict_iterator oci
;
1371 ira_object_t conflict_obj
;
1372 FOR_EACH_OBJECT_CONFLICT (obj
, conflict_obj
, oci
)
1373 if (OBJECT_ALLOCNO (conflict_obj
) == a
)
1379 /* Return TRUE if allocnos A1 and A2 conflicts. Here we are
1380 interested only in conflicts of allocnos with intersecting allocno
1383 allocnos_conflict_p (ira_allocno_t a1
, ira_allocno_t a2
)
1385 /* Compute the upper bound for the linear iteration when the object
1386 conflicts are represented as a sparse vector. In particular this
1387 will make sure we prefer O(1) bitvector testing. */
1388 int num_conflicts_in_vec1
= 0, num_conflicts_in_vec2
= 0;
1389 for (int word
= 0; word
< ALLOCNO_NUM_OBJECTS (a1
); ++word
)
1390 if (OBJECT_CONFLICT_VEC_P (ALLOCNO_OBJECT (a1
, word
)))
1391 num_conflicts_in_vec1
+= OBJECT_NUM_CONFLICTS (ALLOCNO_OBJECT (a1
, word
));
1392 for (int word
= 0; word
< ALLOCNO_NUM_OBJECTS (a2
); ++word
)
1393 if (OBJECT_CONFLICT_VEC_P (ALLOCNO_OBJECT (a2
, word
)))
1394 num_conflicts_in_vec2
+= OBJECT_NUM_CONFLICTS (ALLOCNO_OBJECT (a2
, word
));
1395 if (num_conflicts_in_vec2
< num_conflicts_in_vec1
)
1398 for (int word
= 0; word
< ALLOCNO_NUM_OBJECTS (a1
); word
++)
1400 ira_object_t obj
= ALLOCNO_OBJECT (a1
, word
);
1401 /* Take preferences of conflicting allocnos into account. */
1402 if (object_conflicts_with_allocno_p (obj
, a2
))
1408 /* Update (decrease if DECR_P) HARD_REGNO cost of allocnos connected
1409 by copies to ALLOCNO to increase chances to remove some copies as
1410 the result of subsequent assignment. Update conflict costs.
1411 Record cost updates if RECORD_P is true. */
1413 update_costs_from_allocno (ira_allocno_t allocno
, int hard_regno
,
1414 int divisor
, bool decr_p
, bool record_p
)
1416 int cost
, update_cost
, update_conflict_cost
;
1418 enum reg_class rclass
, aclass
;
1419 ira_allocno_t another_allocno
, start
= allocno
, from
= NULL
;
1420 ira_copy_t cp
, next_cp
;
1422 rclass
= REGNO_REG_CLASS (hard_regno
);
1425 mode
= ALLOCNO_MODE (allocno
);
1426 ira_init_register_move_cost_if_necessary (mode
);
1427 for (cp
= ALLOCNO_COPIES (allocno
); cp
!= NULL
; cp
= next_cp
)
1429 if (cp
->first
== allocno
)
1431 next_cp
= cp
->next_first_allocno_copy
;
1432 another_allocno
= cp
->second
;
1434 else if (cp
->second
== allocno
)
1436 next_cp
= cp
->next_second_allocno_copy
;
1437 another_allocno
= cp
->first
;
1442 if (another_allocno
== from
1443 || (ALLOCNO_COLOR_DATA (another_allocno
) != NULL
1444 && (ALLOCNO_COLOR_DATA (allocno
)->first_thread_allocno
1445 != ALLOCNO_COLOR_DATA (another_allocno
)->first_thread_allocno
)))
1448 aclass
= ALLOCNO_CLASS (another_allocno
);
1449 if (! TEST_HARD_REG_BIT (reg_class_contents
[aclass
],
1451 || ALLOCNO_ASSIGNED_P (another_allocno
))
1454 /* If we have different modes use the smallest one. It is
1455 a sub-register move. It is hard to predict what LRA
1456 will reload (the pseudo or its sub-register) but LRA
1457 will try to minimize the data movement. Also for some
1458 register classes bigger modes might be invalid,
1459 e.g. DImode for AREG on x86. For such cases the
1460 register move cost will be maximal. */
1461 mode
= narrower_subreg_mode (ALLOCNO_MODE (cp
->first
),
1462 ALLOCNO_MODE (cp
->second
));
1464 ira_init_register_move_cost_if_necessary (mode
);
1466 cost
= (cp
->second
== allocno
1467 ? ira_register_move_cost
[mode
][rclass
][aclass
]
1468 : ira_register_move_cost
[mode
][aclass
][rclass
]);
1472 update_cost
= cp
->freq
* cost
/ divisor
;
1473 update_conflict_cost
= update_cost
;
1475 if (internal_flag_ira_verbose
> 5 && ira_dump_file
!= NULL
)
1476 fprintf (ira_dump_file
,
1477 " a%dr%d (hr%d): update cost by %d, conflict cost by %d\n",
1478 ALLOCNO_NUM (another_allocno
), ALLOCNO_REGNO (another_allocno
),
1479 hard_regno
, update_cost
, update_conflict_cost
);
1480 if (update_cost
== 0)
1483 if (! update_allocno_cost (another_allocno
, hard_regno
,
1484 update_cost
, update_conflict_cost
))
1486 queue_update_cost (another_allocno
, start
, allocno
,
1487 divisor
* COST_HOP_DIVISOR
);
1488 if (record_p
&& ALLOCNO_COLOR_DATA (another_allocno
) != NULL
)
1489 ALLOCNO_COLOR_DATA (another_allocno
)->update_cost_records
1490 = get_update_cost_record (hard_regno
, divisor
,
1491 ALLOCNO_COLOR_DATA (another_allocno
)
1492 ->update_cost_records
);
1495 while (get_next_update_cost (&allocno
, &start
, &from
, &divisor
));
1498 /* Decrease preferred ALLOCNO hard register costs and costs of
1499 allocnos connected to ALLOCNO through copy. */
1501 update_costs_from_prefs (ira_allocno_t allocno
)
1505 start_update_cost ();
1506 for (pref
= ALLOCNO_PREFS (allocno
); pref
!= NULL
; pref
= pref
->next_pref
)
1508 if (internal_flag_ira_verbose
> 5 && ira_dump_file
!= NULL
)
1509 fprintf (ira_dump_file
, " Start updating from pref of hr%d for a%dr%d:\n",
1510 pref
->hard_regno
, ALLOCNO_NUM (allocno
), ALLOCNO_REGNO (allocno
));
1511 update_costs_from_allocno (allocno
, pref
->hard_regno
,
1512 COST_HOP_DIVISOR
, true, true);
1516 /* Update (decrease if DECR_P) the cost of allocnos connected to
1517 ALLOCNO through copies to increase chances to remove some copies as
1518 the result of subsequent assignment. ALLOCNO was just assigned to
1519 a hard register. Record cost updates if RECORD_P is true. */
1521 update_costs_from_copies (ira_allocno_t allocno
, bool decr_p
, bool record_p
)
1525 hard_regno
= ALLOCNO_HARD_REGNO (allocno
);
1526 ira_assert (hard_regno
>= 0 && ALLOCNO_CLASS (allocno
) != NO_REGS
);
1527 start_update_cost ();
1528 if (internal_flag_ira_verbose
> 5 && ira_dump_file
!= NULL
)
1529 fprintf (ira_dump_file
, " Start updating from a%dr%d by copies:\n",
1530 ALLOCNO_NUM (allocno
), ALLOCNO_REGNO (allocno
));
1531 update_costs_from_allocno (allocno
, hard_regno
, 1, decr_p
, record_p
);
1534 /* Update conflict_allocno_hard_prefs of allocnos conflicting with
1537 update_conflict_allocno_hard_prefs (ira_allocno_t allocno
)
1539 int l
, nr
= ALLOCNO_NUM_OBJECTS (allocno
);
1541 for (l
= 0; l
< nr
; l
++)
1543 ira_object_t conflict_obj
, obj
= ALLOCNO_OBJECT (allocno
, l
);
1544 ira_object_conflict_iterator oci
;
1546 FOR_EACH_OBJECT_CONFLICT (obj
, conflict_obj
, oci
)
1548 ira_allocno_t conflict_a
= OBJECT_ALLOCNO (conflict_obj
);
1549 allocno_color_data_t conflict_data
= ALLOCNO_COLOR_DATA (conflict_a
);
1552 if (!(hard_reg_set_intersect_p
1553 (ALLOCNO_COLOR_DATA (allocno
)->profitable_hard_regs
,
1554 conflict_data
->profitable_hard_regs
)))
1556 for (pref
= ALLOCNO_PREFS (allocno
);
1558 pref
= pref
->next_pref
)
1559 conflict_data
->conflict_allocno_hard_prefs
+= pref
->freq
;
1564 /* Restore costs of allocnos connected to ALLOCNO by copies as it was
1565 before updating costs of these allocnos from given allocno. This
1566 is a wise thing to do as if given allocno did not get an expected
1567 hard reg, using smaller cost of the hard reg for allocnos connected
1568 by copies to given allocno becomes actually misleading. Free all
1569 update cost records for ALLOCNO as we don't need them anymore. */
1571 restore_costs_from_copies (ira_allocno_t allocno
)
1573 struct update_cost_record
*records
, *curr
;
1575 if (ALLOCNO_COLOR_DATA (allocno
) == NULL
)
1577 records
= ALLOCNO_COLOR_DATA (allocno
)->update_cost_records
;
1578 start_update_cost ();
1579 if (internal_flag_ira_verbose
> 5 && ira_dump_file
!= NULL
)
1580 fprintf (ira_dump_file
, " Start restoring from a%dr%d:\n",
1581 ALLOCNO_NUM (allocno
), ALLOCNO_REGNO (allocno
));
1582 for (curr
= records
; curr
!= NULL
; curr
= curr
->next
)
1583 update_costs_from_allocno (allocno
, curr
->hard_regno
,
1584 curr
->divisor
, true, false);
1585 free_update_cost_record_list (records
);
1586 ALLOCNO_COLOR_DATA (allocno
)->update_cost_records
= NULL
;
1589 /* This function updates COSTS (decrease if DECR_P) for hard_registers
1590 of ACLASS by conflict costs of the unassigned allocnos
1591 connected by copies with allocnos in update_cost_queue. This
1592 update increases chances to remove some copies. */
1594 update_conflict_hard_regno_costs (int *costs
, enum reg_class aclass
,
1597 int i
, cost
, class_size
, freq
, mult
, div
, divisor
;
1598 int index
, hard_regno
;
1599 int *conflict_costs
;
1601 enum reg_class another_aclass
;
1602 ira_allocno_t allocno
, another_allocno
, start
, from
;
1603 ira_copy_t cp
, next_cp
;
1605 while (get_next_update_cost (&allocno
, &start
, &from
, &divisor
))
1606 for (cp
= ALLOCNO_COPIES (allocno
); cp
!= NULL
; cp
= next_cp
)
1608 if (cp
->first
== allocno
)
1610 next_cp
= cp
->next_first_allocno_copy
;
1611 another_allocno
= cp
->second
;
1613 else if (cp
->second
== allocno
)
1615 next_cp
= cp
->next_second_allocno_copy
;
1616 another_allocno
= cp
->first
;
1621 another_aclass
= ALLOCNO_CLASS (another_allocno
);
1622 if (another_allocno
== from
1623 || ALLOCNO_ASSIGNED_P (another_allocno
)
1624 || ALLOCNO_COLOR_DATA (another_allocno
)->may_be_spilled_p
1625 || ! ira_reg_classes_intersect_p
[aclass
][another_aclass
])
1627 if (allocnos_conflict_p (another_allocno
, start
))
1630 class_size
= ira_class_hard_regs_num
[another_aclass
];
1631 ira_allocate_and_copy_costs
1632 (&ALLOCNO_UPDATED_CONFLICT_HARD_REG_COSTS (another_allocno
),
1633 another_aclass
, ALLOCNO_CONFLICT_HARD_REG_COSTS (another_allocno
));
1635 = ALLOCNO_UPDATED_CONFLICT_HARD_REG_COSTS (another_allocno
);
1636 if (conflict_costs
== NULL
)
1641 freq
= ALLOCNO_FREQ (another_allocno
);
1644 div
= freq
* divisor
;
1646 for (i
= class_size
- 1; i
>= 0; i
--)
1648 hard_regno
= ira_class_hard_regs
[another_aclass
][i
];
1649 ira_assert (hard_regno
>= 0);
1650 index
= ira_class_hard_reg_index
[aclass
][hard_regno
];
1653 cost
= (int) (((int64_t) conflict_costs
[i
] * mult
) / div
);
1659 costs
[index
] += cost
;
1662 /* Probably 5 hops will be enough. */
1664 && divisor
<= (COST_HOP_DIVISOR
1667 * COST_HOP_DIVISOR
))
1668 queue_update_cost (another_allocno
, start
, from
, divisor
* COST_HOP_DIVISOR
);
1672 /* Set up conflicting (through CONFLICT_REGS) for each object of
1673 allocno A and the start allocno profitable regs (through
1674 START_PROFITABLE_REGS). Remember that the start profitable regs
1675 exclude hard regs which cannot hold value of mode of allocno A.
1676 This covers mostly cases when multi-register value should be
1679 get_conflict_and_start_profitable_regs (ira_allocno_t a
, bool retry_p
,
1680 HARD_REG_SET
*conflict_regs
,
1681 HARD_REG_SET
*start_profitable_regs
)
1686 nwords
= ALLOCNO_NUM_OBJECTS (a
);
1687 for (i
= 0; i
< nwords
; i
++)
1689 obj
= ALLOCNO_OBJECT (a
, i
);
1690 conflict_regs
[i
] = OBJECT_TOTAL_CONFLICT_HARD_REGS (obj
);
1693 *start_profitable_regs
1694 = (reg_class_contents
[ALLOCNO_CLASS (a
)]
1695 &~ (ira_prohibited_class_mode_regs
1696 [ALLOCNO_CLASS (a
)][ALLOCNO_MODE (a
)]));
1698 *start_profitable_regs
= ALLOCNO_COLOR_DATA (a
)->profitable_hard_regs
;
1701 /* Return true if HARD_REGNO is ok for assigning to allocno A with
1702 PROFITABLE_REGS and whose objects have CONFLICT_REGS. */
1704 check_hard_reg_p (ira_allocno_t a
, int hard_regno
,
1705 HARD_REG_SET
*conflict_regs
, HARD_REG_SET profitable_regs
)
1707 int j
, nwords
, nregs
;
1708 enum reg_class aclass
;
1711 aclass
= ALLOCNO_CLASS (a
);
1712 mode
= ALLOCNO_MODE (a
);
1713 if (TEST_HARD_REG_BIT (ira_prohibited_class_mode_regs
[aclass
][mode
],
1716 /* Checking only profitable hard regs. */
1717 if (! TEST_HARD_REG_BIT (profitable_regs
, hard_regno
))
1719 nregs
= hard_regno_nregs (hard_regno
, mode
);
1720 nwords
= ALLOCNO_NUM_OBJECTS (a
);
1721 for (j
= 0; j
< nregs
; j
++)
1724 int set_to_test_start
= 0, set_to_test_end
= nwords
;
1726 if (nregs
== nwords
)
1728 if (REG_WORDS_BIG_ENDIAN
)
1729 set_to_test_start
= nwords
- j
- 1;
1731 set_to_test_start
= j
;
1732 set_to_test_end
= set_to_test_start
+ 1;
1734 for (k
= set_to_test_start
; k
< set_to_test_end
; k
++)
1735 if (TEST_HARD_REG_BIT (conflict_regs
[k
], hard_regno
+ j
))
1737 if (k
!= set_to_test_end
)
1743 /* Return number of registers needed to be saved and restored at
1744 function prologue/epilogue if we allocate HARD_REGNO to hold value
1747 calculate_saved_nregs (int hard_regno
, machine_mode mode
)
1752 ira_assert (hard_regno
>= 0);
1753 for (i
= hard_regno_nregs (hard_regno
, mode
) - 1; i
>= 0; i
--)
1754 if (!allocated_hardreg_p
[hard_regno
+ i
]
1755 && !crtl
->abi
->clobbers_full_reg_p (hard_regno
+ i
)
1756 && !LOCAL_REGNO (hard_regno
+ i
))
1761 /* Allocnos A1 and A2 are known to conflict. Check whether, in some loop L
1762 that is either the current loop or a nested subloop, the conflict is of
1765 - One allocno (X) is a cap allocno for some non-cap allocno X2.
1767 - X2 belongs to some loop L2.
1769 - The other allocno (Y) is a non-cap allocno.
1771 - Y is an ancestor of some allocno Y2 in L2. (Note that such a Y2
1772 must exist, given that X and Y conflict.)
1774 - Y2 is not referenced in L2 (that is, ALLOCNO_NREFS (Y2) == 0).
1776 - Y can use a different allocation from Y2.
1778 In this case, Y's register is live across L2 but is not used within it,
1779 whereas X's register is used only within L2. The conflict is therefore
1780 only "soft", in that it can easily be avoided by spilling Y2 inside L2
1781 without affecting any insn references.
1783 If the conflict does have this form, return the Y2 that would need to be
1784 spilled in order to allow X and Y (and thus A1 and A2) to use the same
1785 register. Return null otherwise. Returning null is conservatively correct;
1786 any nonnnull return value is an optimization. */
1788 ira_soft_conflict (ira_allocno_t a1
, ira_allocno_t a2
)
1790 /* Search for the loop L and its associated allocnos X and Y. */
1791 int search_depth
= 0;
1792 while (ALLOCNO_CAP_MEMBER (a1
) && ALLOCNO_CAP_MEMBER (a2
))
1794 a1
= ALLOCNO_CAP_MEMBER (a1
);
1795 a2
= ALLOCNO_CAP_MEMBER (a2
);
1796 if (search_depth
++ > max_soft_conflict_loop_depth
)
1799 /* This must be true if A1 and A2 conflict. */
1800 ira_assert (ALLOCNO_LOOP_TREE_NODE (a1
) == ALLOCNO_LOOP_TREE_NODE (a2
));
1802 /* Make A1 the cap allocno (X in the comment above) and A2 the
1803 non-cap allocno (Y in the comment above). */
1804 if (ALLOCNO_CAP_MEMBER (a2
))
1806 if (!ALLOCNO_CAP_MEMBER (a1
))
1809 /* Search for the real allocno that A1 caps (X2 in the comment above). */
1812 a1
= ALLOCNO_CAP_MEMBER (a1
);
1813 if (search_depth
++ > max_soft_conflict_loop_depth
)
1816 while (ALLOCNO_CAP_MEMBER (a1
));
1818 /* Find the associated allocno for A2 (Y2 in the comment above). */
1819 auto node
= ALLOCNO_LOOP_TREE_NODE (a1
);
1820 auto local_a2
= node
->regno_allocno_map
[ALLOCNO_REGNO (a2
)];
1822 /* Find the parent of LOCAL_A2/Y2. LOCAL_A2 must be a descendant of A2
1823 for the conflict query to make sense, so this parent lookup must succeed.
1825 If the parent allocno has no references, it is usually cheaper to
1826 spill at that loop level instead. Keep searching until we find
1827 a parent allocno that does have references (but don't look past
1828 the starting allocno). */
1829 ira_allocno_t local_parent_a2
;
1832 local_parent_a2
= ira_parent_allocno (local_a2
);
1833 if (local_parent_a2
== a2
|| ALLOCNO_NREFS (local_parent_a2
) != 0)
1835 local_a2
= local_parent_a2
;
1839 /* Sanity check to make sure that the conflict we've been given
1841 auto test_a2
= local_parent_a2
;
1842 while (test_a2
!= a2
)
1844 test_a2
= ira_parent_allocno (test_a2
);
1845 ira_assert (test_a2
);
1849 && ALLOCNO_NREFS (local_a2
) == 0
1850 && ira_subloop_allocnos_can_differ_p (local_parent_a2
))
1855 /* The caller has decided to allocate HREGNO to A and has proved that
1856 this is safe. However, the allocation might require the kind of
1857 spilling described in the comment above ira_soft_conflict.
1858 The caller has recorded that:
1860 - The allocnos in ALLOCNOS_TO_SPILL are the ones that would need
1861 to be spilled to satisfy soft conflicts for at least one allocation
1862 (not necessarily HREGNO).
1864 - The soft conflicts apply only to A allocations that overlap
1867 If allocating HREGNO is subject to any soft conflicts, record the
1868 subloop allocnos that need to be spilled. */
1870 spill_soft_conflicts (ira_allocno_t a
, bitmap allocnos_to_spill
,
1871 HARD_REG_SET soft_conflict_regs
, int hregno
)
1873 auto nregs
= hard_regno_nregs (hregno
, ALLOCNO_MODE (a
));
1876 EXECUTE_IF_SET_IN_BITMAP (allocnos_to_spill
, 0, i
, bi
)
1878 /* SPILL_A needs to be spilled for at least one allocation
1879 (not necessarily this one). */
1880 auto spill_a
= ira_allocnos
[i
];
1882 /* Find the corresponding allocno for this loop. */
1883 auto conflict_a
= spill_a
;
1886 conflict_a
= ira_parent_or_cap_allocno (conflict_a
);
1887 ira_assert (conflict_a
);
1889 while (ALLOCNO_LOOP_TREE_NODE (conflict_a
)->level
1890 > ALLOCNO_LOOP_TREE_NODE (a
)->level
);
1892 ira_assert (ALLOCNO_LOOP_TREE_NODE (conflict_a
)
1893 == ALLOCNO_LOOP_TREE_NODE (a
));
1895 if (conflict_a
== a
)
1897 /* SPILL_A is a descendant of A. We don't know (and don't need
1898 to know) which cap allocnos have a soft conflict with A.
1899 All we need to do is test whether the soft conflict applies
1900 to the chosen allocation. */
1901 if (ira_hard_reg_set_intersection_p (hregno
, ALLOCNO_MODE (a
),
1902 soft_conflict_regs
))
1903 ALLOCNO_MIGHT_CONFLICT_WITH_PARENT_P (spill_a
) = true;
1907 /* SPILL_A is a descendant of CONFLICT_A, which has a soft conflict
1908 with A. Test whether the soft conflict applies to the current
1910 ira_assert (ira_soft_conflict (a
, conflict_a
) == spill_a
);
1911 auto conflict_hregno
= ALLOCNO_HARD_REGNO (conflict_a
);
1912 ira_assert (conflict_hregno
>= 0);
1913 auto conflict_nregs
= hard_regno_nregs (conflict_hregno
,
1914 ALLOCNO_MODE (conflict_a
));
1915 if (hregno
+ nregs
> conflict_hregno
1916 && conflict_hregno
+ conflict_nregs
> hregno
)
1917 ALLOCNO_MIGHT_CONFLICT_WITH_PARENT_P (spill_a
) = true;
1922 /* Choose a hard register for allocno A. If RETRY_P is TRUE, it means
1923 that the function called from function
1924 `ira_reassign_conflict_allocnos' and `allocno_reload_assign'. In
1925 this case some allocno data are not defined or updated and we
1926 should not touch these data. The function returns true if we
1927 managed to assign a hard register to the allocno.
1929 To assign a hard register, first of all we calculate all conflict
1930 hard registers which can come from conflicting allocnos with
1931 already assigned hard registers. After that we find first free
1932 hard register with the minimal cost. During hard register cost
1933 calculation we take conflict hard register costs into account to
1934 give a chance for conflicting allocnos to get a better hard
1935 register in the future.
1937 If the best hard register cost is bigger than cost of memory usage
1938 for the allocno, we don't assign a hard register to given allocno
1941 If we assign a hard register to the allocno, we update costs of the
1942 hard register for allocnos connected by copies to improve a chance
1943 to coalesce insns represented by the copies when we assign hard
1944 registers to the allocnos connected by the copies. */
1946 assign_hard_reg (ira_allocno_t a
, bool retry_p
)
1948 HARD_REG_SET conflicting_regs
[2], profitable_hard_regs
;
1949 int i
, j
, hard_regno
, best_hard_regno
, class_size
;
1950 int cost
, mem_cost
, min_cost
, full_cost
, min_full_cost
, nwords
, word
;
1952 enum reg_class aclass
;
1954 static int costs
[FIRST_PSEUDO_REGISTER
], full_costs
[FIRST_PSEUDO_REGISTER
];
1956 enum reg_class rclass
;
1959 bool no_stack_reg_p
;
1961 auto_bitmap allocnos_to_spill
;
1962 HARD_REG_SET soft_conflict_regs
= {};
1964 ira_assert (! ALLOCNO_ASSIGNED_P (a
));
1965 get_conflict_and_start_profitable_regs (a
, retry_p
,
1967 &profitable_hard_regs
);
1968 aclass
= ALLOCNO_CLASS (a
);
1969 class_size
= ira_class_hard_regs_num
[aclass
];
1970 best_hard_regno
= -1;
1972 memset (costs
, 0, sizeof (int) * class_size
);
1973 memset (full_costs
, 0, sizeof (int) * class_size
);
1975 no_stack_reg_p
= false;
1978 start_update_cost ();
1979 mem_cost
+= ALLOCNO_UPDATED_MEMORY_COST (a
);
1981 ira_allocate_and_copy_costs (&ALLOCNO_UPDATED_HARD_REG_COSTS (a
),
1982 aclass
, ALLOCNO_HARD_REG_COSTS (a
));
1983 a_costs
= ALLOCNO_UPDATED_HARD_REG_COSTS (a
);
1985 no_stack_reg_p
= no_stack_reg_p
|| ALLOCNO_TOTAL_NO_STACK_REG_P (a
);
1987 cost
= ALLOCNO_UPDATED_CLASS_COST (a
);
1988 for (i
= 0; i
< class_size
; i
++)
1989 if (a_costs
!= NULL
)
1991 costs
[i
] += a_costs
[i
];
1992 full_costs
[i
] += a_costs
[i
];
1997 full_costs
[i
] += cost
;
1999 nwords
= ALLOCNO_NUM_OBJECTS (a
);
2000 curr_allocno_process
++;
2001 for (word
= 0; word
< nwords
; word
++)
2003 ira_object_t conflict_obj
;
2004 ira_object_t obj
= ALLOCNO_OBJECT (a
, word
);
2005 ira_object_conflict_iterator oci
;
2007 /* Take preferences of conflicting allocnos into account. */
2008 FOR_EACH_OBJECT_CONFLICT (obj
, conflict_obj
, oci
)
2010 ira_allocno_t conflict_a
= OBJECT_ALLOCNO (conflict_obj
);
2011 enum reg_class conflict_aclass
;
2012 allocno_color_data_t data
= ALLOCNO_COLOR_DATA (conflict_a
);
2014 /* Reload can give another class so we need to check all
2017 && ((!ALLOCNO_ASSIGNED_P (conflict_a
)
2018 || ALLOCNO_HARD_REGNO (conflict_a
) < 0)
2019 && !(hard_reg_set_intersect_p
2020 (profitable_hard_regs
,
2022 (conflict_a
)->profitable_hard_regs
))))
2024 /* All conflict allocnos are in consideration bitmap
2025 when retry_p is false. It might change in future and
2026 if it happens the assert will be broken. It means
2027 the code should be modified for the new
2029 ira_assert (bitmap_bit_p (consideration_allocno_bitmap
,
2030 ALLOCNO_NUM (conflict_a
)));
2033 conflict_aclass
= ALLOCNO_CLASS (conflict_a
);
2034 ira_assert (ira_reg_classes_intersect_p
2035 [aclass
][conflict_aclass
]);
2036 if (ALLOCNO_ASSIGNED_P (conflict_a
))
2038 hard_regno
= ALLOCNO_HARD_REGNO (conflict_a
);
2040 && (ira_hard_reg_set_intersection_p
2041 (hard_regno
, ALLOCNO_MODE (conflict_a
),
2042 reg_class_contents
[aclass
])))
2044 int n_objects
= ALLOCNO_NUM_OBJECTS (conflict_a
);
2047 mode
= ALLOCNO_MODE (conflict_a
);
2048 conflict_nregs
= hard_regno_nregs (hard_regno
, mode
);
2049 auto spill_a
= (retry_p
2051 : ira_soft_conflict (a
, conflict_a
));
2054 if (bitmap_set_bit (allocnos_to_spill
,
2055 ALLOCNO_NUM (spill_a
)))
2057 ira_loop_border_costs
border_costs (spill_a
);
2058 auto cost
= border_costs
.spill_inside_loop_cost ();
2059 auto note_conflict
= [&](int r
)
2061 SET_HARD_REG_BIT (soft_conflict_regs
, r
);
2062 auto hri
= ira_class_hard_reg_index
[aclass
][r
];
2066 full_costs
[hri
] += cost
;
2069 for (int r
= hard_regno
;
2070 r
>= 0 && (int) end_hard_regno (mode
, r
) > hard_regno
;
2073 for (int r
= hard_regno
+ 1;
2074 r
< hard_regno
+ conflict_nregs
;
2081 if (conflict_nregs
== n_objects
&& conflict_nregs
> 1)
2083 int num
= OBJECT_SUBWORD (conflict_obj
);
2085 if (REG_WORDS_BIG_ENDIAN
)
2086 SET_HARD_REG_BIT (conflicting_regs
[word
],
2087 hard_regno
+ n_objects
- num
- 1);
2089 SET_HARD_REG_BIT (conflicting_regs
[word
],
2093 conflicting_regs
[word
]
2094 |= ira_reg_mode_hard_regset
[hard_regno
][mode
];
2095 if (hard_reg_set_subset_p (profitable_hard_regs
,
2096 conflicting_regs
[word
]))
2102 && ! ALLOCNO_COLOR_DATA (conflict_a
)->may_be_spilled_p
2103 /* Don't process the conflict allocno twice. */
2104 && (ALLOCNO_COLOR_DATA (conflict_a
)->last_process
2105 != curr_allocno_process
))
2107 int k
, *conflict_costs
;
2109 ALLOCNO_COLOR_DATA (conflict_a
)->last_process
2110 = curr_allocno_process
;
2111 ira_allocate_and_copy_costs
2112 (&ALLOCNO_UPDATED_CONFLICT_HARD_REG_COSTS (conflict_a
),
2114 ALLOCNO_CONFLICT_HARD_REG_COSTS (conflict_a
));
2116 = ALLOCNO_UPDATED_CONFLICT_HARD_REG_COSTS (conflict_a
);
2117 if (conflict_costs
!= NULL
)
2118 for (j
= class_size
- 1; j
>= 0; j
--)
2120 hard_regno
= ira_class_hard_regs
[aclass
][j
];
2121 ira_assert (hard_regno
>= 0);
2122 k
= ira_class_hard_reg_index
[conflict_aclass
][hard_regno
];
2124 /* If HARD_REGNO is not available for CONFLICT_A,
2125 the conflict would be ignored, since HARD_REGNO
2126 will never be assigned to CONFLICT_A. */
2127 || !TEST_HARD_REG_BIT (data
->profitable_hard_regs
,
2130 full_costs
[j
] -= conflict_costs
[k
];
2132 queue_update_cost (conflict_a
, conflict_a
, NULL
, COST_HOP_DIVISOR
);
2137 /* Take into account preferences of allocnos connected by copies to
2138 the conflict allocnos. */
2139 update_conflict_hard_regno_costs (full_costs
, aclass
, true);
2141 /* Take preferences of allocnos connected by copies into
2145 start_update_cost ();
2146 queue_update_cost (a
, a
, NULL
, COST_HOP_DIVISOR
);
2147 update_conflict_hard_regno_costs (full_costs
, aclass
, false);
2149 min_cost
= min_full_cost
= INT_MAX
;
2150 /* We don't care about giving callee saved registers to allocnos no
2151 living through calls because call clobbered registers are
2152 allocated first (it is usual practice to put them first in
2153 REG_ALLOC_ORDER). */
2154 mode
= ALLOCNO_MODE (a
);
2155 for (i
= 0; i
< class_size
; i
++)
2157 hard_regno
= ira_class_hard_regs
[aclass
][i
];
2160 && FIRST_STACK_REG
<= hard_regno
&& hard_regno
<= LAST_STACK_REG
)
2163 if (! check_hard_reg_p (a
, hard_regno
,
2164 conflicting_regs
, profitable_hard_regs
))
2166 if (NUM_REGISTER_FILTERS
2167 && !test_register_filters (ALLOCNO_REGISTER_FILTERS (a
), hard_regno
))
2170 full_cost
= full_costs
[i
];
2171 if (!HONOR_REG_ALLOC_ORDER
)
2173 if ((saved_nregs
= calculate_saved_nregs (hard_regno
, mode
)) != 0)
2174 /* We need to save/restore the hard register in
2175 epilogue/prologue. Therefore we increase the cost. */
2177 rclass
= REGNO_REG_CLASS (hard_regno
);
2178 add_cost
= ((ira_memory_move_cost
[mode
][rclass
][0]
2179 + ira_memory_move_cost
[mode
][rclass
][1])
2180 * saved_nregs
/ hard_regno_nregs (hard_regno
,
2182 * (optimize_size
? 1 :
2183 REG_FREQ_FROM_BB (ENTRY_BLOCK_PTR_FOR_FN (cfun
)));
2185 full_cost
+= add_cost
;
2188 if (min_cost
> cost
)
2190 if (min_full_cost
> full_cost
)
2192 min_full_cost
= full_cost
;
2193 best_hard_regno
= hard_regno
;
2194 ira_assert (hard_regno
>= 0);
2196 if (internal_flag_ira_verbose
> 5 && ira_dump_file
!= NULL
)
2197 fprintf (ira_dump_file
, "(%d=%d,%d) ", hard_regno
, cost
, full_cost
);
2199 if (internal_flag_ira_verbose
> 5 && ira_dump_file
!= NULL
)
2200 fprintf (ira_dump_file
, "\n");
2201 if (min_full_cost
> mem_cost
2202 /* Do not spill static chain pointer pseudo when non-local goto
2204 && ! non_spilled_static_chain_regno_p (ALLOCNO_REGNO (a
)))
2206 if (! retry_p
&& internal_flag_ira_verbose
> 3 && ira_dump_file
!= NULL
)
2207 fprintf (ira_dump_file
, "(memory is more profitable %d vs %d) ",
2208 mem_cost
, min_full_cost
);
2209 best_hard_regno
= -1;
2212 if (best_hard_regno
>= 0)
2214 for (i
= hard_regno_nregs (best_hard_regno
, mode
) - 1; i
>= 0; i
--)
2215 allocated_hardreg_p
[best_hard_regno
+ i
] = true;
2216 spill_soft_conflicts (a
, allocnos_to_spill
, soft_conflict_regs
,
2220 restore_costs_from_copies (a
);
2221 ALLOCNO_HARD_REGNO (a
) = best_hard_regno
;
2222 ALLOCNO_ASSIGNED_P (a
) = true;
2223 if (best_hard_regno
>= 0 && !retry_p
)
2224 update_costs_from_copies (a
, true, true);
2225 ira_assert (ALLOCNO_CLASS (a
) == aclass
);
2226 /* We don't need updated costs anymore. */
2227 ira_free_allocno_updated_costs (a
);
2228 return best_hard_regno
>= 0;
2233 /* An array used to sort copies. */
2234 static ira_copy_t
*sorted_copies
;
2236 /* If allocno A is a cap, return non-cap allocno from which A is
2237 created. Otherwise, return A. */
2238 static ira_allocno_t
2239 get_cap_member (ira_allocno_t a
)
2241 ira_allocno_t member
;
2243 while ((member
= ALLOCNO_CAP_MEMBER (a
)) != NULL
)
2248 /* Return TRUE if live ranges of allocnos A1 and A2 intersect. It is
2249 used to find a conflict for new allocnos or allocnos with the
2250 different allocno classes. */
2252 allocnos_conflict_by_live_ranges_p (ira_allocno_t a1
, ira_allocno_t a2
)
2256 int n1
= ALLOCNO_NUM_OBJECTS (a1
);
2257 int n2
= ALLOCNO_NUM_OBJECTS (a2
);
2261 reg1
= regno_reg_rtx
[ALLOCNO_REGNO (a1
)];
2262 reg2
= regno_reg_rtx
[ALLOCNO_REGNO (a2
)];
2263 if (reg1
!= NULL
&& reg2
!= NULL
2264 && ORIGINAL_REGNO (reg1
) == ORIGINAL_REGNO (reg2
))
2267 /* We don't keep live ranges for caps because they can be quite big.
2268 Use ranges of non-cap allocno from which caps are created. */
2269 a1
= get_cap_member (a1
);
2270 a2
= get_cap_member (a2
);
2271 for (i
= 0; i
< n1
; i
++)
2273 ira_object_t c1
= ALLOCNO_OBJECT (a1
, i
);
2275 for (j
= 0; j
< n2
; j
++)
2277 ira_object_t c2
= ALLOCNO_OBJECT (a2
, j
);
2279 if (ira_live_ranges_intersect_p (OBJECT_LIVE_RANGES (c1
),
2280 OBJECT_LIVE_RANGES (c2
)))
2287 /* The function is used to sort copies according to their execution
2290 copy_freq_compare_func (const void *v1p
, const void *v2p
)
2292 ira_copy_t cp1
= *(const ira_copy_t
*) v1p
, cp2
= *(const ira_copy_t
*) v2p
;
2300 /* If frequencies are equal, sort by copies, so that the results of
2301 qsort leave nothing to chance. */
2302 return cp1
->num
- cp2
->num
;
2307 /* Return true if any allocno from thread of A1 conflicts with any
2308 allocno from thread A2. */
2310 allocno_thread_conflict_p (ira_allocno_t a1
, ira_allocno_t a2
)
2312 ira_allocno_t a
, conflict_a
;
2314 for (a
= ALLOCNO_COLOR_DATA (a2
)->next_thread_allocno
;;
2315 a
= ALLOCNO_COLOR_DATA (a
)->next_thread_allocno
)
2317 for (conflict_a
= ALLOCNO_COLOR_DATA (a1
)->next_thread_allocno
;;
2318 conflict_a
= ALLOCNO_COLOR_DATA (conflict_a
)->next_thread_allocno
)
2320 if (allocnos_conflict_by_live_ranges_p (a
, conflict_a
))
2322 if (conflict_a
== a1
)
2331 /* Merge two threads given correspondingly by their first allocnos T1
2332 and T2 (more accurately merging T2 into T1). */
2334 merge_threads (ira_allocno_t t1
, ira_allocno_t t2
)
2336 ira_allocno_t a
, next
, last
;
2338 gcc_assert (t1
!= t2
2339 && ALLOCNO_COLOR_DATA (t1
)->first_thread_allocno
== t1
2340 && ALLOCNO_COLOR_DATA (t2
)->first_thread_allocno
== t2
);
2341 for (last
= t2
, a
= ALLOCNO_COLOR_DATA (t2
)->next_thread_allocno
;;
2342 a
= ALLOCNO_COLOR_DATA (a
)->next_thread_allocno
)
2344 ALLOCNO_COLOR_DATA (a
)->first_thread_allocno
= t1
;
2349 next
= ALLOCNO_COLOR_DATA (t1
)->next_thread_allocno
;
2350 ALLOCNO_COLOR_DATA (t1
)->next_thread_allocno
= t2
;
2351 ALLOCNO_COLOR_DATA (last
)->next_thread_allocno
= next
;
2352 ALLOCNO_COLOR_DATA (t1
)->thread_freq
+= ALLOCNO_COLOR_DATA (t2
)->thread_freq
;
2355 /* Create threads by processing CP_NUM copies from sorted copies. We
2356 process the most expensive copies first. */
2358 form_threads_from_copies (int cp_num
)
2360 ira_allocno_t a
, thread1
, thread2
;
2363 qsort (sorted_copies
, cp_num
, sizeof (ira_copy_t
), copy_freq_compare_func
);
2364 /* Form threads processing copies, most frequently executed
2366 for (int i
= 0; i
< cp_num
; i
++)
2368 cp
= sorted_copies
[i
];
2369 thread1
= ALLOCNO_COLOR_DATA (cp
->first
)->first_thread_allocno
;
2370 thread2
= ALLOCNO_COLOR_DATA (cp
->second
)->first_thread_allocno
;
2371 if (thread1
== thread2
)
2373 if (! allocno_thread_conflict_p (thread1
, thread2
))
2375 if (internal_flag_ira_verbose
> 3 && ira_dump_file
!= NULL
)
2378 " Forming thread by copy %d:a%dr%d-a%dr%d (freq=%d):\n",
2379 cp
->num
, ALLOCNO_NUM (cp
->first
), ALLOCNO_REGNO (cp
->first
),
2380 ALLOCNO_NUM (cp
->second
), ALLOCNO_REGNO (cp
->second
),
2382 merge_threads (thread1
, thread2
);
2383 if (internal_flag_ira_verbose
> 3 && ira_dump_file
!= NULL
)
2385 thread1
= ALLOCNO_COLOR_DATA (thread1
)->first_thread_allocno
;
2386 fprintf (ira_dump_file
, " Result (freq=%d): a%dr%d(%d)",
2387 ALLOCNO_COLOR_DATA (thread1
)->thread_freq
,
2388 ALLOCNO_NUM (thread1
), ALLOCNO_REGNO (thread1
),
2389 ALLOCNO_FREQ (thread1
));
2390 for (a
= ALLOCNO_COLOR_DATA (thread1
)->next_thread_allocno
;
2392 a
= ALLOCNO_COLOR_DATA (a
)->next_thread_allocno
)
2393 fprintf (ira_dump_file
, " a%dr%d(%d)",
2394 ALLOCNO_NUM (a
), ALLOCNO_REGNO (a
),
2396 fprintf (ira_dump_file
, "\n");
2402 /* Create threads by processing copies of all alocnos from BUCKET. We
2403 process the most expensive copies first. */
2405 form_threads_from_bucket (ira_allocno_t bucket
)
2408 ira_copy_t cp
, next_cp
;
2411 for (a
= bucket
; a
!= NULL
; a
= ALLOCNO_COLOR_DATA (a
)->next_bucket_allocno
)
2413 for (cp
= ALLOCNO_COPIES (a
); cp
!= NULL
; cp
= next_cp
)
2417 next_cp
= cp
->next_first_allocno_copy
;
2418 sorted_copies
[cp_num
++] = cp
;
2420 else if (cp
->second
== a
)
2421 next_cp
= cp
->next_second_allocno_copy
;
2426 form_threads_from_copies (cp_num
);
2429 /* Create threads by processing copies of colorable allocno A. We
2430 process most expensive copies first. */
2432 form_threads_from_colorable_allocno (ira_allocno_t a
)
2434 ira_allocno_t another_a
;
2435 ira_copy_t cp
, next_cp
;
2438 if (internal_flag_ira_verbose
> 3 && ira_dump_file
!= NULL
)
2439 fprintf (ira_dump_file
, " Forming thread from allocno a%dr%d:\n",
2440 ALLOCNO_NUM (a
), ALLOCNO_REGNO (a
));
2441 for (cp
= ALLOCNO_COPIES (a
); cp
!= NULL
; cp
= next_cp
)
2445 next_cp
= cp
->next_first_allocno_copy
;
2446 another_a
= cp
->second
;
2448 else if (cp
->second
== a
)
2450 next_cp
= cp
->next_second_allocno_copy
;
2451 another_a
= cp
->first
;
2455 if ((! ALLOCNO_COLOR_DATA (another_a
)->in_graph_p
2456 && !ALLOCNO_COLOR_DATA (another_a
)->may_be_spilled_p
)
2457 || ALLOCNO_COLOR_DATA (another_a
)->colorable_p
)
2458 sorted_copies
[cp_num
++] = cp
;
2460 form_threads_from_copies (cp_num
);
2463 /* Form initial threads which contain only one allocno. */
2465 init_allocno_threads (void)
2472 EXECUTE_IF_SET_IN_BITMAP (consideration_allocno_bitmap
, 0, j
, bi
)
2474 a
= ira_allocnos
[j
];
2475 /* Set up initial thread data: */
2476 ALLOCNO_COLOR_DATA (a
)->first_thread_allocno
2477 = ALLOCNO_COLOR_DATA (a
)->next_thread_allocno
= a
;
2478 ALLOCNO_COLOR_DATA (a
)->thread_freq
= ALLOCNO_FREQ (a
);
2479 ALLOCNO_COLOR_DATA (a
)->hard_reg_prefs
= 0;
2480 for (pref
= ALLOCNO_PREFS (a
); pref
!= NULL
; pref
= pref
->next_pref
)
2481 ALLOCNO_COLOR_DATA (a
)->hard_reg_prefs
+= pref
->freq
;
2487 /* This page contains the allocator based on the Chaitin-Briggs algorithm. */
2489 /* Bucket of allocnos that can colored currently without spilling. */
2490 static ira_allocno_t colorable_allocno_bucket
;
2492 /* Bucket of allocnos that might be not colored currently without
2494 static ira_allocno_t uncolorable_allocno_bucket
;
2496 /* The current number of allocnos in the uncolorable_bucket. */
2497 static int uncolorable_allocnos_num
;
2499 /* Return the current spill priority of allocno A. The less the
2500 number, the more preferable the allocno for spilling. */
2502 allocno_spill_priority (ira_allocno_t a
)
2504 allocno_color_data_t data
= ALLOCNO_COLOR_DATA (a
);
2507 / (ALLOCNO_EXCESS_PRESSURE_POINTS_NUM (a
)
2508 * ira_reg_class_max_nregs
[ALLOCNO_CLASS (a
)][ALLOCNO_MODE (a
)]
2512 /* Add allocno A to bucket *BUCKET_PTR. A should be not in a bucket
2515 add_allocno_to_bucket (ira_allocno_t a
, ira_allocno_t
*bucket_ptr
)
2517 ira_allocno_t first_a
;
2518 allocno_color_data_t data
;
2520 if (bucket_ptr
== &uncolorable_allocno_bucket
2521 && ALLOCNO_CLASS (a
) != NO_REGS
)
2523 uncolorable_allocnos_num
++;
2524 ira_assert (uncolorable_allocnos_num
> 0);
2526 first_a
= *bucket_ptr
;
2527 data
= ALLOCNO_COLOR_DATA (a
);
2528 data
->next_bucket_allocno
= first_a
;
2529 data
->prev_bucket_allocno
= NULL
;
2530 if (first_a
!= NULL
)
2531 ALLOCNO_COLOR_DATA (first_a
)->prev_bucket_allocno
= a
;
2535 /* Compare two allocnos to define which allocno should be pushed first
2536 into the coloring stack. If the return is a negative number, the
2537 allocno given by the first parameter will be pushed first. In this
2538 case such allocno has less priority than the second one and the
2539 hard register will be assigned to it after assignment to the second
2540 one. As the result of such assignment order, the second allocno
2541 has a better chance to get the best hard register. */
2543 bucket_allocno_compare_func (const void *v1p
, const void *v2p
)
2545 ira_allocno_t a1
= *(const ira_allocno_t
*) v1p
;
2546 ira_allocno_t a2
= *(const ira_allocno_t
*) v2p
;
2547 int diff
, freq1
, freq2
, a1_num
, a2_num
, pref1
, pref2
;
2548 ira_allocno_t t1
= ALLOCNO_COLOR_DATA (a1
)->first_thread_allocno
;
2549 ira_allocno_t t2
= ALLOCNO_COLOR_DATA (a2
)->first_thread_allocno
;
2550 int cl1
= ALLOCNO_CLASS (a1
), cl2
= ALLOCNO_CLASS (a2
);
2552 freq1
= ALLOCNO_COLOR_DATA (t1
)->thread_freq
;
2553 freq2
= ALLOCNO_COLOR_DATA (t2
)->thread_freq
;
2554 if ((diff
= freq1
- freq2
) != 0)
2557 if ((diff
= ALLOCNO_NUM (t2
) - ALLOCNO_NUM (t1
)) != 0)
2560 /* Push pseudos requiring less hard registers first. It means that
2561 we will assign pseudos requiring more hard registers first
2562 avoiding creation small holes in free hard register file into
2563 which the pseudos requiring more hard registers cannot fit. */
2564 if ((diff
= (ira_reg_class_max_nregs
[cl1
][ALLOCNO_MODE (a1
)]
2565 - ira_reg_class_max_nregs
[cl2
][ALLOCNO_MODE (a2
)])) != 0)
2568 freq1
= ALLOCNO_FREQ (a1
);
2569 freq2
= ALLOCNO_FREQ (a2
);
2570 if ((diff
= freq1
- freq2
) != 0)
2573 a1_num
= ALLOCNO_COLOR_DATA (a1
)->available_regs_num
;
2574 a2_num
= ALLOCNO_COLOR_DATA (a2
)->available_regs_num
;
2575 if ((diff
= a2_num
- a1_num
) != 0)
2577 /* Push allocnos with minimal conflict_allocno_hard_prefs first. */
2578 pref1
= ALLOCNO_COLOR_DATA (a1
)->conflict_allocno_hard_prefs
;
2579 pref2
= ALLOCNO_COLOR_DATA (a2
)->conflict_allocno_hard_prefs
;
2580 if ((diff
= pref1
- pref2
) != 0)
2582 return ALLOCNO_NUM (a2
) - ALLOCNO_NUM (a1
);
2585 /* Sort bucket *BUCKET_PTR and return the result through
2588 sort_bucket (ira_allocno_t
*bucket_ptr
,
2589 int (*compare_func
) (const void *, const void *))
2591 ira_allocno_t a
, head
;
2594 for (n
= 0, a
= *bucket_ptr
;
2596 a
= ALLOCNO_COLOR_DATA (a
)->next_bucket_allocno
)
2597 sorted_allocnos
[n
++] = a
;
2600 qsort (sorted_allocnos
, n
, sizeof (ira_allocno_t
), compare_func
);
2602 for (n
--; n
>= 0; n
--)
2604 a
= sorted_allocnos
[n
];
2605 ALLOCNO_COLOR_DATA (a
)->next_bucket_allocno
= head
;
2606 ALLOCNO_COLOR_DATA (a
)->prev_bucket_allocno
= NULL
;
2608 ALLOCNO_COLOR_DATA (head
)->prev_bucket_allocno
= a
;
2614 /* Add ALLOCNO to colorable bucket maintaining the order according
2615 their priority. ALLOCNO should be not in a bucket before the
2618 add_allocno_to_ordered_colorable_bucket (ira_allocno_t allocno
)
2620 ira_allocno_t before
, after
;
2622 form_threads_from_colorable_allocno (allocno
);
2623 for (before
= colorable_allocno_bucket
, after
= NULL
;
2626 before
= ALLOCNO_COLOR_DATA (before
)->next_bucket_allocno
)
2627 if (bucket_allocno_compare_func (&allocno
, &before
) < 0)
2629 ALLOCNO_COLOR_DATA (allocno
)->next_bucket_allocno
= before
;
2630 ALLOCNO_COLOR_DATA (allocno
)->prev_bucket_allocno
= after
;
2632 colorable_allocno_bucket
= allocno
;
2634 ALLOCNO_COLOR_DATA (after
)->next_bucket_allocno
= allocno
;
2636 ALLOCNO_COLOR_DATA (before
)->prev_bucket_allocno
= allocno
;
2639 /* Delete ALLOCNO from bucket *BUCKET_PTR. It should be there before
2642 delete_allocno_from_bucket (ira_allocno_t allocno
, ira_allocno_t
*bucket_ptr
)
2644 ira_allocno_t prev_allocno
, next_allocno
;
2646 if (bucket_ptr
== &uncolorable_allocno_bucket
2647 && ALLOCNO_CLASS (allocno
) != NO_REGS
)
2649 uncolorable_allocnos_num
--;
2650 ira_assert (uncolorable_allocnos_num
>= 0);
2652 prev_allocno
= ALLOCNO_COLOR_DATA (allocno
)->prev_bucket_allocno
;
2653 next_allocno
= ALLOCNO_COLOR_DATA (allocno
)->next_bucket_allocno
;
2654 if (prev_allocno
!= NULL
)
2655 ALLOCNO_COLOR_DATA (prev_allocno
)->next_bucket_allocno
= next_allocno
;
2658 ira_assert (*bucket_ptr
== allocno
);
2659 *bucket_ptr
= next_allocno
;
2661 if (next_allocno
!= NULL
)
2662 ALLOCNO_COLOR_DATA (next_allocno
)->prev_bucket_allocno
= prev_allocno
;
2665 /* Put allocno A onto the coloring stack without removing it from its
2666 bucket. Pushing allocno to the coloring stack can result in moving
2667 conflicting allocnos from the uncolorable bucket to the colorable
2668 one. Update conflict_allocno_hard_prefs of the conflicting
2669 allocnos which are not on stack yet. */
2671 push_allocno_to_stack (ira_allocno_t a
)
2673 enum reg_class aclass
;
2674 allocno_color_data_t data
, conflict_data
;
2675 int size
, i
, n
= ALLOCNO_NUM_OBJECTS (a
);
2677 data
= ALLOCNO_COLOR_DATA (a
);
2678 data
->in_graph_p
= false;
2679 allocno_stack_vec
.safe_push (a
);
2680 aclass
= ALLOCNO_CLASS (a
);
2681 if (aclass
== NO_REGS
)
2683 size
= ira_reg_class_max_nregs
[aclass
][ALLOCNO_MODE (a
)];
2686 /* We will deal with the subwords individually. */
2687 gcc_assert (size
== ALLOCNO_NUM_OBJECTS (a
));
2690 for (i
= 0; i
< n
; i
++)
2692 ira_object_t obj
= ALLOCNO_OBJECT (a
, i
);
2693 ira_object_t conflict_obj
;
2694 ira_object_conflict_iterator oci
;
2696 FOR_EACH_OBJECT_CONFLICT (obj
, conflict_obj
, oci
)
2698 ira_allocno_t conflict_a
= OBJECT_ALLOCNO (conflict_obj
);
2701 conflict_data
= ALLOCNO_COLOR_DATA (conflict_a
);
2702 if (! conflict_data
->in_graph_p
2703 || ALLOCNO_ASSIGNED_P (conflict_a
)
2704 || !(hard_reg_set_intersect_p
2705 (ALLOCNO_COLOR_DATA (a
)->profitable_hard_regs
,
2706 conflict_data
->profitable_hard_regs
)))
2708 for (pref
= ALLOCNO_PREFS (a
); pref
!= NULL
; pref
= pref
->next_pref
)
2709 conflict_data
->conflict_allocno_hard_prefs
-= pref
->freq
;
2710 if (conflict_data
->colorable_p
)
2712 ira_assert (bitmap_bit_p (coloring_allocno_bitmap
,
2713 ALLOCNO_NUM (conflict_a
)));
2714 if (update_left_conflict_sizes_p (conflict_a
, a
, size
))
2716 delete_allocno_from_bucket
2717 (conflict_a
, &uncolorable_allocno_bucket
);
2718 add_allocno_to_ordered_colorable_bucket (conflict_a
);
2719 if (internal_flag_ira_verbose
> 4 && ira_dump_file
!= NULL
)
2721 fprintf (ira_dump_file
, " Making");
2722 ira_print_expanded_allocno (conflict_a
);
2723 fprintf (ira_dump_file
, " colorable\n");
2731 /* Put ALLOCNO onto the coloring stack and remove it from its bucket.
2732 The allocno is in the colorable bucket if COLORABLE_P is TRUE. */
2734 remove_allocno_from_bucket_and_push (ira_allocno_t allocno
, bool colorable_p
)
2737 delete_allocno_from_bucket (allocno
, &colorable_allocno_bucket
);
2739 delete_allocno_from_bucket (allocno
, &uncolorable_allocno_bucket
);
2740 if (internal_flag_ira_verbose
> 3 && ira_dump_file
!= NULL
)
2742 fprintf (ira_dump_file
, " Pushing");
2743 ira_print_expanded_allocno (allocno
);
2745 fprintf (ira_dump_file
, "(cost %d)\n",
2746 ALLOCNO_COLOR_DATA (allocno
)->temp
);
2748 fprintf (ira_dump_file
, "(potential spill: %spri=%d, cost=%d)\n",
2749 ALLOCNO_BAD_SPILL_P (allocno
) ? "bad spill, " : "",
2750 allocno_spill_priority (allocno
),
2751 ALLOCNO_COLOR_DATA (allocno
)->temp
);
2754 ALLOCNO_COLOR_DATA (allocno
)->may_be_spilled_p
= true;
2755 push_allocno_to_stack (allocno
);
2758 /* Put all allocnos from colorable bucket onto the coloring stack. */
2760 push_only_colorable (void)
2762 if (internal_flag_ira_verbose
> 3 && ira_dump_file
!= NULL
)
2763 fprintf (ira_dump_file
, " Forming thread from colorable bucket:\n");
2764 form_threads_from_bucket (colorable_allocno_bucket
);
2765 for (ira_allocno_t a
= colorable_allocno_bucket
;
2767 a
= ALLOCNO_COLOR_DATA (a
)->next_bucket_allocno
)
2768 update_costs_from_prefs (a
);
2769 sort_bucket (&colorable_allocno_bucket
, bucket_allocno_compare_func
);
2770 for (;colorable_allocno_bucket
!= NULL
;)
2771 remove_allocno_from_bucket_and_push (colorable_allocno_bucket
, true);
2774 /* Return the frequency of exit edges (if EXIT_P) or entry from/to the
2775 loop given by its LOOP_NODE. */
2777 ira_loop_edge_freq (ira_loop_tree_node_t loop_node
, int regno
, bool exit_p
)
2783 ira_assert (current_loops
!= NULL
&& loop_node
->loop
!= NULL
2784 && (regno
< 0 || regno
>= FIRST_PSEUDO_REGISTER
));
2788 FOR_EACH_EDGE (e
, ei
, loop_node
->loop
->header
->preds
)
2789 if (e
->src
!= loop_node
->loop
->latch
2791 || (bitmap_bit_p (df_get_live_out (e
->src
), regno
)
2792 && bitmap_bit_p (df_get_live_in (e
->dest
), regno
))))
2793 freq
+= EDGE_FREQUENCY (e
);
2797 auto_vec
<edge
> edges
= get_loop_exit_edges (loop_node
->loop
);
2798 FOR_EACH_VEC_ELT (edges
, i
, e
)
2800 || (bitmap_bit_p (df_get_live_out (e
->src
), regno
)
2801 && bitmap_bit_p (df_get_live_in (e
->dest
), regno
)))
2802 freq
+= EDGE_FREQUENCY (e
);
2805 return REG_FREQ_FROM_EDGE_FREQ (freq
);
2808 /* Construct an object that describes the boundary between A and its
2810 ira_loop_border_costs::ira_loop_border_costs (ira_allocno_t a
)
2811 : m_mode (ALLOCNO_MODE (a
)),
2812 m_class (ALLOCNO_CLASS (a
)),
2813 m_entry_freq (ira_loop_edge_freq (ALLOCNO_LOOP_TREE_NODE (a
),
2814 ALLOCNO_REGNO (a
), false)),
2815 m_exit_freq (ira_loop_edge_freq (ALLOCNO_LOOP_TREE_NODE (a
),
2816 ALLOCNO_REGNO (a
), true))
2820 /* Calculate and return the cost of putting allocno A into memory. */
2822 calculate_allocno_spill_cost (ira_allocno_t a
)
2825 ira_allocno_t parent_allocno
;
2826 ira_loop_tree_node_t parent_node
, loop_node
;
2828 regno
= ALLOCNO_REGNO (a
);
2829 cost
= ALLOCNO_UPDATED_MEMORY_COST (a
) - ALLOCNO_UPDATED_CLASS_COST (a
);
2830 if (ALLOCNO_CAP (a
) != NULL
)
2832 loop_node
= ALLOCNO_LOOP_TREE_NODE (a
);
2833 if ((parent_node
= loop_node
->parent
) == NULL
)
2835 if ((parent_allocno
= parent_node
->regno_allocno_map
[regno
]) == NULL
)
2837 ira_loop_border_costs
border_costs (a
);
2838 if (ALLOCNO_HARD_REGNO (parent_allocno
) < 0)
2839 cost
-= border_costs
.spill_outside_loop_cost ();
2841 cost
+= (border_costs
.spill_inside_loop_cost ()
2842 - border_costs
.move_between_loops_cost ());
2846 /* Used for sorting allocnos for spilling. */
2848 allocno_spill_priority_compare (ira_allocno_t a1
, ira_allocno_t a2
)
2850 int pri1
, pri2
, diff
;
2852 /* Avoid spilling static chain pointer pseudo when non-local goto is
2854 if (non_spilled_static_chain_regno_p (ALLOCNO_REGNO (a1
)))
2856 else if (non_spilled_static_chain_regno_p (ALLOCNO_REGNO (a2
)))
2858 if (ALLOCNO_BAD_SPILL_P (a1
) && ! ALLOCNO_BAD_SPILL_P (a2
))
2860 if (ALLOCNO_BAD_SPILL_P (a2
) && ! ALLOCNO_BAD_SPILL_P (a1
))
2862 pri1
= allocno_spill_priority (a1
);
2863 pri2
= allocno_spill_priority (a2
);
2864 if ((diff
= pri1
- pri2
) != 0)
2867 = ALLOCNO_COLOR_DATA (a1
)->temp
- ALLOCNO_COLOR_DATA (a2
)->temp
) != 0)
2869 return ALLOCNO_NUM (a1
) - ALLOCNO_NUM (a2
);
2872 /* Used for sorting allocnos for spilling. */
2874 allocno_spill_sort_compare (const void *v1p
, const void *v2p
)
2876 ira_allocno_t p1
= *(const ira_allocno_t
*) v1p
;
2877 ira_allocno_t p2
= *(const ira_allocno_t
*) v2p
;
2879 return allocno_spill_priority_compare (p1
, p2
);
2882 /* Push allocnos to the coloring stack. The order of allocnos in the
2883 stack defines the order for the subsequent coloring. */
2885 push_allocnos_to_stack (void)
2890 /* Calculate uncolorable allocno spill costs. */
2891 for (a
= uncolorable_allocno_bucket
;
2893 a
= ALLOCNO_COLOR_DATA (a
)->next_bucket_allocno
)
2894 if (ALLOCNO_CLASS (a
) != NO_REGS
)
2896 cost
= calculate_allocno_spill_cost (a
);
2897 /* ??? Remove cost of copies between the coalesced
2899 ALLOCNO_COLOR_DATA (a
)->temp
= cost
;
2901 sort_bucket (&uncolorable_allocno_bucket
, allocno_spill_sort_compare
);
2904 push_only_colorable ();
2905 a
= uncolorable_allocno_bucket
;
2908 remove_allocno_from_bucket_and_push (a
, false);
2910 ira_assert (colorable_allocno_bucket
== NULL
2911 && uncolorable_allocno_bucket
== NULL
);
2912 ira_assert (uncolorable_allocnos_num
== 0);
2915 /* Pop the coloring stack and assign hard registers to the popped
2918 pop_allocnos_from_stack (void)
2920 ira_allocno_t allocno
;
2921 enum reg_class aclass
;
2923 for (;allocno_stack_vec
.length () != 0;)
2925 allocno
= allocno_stack_vec
.pop ();
2926 aclass
= ALLOCNO_CLASS (allocno
);
2927 if (internal_flag_ira_verbose
> 3 && ira_dump_file
!= NULL
)
2929 fprintf (ira_dump_file
, " Popping");
2930 ira_print_expanded_allocno (allocno
);
2931 fprintf (ira_dump_file
, " -- ");
2933 if (aclass
== NO_REGS
)
2935 ALLOCNO_HARD_REGNO (allocno
) = -1;
2936 ALLOCNO_ASSIGNED_P (allocno
) = true;
2937 ira_assert (ALLOCNO_UPDATED_HARD_REG_COSTS (allocno
) == NULL
);
2939 (ALLOCNO_UPDATED_CONFLICT_HARD_REG_COSTS (allocno
) == NULL
);
2940 if (internal_flag_ira_verbose
> 3 && ira_dump_file
!= NULL
)
2941 fprintf (ira_dump_file
, "assign memory\n");
2943 else if (assign_hard_reg (allocno
, false))
2945 if (internal_flag_ira_verbose
> 3 && ira_dump_file
!= NULL
)
2946 fprintf (ira_dump_file
, " assign reg %d\n",
2947 ALLOCNO_HARD_REGNO (allocno
));
2949 else if (ALLOCNO_ASSIGNED_P (allocno
))
2951 if (internal_flag_ira_verbose
> 3 && ira_dump_file
!= NULL
)
2952 fprintf (ira_dump_file
, "spill%s\n",
2953 ALLOCNO_COLOR_DATA (allocno
)->may_be_spilled_p
2956 ALLOCNO_COLOR_DATA (allocno
)->in_graph_p
= true;
2960 /* Set up number of available hard registers for allocno A. */
2962 setup_allocno_available_regs_num (ira_allocno_t a
)
2964 int i
, n
, hard_regno
, hard_regs_num
, nwords
;
2965 enum reg_class aclass
;
2966 allocno_color_data_t data
;
2968 aclass
= ALLOCNO_CLASS (a
);
2969 data
= ALLOCNO_COLOR_DATA (a
);
2970 data
->available_regs_num
= 0;
2971 if (aclass
== NO_REGS
)
2973 hard_regs_num
= ira_class_hard_regs_num
[aclass
];
2974 nwords
= ALLOCNO_NUM_OBJECTS (a
);
2975 for (n
= 0, i
= hard_regs_num
- 1; i
>= 0; i
--)
2977 hard_regno
= ira_class_hard_regs
[aclass
][i
];
2978 /* Checking only profitable hard regs. */
2979 if (TEST_HARD_REG_BIT (data
->profitable_hard_regs
, hard_regno
))
2982 data
->available_regs_num
= n
;
2983 if (internal_flag_ira_verbose
<= 2 || ira_dump_file
== NULL
)
2987 " Allocno a%dr%d of %s(%d) has %d avail. regs ",
2988 ALLOCNO_NUM (a
), ALLOCNO_REGNO (a
),
2989 reg_class_names
[aclass
], ira_class_hard_regs_num
[aclass
], n
);
2990 print_hard_reg_set (ira_dump_file
, data
->profitable_hard_regs
, false);
2991 fprintf (ira_dump_file
, ", %snode: ",
2992 data
->profitable_hard_regs
== data
->hard_regs_node
->hard_regs
->set
2994 print_hard_reg_set (ira_dump_file
,
2995 data
->hard_regs_node
->hard_regs
->set
, false);
2996 for (i
= 0; i
< nwords
; i
++)
2998 ira_object_t obj
= ALLOCNO_OBJECT (a
, i
);
3003 fprintf (ira_dump_file
, ", ");
3004 fprintf (ira_dump_file
, " obj %d", i
);
3006 fprintf (ira_dump_file
, " (confl regs = ");
3007 print_hard_reg_set (ira_dump_file
, OBJECT_TOTAL_CONFLICT_HARD_REGS (obj
),
3009 fprintf (ira_dump_file
, ")");
3011 fprintf (ira_dump_file
, "\n");
3014 /* Put ALLOCNO in a bucket corresponding to its number and size of its
3015 conflicting allocnos and hard registers. */
3017 put_allocno_into_bucket (ira_allocno_t allocno
)
3019 ALLOCNO_COLOR_DATA (allocno
)->in_graph_p
= true;
3020 setup_allocno_available_regs_num (allocno
);
3021 if (setup_left_conflict_sizes_p (allocno
))
3022 add_allocno_to_bucket (allocno
, &colorable_allocno_bucket
);
3024 add_allocno_to_bucket (allocno
, &uncolorable_allocno_bucket
);
3027 /* Map: allocno number -> allocno priority. */
3028 static int *allocno_priorities
;
3030 /* Set up priorities for N allocnos in array
3031 CONSIDERATION_ALLOCNOS. */
3033 setup_allocno_priorities (ira_allocno_t
*consideration_allocnos
, int n
)
3035 int i
, length
, nrefs
, priority
, max_priority
, mult
, diff
;
3039 for (i
= 0; i
< n
; i
++)
3041 a
= consideration_allocnos
[i
];
3042 nrefs
= ALLOCNO_NREFS (a
);
3043 ira_assert (nrefs
>= 0);
3044 mult
= floor_log2 (ALLOCNO_NREFS (a
)) + 1;
3045 ira_assert (mult
>= 0);
3046 mult
*= ira_reg_class_max_nregs
[ALLOCNO_CLASS (a
)][ALLOCNO_MODE (a
)];
3047 diff
= ALLOCNO_MEMORY_COST (a
) - ALLOCNO_CLASS_COST (a
);
3048 #ifdef __has_builtin
3049 #if __has_builtin(__builtin_smul_overflow)
3050 #define HAS_SMUL_OVERFLOW
3053 /* Multiplication can overflow for very large functions.
3054 Check the overflow and constrain the result if necessary: */
3055 #ifdef HAS_SMUL_OVERFLOW
3056 if (__builtin_smul_overflow (mult
, diff
, &priority
)
3057 || priority
< -INT_MAX
)
3058 priority
= diff
>= 0 ? INT_MAX
: -INT_MAX
;
3061 (sizeof (long long) >= 2 * sizeof (int),
3062 "overflow code does not work for such int and long long sizes");
3063 long long priorityll
= (long long) mult
* diff
;
3064 if (priorityll
< -INT_MAX
|| priorityll
> INT_MAX
)
3065 priority
= diff
>= 0 ? INT_MAX
: -INT_MAX
;
3067 priority
= priorityll
;
3069 allocno_priorities
[ALLOCNO_NUM (a
)] = priority
;
3071 priority
= -priority
;
3072 if (max_priority
< priority
)
3073 max_priority
= priority
;
3075 mult
= max_priority
== 0 ? 1 : INT_MAX
/ max_priority
;
3076 for (i
= 0; i
< n
; i
++)
3078 a
= consideration_allocnos
[i
];
3079 length
= ALLOCNO_EXCESS_PRESSURE_POINTS_NUM (a
);
3080 if (ALLOCNO_NUM_OBJECTS (a
) > 1)
3081 length
/= ALLOCNO_NUM_OBJECTS (a
);
3084 allocno_priorities
[ALLOCNO_NUM (a
)]
3085 = allocno_priorities
[ALLOCNO_NUM (a
)] * mult
/ length
;
3089 /* Sort allocnos according to the profit of usage of a hard register
3090 instead of memory for them. */
3092 allocno_cost_compare_func (const void *v1p
, const void *v2p
)
3094 ira_allocno_t p1
= *(const ira_allocno_t
*) v1p
;
3095 ira_allocno_t p2
= *(const ira_allocno_t
*) v2p
;
3098 c1
= ALLOCNO_UPDATED_MEMORY_COST (p1
) - ALLOCNO_UPDATED_CLASS_COST (p1
);
3099 c2
= ALLOCNO_UPDATED_MEMORY_COST (p2
) - ALLOCNO_UPDATED_CLASS_COST (p2
);
3103 /* If regs are equally good, sort by allocno numbers, so that the
3104 results of qsort leave nothing to chance. */
3105 return ALLOCNO_NUM (p1
) - ALLOCNO_NUM (p2
);
3108 /* Return savings on removed copies when ALLOCNO is assigned to
3111 allocno_copy_cost_saving (ira_allocno_t allocno
, int hard_regno
)
3114 machine_mode allocno_mode
= ALLOCNO_MODE (allocno
);
3115 enum reg_class rclass
;
3116 ira_copy_t cp
, next_cp
;
3118 rclass
= REGNO_REG_CLASS (hard_regno
);
3119 if (ira_reg_class_max_nregs
[rclass
][allocno_mode
]
3120 > ira_class_hard_regs_num
[rclass
])
3121 /* For the above condition the cost can be wrong. Use the allocno
3122 class in this case. */
3123 rclass
= ALLOCNO_CLASS (allocno
);
3124 for (cp
= ALLOCNO_COPIES (allocno
); cp
!= NULL
; cp
= next_cp
)
3126 if (cp
->first
== allocno
)
3128 next_cp
= cp
->next_first_allocno_copy
;
3129 if (ALLOCNO_HARD_REGNO (cp
->second
) != hard_regno
)
3132 else if (cp
->second
== allocno
)
3134 next_cp
= cp
->next_second_allocno_copy
;
3135 if (ALLOCNO_HARD_REGNO (cp
->first
) != hard_regno
)
3140 ira_init_register_move_cost_if_necessary (allocno_mode
);
3141 cost
+= cp
->freq
* ira_register_move_cost
[allocno_mode
][rclass
][rclass
];
3146 /* We used Chaitin-Briggs coloring to assign as many pseudos as
3147 possible to hard registers. Let us try to improve allocation with
3148 cost point of view. This function improves the allocation by
3149 spilling some allocnos and assigning the freed hard registers to
3150 other allocnos if it decreases the overall allocation cost. */
3152 improve_allocation (void)
3155 int j
, k
, n
, hregno
, conflict_hregno
, base_cost
, class_size
, word
, nwords
;
3156 int check
, spill_cost
, min_cost
, nregs
, conflict_nregs
, r
, best
;
3158 enum reg_class aclass
, rclass
;
3161 int costs
[FIRST_PSEUDO_REGISTER
];
3162 HARD_REG_SET conflicting_regs
[2], profitable_hard_regs
;
3168 /* Don't bother to optimize the code with static chain pointer and
3169 non-local goto in order not to spill the chain pointer
3171 if (cfun
->static_chain_decl
&& crtl
->has_nonlocal_goto
)
3173 /* Clear counts used to process conflicting allocnos only once for
3175 EXECUTE_IF_SET_IN_BITMAP (coloring_allocno_bitmap
, 0, i
, bi
)
3176 ALLOCNO_COLOR_DATA (ira_allocnos
[i
])->temp
= 0;
3178 /* Process each allocno and try to assign a hard register to it by
3179 spilling some its conflicting allocnos. */
3180 EXECUTE_IF_SET_IN_BITMAP (coloring_allocno_bitmap
, 0, i
, bi
)
3182 a
= ira_allocnos
[i
];
3183 ALLOCNO_COLOR_DATA (a
)->temp
= 0;
3184 if (empty_profitable_hard_regs (a
))
3187 aclass
= ALLOCNO_CLASS (a
);
3188 allocno_costs
= ALLOCNO_HARD_REG_COSTS (a
);
3189 if ((hregno
= ALLOCNO_HARD_REGNO (a
)) < 0)
3190 base_cost
= ALLOCNO_UPDATED_MEMORY_COST (a
);
3191 else if (allocno_costs
== NULL
)
3192 /* It means that assigning a hard register is not profitable
3193 (we don't waste memory for hard register costs in this
3197 base_cost
= (allocno_costs
[ira_class_hard_reg_index
[aclass
][hregno
]]
3198 - allocno_copy_cost_saving (a
, hregno
));
3200 get_conflict_and_start_profitable_regs (a
, false,
3202 &profitable_hard_regs
);
3203 class_size
= ira_class_hard_regs_num
[aclass
];
3204 mode
= ALLOCNO_MODE (a
);
3205 /* Set up cost improvement for usage of each profitable hard
3206 register for allocno A. */
3207 for (j
= 0; j
< class_size
; j
++)
3209 hregno
= ira_class_hard_regs
[aclass
][j
];
3210 if (! check_hard_reg_p (a
, hregno
,
3211 conflicting_regs
, profitable_hard_regs
))
3213 if (NUM_REGISTER_FILTERS
3214 && !test_register_filters (ALLOCNO_REGISTER_FILTERS (a
), hregno
))
3216 ira_assert (ira_class_hard_reg_index
[aclass
][hregno
] == j
);
3217 k
= allocno_costs
== NULL
? 0 : j
;
3218 costs
[hregno
] = (allocno_costs
== NULL
3219 ? ALLOCNO_UPDATED_CLASS_COST (a
) : allocno_costs
[k
]);
3220 costs
[hregno
] -= allocno_copy_cost_saving (a
, hregno
);
3222 if ((saved_nregs
= calculate_saved_nregs (hregno
, mode
)) != 0)
3224 /* We need to save/restore the hard register in
3225 epilogue/prologue. Therefore we increase the cost.
3226 Since the prolog is placed in the entry BB, the frequency
3227 of the entry BB is considered while computing the cost. */
3228 rclass
= REGNO_REG_CLASS (hregno
);
3229 add_cost
= ((ira_memory_move_cost
[mode
][rclass
][0]
3230 + ira_memory_move_cost
[mode
][rclass
][1])
3231 * saved_nregs
/ hard_regno_nregs (hregno
,
3233 * REG_FREQ_FROM_BB (ENTRY_BLOCK_PTR_FOR_FN (cfun
));
3234 costs
[hregno
] += add_cost
;
3237 costs
[hregno
] -= base_cost
;
3238 if (costs
[hregno
] < 0)
3242 /* There is no chance to improve the allocation cost by
3243 assigning hard register to allocno A even without spilling
3244 conflicting allocnos. */
3246 auto_bitmap allocnos_to_spill
;
3247 HARD_REG_SET soft_conflict_regs
= {};
3248 mode
= ALLOCNO_MODE (a
);
3249 nwords
= ALLOCNO_NUM_OBJECTS (a
);
3250 /* Process each allocno conflicting with A and update the cost
3251 improvement for profitable hard registers of A. To use a
3252 hard register for A we need to spill some conflicting
3253 allocnos and that creates penalty for the cost
3255 for (word
= 0; word
< nwords
; word
++)
3257 ira_object_t conflict_obj
;
3258 ira_object_t obj
= ALLOCNO_OBJECT (a
, word
);
3259 ira_object_conflict_iterator oci
;
3261 FOR_EACH_OBJECT_CONFLICT (obj
, conflict_obj
, oci
)
3263 ira_allocno_t conflict_a
= OBJECT_ALLOCNO (conflict_obj
);
3265 if (ALLOCNO_COLOR_DATA (conflict_a
)->temp
== check
)
3266 /* We already processed this conflicting allocno
3267 because we processed earlier another object of the
3268 conflicting allocno. */
3270 ALLOCNO_COLOR_DATA (conflict_a
)->temp
= check
;
3271 if ((conflict_hregno
= ALLOCNO_HARD_REGNO (conflict_a
)) < 0)
3273 auto spill_a
= ira_soft_conflict (a
, conflict_a
);
3276 if (!bitmap_set_bit (allocnos_to_spill
,
3277 ALLOCNO_NUM (spill_a
)))
3279 ira_loop_border_costs
border_costs (spill_a
);
3280 spill_cost
= border_costs
.spill_inside_loop_cost ();
3284 spill_cost
= ALLOCNO_UPDATED_MEMORY_COST (conflict_a
);
3285 k
= (ira_class_hard_reg_index
3286 [ALLOCNO_CLASS (conflict_a
)][conflict_hregno
]);
3287 ira_assert (k
>= 0);
3288 if ((allocno_costs
= ALLOCNO_HARD_REG_COSTS (conflict_a
))
3290 spill_cost
-= allocno_costs
[k
];
3292 spill_cost
-= ALLOCNO_UPDATED_CLASS_COST (conflict_a
);
3294 += allocno_copy_cost_saving (conflict_a
, conflict_hregno
);
3296 conflict_nregs
= hard_regno_nregs (conflict_hregno
,
3297 ALLOCNO_MODE (conflict_a
));
3298 auto note_conflict
= [&](int r
)
3300 if (check_hard_reg_p (a
, r
,
3301 conflicting_regs
, profitable_hard_regs
))
3304 SET_HARD_REG_BIT (soft_conflict_regs
, r
);
3305 costs
[r
] += spill_cost
;
3308 for (r
= conflict_hregno
;
3309 r
>= 0 && (int) end_hard_regno (mode
, r
) > conflict_hregno
;
3312 for (r
= conflict_hregno
+ 1;
3313 r
< conflict_hregno
+ conflict_nregs
;
3320 /* Now we choose hard register for A which results in highest
3321 allocation cost improvement. */
3322 for (j
= 0; j
< class_size
; j
++)
3324 hregno
= ira_class_hard_regs
[aclass
][j
];
3325 if (check_hard_reg_p (a
, hregno
,
3326 conflicting_regs
, profitable_hard_regs
)
3327 && min_cost
> costs
[hregno
])
3330 min_cost
= costs
[hregno
];
3334 /* We are in a situation when assigning any hard register to A
3335 by spilling some conflicting allocnos does not improve the
3338 spill_soft_conflicts (a
, allocnos_to_spill
, soft_conflict_regs
, best
);
3339 nregs
= hard_regno_nregs (best
, mode
);
3340 /* Now spill conflicting allocnos which contain a hard register
3341 of A when we assign the best chosen hard register to it. */
3342 for (word
= 0; word
< nwords
; word
++)
3344 ira_object_t conflict_obj
;
3345 ira_object_t obj
= ALLOCNO_OBJECT (a
, word
);
3346 ira_object_conflict_iterator oci
;
3348 FOR_EACH_OBJECT_CONFLICT (obj
, conflict_obj
, oci
)
3350 ira_allocno_t conflict_a
= OBJECT_ALLOCNO (conflict_obj
);
3352 if ((conflict_hregno
= ALLOCNO_HARD_REGNO (conflict_a
)) < 0)
3354 conflict_nregs
= hard_regno_nregs (conflict_hregno
,
3355 ALLOCNO_MODE (conflict_a
));
3356 if (best
+ nregs
<= conflict_hregno
3357 || conflict_hregno
+ conflict_nregs
<= best
)
3358 /* No intersection. */
3360 ALLOCNO_HARD_REGNO (conflict_a
) = -1;
3361 sorted_allocnos
[n
++] = conflict_a
;
3362 if (internal_flag_ira_verbose
> 2 && ira_dump_file
!= NULL
)
3363 fprintf (ira_dump_file
, "Spilling a%dr%d for a%dr%d\n",
3364 ALLOCNO_NUM (conflict_a
), ALLOCNO_REGNO (conflict_a
),
3365 ALLOCNO_NUM (a
), ALLOCNO_REGNO (a
));
3368 /* Assign the best chosen hard register to A. */
3369 ALLOCNO_HARD_REGNO (a
) = best
;
3371 for (j
= nregs
- 1; j
>= 0; j
--)
3372 allocated_hardreg_p
[best
+ j
] = true;
3374 if (internal_flag_ira_verbose
> 2 && ira_dump_file
!= NULL
)
3375 fprintf (ira_dump_file
, "Assigning %d to a%dr%d\n",
3376 best
, ALLOCNO_NUM (a
), ALLOCNO_REGNO (a
));
3380 /* We spilled some allocnos to assign their hard registers to other
3381 allocnos. The spilled allocnos are now in array
3382 'sorted_allocnos'. There is still a possibility that some of the
3383 spilled allocnos can get hard registers. So let us try assign
3384 them hard registers again (just a reminder -- function
3385 'assign_hard_reg' assigns hard registers only if it is possible
3386 and profitable). We process the spilled allocnos with biggest
3387 benefit to get hard register first -- see function
3388 'allocno_cost_compare_func'. */
3389 qsort (sorted_allocnos
, n
, sizeof (ira_allocno_t
),
3390 allocno_cost_compare_func
);
3391 for (j
= 0; j
< n
; j
++)
3393 a
= sorted_allocnos
[j
];
3394 ALLOCNO_ASSIGNED_P (a
) = false;
3395 if (internal_flag_ira_verbose
> 3 && ira_dump_file
!= NULL
)
3397 fprintf (ira_dump_file
, " ");
3398 ira_print_expanded_allocno (a
);
3399 fprintf (ira_dump_file
, " -- ");
3401 if (assign_hard_reg (a
, false))
3403 if (internal_flag_ira_verbose
> 3 && ira_dump_file
!= NULL
)
3404 fprintf (ira_dump_file
, "assign hard reg %d\n",
3405 ALLOCNO_HARD_REGNO (a
));
3409 if (internal_flag_ira_verbose
> 3 && ira_dump_file
!= NULL
)
3410 fprintf (ira_dump_file
, "assign memory\n");
3415 /* Sort allocnos according to their priorities. */
3417 allocno_priority_compare_func (const void *v1p
, const void *v2p
)
3419 ira_allocno_t a1
= *(const ira_allocno_t
*) v1p
;
3420 ira_allocno_t a2
= *(const ira_allocno_t
*) v2p
;
3421 int pri1
, pri2
, diff
;
3423 /* Assign hard reg to static chain pointer pseudo first when
3424 non-local goto is used. */
3425 if ((diff
= (non_spilled_static_chain_regno_p (ALLOCNO_REGNO (a2
))
3426 - non_spilled_static_chain_regno_p (ALLOCNO_REGNO (a1
)))) != 0)
3428 pri1
= allocno_priorities
[ALLOCNO_NUM (a1
)];
3429 pri2
= allocno_priorities
[ALLOCNO_NUM (a2
)];
3431 return SORTGT (pri2
, pri1
);
3433 /* If regs are equally good, sort by allocnos, so that the results of
3434 qsort leave nothing to chance. */
3435 return ALLOCNO_NUM (a1
) - ALLOCNO_NUM (a2
);
3438 /* Chaitin-Briggs coloring for allocnos in COLORING_ALLOCNO_BITMAP
3439 taking into account allocnos in CONSIDERATION_ALLOCNO_BITMAP. */
3441 color_allocnos (void)
3447 setup_profitable_hard_regs ();
3448 EXECUTE_IF_SET_IN_BITMAP (coloring_allocno_bitmap
, 0, i
, bi
)
3450 allocno_color_data_t data
;
3451 ira_pref_t pref
, next_pref
;
3453 a
= ira_allocnos
[i
];
3454 data
= ALLOCNO_COLOR_DATA (a
);
3455 data
->conflict_allocno_hard_prefs
= 0;
3456 for (pref
= ALLOCNO_PREFS (a
); pref
!= NULL
; pref
= next_pref
)
3458 next_pref
= pref
->next_pref
;
3459 if (! ira_hard_reg_in_set_p (pref
->hard_regno
,
3461 data
->profitable_hard_regs
))
3462 ira_remove_pref (pref
);
3466 if (flag_ira_algorithm
== IRA_ALGORITHM_PRIORITY
)
3469 EXECUTE_IF_SET_IN_BITMAP (coloring_allocno_bitmap
, 0, i
, bi
)
3471 a
= ira_allocnos
[i
];
3472 if (ALLOCNO_CLASS (a
) == NO_REGS
)
3474 ALLOCNO_HARD_REGNO (a
) = -1;
3475 ALLOCNO_ASSIGNED_P (a
) = true;
3476 ira_assert (ALLOCNO_UPDATED_HARD_REG_COSTS (a
) == NULL
);
3477 ira_assert (ALLOCNO_UPDATED_CONFLICT_HARD_REG_COSTS (a
) == NULL
);
3478 if (internal_flag_ira_verbose
> 3 && ira_dump_file
!= NULL
)
3480 fprintf (ira_dump_file
, " Spill");
3481 ira_print_expanded_allocno (a
);
3482 fprintf (ira_dump_file
, "\n");
3486 sorted_allocnos
[n
++] = a
;
3490 setup_allocno_priorities (sorted_allocnos
, n
);
3491 qsort (sorted_allocnos
, n
, sizeof (ira_allocno_t
),
3492 allocno_priority_compare_func
);
3493 for (i
= 0; i
< n
; i
++)
3495 a
= sorted_allocnos
[i
];
3496 if (internal_flag_ira_verbose
> 3 && ira_dump_file
!= NULL
)
3498 fprintf (ira_dump_file
, " ");
3499 ira_print_expanded_allocno (a
);
3500 fprintf (ira_dump_file
, " -- ");
3502 if (assign_hard_reg (a
, false))
3504 if (internal_flag_ira_verbose
> 3 && ira_dump_file
!= NULL
)
3505 fprintf (ira_dump_file
, "assign hard reg %d\n",
3506 ALLOCNO_HARD_REGNO (a
));
3510 if (internal_flag_ira_verbose
> 3 && ira_dump_file
!= NULL
)
3511 fprintf (ira_dump_file
, "assign memory\n");
3518 form_allocno_hard_regs_nodes_forest ();
3519 if (internal_flag_ira_verbose
> 2 && ira_dump_file
!= NULL
)
3520 print_hard_regs_forest (ira_dump_file
);
3521 EXECUTE_IF_SET_IN_BITMAP (coloring_allocno_bitmap
, 0, i
, bi
)
3523 a
= ira_allocnos
[i
];
3524 if (ALLOCNO_CLASS (a
) != NO_REGS
&& ! empty_profitable_hard_regs (a
))
3526 ALLOCNO_COLOR_DATA (a
)->in_graph_p
= true;
3527 update_conflict_allocno_hard_prefs (a
);
3531 ALLOCNO_HARD_REGNO (a
) = -1;
3532 ALLOCNO_ASSIGNED_P (a
) = true;
3533 /* We don't need updated costs anymore. */
3534 ira_free_allocno_updated_costs (a
);
3535 if (internal_flag_ira_verbose
> 3 && ira_dump_file
!= NULL
)
3537 fprintf (ira_dump_file
, " Spill");
3538 ira_print_expanded_allocno (a
);
3539 fprintf (ira_dump_file
, "\n");
3543 /* Put the allocnos into the corresponding buckets. */
3544 colorable_allocno_bucket
= NULL
;
3545 uncolorable_allocno_bucket
= NULL
;
3546 EXECUTE_IF_SET_IN_BITMAP (coloring_allocno_bitmap
, 0, i
, bi
)
3548 a
= ira_allocnos
[i
];
3549 if (ALLOCNO_COLOR_DATA (a
)->in_graph_p
)
3550 put_allocno_into_bucket (a
);
3552 push_allocnos_to_stack ();
3553 pop_allocnos_from_stack ();
3554 finish_allocno_hard_regs_nodes_forest ();
3556 improve_allocation ();
3561 /* Output information about the loop given by its LOOP_TREE_NODE. */
3563 print_loop_title (ira_loop_tree_node_t loop_tree_node
)
3567 ira_loop_tree_node_t subloop_node
, dest_loop_node
;
3571 if (loop_tree_node
->parent
== NULL
)
3572 fprintf (ira_dump_file
,
3573 "\n Loop 0 (parent -1, header bb%d, depth 0)\n bbs:",
3577 ira_assert (current_loops
!= NULL
&& loop_tree_node
->loop
!= NULL
);
3578 fprintf (ira_dump_file
,
3579 "\n Loop %d (parent %d, header bb%d, depth %d)\n bbs:",
3580 loop_tree_node
->loop_num
, loop_tree_node
->parent
->loop_num
,
3581 loop_tree_node
->loop
->header
->index
,
3582 loop_depth (loop_tree_node
->loop
));
3584 for (subloop_node
= loop_tree_node
->children
;
3585 subloop_node
!= NULL
;
3586 subloop_node
= subloop_node
->next
)
3587 if (subloop_node
->bb
!= NULL
)
3589 fprintf (ira_dump_file
, " %d", subloop_node
->bb
->index
);
3590 FOR_EACH_EDGE (e
, ei
, subloop_node
->bb
->succs
)
3591 if (e
->dest
!= EXIT_BLOCK_PTR_FOR_FN (cfun
)
3592 && ((dest_loop_node
= IRA_BB_NODE (e
->dest
)->parent
)
3594 fprintf (ira_dump_file
, "(->%d:l%d)",
3595 e
->dest
->index
, dest_loop_node
->loop_num
);
3597 fprintf (ira_dump_file
, "\n all:");
3598 EXECUTE_IF_SET_IN_BITMAP (loop_tree_node
->all_allocnos
, 0, j
, bi
)
3599 fprintf (ira_dump_file
, " %dr%d", j
, ALLOCNO_REGNO (ira_allocnos
[j
]));
3600 fprintf (ira_dump_file
, "\n modified regnos:");
3601 EXECUTE_IF_SET_IN_BITMAP (loop_tree_node
->modified_regnos
, 0, j
, bi
)
3602 fprintf (ira_dump_file
, " %d", j
);
3603 fprintf (ira_dump_file
, "\n border:");
3604 EXECUTE_IF_SET_IN_BITMAP (loop_tree_node
->border_allocnos
, 0, j
, bi
)
3605 fprintf (ira_dump_file
, " %dr%d", j
, ALLOCNO_REGNO (ira_allocnos
[j
]));
3606 fprintf (ira_dump_file
, "\n Pressure:");
3607 for (j
= 0; (int) j
< ira_pressure_classes_num
; j
++)
3609 enum reg_class pclass
;
3611 pclass
= ira_pressure_classes
[j
];
3612 if (loop_tree_node
->reg_pressure
[pclass
] == 0)
3614 fprintf (ira_dump_file
, " %s=%d", reg_class_names
[pclass
],
3615 loop_tree_node
->reg_pressure
[pclass
]);
3617 fprintf (ira_dump_file
, "\n");
3620 /* Color the allocnos inside loop (in the extreme case it can be all
3621 of the function) given the corresponding LOOP_TREE_NODE. The
3622 function is called for each loop during top-down traverse of the
3625 color_pass (ira_loop_tree_node_t loop_tree_node
)
3627 int regno
, hard_regno
, index
= -1, n
;
3632 enum reg_class rclass
, aclass
;
3633 ira_allocno_t a
, subloop_allocno
;
3634 ira_loop_tree_node_t subloop_node
;
3636 ira_assert (loop_tree_node
->bb
== NULL
);
3637 if (internal_flag_ira_verbose
> 1 && ira_dump_file
!= NULL
)
3638 print_loop_title (loop_tree_node
);
3640 bitmap_copy (coloring_allocno_bitmap
, loop_tree_node
->all_allocnos
);
3641 bitmap_copy (consideration_allocno_bitmap
, coloring_allocno_bitmap
);
3643 EXECUTE_IF_SET_IN_BITMAP (consideration_allocno_bitmap
, 0, j
, bi
)
3645 a
= ira_allocnos
[j
];
3647 if (! ALLOCNO_ASSIGNED_P (a
))
3649 bitmap_clear_bit (coloring_allocno_bitmap
, ALLOCNO_NUM (a
));
3652 = (allocno_color_data_t
) ira_allocate (sizeof (struct allocno_color_data
)
3654 memset (allocno_color_data
, 0, sizeof (struct allocno_color_data
) * n
);
3655 curr_allocno_process
= 0;
3657 EXECUTE_IF_SET_IN_BITMAP (consideration_allocno_bitmap
, 0, j
, bi
)
3659 a
= ira_allocnos
[j
];
3660 ALLOCNO_ADD_DATA (a
) = allocno_color_data
+ n
;
3663 init_allocno_threads ();
3664 /* Color all mentioned allocnos including transparent ones. */
3666 /* Process caps. They are processed just once. */
3667 if (flag_ira_region
== IRA_REGION_MIXED
3668 || flag_ira_region
== IRA_REGION_ALL
)
3669 EXECUTE_IF_SET_IN_BITMAP (loop_tree_node
->all_allocnos
, 0, j
, bi
)
3671 a
= ira_allocnos
[j
];
3672 if (ALLOCNO_CAP_MEMBER (a
) == NULL
)
3674 /* Remove from processing in the next loop. */
3675 bitmap_clear_bit (consideration_allocno_bitmap
, j
);
3676 rclass
= ALLOCNO_CLASS (a
);
3677 subloop_allocno
= ALLOCNO_CAP_MEMBER (a
);
3678 subloop_node
= ALLOCNO_LOOP_TREE_NODE (subloop_allocno
);
3679 if (ira_single_region_allocno_p (a
, subloop_allocno
))
3681 mode
= ALLOCNO_MODE (a
);
3682 hard_regno
= ALLOCNO_HARD_REGNO (a
);
3683 if (hard_regno
>= 0)
3685 index
= ira_class_hard_reg_index
[rclass
][hard_regno
];
3686 ira_assert (index
>= 0);
3688 regno
= ALLOCNO_REGNO (a
);
3689 ira_assert (!ALLOCNO_ASSIGNED_P (subloop_allocno
));
3690 ALLOCNO_HARD_REGNO (subloop_allocno
) = hard_regno
;
3691 ALLOCNO_ASSIGNED_P (subloop_allocno
) = true;
3692 if (hard_regno
>= 0)
3693 update_costs_from_copies (subloop_allocno
, true, true);
3694 /* We don't need updated costs anymore. */
3695 ira_free_allocno_updated_costs (subloop_allocno
);
3698 /* Update costs of the corresponding allocnos (not caps) in the
3700 for (subloop_node
= loop_tree_node
->subloops
;
3701 subloop_node
!= NULL
;
3702 subloop_node
= subloop_node
->subloop_next
)
3704 ira_assert (subloop_node
->bb
== NULL
);
3705 EXECUTE_IF_SET_IN_BITMAP (consideration_allocno_bitmap
, 0, j
, bi
)
3707 a
= ira_allocnos
[j
];
3708 ira_assert (ALLOCNO_CAP_MEMBER (a
) == NULL
);
3709 mode
= ALLOCNO_MODE (a
);
3710 rclass
= ALLOCNO_CLASS (a
);
3711 hard_regno
= ALLOCNO_HARD_REGNO (a
);
3712 /* Use hard register class here. ??? */
3713 if (hard_regno
>= 0)
3715 index
= ira_class_hard_reg_index
[rclass
][hard_regno
];
3716 ira_assert (index
>= 0);
3718 regno
= ALLOCNO_REGNO (a
);
3719 /* ??? conflict costs */
3720 subloop_allocno
= subloop_node
->regno_allocno_map
[regno
];
3721 if (subloop_allocno
== NULL
3722 || ALLOCNO_CAP (subloop_allocno
) != NULL
)
3724 ira_assert (ALLOCNO_CLASS (subloop_allocno
) == rclass
);
3725 ira_assert (bitmap_bit_p (subloop_node
->all_allocnos
,
3726 ALLOCNO_NUM (subloop_allocno
)));
3727 if (ira_single_region_allocno_p (a
, subloop_allocno
)
3728 || !ira_subloop_allocnos_can_differ_p (a
, hard_regno
>= 0,
3731 gcc_assert (!ALLOCNO_MIGHT_CONFLICT_WITH_PARENT_P
3733 if (! ALLOCNO_ASSIGNED_P (subloop_allocno
))
3735 ALLOCNO_HARD_REGNO (subloop_allocno
) = hard_regno
;
3736 ALLOCNO_ASSIGNED_P (subloop_allocno
) = true;
3737 if (hard_regno
>= 0)
3738 update_costs_from_copies (subloop_allocno
, true, true);
3739 /* We don't need updated costs anymore. */
3740 ira_free_allocno_updated_costs (subloop_allocno
);
3743 else if (hard_regno
< 0)
3745 /* If we allocate a register to SUBLOOP_ALLOCNO, we'll need
3746 to load the register on entry to the subloop and store
3747 the register back on exit from the subloop. This incurs
3748 a fixed cost for all registers. Since UPDATED_MEMORY_COST
3749 is (and should only be) used relative to the register costs
3750 for the same allocno, we can subtract this shared register
3751 cost from the memory cost. */
3752 ira_loop_border_costs
border_costs (subloop_allocno
);
3753 ALLOCNO_UPDATED_MEMORY_COST (subloop_allocno
)
3754 -= border_costs
.spill_outside_loop_cost ();
3758 ira_loop_border_costs
border_costs (subloop_allocno
);
3759 aclass
= ALLOCNO_CLASS (subloop_allocno
);
3760 ira_init_register_move_cost_if_necessary (mode
);
3761 cost
= border_costs
.move_between_loops_cost ();
3762 ira_allocate_and_set_or_copy_costs
3763 (&ALLOCNO_UPDATED_HARD_REG_COSTS (subloop_allocno
), aclass
,
3764 ALLOCNO_UPDATED_CLASS_COST (subloop_allocno
),
3765 ALLOCNO_HARD_REG_COSTS (subloop_allocno
));
3766 ira_allocate_and_set_or_copy_costs
3767 (&ALLOCNO_UPDATED_CONFLICT_HARD_REG_COSTS (subloop_allocno
),
3768 aclass
, 0, ALLOCNO_CONFLICT_HARD_REG_COSTS (subloop_allocno
));
3769 ALLOCNO_UPDATED_HARD_REG_COSTS (subloop_allocno
)[index
] -= cost
;
3770 ALLOCNO_UPDATED_CONFLICT_HARD_REG_COSTS (subloop_allocno
)[index
]
3772 if (ALLOCNO_UPDATED_CLASS_COST (subloop_allocno
)
3773 > ALLOCNO_UPDATED_HARD_REG_COSTS (subloop_allocno
)[index
])
3774 ALLOCNO_UPDATED_CLASS_COST (subloop_allocno
)
3775 = ALLOCNO_UPDATED_HARD_REG_COSTS (subloop_allocno
)[index
];
3776 /* If we spill SUBLOOP_ALLOCNO, we'll need to store HARD_REGNO
3777 on entry to the subloop and restore HARD_REGNO on exit from
3779 ALLOCNO_UPDATED_MEMORY_COST (subloop_allocno
)
3780 += border_costs
.spill_inside_loop_cost ();
3784 ira_free (allocno_color_data
);
3785 EXECUTE_IF_SET_IN_BITMAP (consideration_allocno_bitmap
, 0, j
, bi
)
3787 a
= ira_allocnos
[j
];
3788 ALLOCNO_ADD_DATA (a
) = NULL
;
3792 /* Initialize the common data for coloring and calls functions to do
3793 Chaitin-Briggs and regional coloring. */
3797 coloring_allocno_bitmap
= ira_allocate_bitmap ();
3798 if (internal_flag_ira_verbose
> 0 && ira_dump_file
!= NULL
)
3799 fprintf (ira_dump_file
, "\n**** Allocnos coloring:\n\n");
3801 ira_traverse_loop_tree (false, ira_loop_tree_root
, color_pass
, NULL
);
3803 if (internal_flag_ira_verbose
> 1 && ira_dump_file
!= NULL
)
3804 ira_print_disposition (ira_dump_file
);
3806 ira_free_bitmap (coloring_allocno_bitmap
);
3811 /* Move spill/restore code, which are to be generated in ira-emit.cc,
3812 to less frequent points (if it is profitable) by reassigning some
3813 allocnos (in loop with subloops containing in another loop) to
3814 memory which results in longer live-range where the corresponding
3815 pseudo-registers will be in memory. */
3817 move_spill_restore (void)
3819 int cost
, regno
, hard_regno
, hard_regno2
, index
;
3822 enum reg_class rclass
;
3823 ira_allocno_t a
, parent_allocno
, subloop_allocno
;
3824 ira_loop_tree_node_t parent
, loop_node
, subloop_node
;
3825 ira_allocno_iterator ai
;
3830 if (internal_flag_ira_verbose
> 0 && ira_dump_file
!= NULL
)
3831 fprintf (ira_dump_file
, "New iteration of spill/restore move\n");
3832 FOR_EACH_ALLOCNO (a
, ai
)
3834 regno
= ALLOCNO_REGNO (a
);
3835 loop_node
= ALLOCNO_LOOP_TREE_NODE (a
);
3836 if (ALLOCNO_CAP_MEMBER (a
) != NULL
3837 || ALLOCNO_CAP (a
) != NULL
3838 || (hard_regno
= ALLOCNO_HARD_REGNO (a
)) < 0
3839 || loop_node
->children
== NULL
3840 /* don't do the optimization because it can create
3841 copies and the reload pass can spill the allocno set
3842 by copy although the allocno will not get memory
3844 || ira_equiv_no_lvalue_p (regno
)
3845 || !bitmap_bit_p (loop_node
->border_allocnos
, ALLOCNO_NUM (a
))
3846 /* Do not spill static chain pointer pseudo when
3847 non-local goto is used. */
3848 || non_spilled_static_chain_regno_p (regno
))
3850 mode
= ALLOCNO_MODE (a
);
3851 rclass
= ALLOCNO_CLASS (a
);
3852 index
= ira_class_hard_reg_index
[rclass
][hard_regno
];
3853 ira_assert (index
>= 0);
3854 cost
= (ALLOCNO_MEMORY_COST (a
)
3855 - (ALLOCNO_HARD_REG_COSTS (a
) == NULL
3856 ? ALLOCNO_CLASS_COST (a
)
3857 : ALLOCNO_HARD_REG_COSTS (a
)[index
]));
3858 ira_init_register_move_cost_if_necessary (mode
);
3859 for (subloop_node
= loop_node
->subloops
;
3860 subloop_node
!= NULL
;
3861 subloop_node
= subloop_node
->subloop_next
)
3863 ira_assert (subloop_node
->bb
== NULL
);
3864 subloop_allocno
= subloop_node
->regno_allocno_map
[regno
];
3865 if (subloop_allocno
== NULL
)
3867 ira_assert (rclass
== ALLOCNO_CLASS (subloop_allocno
));
3868 ira_loop_border_costs
border_costs (subloop_allocno
);
3870 /* We have accumulated cost. To get the real cost of
3871 allocno usage in the loop we should subtract the costs
3872 added by propagate_allocno_info for the subloop allocnos. */
3874 = (ALLOCNO_HARD_REG_COSTS (subloop_allocno
) == NULL
3875 ? ALLOCNO_CLASS_COST (subloop_allocno
)
3876 : ALLOCNO_HARD_REG_COSTS (subloop_allocno
)[index
]);
3879 = (border_costs
.spill_inside_loop_cost ()
3880 + ALLOCNO_MEMORY_COST (subloop_allocno
));
3882 /* If HARD_REGNO conflicts with SUBLOOP_A then
3883 propagate_allocno_info will have propagated
3884 the cost of spilling HARD_REGNO in SUBLOOP_NODE.
3885 (ira_subloop_allocnos_can_differ_p must be true
3886 in that case.) If HARD_REGNO is a caller-saved
3887 register, we might have modelled it in the same way.
3889 Otherwise, SPILL_COST acted as a cap on the propagated
3890 register cost, in cases where the allocations can differ. */
3891 auto conflicts
= ira_total_conflict_hard_regs (subloop_allocno
);
3892 if (TEST_HARD_REG_BIT (conflicts
, hard_regno
)
3893 || (ira_need_caller_save_p (subloop_allocno
, hard_regno
)
3894 && ira_caller_save_loop_spill_p (a
, subloop_allocno
,
3896 reg_cost
= spill_cost
;
3897 else if (ira_subloop_allocnos_can_differ_p (a
))
3898 reg_cost
= MIN (reg_cost
, spill_cost
);
3900 cost
-= ALLOCNO_MEMORY_COST (subloop_allocno
) - reg_cost
;
3902 if ((hard_regno2
= ALLOCNO_HARD_REGNO (subloop_allocno
)) < 0)
3903 /* The register was spilled in the subloop. If we spill
3904 it in the outer loop too then we'll no longer need to
3905 save the register on entry to the subloop and restore
3906 the register on exit from the subloop. */
3907 cost
-= border_costs
.spill_inside_loop_cost ();
3910 /* The register was also allocated in the subloop. If we
3911 spill it in the outer loop then we'll need to load the
3912 register on entry to the subloop and store the register
3913 back on exit from the subloop. */
3914 cost
+= border_costs
.spill_outside_loop_cost ();
3915 if (hard_regno2
!= hard_regno
)
3916 cost
-= border_costs
.move_between_loops_cost ();
3919 if ((parent
= loop_node
->parent
) != NULL
3920 && (parent_allocno
= parent
->regno_allocno_map
[regno
]) != NULL
)
3922 ira_assert (rclass
== ALLOCNO_CLASS (parent_allocno
));
3923 ira_loop_border_costs
border_costs (a
);
3924 if ((hard_regno2
= ALLOCNO_HARD_REGNO (parent_allocno
)) < 0)
3925 /* The register was spilled in the parent loop. If we spill
3926 it in this loop too then we'll no longer need to load the
3927 register on entry to this loop and save the register back
3928 on exit from this loop. */
3929 cost
-= border_costs
.spill_outside_loop_cost ();
3932 /* The register was also allocated in the parent loop.
3933 If we spill it in this loop then we'll need to save
3934 the register on entry to this loop and restore the
3935 register on exit from this loop. */
3936 cost
+= border_costs
.spill_inside_loop_cost ();
3937 if (hard_regno2
!= hard_regno
)
3938 cost
-= border_costs
.move_between_loops_cost ();
3943 ALLOCNO_HARD_REGNO (a
) = -1;
3944 if (internal_flag_ira_verbose
> 3 && ira_dump_file
!= NULL
)
3948 " Moving spill/restore for a%dr%d up from loop %d",
3949 ALLOCNO_NUM (a
), regno
, loop_node
->loop_num
);
3950 fprintf (ira_dump_file
, " - profit %d\n", -cost
);
3962 /* Update current hard reg costs and current conflict hard reg costs
3963 for allocno A. It is done by processing its copies containing
3964 other allocnos already assigned. */
3966 update_curr_costs (ira_allocno_t a
)
3968 int i
, hard_regno
, cost
;
3970 enum reg_class aclass
, rclass
;
3971 ira_allocno_t another_a
;
3972 ira_copy_t cp
, next_cp
;
3974 ira_free_allocno_updated_costs (a
);
3975 ira_assert (! ALLOCNO_ASSIGNED_P (a
));
3976 aclass
= ALLOCNO_CLASS (a
);
3977 if (aclass
== NO_REGS
)
3979 mode
= ALLOCNO_MODE (a
);
3980 ira_init_register_move_cost_if_necessary (mode
);
3981 for (cp
= ALLOCNO_COPIES (a
); cp
!= NULL
; cp
= next_cp
)
3985 next_cp
= cp
->next_first_allocno_copy
;
3986 another_a
= cp
->second
;
3988 else if (cp
->second
== a
)
3990 next_cp
= cp
->next_second_allocno_copy
;
3991 another_a
= cp
->first
;
3995 if (! ira_reg_classes_intersect_p
[aclass
][ALLOCNO_CLASS (another_a
)]
3996 || ! ALLOCNO_ASSIGNED_P (another_a
)
3997 || (hard_regno
= ALLOCNO_HARD_REGNO (another_a
)) < 0)
3999 rclass
= REGNO_REG_CLASS (hard_regno
);
4000 i
= ira_class_hard_reg_index
[aclass
][hard_regno
];
4003 cost
= (cp
->first
== a
4004 ? ira_register_move_cost
[mode
][rclass
][aclass
]
4005 : ira_register_move_cost
[mode
][aclass
][rclass
]);
4006 ira_allocate_and_set_or_copy_costs
4007 (&ALLOCNO_UPDATED_HARD_REG_COSTS (a
), aclass
, ALLOCNO_CLASS_COST (a
),
4008 ALLOCNO_HARD_REG_COSTS (a
));
4009 ira_allocate_and_set_or_copy_costs
4010 (&ALLOCNO_UPDATED_CONFLICT_HARD_REG_COSTS (a
),
4011 aclass
, 0, ALLOCNO_CONFLICT_HARD_REG_COSTS (a
));
4012 ALLOCNO_UPDATED_HARD_REG_COSTS (a
)[i
] -= cp
->freq
* cost
;
4013 ALLOCNO_UPDATED_CONFLICT_HARD_REG_COSTS (a
)[i
] -= cp
->freq
* cost
;
4017 /* Try to assign hard registers to the unassigned allocnos and
4018 allocnos conflicting with them or conflicting with allocnos whose
4019 regno >= START_REGNO. The function is called after ira_flattening,
4020 so more allocnos (including ones created in ira-emit.cc) will have a
4021 chance to get a hard register. We use simple assignment algorithm
4022 based on priorities. */
4024 ira_reassign_conflict_allocnos (int start_regno
)
4026 int i
, allocnos_to_color_num
;
4028 enum reg_class aclass
;
4029 bitmap allocnos_to_color
;
4030 ira_allocno_iterator ai
;
4032 allocnos_to_color
= ira_allocate_bitmap ();
4033 allocnos_to_color_num
= 0;
4034 FOR_EACH_ALLOCNO (a
, ai
)
4036 int n
= ALLOCNO_NUM_OBJECTS (a
);
4038 if (! ALLOCNO_ASSIGNED_P (a
)
4039 && ! bitmap_bit_p (allocnos_to_color
, ALLOCNO_NUM (a
)))
4041 if (ALLOCNO_CLASS (a
) != NO_REGS
)
4042 sorted_allocnos
[allocnos_to_color_num
++] = a
;
4045 ALLOCNO_ASSIGNED_P (a
) = true;
4046 ALLOCNO_HARD_REGNO (a
) = -1;
4047 ira_assert (ALLOCNO_UPDATED_HARD_REG_COSTS (a
) == NULL
);
4048 ira_assert (ALLOCNO_UPDATED_CONFLICT_HARD_REG_COSTS (a
) == NULL
);
4050 bitmap_set_bit (allocnos_to_color
, ALLOCNO_NUM (a
));
4052 if (ALLOCNO_REGNO (a
) < start_regno
4053 || (aclass
= ALLOCNO_CLASS (a
)) == NO_REGS
)
4055 for (i
= 0; i
< n
; i
++)
4057 ira_object_t obj
= ALLOCNO_OBJECT (a
, i
);
4058 ira_object_t conflict_obj
;
4059 ira_object_conflict_iterator oci
;
4061 FOR_EACH_OBJECT_CONFLICT (obj
, conflict_obj
, oci
)
4063 ira_allocno_t conflict_a
= OBJECT_ALLOCNO (conflict_obj
);
4065 ira_assert (ira_reg_classes_intersect_p
4066 [aclass
][ALLOCNO_CLASS (conflict_a
)]);
4067 if (!bitmap_set_bit (allocnos_to_color
, ALLOCNO_NUM (conflict_a
)))
4069 sorted_allocnos
[allocnos_to_color_num
++] = conflict_a
;
4073 ira_free_bitmap (allocnos_to_color
);
4074 if (allocnos_to_color_num
> 1)
4076 setup_allocno_priorities (sorted_allocnos
, allocnos_to_color_num
);
4077 qsort (sorted_allocnos
, allocnos_to_color_num
, sizeof (ira_allocno_t
),
4078 allocno_priority_compare_func
);
4080 for (i
= 0; i
< allocnos_to_color_num
; i
++)
4082 a
= sorted_allocnos
[i
];
4083 ALLOCNO_ASSIGNED_P (a
) = false;
4084 update_curr_costs (a
);
4086 for (i
= 0; i
< allocnos_to_color_num
; i
++)
4088 a
= sorted_allocnos
[i
];
4089 if (assign_hard_reg (a
, true))
4091 if (internal_flag_ira_verbose
> 3 && ira_dump_file
!= NULL
)
4094 " Secondary allocation: assign hard reg %d to reg %d\n",
4095 ALLOCNO_HARD_REGNO (a
), ALLOCNO_REGNO (a
));
4102 /* This page contains functions used to find conflicts using allocno
4105 #ifdef ENABLE_IRA_CHECKING
4107 /* Return TRUE if live ranges of pseudo-registers REGNO1 and REGNO2
4108 intersect. This should be used when there is only one region.
4109 Currently this is used during reload. */
4111 conflict_by_live_ranges_p (int regno1
, int regno2
)
4113 ira_allocno_t a1
, a2
;
4115 ira_assert (regno1
>= FIRST_PSEUDO_REGISTER
4116 && regno2
>= FIRST_PSEUDO_REGISTER
);
4117 /* Reg info calculated by dataflow infrastructure can be different
4118 from one calculated by regclass. */
4119 if ((a1
= ira_loop_tree_root
->regno_allocno_map
[regno1
]) == NULL
4120 || (a2
= ira_loop_tree_root
->regno_allocno_map
[regno2
]) == NULL
)
4122 return allocnos_conflict_by_live_ranges_p (a1
, a2
);
4129 /* This page contains code to coalesce memory stack slots used by
4130 spilled allocnos. This results in smaller stack frame, better data
4131 locality, and in smaller code for some architectures like
4132 x86/x86_64 where insn size depends on address displacement value.
4133 On the other hand, it can worsen insn scheduling after the RA but
4134 in practice it is less important than smaller stack frames. */
4136 /* TRUE if we coalesced some allocnos. In other words, if we got
4137 loops formed by members first_coalesced_allocno and
4138 next_coalesced_allocno containing more one allocno. */
4139 static bool allocno_coalesced_p
;
4141 /* Bitmap used to prevent a repeated allocno processing because of
4143 static bitmap processed_coalesced_allocno_bitmap
;
4146 typedef struct coalesce_data
*coalesce_data_t
;
4148 /* To decrease footprint of ira_allocno structure we store all data
4149 needed only for coalescing in the following structure. */
4150 struct coalesce_data
4152 /* Coalesced allocnos form a cyclic list. One allocno given by
4153 FIRST represents all coalesced allocnos. The
4154 list is chained by NEXT. */
4155 ira_allocno_t first
;
4160 /* Container for storing allocno data concerning coalescing. */
4161 static coalesce_data_t allocno_coalesce_data
;
4163 /* Macro to access the data concerning coalescing. */
4164 #define ALLOCNO_COALESCE_DATA(a) ((coalesce_data_t) ALLOCNO_ADD_DATA (a))
4166 /* Merge two sets of coalesced allocnos given correspondingly by
4167 allocnos A1 and A2 (more accurately merging A2 set into A1
4170 merge_allocnos (ira_allocno_t a1
, ira_allocno_t a2
)
4172 ira_allocno_t a
, first
, last
, next
;
4174 first
= ALLOCNO_COALESCE_DATA (a1
)->first
;
4175 a
= ALLOCNO_COALESCE_DATA (a2
)->first
;
4178 for (last
= a2
, a
= ALLOCNO_COALESCE_DATA (a2
)->next
;;
4179 a
= ALLOCNO_COALESCE_DATA (a
)->next
)
4181 ALLOCNO_COALESCE_DATA (a
)->first
= first
;
4186 next
= allocno_coalesce_data
[ALLOCNO_NUM (first
)].next
;
4187 allocno_coalesce_data
[ALLOCNO_NUM (first
)].next
= a2
;
4188 allocno_coalesce_data
[ALLOCNO_NUM (last
)].next
= next
;
4191 /* Return TRUE if there are conflicting allocnos from two sets of
4192 coalesced allocnos given correspondingly by allocnos A1 and A2. We
4193 use live ranges to find conflicts because conflicts are represented
4194 only for allocnos of the same allocno class and during the reload
4195 pass we coalesce allocnos for sharing stack memory slots. */
4197 coalesced_allocno_conflict_p (ira_allocno_t a1
, ira_allocno_t a2
)
4199 ira_allocno_t a
, conflict_a
;
4201 if (allocno_coalesced_p
)
4203 bitmap_clear (processed_coalesced_allocno_bitmap
);
4204 for (a
= ALLOCNO_COALESCE_DATA (a1
)->next
;;
4205 a
= ALLOCNO_COALESCE_DATA (a
)->next
)
4207 bitmap_set_bit (processed_coalesced_allocno_bitmap
, ALLOCNO_NUM (a
));
4212 for (a
= ALLOCNO_COALESCE_DATA (a2
)->next
;;
4213 a
= ALLOCNO_COALESCE_DATA (a
)->next
)
4215 for (conflict_a
= ALLOCNO_COALESCE_DATA (a1
)->next
;;
4216 conflict_a
= ALLOCNO_COALESCE_DATA (conflict_a
)->next
)
4218 if (allocnos_conflict_by_live_ranges_p (a
, conflict_a
))
4220 if (conflict_a
== a1
)
4229 /* The major function for aggressive allocno coalescing. We coalesce
4230 only spilled allocnos. If some allocnos have been coalesced, we
4231 set up flag allocno_coalesced_p. */
4233 coalesce_allocnos (void)
4236 ira_copy_t cp
, next_cp
;
4238 int i
, n
, cp_num
, regno
;
4242 /* Collect copies. */
4243 EXECUTE_IF_SET_IN_BITMAP (coloring_allocno_bitmap
, 0, j
, bi
)
4245 a
= ira_allocnos
[j
];
4246 regno
= ALLOCNO_REGNO (a
);
4247 if (! ALLOCNO_ASSIGNED_P (a
) || ALLOCNO_HARD_REGNO (a
) >= 0
4248 || ira_equiv_no_lvalue_p (regno
))
4250 for (cp
= ALLOCNO_COPIES (a
); cp
!= NULL
; cp
= next_cp
)
4254 next_cp
= cp
->next_first_allocno_copy
;
4255 regno
= ALLOCNO_REGNO (cp
->second
);
4256 /* For priority coloring we coalesce allocnos only with
4257 the same allocno class not with intersected allocno
4258 classes as it were possible. It is done for
4260 if ((cp
->insn
!= NULL
|| cp
->constraint_p
)
4261 && ALLOCNO_ASSIGNED_P (cp
->second
)
4262 && ALLOCNO_HARD_REGNO (cp
->second
) < 0
4263 && ! ira_equiv_no_lvalue_p (regno
))
4264 sorted_copies
[cp_num
++] = cp
;
4266 else if (cp
->second
== a
)
4267 next_cp
= cp
->next_second_allocno_copy
;
4272 qsort (sorted_copies
, cp_num
, sizeof (ira_copy_t
), copy_freq_compare_func
);
4273 /* Coalesced copies, most frequently executed first. */
4274 for (; cp_num
!= 0;)
4276 for (i
= 0; i
< cp_num
; i
++)
4278 cp
= sorted_copies
[i
];
4279 if (! coalesced_allocno_conflict_p (cp
->first
, cp
->second
))
4281 allocno_coalesced_p
= true;
4282 if (internal_flag_ira_verbose
> 3 && ira_dump_file
!= NULL
)
4285 " Coalescing copy %d:a%dr%d-a%dr%d (freq=%d)\n",
4286 cp
->num
, ALLOCNO_NUM (cp
->first
), ALLOCNO_REGNO (cp
->first
),
4287 ALLOCNO_NUM (cp
->second
), ALLOCNO_REGNO (cp
->second
),
4289 merge_allocnos (cp
->first
, cp
->second
);
4294 /* Collect the rest of copies. */
4295 for (n
= 0; i
< cp_num
; i
++)
4297 cp
= sorted_copies
[i
];
4298 if (allocno_coalesce_data
[ALLOCNO_NUM (cp
->first
)].first
4299 != allocno_coalesce_data
[ALLOCNO_NUM (cp
->second
)].first
)
4300 sorted_copies
[n
++] = cp
;
4306 /* Usage cost and order number of coalesced allocno set to which
4307 given pseudo register belongs to. */
4308 static int *regno_coalesced_allocno_cost
;
4309 static int *regno_coalesced_allocno_num
;
4311 /* Sort pseudos according frequencies of coalesced allocno sets they
4312 belong to (putting most frequently ones first), and according to
4313 coalesced allocno set order numbers. */
4315 coalesced_pseudo_reg_freq_compare (const void *v1p
, const void *v2p
)
4317 const int regno1
= *(const int *) v1p
;
4318 const int regno2
= *(const int *) v2p
;
4321 if ((diff
= (regno_coalesced_allocno_cost
[regno2
]
4322 - regno_coalesced_allocno_cost
[regno1
])) != 0)
4324 if ((diff
= (regno_coalesced_allocno_num
[regno1
]
4325 - regno_coalesced_allocno_num
[regno2
])) != 0)
4327 return regno1
- regno2
;
4330 /* Widest width in which each pseudo reg is referred to (via subreg).
4331 It is used for sorting pseudo registers. */
4332 static machine_mode
*regno_max_ref_mode
;
4334 /* Sort pseudos according their slot numbers (putting ones with
4335 smaller numbers first, or last when the frame pointer is not
4338 coalesced_pseudo_reg_slot_compare (const void *v1p
, const void *v2p
)
4340 const int regno1
= *(const int *) v1p
;
4341 const int regno2
= *(const int *) v2p
;
4342 ira_allocno_t a1
= ira_regno_allocno_map
[regno1
];
4343 ira_allocno_t a2
= ira_regno_allocno_map
[regno2
];
4344 int diff
, slot_num1
, slot_num2
;
4345 machine_mode mode1
, mode2
;
4347 if (a1
== NULL
|| ALLOCNO_HARD_REGNO (a1
) >= 0)
4349 if (a2
== NULL
|| ALLOCNO_HARD_REGNO (a2
) >= 0)
4350 return regno1
- regno2
;
4353 else if (a2
== NULL
|| ALLOCNO_HARD_REGNO (a2
) >= 0)
4355 slot_num1
= -ALLOCNO_HARD_REGNO (a1
);
4356 slot_num2
= -ALLOCNO_HARD_REGNO (a2
);
4357 if ((diff
= slot_num1
- slot_num2
) != 0)
4358 return (frame_pointer_needed
4359 || (!FRAME_GROWS_DOWNWARD
) == STACK_GROWS_DOWNWARD
? diff
: -diff
);
4360 mode1
= wider_subreg_mode (PSEUDO_REGNO_MODE (regno1
),
4361 regno_max_ref_mode
[regno1
]);
4362 mode2
= wider_subreg_mode (PSEUDO_REGNO_MODE (regno2
),
4363 regno_max_ref_mode
[regno2
]);
4364 if ((diff
= compare_sizes_for_sort (GET_MODE_SIZE (mode2
),
4365 GET_MODE_SIZE (mode1
))) != 0)
4367 return regno1
- regno2
;
4370 /* Setup REGNO_COALESCED_ALLOCNO_COST and REGNO_COALESCED_ALLOCNO_NUM
4371 for coalesced allocno sets containing allocnos with their regnos
4372 given in array PSEUDO_REGNOS of length N. */
4374 setup_coalesced_allocno_costs_and_nums (int *pseudo_regnos
, int n
)
4376 int i
, num
, regno
, cost
;
4377 ira_allocno_t allocno
, a
;
4379 for (num
= i
= 0; i
< n
; i
++)
4381 regno
= pseudo_regnos
[i
];
4382 allocno
= ira_regno_allocno_map
[regno
];
4383 if (allocno
== NULL
)
4385 regno_coalesced_allocno_cost
[regno
] = 0;
4386 regno_coalesced_allocno_num
[regno
] = ++num
;
4389 if (ALLOCNO_COALESCE_DATA (allocno
)->first
!= allocno
)
4392 for (cost
= 0, a
= ALLOCNO_COALESCE_DATA (allocno
)->next
;;
4393 a
= ALLOCNO_COALESCE_DATA (a
)->next
)
4395 cost
+= ALLOCNO_FREQ (a
);
4399 for (a
= ALLOCNO_COALESCE_DATA (allocno
)->next
;;
4400 a
= ALLOCNO_COALESCE_DATA (a
)->next
)
4402 regno_coalesced_allocno_num
[ALLOCNO_REGNO (a
)] = num
;
4403 regno_coalesced_allocno_cost
[ALLOCNO_REGNO (a
)] = cost
;
4410 /* Collect spilled allocnos representing coalesced allocno sets (the
4411 first coalesced allocno). The collected allocnos are returned
4412 through array SPILLED_COALESCED_ALLOCNOS. The function returns the
4413 number of the collected allocnos. The allocnos are given by their
4414 regnos in array PSEUDO_REGNOS of length N. */
4416 collect_spilled_coalesced_allocnos (int *pseudo_regnos
, int n
,
4417 ira_allocno_t
*spilled_coalesced_allocnos
)
4420 ira_allocno_t allocno
;
4422 for (num
= i
= 0; i
< n
; i
++)
4424 regno
= pseudo_regnos
[i
];
4425 allocno
= ira_regno_allocno_map
[regno
];
4426 if (allocno
== NULL
|| ALLOCNO_HARD_REGNO (allocno
) >= 0
4427 || ALLOCNO_COALESCE_DATA (allocno
)->first
!= allocno
)
4429 spilled_coalesced_allocnos
[num
++] = allocno
;
4434 /* Array of live ranges of size IRA_ALLOCNOS_NUM. Live range for
4435 given slot contains live ranges of coalesced allocnos assigned to
4437 static live_range_t
*slot_coalesced_allocnos_live_ranges
;
4439 /* Return TRUE if coalesced allocnos represented by ALLOCNO has live
4440 ranges intersected with live ranges of coalesced allocnos assigned
4441 to slot with number N. */
4443 slot_coalesced_allocno_live_ranges_intersect_p (ira_allocno_t allocno
, int n
)
4447 for (a
= ALLOCNO_COALESCE_DATA (allocno
)->next
;;
4448 a
= ALLOCNO_COALESCE_DATA (a
)->next
)
4451 int nr
= ALLOCNO_NUM_OBJECTS (a
);
4452 gcc_assert (ALLOCNO_CAP_MEMBER (a
) == NULL
);
4453 for (i
= 0; i
< nr
; i
++)
4455 ira_object_t obj
= ALLOCNO_OBJECT (a
, i
);
4457 if (ira_live_ranges_intersect_p
4458 (slot_coalesced_allocnos_live_ranges
[n
],
4459 OBJECT_LIVE_RANGES (obj
)))
4468 /* Update live ranges of slot to which coalesced allocnos represented
4469 by ALLOCNO were assigned. */
4471 setup_slot_coalesced_allocno_live_ranges (ira_allocno_t allocno
)
4477 n
= ALLOCNO_COALESCE_DATA (allocno
)->temp
;
4478 for (a
= ALLOCNO_COALESCE_DATA (allocno
)->next
;;
4479 a
= ALLOCNO_COALESCE_DATA (a
)->next
)
4481 int nr
= ALLOCNO_NUM_OBJECTS (a
);
4482 gcc_assert (ALLOCNO_CAP_MEMBER (a
) == NULL
);
4483 for (i
= 0; i
< nr
; i
++)
4485 ira_object_t obj
= ALLOCNO_OBJECT (a
, i
);
4487 r
= ira_copy_live_range_list (OBJECT_LIVE_RANGES (obj
));
4488 slot_coalesced_allocnos_live_ranges
[n
]
4489 = ira_merge_live_ranges
4490 (slot_coalesced_allocnos_live_ranges
[n
], r
);
4497 /* We have coalesced allocnos involving in copies. Coalesce allocnos
4498 further in order to share the same memory stack slot. Allocnos
4499 representing sets of allocnos coalesced before the call are given
4500 in array SPILLED_COALESCED_ALLOCNOS of length NUM. Return TRUE if
4501 some allocnos were coalesced in the function. */
4503 coalesce_spill_slots (ira_allocno_t
*spilled_coalesced_allocnos
, int num
)
4505 int i
, j
, n
, last_coalesced_allocno_num
;
4506 ira_allocno_t allocno
, a
;
4507 bool merged_p
= false;
4508 bitmap set_jump_crosses
= regstat_get_setjmp_crosses ();
4510 slot_coalesced_allocnos_live_ranges
4511 = (live_range_t
*) ira_allocate (sizeof (live_range_t
) * ira_allocnos_num
);
4512 memset (slot_coalesced_allocnos_live_ranges
, 0,
4513 sizeof (live_range_t
) * ira_allocnos_num
);
4514 last_coalesced_allocno_num
= 0;
4515 /* Coalesce non-conflicting spilled allocnos preferring most
4517 for (i
= 0; i
< num
; i
++)
4519 allocno
= spilled_coalesced_allocnos
[i
];
4520 if (ALLOCNO_COALESCE_DATA (allocno
)->first
!= allocno
4521 || bitmap_bit_p (set_jump_crosses
, ALLOCNO_REGNO (allocno
))
4522 || ira_equiv_no_lvalue_p (ALLOCNO_REGNO (allocno
)))
4524 for (j
= 0; j
< i
; j
++)
4526 a
= spilled_coalesced_allocnos
[j
];
4527 n
= ALLOCNO_COALESCE_DATA (a
)->temp
;
4528 if (ALLOCNO_COALESCE_DATA (a
)->first
== a
4529 && ! bitmap_bit_p (set_jump_crosses
, ALLOCNO_REGNO (a
))
4530 && ! ira_equiv_no_lvalue_p (ALLOCNO_REGNO (a
))
4531 && ! slot_coalesced_allocno_live_ranges_intersect_p (allocno
, n
))
4536 /* No coalescing: set up number for coalesced allocnos
4537 represented by ALLOCNO. */
4538 ALLOCNO_COALESCE_DATA (allocno
)->temp
= last_coalesced_allocno_num
++;
4539 setup_slot_coalesced_allocno_live_ranges (allocno
);
4543 allocno_coalesced_p
= true;
4545 if (internal_flag_ira_verbose
> 3 && ira_dump_file
!= NULL
)
4546 fprintf (ira_dump_file
,
4547 " Coalescing spilled allocnos a%dr%d->a%dr%d\n",
4548 ALLOCNO_NUM (allocno
), ALLOCNO_REGNO (allocno
),
4549 ALLOCNO_NUM (a
), ALLOCNO_REGNO (a
));
4550 ALLOCNO_COALESCE_DATA (allocno
)->temp
4551 = ALLOCNO_COALESCE_DATA (a
)->temp
;
4552 setup_slot_coalesced_allocno_live_ranges (allocno
);
4553 merge_allocnos (a
, allocno
);
4554 ira_assert (ALLOCNO_COALESCE_DATA (a
)->first
== a
);
4557 for (i
= 0; i
< ira_allocnos_num
; i
++)
4558 ira_finish_live_range_list (slot_coalesced_allocnos_live_ranges
[i
]);
4559 ira_free (slot_coalesced_allocnos_live_ranges
);
4563 /* Sort pseudo-register numbers in array PSEUDO_REGNOS of length N for
4564 subsequent assigning stack slots to them in the reload pass. To do
4565 this we coalesce spilled allocnos first to decrease the number of
4566 memory-memory move insns. This function is called by the
4569 ira_sort_regnos_for_alter_reg (int *pseudo_regnos
, int n
,
4570 machine_mode
*reg_max_ref_mode
)
4572 int max_regno
= max_reg_num ();
4573 int i
, regno
, num
, slot_num
;
4574 ira_allocno_t allocno
, a
;
4575 ira_allocno_iterator ai
;
4576 ira_allocno_t
*spilled_coalesced_allocnos
;
4578 ira_assert (! ira_use_lra_p
);
4580 /* Set up allocnos can be coalesced. */
4581 coloring_allocno_bitmap
= ira_allocate_bitmap ();
4582 for (i
= 0; i
< n
; i
++)
4584 regno
= pseudo_regnos
[i
];
4585 allocno
= ira_regno_allocno_map
[regno
];
4586 if (allocno
!= NULL
)
4587 bitmap_set_bit (coloring_allocno_bitmap
, ALLOCNO_NUM (allocno
));
4589 allocno_coalesced_p
= false;
4590 processed_coalesced_allocno_bitmap
= ira_allocate_bitmap ();
4591 allocno_coalesce_data
4592 = (coalesce_data_t
) ira_allocate (sizeof (struct coalesce_data
)
4593 * ira_allocnos_num
);
4594 /* Initialize coalesce data for allocnos. */
4595 FOR_EACH_ALLOCNO (a
, ai
)
4597 ALLOCNO_ADD_DATA (a
) = allocno_coalesce_data
+ ALLOCNO_NUM (a
);
4598 ALLOCNO_COALESCE_DATA (a
)->first
= a
;
4599 ALLOCNO_COALESCE_DATA (a
)->next
= a
;
4601 coalesce_allocnos ();
4602 ira_free_bitmap (coloring_allocno_bitmap
);
4603 regno_coalesced_allocno_cost
4604 = (int *) ira_allocate (max_regno
* sizeof (int));
4605 regno_coalesced_allocno_num
4606 = (int *) ira_allocate (max_regno
* sizeof (int));
4607 memset (regno_coalesced_allocno_num
, 0, max_regno
* sizeof (int));
4608 setup_coalesced_allocno_costs_and_nums (pseudo_regnos
, n
);
4609 /* Sort regnos according frequencies of the corresponding coalesced
4611 qsort (pseudo_regnos
, n
, sizeof (int), coalesced_pseudo_reg_freq_compare
);
4612 spilled_coalesced_allocnos
4613 = (ira_allocno_t
*) ira_allocate (ira_allocnos_num
4614 * sizeof (ira_allocno_t
));
4615 /* Collect allocnos representing the spilled coalesced allocno
4617 num
= collect_spilled_coalesced_allocnos (pseudo_regnos
, n
,
4618 spilled_coalesced_allocnos
);
4619 if (flag_ira_share_spill_slots
4620 && coalesce_spill_slots (spilled_coalesced_allocnos
, num
))
4622 setup_coalesced_allocno_costs_and_nums (pseudo_regnos
, n
);
4623 qsort (pseudo_regnos
, n
, sizeof (int),
4624 coalesced_pseudo_reg_freq_compare
);
4625 num
= collect_spilled_coalesced_allocnos (pseudo_regnos
, n
,
4626 spilled_coalesced_allocnos
);
4628 ira_free_bitmap (processed_coalesced_allocno_bitmap
);
4629 allocno_coalesced_p
= false;
4630 /* Assign stack slot numbers to spilled allocno sets, use smaller
4631 numbers for most frequently used coalesced allocnos. -1 is
4632 reserved for dynamic search of stack slots for pseudos spilled by
4635 for (i
= 0; i
< num
; i
++)
4637 allocno
= spilled_coalesced_allocnos
[i
];
4638 if (ALLOCNO_COALESCE_DATA (allocno
)->first
!= allocno
4639 || ALLOCNO_HARD_REGNO (allocno
) >= 0
4640 || ira_equiv_no_lvalue_p (ALLOCNO_REGNO (allocno
)))
4642 if (internal_flag_ira_verbose
> 3 && ira_dump_file
!= NULL
)
4643 fprintf (ira_dump_file
, " Slot %d (freq,size):", slot_num
);
4645 for (a
= ALLOCNO_COALESCE_DATA (allocno
)->next
;;
4646 a
= ALLOCNO_COALESCE_DATA (a
)->next
)
4648 ira_assert (ALLOCNO_HARD_REGNO (a
) < 0);
4649 ALLOCNO_HARD_REGNO (a
) = -slot_num
;
4650 if (internal_flag_ira_verbose
> 3 && ira_dump_file
!= NULL
)
4652 machine_mode mode
= wider_subreg_mode
4653 (PSEUDO_REGNO_MODE (ALLOCNO_REGNO (a
)),
4654 reg_max_ref_mode
[ALLOCNO_REGNO (a
)]);
4655 fprintf (ira_dump_file
, " a%dr%d(%d,",
4656 ALLOCNO_NUM (a
), ALLOCNO_REGNO (a
), ALLOCNO_FREQ (a
));
4657 print_dec (GET_MODE_SIZE (mode
), ira_dump_file
, SIGNED
);
4658 fprintf (ira_dump_file
, ")\n");
4664 if (internal_flag_ira_verbose
> 3 && ira_dump_file
!= NULL
)
4665 fprintf (ira_dump_file
, "\n");
4667 ira_spilled_reg_stack_slots_num
= slot_num
- 1;
4668 ira_free (spilled_coalesced_allocnos
);
4669 /* Sort regnos according the slot numbers. */
4670 regno_max_ref_mode
= reg_max_ref_mode
;
4671 qsort (pseudo_regnos
, n
, sizeof (int), coalesced_pseudo_reg_slot_compare
);
4672 FOR_EACH_ALLOCNO (a
, ai
)
4673 ALLOCNO_ADD_DATA (a
) = NULL
;
4674 ira_free (allocno_coalesce_data
);
4675 ira_free (regno_coalesced_allocno_num
);
4676 ira_free (regno_coalesced_allocno_cost
);
4681 /* This page contains code used by the reload pass to improve the
4684 /* The function is called from reload to mark changes in the
4685 allocation of REGNO made by the reload. Remember that reg_renumber
4686 reflects the change result. */
4688 ira_mark_allocation_change (int regno
)
4690 ira_allocno_t a
= ira_regno_allocno_map
[regno
];
4691 int old_hard_regno
, hard_regno
, cost
;
4692 enum reg_class aclass
= ALLOCNO_CLASS (a
);
4694 ira_assert (a
!= NULL
);
4695 hard_regno
= reg_renumber
[regno
];
4696 if ((old_hard_regno
= ALLOCNO_HARD_REGNO (a
)) == hard_regno
)
4698 if (old_hard_regno
< 0)
4699 cost
= -ALLOCNO_MEMORY_COST (a
);
4702 ira_assert (ira_class_hard_reg_index
[aclass
][old_hard_regno
] >= 0);
4703 cost
= -(ALLOCNO_HARD_REG_COSTS (a
) == NULL
4704 ? ALLOCNO_CLASS_COST (a
)
4705 : ALLOCNO_HARD_REG_COSTS (a
)
4706 [ira_class_hard_reg_index
[aclass
][old_hard_regno
]]);
4707 update_costs_from_copies (a
, false, false);
4709 ira_overall_cost
-= cost
;
4710 ALLOCNO_HARD_REGNO (a
) = hard_regno
;
4713 ALLOCNO_HARD_REGNO (a
) = -1;
4714 cost
+= ALLOCNO_MEMORY_COST (a
);
4716 else if (ira_class_hard_reg_index
[aclass
][hard_regno
] >= 0)
4718 cost
+= (ALLOCNO_HARD_REG_COSTS (a
) == NULL
4719 ? ALLOCNO_CLASS_COST (a
)
4720 : ALLOCNO_HARD_REG_COSTS (a
)
4721 [ira_class_hard_reg_index
[aclass
][hard_regno
]]);
4722 update_costs_from_copies (a
, true, false);
4725 /* Reload changed class of the allocno. */
4727 ira_overall_cost
+= cost
;
4730 /* This function is called when reload deletes memory-memory move. In
4731 this case we marks that the allocation of the corresponding
4732 allocnos should be not changed in future. Otherwise we risk to get
4735 ira_mark_memory_move_deletion (int dst_regno
, int src_regno
)
4737 ira_allocno_t dst
= ira_regno_allocno_map
[dst_regno
];
4738 ira_allocno_t src
= ira_regno_allocno_map
[src_regno
];
4740 ira_assert (dst
!= NULL
&& src
!= NULL
4741 && ALLOCNO_HARD_REGNO (dst
) < 0
4742 && ALLOCNO_HARD_REGNO (src
) < 0);
4743 ALLOCNO_DONT_REASSIGN_P (dst
) = true;
4744 ALLOCNO_DONT_REASSIGN_P (src
) = true;
4747 /* Try to assign a hard register (except for FORBIDDEN_REGS) to
4748 allocno A and return TRUE in the case of success. */
4750 allocno_reload_assign (ira_allocno_t a
, HARD_REG_SET forbidden_regs
)
4753 enum reg_class aclass
;
4754 int regno
= ALLOCNO_REGNO (a
);
4755 HARD_REG_SET saved
[2];
4758 n
= ALLOCNO_NUM_OBJECTS (a
);
4759 for (i
= 0; i
< n
; i
++)
4761 ira_object_t obj
= ALLOCNO_OBJECT (a
, i
);
4762 saved
[i
] = OBJECT_TOTAL_CONFLICT_HARD_REGS (obj
);
4763 OBJECT_TOTAL_CONFLICT_HARD_REGS (obj
) |= forbidden_regs
;
4764 if (! flag_caller_saves
&& ALLOCNO_CALLS_CROSSED_NUM (a
) != 0)
4765 OBJECT_TOTAL_CONFLICT_HARD_REGS (obj
) |= ira_need_caller_save_regs (a
);
4767 ALLOCNO_ASSIGNED_P (a
) = false;
4768 aclass
= ALLOCNO_CLASS (a
);
4769 update_curr_costs (a
);
4770 assign_hard_reg (a
, true);
4771 hard_regno
= ALLOCNO_HARD_REGNO (a
);
4772 reg_renumber
[regno
] = hard_regno
;
4774 ALLOCNO_HARD_REGNO (a
) = -1;
4777 ira_assert (ira_class_hard_reg_index
[aclass
][hard_regno
] >= 0);
4779 -= (ALLOCNO_MEMORY_COST (a
)
4780 - (ALLOCNO_HARD_REG_COSTS (a
) == NULL
4781 ? ALLOCNO_CLASS_COST (a
)
4782 : ALLOCNO_HARD_REG_COSTS (a
)[ira_class_hard_reg_index
4783 [aclass
][hard_regno
]]));
4784 if (ira_need_caller_save_p (a
, hard_regno
))
4786 ira_assert (flag_caller_saves
);
4787 caller_save_needed
= 1;
4791 /* If we found a hard register, modify the RTL for the pseudo
4792 register to show the hard register, and mark the pseudo register
4794 if (reg_renumber
[regno
] >= 0)
4796 if (internal_flag_ira_verbose
> 3 && ira_dump_file
!= NULL
)
4797 fprintf (ira_dump_file
, ": reassign to %d\n", reg_renumber
[regno
]);
4798 SET_REGNO (regno_reg_rtx
[regno
], reg_renumber
[regno
]);
4799 mark_home_live (regno
);
4801 else if (internal_flag_ira_verbose
> 3 && ira_dump_file
!= NULL
)
4802 fprintf (ira_dump_file
, "\n");
4803 for (i
= 0; i
< n
; i
++)
4805 ira_object_t obj
= ALLOCNO_OBJECT (a
, i
);
4806 OBJECT_TOTAL_CONFLICT_HARD_REGS (obj
) = saved
[i
];
4808 return reg_renumber
[regno
] >= 0;
4811 /* Sort pseudos according their usage frequencies (putting most
4812 frequently ones first). */
4814 pseudo_reg_compare (const void *v1p
, const void *v2p
)
4816 int regno1
= *(const int *) v1p
;
4817 int regno2
= *(const int *) v2p
;
4820 if ((diff
= REG_FREQ (regno2
) - REG_FREQ (regno1
)) != 0)
4822 return regno1
- regno2
;
4825 /* Try to allocate hard registers to SPILLED_PSEUDO_REGS (there are
4826 NUM of them) or spilled pseudos conflicting with pseudos in
4827 SPILLED_PSEUDO_REGS. Return TRUE and update SPILLED, if the
4828 allocation has been changed. The function doesn't use
4829 BAD_SPILL_REGS and hard registers in PSEUDO_FORBIDDEN_REGS and
4830 PSEUDO_PREVIOUS_REGS for the corresponding pseudos. The function
4831 is called by the reload pass at the end of each reload
4834 ira_reassign_pseudos (int *spilled_pseudo_regs
, int num
,
4835 HARD_REG_SET bad_spill_regs
,
4836 HARD_REG_SET
*pseudo_forbidden_regs
,
4837 HARD_REG_SET
*pseudo_previous_regs
,
4843 HARD_REG_SET forbidden_regs
;
4844 bitmap temp
= BITMAP_ALLOC (NULL
);
4846 /* Add pseudos which conflict with pseudos already in
4847 SPILLED_PSEUDO_REGS to SPILLED_PSEUDO_REGS. This is preferable
4848 to allocating in two steps as some of the conflicts might have
4849 a higher priority than the pseudos passed in SPILLED_PSEUDO_REGS. */
4850 for (i
= 0; i
< num
; i
++)
4851 bitmap_set_bit (temp
, spilled_pseudo_regs
[i
]);
4853 for (i
= 0, n
= num
; i
< n
; i
++)
4856 int regno
= spilled_pseudo_regs
[i
];
4857 bitmap_set_bit (temp
, regno
);
4859 a
= ira_regno_allocno_map
[regno
];
4860 nr
= ALLOCNO_NUM_OBJECTS (a
);
4861 for (j
= 0; j
< nr
; j
++)
4863 ira_object_t conflict_obj
;
4864 ira_object_t obj
= ALLOCNO_OBJECT (a
, j
);
4865 ira_object_conflict_iterator oci
;
4867 FOR_EACH_OBJECT_CONFLICT (obj
, conflict_obj
, oci
)
4869 ira_allocno_t conflict_a
= OBJECT_ALLOCNO (conflict_obj
);
4870 if (ALLOCNO_HARD_REGNO (conflict_a
) < 0
4871 && ! ALLOCNO_DONT_REASSIGN_P (conflict_a
)
4872 && bitmap_set_bit (temp
, ALLOCNO_REGNO (conflict_a
)))
4874 spilled_pseudo_regs
[num
++] = ALLOCNO_REGNO (conflict_a
);
4875 /* ?!? This seems wrong. */
4876 bitmap_set_bit (consideration_allocno_bitmap
,
4877 ALLOCNO_NUM (conflict_a
));
4884 qsort (spilled_pseudo_regs
, num
, sizeof (int), pseudo_reg_compare
);
4886 /* Try to assign hard registers to pseudos from
4887 SPILLED_PSEUDO_REGS. */
4888 for (i
= 0; i
< num
; i
++)
4890 regno
= spilled_pseudo_regs
[i
];
4891 forbidden_regs
= (bad_spill_regs
4892 | pseudo_forbidden_regs
[regno
]
4893 | pseudo_previous_regs
[regno
]);
4894 gcc_assert (reg_renumber
[regno
] < 0);
4895 a
= ira_regno_allocno_map
[regno
];
4896 ira_mark_allocation_change (regno
);
4897 ira_assert (reg_renumber
[regno
] < 0);
4898 if (internal_flag_ira_verbose
> 3 && ira_dump_file
!= NULL
)
4899 fprintf (ira_dump_file
,
4900 " Try Assign %d(a%d), cost=%d", regno
, ALLOCNO_NUM (a
),
4901 ALLOCNO_MEMORY_COST (a
)
4902 - ALLOCNO_CLASS_COST (a
));
4903 allocno_reload_assign (a
, forbidden_regs
);
4904 if (reg_renumber
[regno
] >= 0)
4906 CLEAR_REGNO_REG_SET (spilled
, regno
);
4914 /* The function is called by reload and returns already allocated
4915 stack slot (if any) for REGNO with given INHERENT_SIZE and
4916 TOTAL_SIZE. In the case of failure to find a slot which can be
4917 used for REGNO, the function returns NULL. */
4919 ira_reuse_stack_slot (int regno
, poly_uint64 inherent_size
,
4920 poly_uint64 total_size
)
4923 int slot_num
, best_slot_num
;
4924 int cost
, best_cost
;
4925 ira_copy_t cp
, next_cp
;
4926 ira_allocno_t another_allocno
, allocno
= ira_regno_allocno_map
[regno
];
4929 class ira_spilled_reg_stack_slot
*slot
= NULL
;
4931 ira_assert (! ira_use_lra_p
);
4933 ira_assert (known_eq (inherent_size
, PSEUDO_REGNO_BYTES (regno
))
4934 && known_le (inherent_size
, total_size
)
4935 && ALLOCNO_HARD_REGNO (allocno
) < 0);
4936 if (! flag_ira_share_spill_slots
)
4938 slot_num
= -ALLOCNO_HARD_REGNO (allocno
) - 2;
4941 slot
= &ira_spilled_reg_stack_slots
[slot_num
];
4946 best_cost
= best_slot_num
= -1;
4948 /* It means that the pseudo was spilled in the reload pass, try
4951 slot_num
< ira_spilled_reg_stack_slots_num
;
4954 slot
= &ira_spilled_reg_stack_slots
[slot_num
];
4955 if (slot
->mem
== NULL_RTX
)
4957 if (maybe_lt (slot
->width
, total_size
)
4958 || maybe_lt (GET_MODE_SIZE (GET_MODE (slot
->mem
)), inherent_size
))
4961 EXECUTE_IF_SET_IN_BITMAP (&slot
->spilled_regs
,
4962 FIRST_PSEUDO_REGISTER
, i
, bi
)
4964 another_allocno
= ira_regno_allocno_map
[i
];
4965 if (allocnos_conflict_by_live_ranges_p (allocno
,
4969 for (cost
= 0, cp
= ALLOCNO_COPIES (allocno
);
4973 if (cp
->first
== allocno
)
4975 next_cp
= cp
->next_first_allocno_copy
;
4976 another_allocno
= cp
->second
;
4978 else if (cp
->second
== allocno
)
4980 next_cp
= cp
->next_second_allocno_copy
;
4981 another_allocno
= cp
->first
;
4985 if (cp
->insn
== NULL_RTX
)
4987 if (bitmap_bit_p (&slot
->spilled_regs
,
4988 ALLOCNO_REGNO (another_allocno
)))
4991 if (cost
> best_cost
)
4994 best_slot_num
= slot_num
;
5001 slot_num
= best_slot_num
;
5002 slot
= &ira_spilled_reg_stack_slots
[slot_num
];
5003 SET_REGNO_REG_SET (&slot
->spilled_regs
, regno
);
5005 ALLOCNO_HARD_REGNO (allocno
) = -slot_num
- 2;
5010 ira_assert (known_ge (slot
->width
, total_size
));
5011 #ifdef ENABLE_IRA_CHECKING
5012 EXECUTE_IF_SET_IN_BITMAP (&slot
->spilled_regs
,
5013 FIRST_PSEUDO_REGISTER
, i
, bi
)
5015 ira_assert (! conflict_by_live_ranges_p (regno
, i
));
5018 SET_REGNO_REG_SET (&slot
->spilled_regs
, regno
);
5019 if (internal_flag_ira_verbose
> 3 && ira_dump_file
)
5021 fprintf (ira_dump_file
, " Assigning %d(freq=%d) slot %d of",
5022 regno
, REG_FREQ (regno
), slot_num
);
5023 EXECUTE_IF_SET_IN_BITMAP (&slot
->spilled_regs
,
5024 FIRST_PSEUDO_REGISTER
, i
, bi
)
5026 if ((unsigned) regno
!= i
)
5027 fprintf (ira_dump_file
, " %d", i
);
5029 fprintf (ira_dump_file
, "\n");
5035 /* This is called by reload every time a new stack slot X with
5036 TOTAL_SIZE was allocated for REGNO. We store this info for
5037 subsequent ira_reuse_stack_slot calls. */
5039 ira_mark_new_stack_slot (rtx x
, int regno
, poly_uint64 total_size
)
5041 class ira_spilled_reg_stack_slot
*slot
;
5043 ira_allocno_t allocno
;
5045 ira_assert (! ira_use_lra_p
);
5047 ira_assert (known_le (PSEUDO_REGNO_BYTES (regno
), total_size
));
5048 allocno
= ira_regno_allocno_map
[regno
];
5049 slot_num
= -ALLOCNO_HARD_REGNO (allocno
) - 2;
5052 slot_num
= ira_spilled_reg_stack_slots_num
++;
5053 ALLOCNO_HARD_REGNO (allocno
) = -slot_num
- 2;
5055 slot
= &ira_spilled_reg_stack_slots
[slot_num
];
5056 INIT_REG_SET (&slot
->spilled_regs
);
5057 SET_REGNO_REG_SET (&slot
->spilled_regs
, regno
);
5059 slot
->width
= total_size
;
5060 if (internal_flag_ira_verbose
> 3 && ira_dump_file
)
5061 fprintf (ira_dump_file
, " Assigning %d(freq=%d) a new slot %d\n",
5062 regno
, REG_FREQ (regno
), slot_num
);
5066 /* Return spill cost for pseudo-registers whose numbers are in array
5067 REGNOS (with a negative number as an end marker) for reload with
5068 given IN and OUT for INSN. Return also number points (through
5069 EXCESS_PRESSURE_LIVE_LENGTH) where the pseudo-register lives and
5070 the register pressure is high, number of references of the
5071 pseudo-registers (through NREFS), the number of psuedo registers
5072 whose allocated register wouldn't need saving in the prologue
5073 (through CALL_USED_COUNT), and the first hard regno occupied by the
5074 pseudo-registers (through FIRST_HARD_REGNO). */
5076 calculate_spill_cost (int *regnos
, rtx in
, rtx out
, rtx_insn
*insn
,
5077 int *excess_pressure_live_length
,
5078 int *nrefs
, int *call_used_count
, int *first_hard_regno
)
5080 int i
, cost
, regno
, hard_regno
, count
, saved_cost
;
5086 for (length
= count
= cost
= i
= 0;; i
++)
5091 *nrefs
+= REG_N_REFS (regno
);
5092 hard_regno
= reg_renumber
[regno
];
5093 ira_assert (hard_regno
>= 0);
5094 a
= ira_regno_allocno_map
[regno
];
5095 length
+= ALLOCNO_EXCESS_PRESSURE_POINTS_NUM (a
) / ALLOCNO_NUM_OBJECTS (a
);
5096 cost
+= ALLOCNO_MEMORY_COST (a
) - ALLOCNO_CLASS_COST (a
);
5097 if (in_hard_reg_set_p (crtl
->abi
->full_reg_clobbers (),
5098 ALLOCNO_MODE (a
), hard_regno
))
5100 in_p
= in
&& REG_P (in
) && (int) REGNO (in
) == hard_regno
;
5101 out_p
= out
&& REG_P (out
) && (int) REGNO (out
) == hard_regno
;
5103 && find_regno_note (insn
, REG_DEAD
, hard_regno
) != NULL_RTX
)
5107 saved_cost
+= ira_memory_move_cost
5108 [ALLOCNO_MODE (a
)][ALLOCNO_CLASS (a
)][1];
5111 += ira_memory_move_cost
5112 [ALLOCNO_MODE (a
)][ALLOCNO_CLASS (a
)][0];
5113 cost
-= REG_FREQ_FROM_BB (BLOCK_FOR_INSN (insn
)) * saved_cost
;
5116 *excess_pressure_live_length
= length
;
5117 *call_used_count
= count
;
5121 hard_regno
= reg_renumber
[regnos
[0]];
5123 *first_hard_regno
= hard_regno
;
5127 /* Return TRUE if spilling pseudo-registers whose numbers are in array
5128 REGNOS is better than spilling pseudo-registers with numbers in
5129 OTHER_REGNOS for reload with given IN and OUT for INSN. The
5130 function used by the reload pass to make better register spilling
5133 ira_better_spill_reload_regno_p (int *regnos
, int *other_regnos
,
5134 rtx in
, rtx out
, rtx_insn
*insn
)
5136 int cost
, other_cost
;
5137 int length
, other_length
;
5138 int nrefs
, other_nrefs
;
5139 int call_used_count
, other_call_used_count
;
5140 int hard_regno
, other_hard_regno
;
5142 cost
= calculate_spill_cost (regnos
, in
, out
, insn
,
5143 &length
, &nrefs
, &call_used_count
, &hard_regno
);
5144 other_cost
= calculate_spill_cost (other_regnos
, in
, out
, insn
,
5145 &other_length
, &other_nrefs
,
5146 &other_call_used_count
,
5148 if (nrefs
== 0 && other_nrefs
!= 0)
5150 if (nrefs
!= 0 && other_nrefs
== 0)
5152 if (cost
!= other_cost
)
5153 return cost
< other_cost
;
5154 if (length
!= other_length
)
5155 return length
> other_length
;
5156 #ifdef REG_ALLOC_ORDER
5157 if (hard_regno
>= 0 && other_hard_regno
>= 0)
5158 return (inv_reg_alloc_order
[hard_regno
]
5159 < inv_reg_alloc_order
[other_hard_regno
]);
5161 if (call_used_count
!= other_call_used_count
)
5162 return call_used_count
> other_call_used_count
;
5169 /* Allocate and initialize data necessary for assign_hard_reg. */
5171 ira_initiate_assign (void)
5174 = (ira_allocno_t
*) ira_allocate (sizeof (ira_allocno_t
)
5175 * ira_allocnos_num
);
5176 consideration_allocno_bitmap
= ira_allocate_bitmap ();
5177 initiate_cost_update ();
5178 allocno_priorities
= (int *) ira_allocate (sizeof (int) * ira_allocnos_num
);
5179 sorted_copies
= (ira_copy_t
*) ira_allocate (ira_copies_num
5180 * sizeof (ira_copy_t
));
5183 /* Deallocate data used by assign_hard_reg. */
5185 ira_finish_assign (void)
5187 ira_free (sorted_allocnos
);
5188 ira_free_bitmap (consideration_allocno_bitmap
);
5189 finish_cost_update ();
5190 ira_free (allocno_priorities
);
5191 ira_free (sorted_copies
);
5196 /* Entry function doing color-based register allocation. */
5200 allocno_stack_vec
.create (ira_allocnos_num
);
5201 memset (allocated_hardreg_p
, 0, sizeof (allocated_hardreg_p
));
5202 ira_initiate_assign ();
5204 ira_finish_assign ();
5205 allocno_stack_vec
.release ();
5206 move_spill_restore ();
5211 /* This page contains a simple register allocator without usage of
5212 allocno conflicts. This is used for fast allocation for -O0. */
5214 /* Do register allocation by not using allocno conflicts. It uses
5215 only allocno live ranges. The algorithm is close to Chow's
5216 priority coloring. */
5218 fast_allocation (void)
5220 int i
, j
, k
, num
, class_size
, hard_regno
, best_hard_regno
, cost
, min_cost
;
5223 bool no_stack_reg_p
;
5225 enum reg_class aclass
;
5228 ira_allocno_iterator ai
;
5230 HARD_REG_SET conflict_hard_regs
, *used_hard_regs
;
5232 sorted_allocnos
= (ira_allocno_t
*) ira_allocate (sizeof (ira_allocno_t
)
5233 * ira_allocnos_num
);
5235 FOR_EACH_ALLOCNO (a
, ai
)
5236 sorted_allocnos
[num
++] = a
;
5237 allocno_priorities
= (int *) ira_allocate (sizeof (int) * ira_allocnos_num
);
5238 setup_allocno_priorities (sorted_allocnos
, num
);
5239 used_hard_regs
= (HARD_REG_SET
*) ira_allocate (sizeof (HARD_REG_SET
)
5241 for (i
= 0; i
< ira_max_point
; i
++)
5242 CLEAR_HARD_REG_SET (used_hard_regs
[i
]);
5243 qsort (sorted_allocnos
, num
, sizeof (ira_allocno_t
),
5244 allocno_priority_compare_func
);
5245 for (i
= 0; i
< num
; i
++)
5249 a
= sorted_allocnos
[i
];
5250 nr
= ALLOCNO_NUM_OBJECTS (a
);
5251 CLEAR_HARD_REG_SET (conflict_hard_regs
);
5252 for (l
= 0; l
< nr
; l
++)
5254 ira_object_t obj
= ALLOCNO_OBJECT (a
, l
);
5255 conflict_hard_regs
|= OBJECT_CONFLICT_HARD_REGS (obj
);
5256 for (r
= OBJECT_LIVE_RANGES (obj
); r
!= NULL
; r
= r
->next
)
5257 for (j
= r
->start
; j
<= r
->finish
; j
++)
5258 conflict_hard_regs
|= used_hard_regs
[j
];
5260 aclass
= ALLOCNO_CLASS (a
);
5261 ALLOCNO_ASSIGNED_P (a
) = true;
5262 ALLOCNO_HARD_REGNO (a
) = -1;
5263 if (hard_reg_set_subset_p (reg_class_contents
[aclass
],
5264 conflict_hard_regs
))
5266 mode
= ALLOCNO_MODE (a
);
5268 no_stack_reg_p
= ALLOCNO_NO_STACK_REG_P (a
);
5270 class_size
= ira_class_hard_regs_num
[aclass
];
5271 costs
= ALLOCNO_HARD_REG_COSTS (a
);
5273 best_hard_regno
= -1;
5274 for (j
= 0; j
< class_size
; j
++)
5276 hard_regno
= ira_class_hard_regs
[aclass
][j
];
5278 if (no_stack_reg_p
&& FIRST_STACK_REG
<= hard_regno
5279 && hard_regno
<= LAST_STACK_REG
)
5282 if (ira_hard_reg_set_intersection_p (hard_regno
, mode
, conflict_hard_regs
)
5283 || (TEST_HARD_REG_BIT
5284 (ira_prohibited_class_mode_regs
[aclass
][mode
], hard_regno
)))
5286 if (NUM_REGISTER_FILTERS
5287 && !test_register_filters (ALLOCNO_REGISTER_FILTERS (a
),
5292 best_hard_regno
= hard_regno
;
5296 if (min_cost
> cost
)
5299 best_hard_regno
= hard_regno
;
5302 if (best_hard_regno
< 0)
5304 ALLOCNO_HARD_REGNO (a
) = hard_regno
= best_hard_regno
;
5305 for (l
= 0; l
< nr
; l
++)
5307 ira_object_t obj
= ALLOCNO_OBJECT (a
, l
);
5308 for (r
= OBJECT_LIVE_RANGES (obj
); r
!= NULL
; r
= r
->next
)
5309 for (k
= r
->start
; k
<= r
->finish
; k
++)
5310 used_hard_regs
[k
] |= ira_reg_mode_hard_regset
[hard_regno
][mode
];
5313 ira_free (sorted_allocnos
);
5314 ira_free (used_hard_regs
);
5315 ira_free (allocno_priorities
);
5316 if (internal_flag_ira_verbose
> 1 && ira_dump_file
!= NULL
)
5317 ira_print_disposition (ira_dump_file
);
5322 /* Entry function doing coloring. */
5327 ira_allocno_iterator ai
;
5329 /* Setup updated costs. */
5330 FOR_EACH_ALLOCNO (a
, ai
)
5332 ALLOCNO_UPDATED_MEMORY_COST (a
) = ALLOCNO_MEMORY_COST (a
);
5333 ALLOCNO_UPDATED_CLASS_COST (a
) = ALLOCNO_CLASS_COST (a
);
5335 if (ira_conflicts_p
)