This bug was causing miscompiled code related to branch delay slots Signed-off-by: Felix Fietkau <nbd@nbd.name>master
parent
38dc678166
commit
25aaff9100
@ -0,0 +1,136 @@ |
||||
From: ebotcazou <ebotcazou@138bc75d-0d04-0410-961f-82ee72b054a4>
|
||||
Date: Mon, 26 Feb 2018 16:29:30 +0000
|
||||
Subject: [PATCH] PR rtl-optimization/83496 * reorg.c
|
||||
(steal_delay_list_from_target): Change REDUNDANT array from booleans to
|
||||
RTXes. Call fix_reg_dead_note on every non-null element.
|
||||
(steal_delay_list_from_fallthrough): Call fix_reg_dead_note on a
|
||||
redundant insn, if any. (relax_delay_slots): Likewise.
|
||||
(update_reg_unused_notes): Rename REDUNDANT_INSN to OTHER_INSN.
|
||||
|
||||
git-svn-id: svn+ssh://gcc.gnu.org/svn/gcc/trunk@257996 138bc75d-0d04-0410-961f-82ee72b054a4
|
||||
---
|
||||
create mode 120000 gcc/testsuite/gcc.c-torture/execute/20180226-1.c
|
||||
|
||||
--- a/gcc/reorg.c
|
||||
+++ b/gcc/reorg.c
|
||||
@@ -1035,7 +1035,8 @@ check_annul_list_true_false (int annul_t
|
||||
|
||||
static void
|
||||
steal_delay_list_from_target (rtx_insn *insn, rtx condition, rtx_sequence *seq,
|
||||
- vec<rtx_insn *> *delay_list, resources *sets,
|
||||
+ vec<rtx_insn *> *delay_list,
|
||||
+ struct resources *sets,
|
||||
struct resources *needed,
|
||||
struct resources *other_needed,
|
||||
int slots_to_fill, int *pslots_filled,
|
||||
@@ -1048,7 +1049,7 @@ steal_delay_list_from_target (rtx_insn *
|
||||
int used_annul = 0;
|
||||
int i;
|
||||
struct resources cc_set;
|
||||
- bool *redundant;
|
||||
+ rtx_insn **redundant;
|
||||
|
||||
/* We can't do anything if there are more delay slots in SEQ than we
|
||||
can handle, or if we don't know that it will be a taken branch.
|
||||
@@ -1087,7 +1088,7 @@ steal_delay_list_from_target (rtx_insn *
|
||||
if (! targetm.can_follow_jump (insn, seq->insn (0)))
|
||||
return;
|
||||
|
||||
- redundant = XALLOCAVEC (bool, XVECLEN (seq, 0));
|
||||
+ redundant = XALLOCAVEC (rtx_insn *, XVECLEN (seq, 0));
|
||||
for (i = 1; i < seq->len (); i++)
|
||||
{
|
||||
rtx_insn *trial = seq->insn (i);
|
||||
@@ -1151,7 +1152,10 @@ steal_delay_list_from_target (rtx_insn *
|
||||
we therefore decided not to copy. */
|
||||
for (i = 1; i < seq->len (); i++)
|
||||
if (redundant[i])
|
||||
- update_block (seq->insn (i), insn);
|
||||
+ {
|
||||
+ fix_reg_dead_note (redundant[i], insn);
|
||||
+ update_block (seq->insn (i), insn);
|
||||
+ }
|
||||
|
||||
/* Show the place to which we will be branching. */
|
||||
*pnew_thread = first_active_target_insn (JUMP_LABEL (seq->insn (0)));
|
||||
@@ -1198,6 +1202,7 @@ steal_delay_list_from_fallthrough (rtx_i
|
||||
for (i = 1; i < seq->len (); i++)
|
||||
{
|
||||
rtx_insn *trial = seq->insn (i);
|
||||
+ rtx_insn *prior_insn;
|
||||
|
||||
/* If TRIAL sets CC0, stealing it will move it too far from the use
|
||||
of CC0. */
|
||||
@@ -1209,8 +1214,9 @@ steal_delay_list_from_fallthrough (rtx_i
|
||||
break;
|
||||
|
||||
/* If this insn was already done, we don't need it. */
|
||||
- if (redundant_insn (trial, insn, *delay_list))
|
||||
+ if ((prior_insn = redundant_insn (trial, insn, *delay_list)))
|
||||
{
|
||||
+ fix_reg_dead_note (prior_insn, insn);
|
||||
update_block (trial, insn);
|
||||
delete_from_delay_slot (trial);
|
||||
continue;
|
||||
@@ -1790,15 +1796,14 @@ fix_reg_dead_note (rtx_insn *start_insn,
|
||||
}
|
||||
}
|
||||
|
||||
-/* Delete any REG_UNUSED notes that exist on INSN but not on REDUNDANT_INSN.
|
||||
+/* Delete any REG_UNUSED notes that exist on INSN but not on OTHER_INSN.
|
||||
|
||||
This handles the case of udivmodXi4 instructions which optimize their
|
||||
- output depending on whether any REG_UNUSED notes are present.
|
||||
- we must make sure that INSN calculates as many results as REDUNDANT_INSN
|
||||
- does. */
|
||||
+ output depending on whether any REG_UNUSED notes are present. We must
|
||||
+ make sure that INSN calculates as many results as OTHER_INSN does. */
|
||||
|
||||
static void
|
||||
-update_reg_unused_notes (rtx_insn *insn, rtx redundant_insn)
|
||||
+update_reg_unused_notes (rtx_insn *insn, rtx other_insn)
|
||||
{
|
||||
rtx link, next;
|
||||
|
||||
@@ -1810,8 +1815,7 @@ update_reg_unused_notes (rtx_insn *insn,
|
||||
|| !REG_P (XEXP (link, 0)))
|
||||
continue;
|
||||
|
||||
- if (! find_regno_note (redundant_insn, REG_UNUSED,
|
||||
- REGNO (XEXP (link, 0))))
|
||||
+ if (!find_regno_note (other_insn, REG_UNUSED, REGNO (XEXP (link, 0))))
|
||||
remove_note (insn, link);
|
||||
}
|
||||
}
|
||||
@@ -2324,9 +2328,8 @@ follow_jumps (rtx label, rtx_insn *jump,
|
||||
taken and THREAD_IF_TRUE is set. This is used for the branch at the
|
||||
end of a loop back up to the top.
|
||||
|
||||
- OWN_THREAD and OWN_OPPOSITE_THREAD are true if we are the only user of the
|
||||
- thread. I.e., it is the fallthrough code of our jump or the target of the
|
||||
- jump when we are the only jump going there.
|
||||
+ OWN_THREAD is true if we are the only user of the thread, i.e. it is
|
||||
+ the target of the jump when we are the only jump going there.
|
||||
|
||||
If OWN_THREAD is false, it must be the "true" thread of a jump. In that
|
||||
case, we can only take insns from the head of the thread for our delay
|
||||
@@ -3117,7 +3120,7 @@ relax_delay_slots (rtx_insn *first)
|
||||
/* Look at every JUMP_INSN and see if we can improve it. */
|
||||
for (insn = first; insn; insn = next)
|
||||
{
|
||||
- rtx_insn *other;
|
||||
+ rtx_insn *other, *prior_insn;
|
||||
bool crossing;
|
||||
|
||||
next = next_active_insn (insn);
|
||||
@@ -3223,8 +3226,9 @@ relax_delay_slots (rtx_insn *first)
|
||||
/* See if the first insn in the delay slot is redundant with some
|
||||
previous insn. Remove it from the delay slot if so; then set up
|
||||
to reprocess this insn. */
|
||||
- if (redundant_insn (pat->insn (1), delay_insn, vNULL))
|
||||
+ if ((prior_insn = redundant_insn (pat->insn (1), delay_insn, vNULL)))
|
||||
{
|
||||
+ fix_reg_dead_note (prior_insn, insn);
|
||||
update_block (pat->insn (1), insn);
|
||||
delete_from_delay_slot (pat->insn (1));
|
||||
next = prev_active_insn (next);
|
Loading…
Reference in new issue