1 /* -*- mode: C; c-file-style: "k&r"; tab-width 4; indent-tabs-mode: t; -*- */
4 * Copyright (C) 2014 Rob Clark <robclark@freedesktop.org>
6 * Permission is hereby granted, free of charge, to any person obtaining a
7 * copy of this software and associated documentation files (the "Software"),
8 * to deal in the Software without restriction, including without limitation
9 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
10 * and/or sell copies of the Software, and to permit persons to whom the
11 * Software is furnished to do so, subject to the following conditions:
13 * The above copyright notice and this permission notice (including the next
14 * paragraph) shall be included in all copies or substantial portions of the
17 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
18 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
19 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
20 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
21 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
22 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
26 * Rob Clark <robclark@freedesktop.org>
30 #include "util/u_math.h"
40 * Instruction Scheduling:
42 * Using the depth sorted list from depth pass, attempt to recursively
43 * schedule deepest unscheduled path. The first instruction that cannot
44 * be scheduled, returns the required delay slots it needs, at which
45 * point we return back up to the top and attempt to schedule by next
46 * highest depth. After a sufficient number of instructions have been
47 * scheduled, return back to beginning of list and start again. If you
48 * reach the end of depth sorted list without being able to insert any
49 * instruction, insert nop's. Repeat until no more unscheduled
52 * There are a few special cases that need to be handled, since sched
53 * is currently independent of register allocation. Usages of address
54 * register (a0.x) or predicate register (p0.x) must be serialized. Ie.
55 * if you have two pairs of instructions that write the same special
56 * register and then read it, then those pairs cannot be interleaved.
57 * To solve this, when we are in such a scheduling "critical section",
58 * and we encounter a conflicting write to a special register, we try
59 * to schedule any remaining instructions that use that value first.
62 struct ir3_sched_ctx
{
63 struct ir3_instruction
*scheduled
; /* last scheduled instr */
64 struct ir3_instruction
*addr
; /* current a0.x user, if any */
65 struct ir3_instruction
*pred
; /* current p0.x user, if any */
70 static struct ir3_instruction
*
71 deepest(struct ir3_instruction
**srcs
, unsigned nsrcs
)
73 struct ir3_instruction
*d
= NULL
;
74 unsigned i
= 0, id
= 0;
76 while ((i
< nsrcs
) && !(d
= srcs
[id
= i
]))
82 for (; i
< nsrcs
; i
++)
83 if (srcs
[i
] && (srcs
[i
]->depth
> d
->depth
))
91 static unsigned distance(struct ir3_sched_ctx
*ctx
,
92 struct ir3_instruction
*instr
, unsigned maxd
)
94 struct ir3_instruction
*n
= ctx
->scheduled
;
96 while (n
&& (n
!= instr
) && (d
< maxd
)) {
97 if (is_alu(n
) || is_flow(n
))
104 /* TODO maybe we want double linked list? */
105 static struct ir3_instruction
* prev(struct ir3_instruction
*instr
)
107 struct ir3_instruction
*p
= instr
->block
->head
;
108 while (p
&& (p
->next
!= instr
))
113 static bool is_sfu_or_mem(struct ir3_instruction
*instr
)
115 return is_sfu(instr
) || is_mem(instr
);
118 static void schedule(struct ir3_sched_ctx
*ctx
,
119 struct ir3_instruction
*instr
, bool remove
)
121 struct ir3_block
*block
= instr
->block
;
123 /* maybe there is a better way to handle this than just stuffing
124 * a nop.. ideally we'd know about this constraint in the
125 * scheduling and depth calculation..
127 if (ctx
->scheduled
&& is_sfu_or_mem(ctx
->scheduled
) && is_sfu_or_mem(instr
))
128 schedule(ctx
, ir3_NOP(block
), false);
130 /* remove from depth list:
133 struct ir3_instruction
*p
= prev(instr
);
135 /* NOTE: this can happen for inputs which are not
136 * read.. in that case there is no need to schedule
137 * the input, so just bail:
139 if (instr
!= (p
? p
->next
: block
->head
))
143 p
->next
= instr
->next
;
145 block
->head
= instr
->next
;
148 if (writes_addr(instr
)) {
149 assert(ctx
->addr
== NULL
);
153 if (writes_pred(instr
)) {
154 assert(ctx
->pred
== NULL
);
158 instr
->flags
|= IR3_INSTR_MARK
;
160 instr
->next
= ctx
->scheduled
;
161 ctx
->scheduled
= instr
;
167 * Delay-slot calculation. Follows fanin/fanout.
170 /* calculate delay for specified src: */
171 static unsigned delay_calc_srcn(struct ir3_sched_ctx
*ctx
,
172 struct ir3_instruction
*assigner
,
173 struct ir3_instruction
*consumer
, unsigned srcn
)
177 if (is_meta(assigner
)) {
178 struct ir3_instruction
*src
;
179 foreach_ssa_src(src
, assigner
) {
180 unsigned d
= delay_calc_srcn(ctx
, src
, consumer
, srcn
);
181 delay
= MAX2(delay
, d
);
184 delay
= ir3_delayslots(assigner
, consumer
, srcn
);
185 delay
-= distance(ctx
, assigner
, delay
);
191 /* calculate delay for instruction (maximum of delay for all srcs): */
192 static unsigned delay_calc(struct ir3_sched_ctx
*ctx
,
193 struct ir3_instruction
*instr
)
196 struct ir3_instruction
*src
;
198 foreach_ssa_src_n(src
, i
, instr
) {
199 unsigned d
= delay_calc_srcn(ctx
, src
, instr
, i
);
200 delay
= MAX2(delay
, d
);
206 /* A negative return value signals that an instruction has been newly
207 * SCHEDULED (or DELAYED due to address or predicate register already
208 * in use), return back up to the top of the stack (to block_sched())
210 static int trysched(struct ir3_sched_ctx
*ctx
,
211 struct ir3_instruction
*instr
)
213 struct ir3_instruction
*srcs
[64];
214 struct ir3_instruction
*src
;
215 unsigned delay
, nsrcs
= 0;
217 /* if already scheduled: */
218 if (instr
->flags
& IR3_INSTR_MARK
)
221 /* figure out our src's, copy 'em out into an array for sorting: */
222 foreach_ssa_src(src
, instr
) {
223 debug_assert(nsrcs
< ARRAY_SIZE(srcs
));
227 /* for each src register in sorted order:
230 while ((src
= deepest(srcs
, nsrcs
))) {
231 delay
= trysched(ctx
, src
);
236 /* all our dependents are scheduled, figure out if
237 * we have enough delay slots to schedule ourself:
239 delay
= delay_calc(ctx
, instr
);
243 /* if the instruction is a kill, we need to ensure *every*
244 * bary.f is scheduled. The hw seems unhappy if the thread
245 * gets killed before the end-input (ei) flag is hit.
247 * We could do this by adding each bary.f instruction as
248 * virtual ssa src for the kill instruction. But we have
249 * fixed length instr->regs[].
251 * TODO this wouldn't be quite right if we had multiple
252 * basic blocks, if any block was conditional. We'd need
253 * to schedule the bary.f's outside of any block which
254 * was conditional that contained a kill.. I think..
256 if (is_kill(instr
)) {
257 struct ir3
*ir
= instr
->block
->shader
;
260 for (i
= 0; i
< ir
->baryfs_count
; i
++) {
261 struct ir3_instruction
*baryf
= ir
->baryfs
[i
];
262 if (baryf
->depth
== DEPTH_UNUSED
)
264 delay
= trysched(ctx
, baryf
);
270 /* if this is a write to address/predicate register, and that
271 * register is currently in use, we need to defer until it is
274 if (writes_addr(instr
) && ctx
->addr
) {
275 assert(ctx
->addr
!= instr
);
278 if (writes_pred(instr
) && ctx
->pred
) {
279 assert(ctx
->pred
!= instr
);
283 schedule(ctx
, instr
, true);
287 static struct ir3_instruction
* reverse(struct ir3_instruction
*instr
)
289 struct ir3_instruction
*reversed
= NULL
;
291 struct ir3_instruction
*next
= instr
->next
;
292 instr
->next
= reversed
;
299 static bool uses_current_addr(struct ir3_sched_ctx
*ctx
,
300 struct ir3_instruction
*instr
)
302 return instr
->address
&& (ctx
->addr
== instr
->address
);
305 static bool uses_current_pred(struct ir3_sched_ctx
*ctx
,
306 struct ir3_instruction
*instr
)
308 struct ir3_instruction
*src
;
309 foreach_ssa_src(src
, instr
)
310 if (ctx
->pred
== src
)
315 /* when we encounter an instruction that writes to the address register
316 * when it is in use, we delay that instruction and try to schedule all
317 * other instructions using the current address register:
319 static int block_sched_undelayed(struct ir3_sched_ctx
*ctx
,
320 struct ir3_block
*block
)
322 struct ir3_instruction
*instr
= block
->head
;
323 bool addr_in_use
= false;
324 bool pred_in_use
= false;
325 bool all_delayed
= true;
326 unsigned cnt
= ~0, attempted
= 0;
329 struct ir3_instruction
*next
= instr
->next
;
330 bool addr
= uses_current_addr(ctx
, instr
);
331 bool pred
= uses_current_pred(ctx
, instr
);
334 int ret
= trysched(ctx
, instr
);
339 if (ret
== SCHEDULED
)
342 cnt
= MIN2(cnt
, ret
);
360 /* detect if we've gotten ourselves into an impossible situation
363 if (all_delayed
&& (attempted
> 0)) {
365 /* TODO we probably need to keep a list of instructions
366 * that reference predicate, similar to indirects
372 struct ir3
*ir
= ctx
->addr
->block
->shader
;
373 struct ir3_instruction
*new_addr
=
374 ir3_instr_clone(ctx
->addr
);
377 /* original addr is scheduled, but new one isn't: */
378 new_addr
->flags
&= ~IR3_INSTR_MARK
;
380 for (i
= 0; i
< ir
->indirects_count
; i
++) {
381 struct ir3_instruction
*indirect
= ir
->indirects
[i
];
383 /* skip instructions already scheduled: */
384 if (indirect
->flags
& IR3_INSTR_MARK
)
387 /* remap remaining instructions using current addr
390 if (indirect
->address
== ctx
->addr
)
391 indirect
->address
= new_addr
;
394 /* all remaining indirects remapped to new addr: */
397 /* not really, but this will trigger us to go back to
398 * main trysched() loop now that we've resolved the
399 * conflict by duplicating the instr that writes to
400 * the address register.
409 static void block_sched(struct ir3_sched_ctx
*ctx
, struct ir3_block
*block
)
411 struct ir3_instruction
*instr
;
413 /* schedule all the shader input's (meta-instr) first so that
414 * the RA step sees that the input registers contain a value
415 * from the start of the shader:
417 if (!block
->parent
) {
419 for (i
= 0; i
< block
->ninputs
; i
++) {
420 struct ir3_instruction
*in
= block
->inputs
[i
];
422 schedule(ctx
, in
, true);
426 while ((instr
= block
->head
) && !ctx
->error
) {
427 /* NOTE: always grab next *before* trysched(), in case the
428 * instruction is actually scheduled (and therefore moved
429 * from depth list into scheduled list)
431 struct ir3_instruction
*next
= instr
->next
;
432 int cnt
= trysched(ctx
, instr
);
435 cnt
= block_sched_undelayed(ctx
, block
);
437 /* -1 is signal to return up stack, but to us means same as 0: */
442 /* if deepest remaining instruction cannot be scheduled, try
443 * the increasingly more shallow instructions until needed
444 * number of delay slots is filled:
446 while (instr
&& (cnt
> ctx
->cnt
)) {
448 trysched(ctx
, instr
);
452 /* and if we run out of instructions that can be scheduled,
453 * then it is time for nop's:
455 while (cnt
> ctx
->cnt
)
456 schedule(ctx
, ir3_NOP(block
), false);
459 /* at this point, scheduled list is in reverse order, so fix that: */
460 block
->head
= reverse(ctx
->scheduled
);
463 int ir3_block_sched(struct ir3_block
*block
)
465 struct ir3_sched_ctx ctx
= {0};
466 ir3_clear_mark(block
->shader
);
467 block_sched(&ctx
, block
);