Merge commit 'origin/master' into radeon-rewrite
[mesa.git] / src / mesa / drivers / dri / radeon / radeon_cs_legacy.c
1 /*
2 * Copyright © 2008 Nicolai Haehnle
3 * Copyright © 2008 Jérôme Glisse
4 * All Rights Reserved.
5 *
6 * Permission is hereby granted, free of charge, to any person obtaining a
7 * copy of this software and associated documentation files (the
8 * "Software"), to deal in the Software without restriction, including
9 * without limitation the rights to use, copy, modify, merge, publish,
10 * distribute, sub license, and/or sell copies of the Software, and to
11 * permit persons to whom the Software is furnished to do so, subject to
12 * the following conditions:
13 *
14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16 * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
17 * THE COPYRIGHT HOLDERS, AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM,
18 * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
19 * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
20 * USE OR OTHER DEALINGS IN THE SOFTWARE.
21 *
22 * The above copyright notice and this permission notice (including the
23 * next paragraph) shall be included in all copies or substantial portions
24 * of the Software.
25 */
26 /*
27 * Authors:
28 * Aapo Tahkola <aet@rasterburn.org>
29 * Nicolai Haehnle <prefect_@gmx.net>
30 * Jérôme Glisse <glisse@freedesktop.org>
31 */
32 #include <errno.h>
33
34 #include "radeon_bocs_wrapper.h"
35
36 struct cs_manager_legacy {
37 struct radeon_cs_manager base;
38 struct radeon_context *ctx;
39 /* hack for scratch stuff */
40 uint32_t pending_age;
41 uint32_t pending_count;
42
43
44 };
45
46 struct cs_reloc_legacy {
47 struct radeon_cs_reloc base;
48 uint32_t cindices;
49 uint32_t *indices;
50 };
51
52
53 static struct radeon_cs *cs_create(struct radeon_cs_manager *csm,
54 uint32_t ndw)
55 {
56 struct radeon_cs *cs;
57
58 cs = (struct radeon_cs*)calloc(1, sizeof(struct radeon_cs));
59 if (cs == NULL) {
60 return NULL;
61 }
62 cs->csm = csm;
63 cs->ndw = (ndw + 0x3FF) & (~0x3FF);
64 cs->packets = (uint32_t*)malloc(4*cs->ndw);
65 if (cs->packets == NULL) {
66 free(cs);
67 return NULL;
68 }
69 cs->relocs_total_size = 0;
70 return cs;
71 }
72
73 static int cs_write_reloc(struct radeon_cs *cs,
74 struct radeon_bo *bo,
75 uint32_t read_domain,
76 uint32_t write_domain,
77 uint32_t flags)
78 {
79 struct cs_reloc_legacy *relocs;
80 int i;
81
82 relocs = (struct cs_reloc_legacy *)cs->relocs;
83 /* check domains */
84 if ((read_domain && write_domain) || (!read_domain && !write_domain)) {
85 /* in one CS a bo can only be in read or write domain but not
86 * in read & write domain at the same sime
87 */
88 return -EINVAL;
89 }
90 if (read_domain == RADEON_GEM_DOMAIN_CPU) {
91 return -EINVAL;
92 }
93 if (write_domain == RADEON_GEM_DOMAIN_CPU) {
94 return -EINVAL;
95 }
96 /* check if bo is already referenced */
97 for(i = 0; i < cs->crelocs; i++) {
98 uint32_t *indices;
99
100 if (relocs[i].base.bo->handle == bo->handle) {
101 /* Check domains must be in read or write. As we check already
102 * checked that in argument one of the read or write domain was
103 * set we only need to check that if previous reloc as the read
104 * domain set then the read_domain should also be set for this
105 * new relocation.
106 */
107 if (relocs[i].base.read_domain && !read_domain) {
108 return -EINVAL;
109 }
110 if (relocs[i].base.write_domain && !write_domain) {
111 return -EINVAL;
112 }
113 relocs[i].base.read_domain |= read_domain;
114 relocs[i].base.write_domain |= write_domain;
115 /* save indice */
116 relocs[i].cindices++;
117 indices = (uint32_t*)realloc(relocs[i].indices,
118 relocs[i].cindices * 4);
119 if (indices == NULL) {
120 relocs[i].cindices -= 1;
121 return -ENOMEM;
122 }
123 relocs[i].indices = indices;
124 relocs[i].indices[relocs[i].cindices - 1] = cs->cdw - 1;
125 return 0;
126 }
127 }
128 /* add bo to reloc */
129 relocs = (struct cs_reloc_legacy*)
130 realloc(cs->relocs,
131 sizeof(struct cs_reloc_legacy) * (cs->crelocs + 1));
132 if (relocs == NULL) {
133 return -ENOMEM;
134 }
135 cs->relocs = relocs;
136 relocs[cs->crelocs].base.bo = bo;
137 relocs[cs->crelocs].base.read_domain = read_domain;
138 relocs[cs->crelocs].base.write_domain = write_domain;
139 relocs[cs->crelocs].base.flags = flags;
140 relocs[cs->crelocs].indices = (uint32_t*)malloc(4);
141 if (relocs[cs->crelocs].indices == NULL) {
142 return -ENOMEM;
143 }
144 relocs[cs->crelocs].indices[0] = cs->cdw - 1;
145 relocs[cs->crelocs].cindices = 1;
146 cs->relocs_total_size += radeon_bo_legacy_relocs_size(bo);
147 cs->crelocs++;
148 radeon_bo_ref(bo);
149 return 0;
150 }
151
152 static int cs_begin(struct radeon_cs *cs,
153 uint32_t ndw,
154 const char *file,
155 const char *func,
156 int line)
157 {
158 if (cs->section) {
159 fprintf(stderr, "CS already in a section(%s,%s,%d)\n",
160 cs->section_file, cs->section_func, cs->section_line);
161 fprintf(stderr, "CS can't start section(%s,%s,%d)\n",
162 file, func, line);
163 return -EPIPE;
164 }
165 cs->section = 1;
166 cs->section_ndw = ndw;
167 cs->section_cdw = 0;
168 cs->section_file = file;
169 cs->section_func = func;
170 cs->section_line = line;
171
172
173 if (cs->cdw + ndw > cs->ndw) {
174 uint32_t tmp, *ptr;
175 int num = (ndw > 0x3FF) ? ndw : 0x3FF;
176
177 tmp = (cs->cdw + 1 + num) & (~num);
178 ptr = (uint32_t*)realloc(cs->packets, 4 * tmp);
179 if (ptr == NULL) {
180 return -ENOMEM;
181 }
182 cs->packets = ptr;
183 cs->ndw = tmp;
184 }
185
186 return 0;
187 }
188
189 static int cs_end(struct radeon_cs *cs,
190 const char *file,
191 const char *func,
192 int line)
193
194 {
195 if (!cs->section) {
196 fprintf(stderr, "CS no section to end at (%s,%s,%d)\n",
197 file, func, line);
198 return -EPIPE;
199 }
200 cs->section = 0;
201 if (cs->section_ndw != cs->section_cdw) {
202 fprintf(stderr, "CS section size missmatch start at (%s,%s,%d) %d vs %d\n",
203 cs->section_file, cs->section_func, cs->section_line, cs->section_ndw, cs->section_cdw);
204 fprintf(stderr, "CS section end at (%s,%s,%d)\n",
205 file, func, line);
206 return -EPIPE;
207 }
208 return 0;
209 }
210
211 static int cs_process_relocs(struct radeon_cs *cs)
212 {
213 struct cs_manager_legacy *csm = (struct cs_manager_legacy*)cs->csm;
214 struct cs_reloc_legacy *relocs;
215 int i, j, r;
216
217 csm = (struct cs_manager_legacy*)cs->csm;
218 relocs = (struct cs_reloc_legacy *)cs->relocs;
219 restart:
220 for (i = 0; i < cs->crelocs; i++) {
221 for (j = 0; j < relocs[i].cindices; j++) {
222 uint32_t soffset, eoffset;
223
224 r = radeon_bo_legacy_validate(relocs[i].base.bo,
225 &soffset, &eoffset);
226 if (r == -EAGAIN)
227 goto restart;
228 if (r) {
229 fprintf(stderr, "validated %p [0x%08X, 0x%08X]\n",
230 relocs[i].base.bo, soffset, eoffset);
231 return r;
232 }
233 cs->packets[relocs[i].indices[j]] += soffset;
234 if (cs->packets[relocs[i].indices[j]] >= eoffset) {
235 /* radeon_bo_debug(relocs[i].base.bo, 12); */
236 fprintf(stderr, "validated %p [0x%08X, 0x%08X]\n",
237 relocs[i].base.bo, soffset, eoffset);
238 fprintf(stderr, "above end: %p 0x%08X 0x%08X\n",
239 relocs[i].base.bo,
240 cs->packets[relocs[i].indices[j]],
241 eoffset);
242 exit(0);
243 return -EINVAL;
244 }
245 }
246 }
247 return 0;
248 }
249
250 static int cs_set_age(struct radeon_cs *cs)
251 {
252 struct cs_manager_legacy *csm = (struct cs_manager_legacy*)cs->csm;
253 struct cs_reloc_legacy *relocs;
254 int i;
255
256 relocs = (struct cs_reloc_legacy *)cs->relocs;
257 for (i = 0; i < cs->crelocs; i++) {
258 radeon_bo_legacy_pending(relocs[i].base.bo, csm->pending_age);
259 radeon_bo_unref(relocs[i].base.bo);
260 }
261 return 0;
262 }
263
264 static int cs_emit(struct radeon_cs *cs)
265 {
266 struct cs_manager_legacy *csm = (struct cs_manager_legacy*)cs->csm;
267 drm_radeon_cmd_buffer_t cmd;
268 drm_r300_cmd_header_t age;
269 uint64_t ull;
270 int r;
271
272 csm->ctx->vtbl.emit_cs_header(cs, csm->ctx);
273
274 /* append buffer age */
275 if (IS_R300_CLASS(csm->ctx->radeonScreen)) {
276 age.scratch.cmd_type = R300_CMD_SCRATCH;
277 /* Scratch register 2 corresponds to what radeonGetAge polls */
278 csm->pending_age = 0;
279 csm->pending_count = 1;
280 ull = (uint64_t) (intptr_t) &csm->pending_age;
281 age.scratch.reg = 2;
282 age.scratch.n_bufs = 1;
283 age.scratch.flags = 0;
284 radeon_cs_write_dword(cs, age.u);
285 radeon_cs_write_qword(cs, ull);
286 radeon_cs_write_dword(cs, 0);
287 }
288
289 r = cs_process_relocs(cs);
290 if (r) {
291 return 0;
292 }
293
294 cmd.buf = (char *)cs->packets;
295 cmd.bufsz = cs->cdw * 4;
296 if (csm->ctx->state.scissor.enabled) {
297 cmd.nbox = csm->ctx->state.scissor.numClipRects;
298 cmd.boxes = (drm_clip_rect_t *) csm->ctx->state.scissor.pClipRects;
299 } else {
300 cmd.nbox = csm->ctx->numClipRects;
301 cmd.boxes = (drm_clip_rect_t *) csm->ctx->pClipRects;
302 }
303
304 //dump_cmdbuf(cs);
305
306 r = drmCommandWrite(cs->csm->fd, DRM_RADEON_CMDBUF, &cmd, sizeof(cmd));
307 if (r) {
308 return r;
309 }
310 if (!IS_R300_CLASS(csm->ctx->radeonScreen)) {
311 drm_radeon_irq_emit_t emit_cmd;
312 emit_cmd.irq_seq = &csm->pending_age;
313 r = drmCommandWrite(cs->csm->fd, DRM_RADEON_IRQ_EMIT, &emit_cmd, sizeof(emit_cmd));
314 if (r) {
315 return r;
316 }
317 }
318 cs_set_age(cs);
319
320 cs->csm->read_used = 0;
321 cs->csm->vram_write_used = 0;
322 cs->csm->gart_write_used = 0;
323 return 0;
324 }
325
326 static void inline cs_free_reloc(void *relocs_p, int crelocs)
327 {
328 struct cs_reloc_legacy *relocs = relocs_p;
329 int i;
330 if (!relocs_p)
331 return;
332 for (i = 0; i < crelocs; i++)
333 free(relocs[i].indices);
334 }
335
336 static int cs_destroy(struct radeon_cs *cs)
337 {
338 cs_free_reloc(cs->relocs, cs->crelocs);
339 free(cs->relocs);
340 free(cs->packets);
341 free(cs);
342 return 0;
343 }
344
345 static int cs_erase(struct radeon_cs *cs)
346 {
347 cs_free_reloc(cs->relocs, cs->crelocs);
348 free(cs->relocs);
349 cs->relocs_total_size = 0;
350 cs->relocs = NULL;
351 cs->crelocs = 0;
352 cs->cdw = 0;
353 cs->section = 0;
354 return 0;
355 }
356
357 static int cs_need_flush(struct radeon_cs *cs)
358 {
359 /* this function used to flush when the BO usage got to
360 * a certain size, now the higher levels handle this better */
361 return 0;
362 }
363
364 static void cs_print(struct radeon_cs *cs, FILE *file)
365 {
366 }
367
368 static int cs_check_space(struct radeon_cs *cs, struct radeon_cs_space_check *bos, int num_bo)
369 {
370 struct radeon_cs_manager *csm = cs->csm;
371 int this_op_read = 0, this_op_gart_write = 0, this_op_vram_write = 0;
372 uint32_t read_domains, write_domain;
373 int i;
374 struct radeon_bo *bo;
375
376 /* check the totals for this operation */
377
378 if (num_bo == 0)
379 return 0;
380
381 /* prepare */
382 for (i = 0; i < num_bo; i++) {
383 bo = bos[i].bo;
384
385 bos[i].new_accounted = 0;
386 read_domains = bos[i].read_domains;
387 write_domain = bos[i].write_domain;
388
389 /* pinned bos don't count */
390 if (radeon_legacy_bo_is_static(bo))
391 continue;
392
393 /* already accounted this bo */
394 if (write_domain && (write_domain == bo->space_accounted))
395 continue;
396
397 if (read_domains && ((read_domains << 16) == bo->space_accounted))
398 continue;
399
400 if (bo->space_accounted == 0) {
401 if (write_domain == RADEON_GEM_DOMAIN_VRAM)
402 this_op_vram_write += bo->size;
403 else if (write_domain == RADEON_GEM_DOMAIN_GTT)
404 this_op_gart_write += bo->size;
405 else
406 this_op_read += bo->size;
407 bos[i].new_accounted = (read_domains << 16) | write_domain;
408 } else {
409 uint16_t old_read, old_write;
410
411 old_read = bo->space_accounted >> 16;
412 old_write = bo->space_accounted & 0xffff;
413
414 if (write_domain && (old_read & write_domain)) {
415 bos[i].new_accounted = write_domain;
416 /* moving from read to a write domain */
417 if (write_domain == RADEON_GEM_DOMAIN_VRAM) {
418 this_op_read -= bo->size;
419 this_op_vram_write += bo->size;
420 } else if (write_domain == RADEON_GEM_DOMAIN_VRAM) {
421 this_op_read -= bo->size;
422 this_op_gart_write += bo->size;
423 }
424 } else if (read_domains & old_write) {
425 bos[i].new_accounted = bo->space_accounted & 0xffff;
426 } else {
427 /* rewrite the domains */
428 if (write_domain != old_write)
429 fprintf(stderr,"WRITE DOMAIN RELOC FAILURE 0x%x %d %d\n", bo->handle, write_domain, old_write);
430 if (read_domains != old_read)
431 fprintf(stderr,"READ DOMAIN RELOC FAILURE 0x%x %d %d\n", bo->handle, read_domains, old_read);
432 return RADEON_CS_SPACE_FLUSH;
433 }
434 }
435 }
436
437 if (this_op_read < 0)
438 this_op_read = 0;
439
440 /* check sizes - operation first */
441 if ((this_op_read + this_op_gart_write > csm->gart_limit) ||
442 (this_op_vram_write > csm->vram_limit)) {
443 return RADEON_CS_SPACE_OP_TO_BIG;
444 }
445
446 if (((csm->vram_write_used + this_op_vram_write) > csm->vram_limit) ||
447 ((csm->read_used + csm->gart_write_used + this_op_gart_write + this_op_read) > csm->gart_limit)) {
448 return RADEON_CS_SPACE_FLUSH;
449 }
450
451 csm->gart_write_used += this_op_gart_write;
452 csm->vram_write_used += this_op_vram_write;
453 csm->read_used += this_op_read;
454 /* commit */
455 for (i = 0; i < num_bo; i++) {
456 bo = bos[i].bo;
457 bo->space_accounted = bos[i].new_accounted;
458 }
459
460 return RADEON_CS_SPACE_OK;
461 }
462
463 static struct radeon_cs_funcs radeon_cs_legacy_funcs = {
464 cs_create,
465 cs_write_reloc,
466 cs_begin,
467 cs_end,
468 cs_emit,
469 cs_destroy,
470 cs_erase,
471 cs_need_flush,
472 cs_print,
473 cs_check_space
474 };
475
476 struct radeon_cs_manager *radeon_cs_manager_legacy_ctor(struct radeon_context *ctx)
477 {
478 struct cs_manager_legacy *csm;
479
480 csm = (struct cs_manager_legacy*)
481 calloc(1, sizeof(struct cs_manager_legacy));
482 if (csm == NULL) {
483 return NULL;
484 }
485 csm->base.funcs = &radeon_cs_legacy_funcs;
486 csm->base.fd = ctx->dri.fd;
487 csm->ctx = ctx;
488 csm->pending_age = 1;
489 return (struct radeon_cs_manager*)csm;
490 }
491
492 void radeon_cs_manager_legacy_dtor(struct radeon_cs_manager *csm)
493 {
494 free(csm);
495 }
496