2 * Copyright (C) 2008 VMware, Inc.
3 * Copyright (C) 2014 Broadcom
4 * Copyright (C) 2018-2019 Alyssa Rosenzweig
5 * Copyright (C) 2019 Collabora, Ltd.
7 * Permission is hereby granted, free of charge, to any person obtaining a
8 * copy of this software and associated documentation files (the "Software"),
9 * to deal in the Software without restriction, including without limitation
10 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
11 * and/or sell copies of the Software, and to permit persons to whom the
12 * Software is furnished to do so, subject to the following conditions:
14 * The above copyright notice and this permission notice (including the next
15 * paragraph) shall be included in all copies or substantial portions of the
18 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
19 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
20 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
21 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
22 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
23 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
26 * Authors (Collabora):
27 * Tomeu Vizoso <tomeu.vizoso@collabora.com>
28 * Alyssa Rosenzweig <alyssa.rosenzweig@collabora.com>
34 #include "drm-uapi/drm_fourcc.h"
36 #include "state_tracker/winsys_handle.h"
37 #include "util/u_format.h"
38 #include "util/u_memory.h"
39 #include "util/u_surface.h"
40 #include "util/u_transfer.h"
41 #include "util/u_transfer_helper.h"
42 #include "util/u_gen_mipmap.h"
44 #include "pan_context.h"
45 #include "pan_screen.h"
46 #include "pan_resource.h"
48 #include "pan_tiling.h"
50 static struct pipe_resource
*
51 panfrost_resource_from_handle(struct pipe_screen
*pscreen
,
52 const struct pipe_resource
*templat
,
53 struct winsys_handle
*whandle
,
56 struct panfrost_screen
*screen
= pan_screen(pscreen
);
57 struct panfrost_resource
*rsc
;
58 struct pipe_resource
*prsc
;
60 assert(whandle
->type
== WINSYS_HANDLE_TYPE_FD
);
62 rsc
= rzalloc(pscreen
, struct panfrost_resource
);
70 pipe_reference_init(&prsc
->reference
, 1);
71 prsc
->screen
= pscreen
;
73 rsc
->bo
= panfrost_drm_import_bo(screen
, whandle
);
74 rsc
->bo
->slices
[0].stride
= whandle
->stride
;
75 rsc
->bo
->slices
[0].initialized
= true;
79 renderonly_create_gpu_import_for_resource(prsc
, screen
->ro
, NULL
);
80 /* failure is expected in some cases.. */
87 panfrost_resource_get_handle(struct pipe_screen
*pscreen
,
88 struct pipe_context
*ctx
,
89 struct pipe_resource
*pt
,
90 struct winsys_handle
*handle
,
93 struct panfrost_screen
*screen
= pan_screen(pscreen
);
94 struct panfrost_resource
*rsrc
= (struct panfrost_resource
*) pt
;
95 struct renderonly_scanout
*scanout
= rsrc
->scanout
;
97 handle
->modifier
= DRM_FORMAT_MOD_INVALID
;
99 if (handle
->type
== WINSYS_HANDLE_TYPE_SHARED
) {
101 } else if (handle
->type
== WINSYS_HANDLE_TYPE_KMS
) {
102 if (renderonly_get_handle(scanout
, handle
))
105 handle
->handle
= rsrc
->bo
->gem_handle
;
106 handle
->stride
= rsrc
->bo
->slices
[0].stride
;
108 } else if (handle
->type
== WINSYS_HANDLE_TYPE_FD
) {
110 struct drm_prime_handle args
= {
111 .handle
= scanout
->handle
,
112 .flags
= DRM_CLOEXEC
,
115 int ret
= drmIoctl(screen
->ro
->kms_fd
, DRM_IOCTL_PRIME_HANDLE_TO_FD
, &args
);
119 handle
->stride
= scanout
->stride
;
120 handle
->handle
= args
.fd
;
124 return panfrost_drm_export_bo(screen
, rsrc
->bo
->gem_handle
,
125 rsrc
->bo
->slices
[0].stride
,
133 panfrost_flush_resource(struct pipe_context
*pctx
, struct pipe_resource
*prsc
)
135 //DBG("TODO %s\n", __func__);
138 static struct pipe_surface
*
139 panfrost_create_surface(struct pipe_context
*pipe
,
140 struct pipe_resource
*pt
,
141 const struct pipe_surface
*surf_tmpl
)
143 struct pipe_surface
*ps
= NULL
;
145 ps
= rzalloc(pipe
, struct pipe_surface
);
148 pipe_reference_init(&ps
->reference
, 1);
149 pipe_resource_reference(&ps
->texture
, pt
);
151 ps
->format
= surf_tmpl
->format
;
153 if (pt
->target
!= PIPE_BUFFER
) {
154 assert(surf_tmpl
->u
.tex
.level
<= pt
->last_level
);
155 ps
->width
= u_minify(pt
->width0
, surf_tmpl
->u
.tex
.level
);
156 ps
->height
= u_minify(pt
->height0
, surf_tmpl
->u
.tex
.level
);
157 ps
->u
.tex
.level
= surf_tmpl
->u
.tex
.level
;
158 ps
->u
.tex
.first_layer
= surf_tmpl
->u
.tex
.first_layer
;
159 ps
->u
.tex
.last_layer
= surf_tmpl
->u
.tex
.last_layer
;
161 /* setting width as number of elements should get us correct renderbuffer width */
162 ps
->width
= surf_tmpl
->u
.buf
.last_element
- surf_tmpl
->u
.buf
.first_element
+ 1;
163 ps
->height
= pt
->height0
;
164 ps
->u
.buf
.first_element
= surf_tmpl
->u
.buf
.first_element
;
165 ps
->u
.buf
.last_element
= surf_tmpl
->u
.buf
.last_element
;
166 assert(ps
->u
.buf
.first_element
<= ps
->u
.buf
.last_element
);
167 assert(ps
->u
.buf
.last_element
< ps
->width
);
175 panfrost_surface_destroy(struct pipe_context
*pipe
,
176 struct pipe_surface
*surf
)
178 assert(surf
->texture
);
179 pipe_resource_reference(&surf
->texture
, NULL
);
183 static struct pipe_resource
*
184 panfrost_create_scanout_res(struct pipe_screen
*screen
,
185 const struct pipe_resource
*template)
187 struct panfrost_screen
*pscreen
= pan_screen(screen
);
188 struct pipe_resource scanout_templat
= *template;
189 struct renderonly_scanout
*scanout
;
190 struct winsys_handle handle
;
191 struct pipe_resource
*res
;
193 scanout
= renderonly_scanout_for_resource(&scanout_templat
,
194 pscreen
->ro
, &handle
);
198 assert(handle
.type
== WINSYS_HANDLE_TYPE_FD
);
199 /* TODO: handle modifiers? */
200 res
= screen
->resource_from_handle(screen
, template, &handle
,
201 PIPE_HANDLE_USAGE_FRAMEBUFFER_WRITE
);
202 close(handle
.handle
);
206 struct panfrost_resource
*pres
= pan_resource(res
);
208 pres
->scanout
= scanout
;
209 pscreen
->display_target
= pres
;
214 /* Computes sizes for checksumming, which is 8 bytes per 16x16 tile */
216 #define CHECKSUM_TILE_WIDTH 16
217 #define CHECKSUM_TILE_HEIGHT 16
218 #define CHECKSUM_BYTES_PER_TILE 8
221 panfrost_compute_checksum_sizes(
222 struct panfrost_slice
*slice
,
226 unsigned aligned_width
= ALIGN(width
, CHECKSUM_TILE_WIDTH
);
227 unsigned aligned_height
= ALIGN(height
, CHECKSUM_TILE_HEIGHT
);
229 unsigned tile_count_x
= aligned_width
/ CHECKSUM_TILE_WIDTH
;
230 unsigned tile_count_y
= aligned_height
/ CHECKSUM_TILE_HEIGHT
;
232 slice
->checksum_stride
= tile_count_x
* CHECKSUM_BYTES_PER_TILE
;
234 return slice
->checksum_stride
* tile_count_y
;
237 /* Setup the mip tree given a particular layout, possibly with checksumming */
240 panfrost_setup_slices(const struct pipe_resource
*tmpl
, struct panfrost_bo
*bo
)
242 unsigned width
= tmpl
->width0
;
243 unsigned height
= tmpl
->height0
;
244 unsigned depth
= tmpl
->depth0
;
245 unsigned bytes_per_pixel
= util_format_get_blocksize(tmpl
->format
);
249 /* Tiled operates blockwise; linear is packed. Also, anything
250 * we render to has to be tile-aligned. Maybe not strictly
251 * necessary, but we're not *that* pressed for memory and it
252 * makes code a lot simpler */
254 bool renderable
= tmpl
->bind
&
255 (PIPE_BIND_RENDER_TARGET
| PIPE_BIND_DEPTH_STENCIL
);
256 bool afbc
= bo
->layout
== PAN_AFBC
;
257 bool tiled
= bo
->layout
== PAN_TILED
;
258 bool should_align
= renderable
|| tiled
;
260 /* We don't know how to specify a 2D stride for 3D textures */
262 bool can_align_stride
=
263 tmpl
->target
!= PIPE_TEXTURE_3D
;
265 should_align
&= can_align_stride
;
268 unsigned size_2d
= 0;
270 for (unsigned l
= 0; l
<= tmpl
->last_level
; ++l
) {
271 struct panfrost_slice
*slice
= &bo
->slices
[l
];
273 unsigned effective_width
= width
;
274 unsigned effective_height
= height
;
275 unsigned effective_depth
= depth
;
278 effective_width
= ALIGN(effective_width
, 16);
279 effective_height
= ALIGN(effective_height
, 16);
281 /* We don't need to align depth */
284 slice
->offset
= offset
;
286 /* Compute the would-be stride */
287 unsigned stride
= bytes_per_pixel
* effective_width
;
289 /* ..but cache-line align it for performance */
290 if (can_align_stride
&& bo
->layout
== PAN_LINEAR
)
291 stride
= ALIGN(stride
, 64);
293 slice
->stride
= stride
;
295 unsigned slice_one_size
= slice
->stride
* effective_height
;
296 unsigned slice_full_size
= slice_one_size
* effective_depth
;
298 /* Report 2D size for 3D texturing */
301 size_2d
= slice_one_size
;
303 /* Compute AFBC sizes if necessary */
306 panfrost_afbc_header_size(width
, height
);
308 offset
+= slice
->header_size
;
311 offset
+= slice_full_size
;
313 /* Add a checksum region if necessary */
314 if (bo
->checksummed
) {
315 slice
->checksum_offset
= offset
;
317 unsigned size
= panfrost_compute_checksum_sizes(
318 slice
, width
, height
);
323 width
= u_minify(width
, 1);
324 height
= u_minify(height
, 1);
325 depth
= u_minify(depth
, 1);
328 assert(tmpl
->array_size
);
330 if (tmpl
->target
!= PIPE_TEXTURE_3D
) {
331 /* Arrays and cubemaps have the entire miptree duplicated */
333 bo
->cubemap_stride
= ALIGN(offset
, 64);
334 bo
->size
= ALIGN(bo
->cubemap_stride
* tmpl
->array_size
, 4096);
336 /* 3D strides across the 2D layers */
337 assert(tmpl
->array_size
== 1);
339 bo
->cubemap_stride
= size_2d
;
340 bo
->size
= ALIGN(offset
, 4096);
344 static struct panfrost_bo
*
345 panfrost_create_bo(struct panfrost_screen
*screen
, const struct pipe_resource
*template)
347 struct panfrost_bo
*bo
= rzalloc(screen
, struct panfrost_bo
);
348 pipe_reference_init(&bo
->reference
, 1);
350 /* Based on the usage, figure out what storing will be used. There are
353 * Linear: the basic format, bad for memory bandwidth, bad for cache
354 * use. Zero-copy, though. Renderable.
356 * Tiled: Not compressed, but cache-optimized. Expensive to write into
357 * (due to software tiling), but cheap to sample from. Ideal for most
360 * AFBC: Compressed and renderable (so always desirable for non-scanout
361 * rendertargets). Cheap to sample from. The format is black box, so we
362 * can't read/write from software.
365 /* Tiling textures is almost always faster, unless we only use it once */
367 bool is_texture
= (template->bind
& PIPE_BIND_SAMPLER_VIEW
);
368 bool is_2d
= template->depth0
== 1 && template->array_size
== 1;
369 bool is_streaming
= (template->usage
!= PIPE_USAGE_STREAM
);
371 bool should_tile
= is_streaming
&& is_texture
&& is_2d
;
373 /* Depth/stencil can't be tiled, only linear or AFBC */
374 should_tile
&= !(template->bind
& PIPE_BIND_DEPTH_STENCIL
);
376 /* FBOs we would like to checksum, if at all possible */
377 bool can_checksum
= !(template->bind
& (PIPE_BIND_SCANOUT
| PIPE_BIND_SHARED
));
378 bool should_checksum
= template->bind
& PIPE_BIND_RENDER_TARGET
;
380 bo
->checksummed
= can_checksum
&& should_checksum
;
382 /* Set the layout appropriately */
383 bo
->layout
= should_tile
? PAN_TILED
: PAN_LINEAR
;
385 panfrost_setup_slices(template, bo
);
387 struct panfrost_memory mem
;
389 panfrost_drm_allocate_slab(screen
, &mem
, bo
->size
/ 4096, true, 0, 0, 0);
393 bo
->gem_handle
= mem
.gem_handle
;
398 static struct pipe_resource
*
399 panfrost_resource_create(struct pipe_screen
*screen
,
400 const struct pipe_resource
*template)
402 /* Make sure we're familiar */
403 switch (template->target
) {
405 case PIPE_TEXTURE_1D
:
406 case PIPE_TEXTURE_2D
:
407 case PIPE_TEXTURE_3D
:
408 case PIPE_TEXTURE_CUBE
:
409 case PIPE_TEXTURE_RECT
:
410 case PIPE_TEXTURE_2D_ARRAY
:
413 DBG("Unknown texture target %d\n", template->target
);
418 (PIPE_BIND_DISPLAY_TARGET
| PIPE_BIND_SCANOUT
| PIPE_BIND_SHARED
))
419 return panfrost_create_scanout_res(screen
, template);
421 struct panfrost_resource
*so
= rzalloc(screen
, struct panfrost_resource
);
422 struct panfrost_screen
*pscreen
= (struct panfrost_screen
*) screen
;
424 so
->base
= *template;
425 so
->base
.screen
= screen
;
427 pipe_reference_init(&so
->base
.reference
, 1);
429 util_range_init(&so
->valid_buffer_range
);
431 so
->bo
= panfrost_create_bo(pscreen
, template);
432 return (struct pipe_resource
*)so
;
436 panfrost_destroy_bo(struct panfrost_screen
*screen
, struct panfrost_bo
*bo
)
438 struct panfrost_memory mem
= {
442 .gem_handle
= bo
->gem_handle
,
445 panfrost_drm_free_slab(screen
, &mem
);
450 panfrost_bo_reference(struct panfrost_bo
*bo
)
452 pipe_reference(NULL
, &bo
->reference
);
456 panfrost_bo_unreference(struct pipe_screen
*screen
, struct panfrost_bo
*bo
)
458 /* When the reference count goes to zero, we need to cleanup */
460 if (pipe_reference(&bo
->reference
, NULL
)) {
461 panfrost_destroy_bo(pan_screen(screen
), bo
);
466 panfrost_resource_destroy(struct pipe_screen
*screen
,
467 struct pipe_resource
*pt
)
469 struct panfrost_screen
*pscreen
= pan_screen(screen
);
470 struct panfrost_resource
*rsrc
= (struct panfrost_resource
*) pt
;
473 renderonly_scanout_destroy(rsrc
->scanout
, pscreen
->ro
);
476 panfrost_bo_unreference(screen
, rsrc
->bo
);
478 util_range_destroy(&rsrc
->valid_buffer_range
);
483 panfrost_transfer_map(struct pipe_context
*pctx
,
484 struct pipe_resource
*resource
,
486 unsigned usage
, /* a combination of PIPE_TRANSFER_x */
487 const struct pipe_box
*box
,
488 struct pipe_transfer
**out_transfer
)
490 int bytes_per_pixel
= util_format_get_blocksize(resource
->format
);
491 struct panfrost_resource
*rsrc
= pan_resource(resource
);
492 struct panfrost_bo
*bo
= rsrc
->bo
;
494 struct panfrost_gtransfer
*transfer
= rzalloc(pctx
, struct panfrost_gtransfer
);
495 transfer
->base
.level
= level
;
496 transfer
->base
.usage
= usage
;
497 transfer
->base
.box
= *box
;
499 pipe_resource_reference(&transfer
->base
.resource
, resource
);
501 *out_transfer
= &transfer
->base
;
503 /* Check if we're bound for rendering and this is a read pixels. If so,
504 * we need to flush */
506 struct panfrost_context
*ctx
= pan_context(pctx
);
507 struct pipe_framebuffer_state
*fb
= &ctx
->pipe_framebuffer
;
509 bool is_bound
= false;
511 for (unsigned c
= 0; c
< fb
->nr_cbufs
; ++c
) {
512 is_bound
|= fb
->cbufs
[c
]->texture
== resource
;
515 if (is_bound
&& (usage
& PIPE_TRANSFER_READ
)) {
517 panfrost_flush(pctx
, NULL
, PIPE_FLUSH_END_OF_FRAME
);
520 /* TODO: Respect usage flags */
522 if (usage
& PIPE_TRANSFER_DISCARD_WHOLE_RESOURCE
) {
523 /* TODO: reallocate */
524 //printf("debug: Missed reallocate\n");
525 } else if ((usage
& PIPE_TRANSFER_WRITE
)
526 && resource
->target
== PIPE_BUFFER
527 && !util_ranges_intersect(&rsrc
->valid_buffer_range
, box
->x
, box
->x
+ box
->width
)) {
528 /* No flush for writes to uninitialized */
529 } else if (!(usage
& PIPE_TRANSFER_UNSYNCHRONIZED
)) {
530 if (usage
& PIPE_TRANSFER_WRITE
) {
531 /* STUB: flush reading */
532 //printf("debug: missed reading flush %d\n", resource->target);
533 } else if (usage
& PIPE_TRANSFER_READ
) {
534 /* STUB: flush writing */
535 //printf("debug: missed writing flush %d (%d-%d)\n", resource->target, box->x, box->x + box->width);
537 /* Why are you even mapping?! */
541 if (bo
->layout
!= PAN_LINEAR
) {
542 /* Non-linear resources need to be indirectly mapped */
544 if (usage
& PIPE_TRANSFER_MAP_DIRECTLY
)
547 transfer
->base
.stride
= box
->width
* bytes_per_pixel
;
548 transfer
->base
.layer_stride
= transfer
->base
.stride
* box
->height
;
549 transfer
->map
= rzalloc_size(transfer
, transfer
->base
.layer_stride
* box
->depth
);
550 assert(box
->depth
== 1);
552 if ((usage
& PIPE_TRANSFER_READ
) && bo
->slices
[level
].initialized
) {
553 if (bo
->layout
== PAN_AFBC
) {
554 DBG("Unimplemented: reads from AFBC");
555 } else if (bo
->layout
== PAN_TILED
) {
556 panfrost_load_tiled_image(
558 bo
->cpu
+ bo
->slices
[level
].offset
,
560 transfer
->base
.stride
,
561 bo
->slices
[level
].stride
,
562 util_format_get_blocksize(resource
->format
));
566 return transfer
->map
;
568 transfer
->base
.stride
= bo
->slices
[level
].stride
;
569 transfer
->base
.layer_stride
= bo
->cubemap_stride
;
571 /* By mapping direct-write, we're implicitly already
572 * initialized (maybe), so be conservative */
574 if ((usage
& PIPE_TRANSFER_WRITE
) && (usage
& PIPE_TRANSFER_MAP_DIRECTLY
))
575 bo
->slices
[level
].initialized
= true;
578 + bo
->slices
[level
].offset
579 + transfer
->base
.box
.z
* bo
->cubemap_stride
580 + transfer
->base
.box
.y
* bo
->slices
[level
].stride
581 + transfer
->base
.box
.x
* bytes_per_pixel
;
586 panfrost_transfer_unmap(struct pipe_context
*pctx
,
587 struct pipe_transfer
*transfer
)
589 /* Gallium expects writeback here, so we tile */
591 struct panfrost_gtransfer
*trans
= pan_transfer(transfer
);
592 struct panfrost_resource
*prsrc
= (struct panfrost_resource
*) transfer
->resource
;
595 struct panfrost_bo
*bo
= prsrc
->bo
;
597 if (transfer
->usage
& PIPE_TRANSFER_WRITE
) {
598 unsigned level
= transfer
->level
;
599 bo
->slices
[level
].initialized
= true;
601 if (bo
->layout
== PAN_AFBC
) {
602 DBG("Unimplemented: writes to AFBC\n");
603 } else if (bo
->layout
== PAN_TILED
) {
604 assert(transfer
->box
.depth
== 1);
606 panfrost_store_tiled_image(
607 bo
->cpu
+ bo
->slices
[level
].offset
,
610 bo
->slices
[level
].stride
,
612 util_format_get_blocksize(prsrc
->base
.format
));
618 util_range_add(&prsrc
->valid_buffer_range
,
620 transfer
->box
.x
+ transfer
->box
.width
);
622 /* Derefence the resource */
623 pipe_resource_reference(&transfer
->resource
, NULL
);
625 /* Transfer itself is RALLOCed at the moment */
626 ralloc_free(transfer
);
630 panfrost_transfer_flush_region(struct pipe_context
*pctx
,
631 struct pipe_transfer
*transfer
,
632 const struct pipe_box
*box
)
634 struct panfrost_resource
*rsc
= pan_resource(transfer
->resource
);
636 if (transfer
->resource
->target
== PIPE_BUFFER
) {
637 util_range_add(&rsc
->valid_buffer_range
,
638 transfer
->box
.x
+ box
->x
,
639 transfer
->box
.x
+ box
->x
+ box
->width
);
643 static struct pb_slab
*
644 panfrost_slab_alloc(void *priv
, unsigned heap
, unsigned entry_size
, unsigned group_index
)
646 struct panfrost_screen
*screen
= (struct panfrost_screen
*) priv
;
647 struct panfrost_memory
*mem
= rzalloc(screen
, struct panfrost_memory
);
649 size_t slab_size
= (1 << (MAX_SLAB_ENTRY_SIZE
+ 1));
651 mem
->slab
.num_entries
= slab_size
/ entry_size
;
652 mem
->slab
.num_free
= mem
->slab
.num_entries
;
654 LIST_INITHEAD(&mem
->slab
.free
);
655 for (unsigned i
= 0; i
< mem
->slab
.num_entries
; ++i
) {
656 /* Create a slab entry */
657 struct panfrost_memory_entry
*entry
= rzalloc(mem
, struct panfrost_memory_entry
);
658 entry
->offset
= entry_size
* i
;
660 entry
->base
.slab
= &mem
->slab
;
661 entry
->base
.group_index
= group_index
;
663 LIST_ADDTAIL(&entry
->base
.head
, &mem
->slab
.free
);
666 /* Actually allocate the memory from kernel-space. Mapped, same_va, no
669 panfrost_drm_allocate_slab(screen
, mem
, slab_size
/ 4096, true, 0, 0, 0);
675 panfrost_slab_can_reclaim(void *priv
, struct pb_slab_entry
*entry
)
677 struct panfrost_memory_entry
*p_entry
= (struct panfrost_memory_entry
*) entry
;
678 return p_entry
->freed
;
682 panfrost_slab_free(void *priv
, struct pb_slab
*slab
)
684 struct panfrost_memory
*mem
= (struct panfrost_memory
*) slab
;
685 struct panfrost_screen
*screen
= (struct panfrost_screen
*) priv
;
687 panfrost_drm_free_slab(screen
, mem
);
692 panfrost_invalidate_resource(struct pipe_context
*pctx
, struct pipe_resource
*prsc
)
694 //DBG("TODO %s\n", __func__);
697 static enum pipe_format
698 panfrost_resource_get_internal_format(struct pipe_resource
*prsrc
)
700 return prsrc
->format
;
704 panfrost_generate_mipmap(
705 struct pipe_context
*pctx
,
706 struct pipe_resource
*prsrc
,
707 enum pipe_format format
,
710 unsigned first_layer
,
713 struct panfrost_context
*ctx
= pan_context(pctx
);
714 struct panfrost_resource
*rsrc
= pan_resource(prsrc
);
716 /* Generating a mipmap invalidates the written levels, so make that
717 * explicit so we don't try to wallpaper them back and end up with
718 * u_blitter recursion */
721 for (unsigned l
= base_level
+ 1; l
<= last_level
; ++l
)
722 rsrc
->bo
->slices
[l
].initialized
= false;
724 /* Beyond that, we just delegate the hard stuff. We're careful to
725 * include flushes on both ends to make sure the data is really valid.
726 * We could be doing a lot better perf-wise, especially once we have
727 * reorder-type optimizations in place. But for now prioritize
730 struct panfrost_job
*job
= panfrost_get_job_for_fbo(ctx
);
731 bool has_draws
= job
->last_job
.gpu
;
734 panfrost_flush(pctx
, NULL
, PIPE_FLUSH_END_OF_FRAME
);
736 /* We've flushed the original buffer if needed, now trigger a blit */
738 bool blit_res
= util_gen_mipmap(
740 base_level
, last_level
,
741 first_layer
, last_layer
,
742 PIPE_TEX_FILTER_LINEAR
);
744 /* If the blit was successful, flush once more. If it wasn't, well, let
745 * the state tracker deal with it. */
748 panfrost_flush(pctx
, NULL
, PIPE_FLUSH_END_OF_FRAME
);
753 /* Computes the address to a texture at a particular slice */
756 panfrost_get_texture_address(
757 struct panfrost_resource
*rsrc
,
758 unsigned level
, unsigned face
)
760 unsigned level_offset
= rsrc
->bo
->slices
[level
].offset
;
761 unsigned face_offset
= face
* rsrc
->bo
->cubemap_stride
;
763 return rsrc
->bo
->gpu
+ level_offset
+ face_offset
;
767 panfrost_resource_set_stencil(struct pipe_resource
*prsrc
,
768 struct pipe_resource
*stencil
)
770 pan_resource(prsrc
)->separate_stencil
= pan_resource(stencil
);
773 static struct pipe_resource
*
774 panfrost_resource_get_stencil(struct pipe_resource
*prsrc
)
776 return &pan_resource(prsrc
)->separate_stencil
->base
;
779 static const struct u_transfer_vtbl transfer_vtbl
= {
780 .resource_create
= panfrost_resource_create
,
781 .resource_destroy
= panfrost_resource_destroy
,
782 .transfer_map
= panfrost_transfer_map
,
783 .transfer_unmap
= panfrost_transfer_unmap
,
784 .transfer_flush_region
= panfrost_transfer_flush_region
,
785 .get_internal_format
= panfrost_resource_get_internal_format
,
786 .set_stencil
= panfrost_resource_set_stencil
,
787 .get_stencil
= panfrost_resource_get_stencil
,
791 panfrost_resource_screen_init(struct panfrost_screen
*pscreen
)
793 //pscreen->base.resource_create_with_modifiers =
794 // panfrost_resource_create_with_modifiers;
795 pscreen
->base
.resource_create
= u_transfer_helper_resource_create
;
796 pscreen
->base
.resource_destroy
= u_transfer_helper_resource_destroy
;
797 pscreen
->base
.resource_from_handle
= panfrost_resource_from_handle
;
798 pscreen
->base
.resource_get_handle
= panfrost_resource_get_handle
;
799 pscreen
->base
.transfer_helper
= u_transfer_helper_create(&transfer_vtbl
,
803 pb_slabs_init(&pscreen
->slabs
,
807 3, /* Number of heaps */
811 panfrost_slab_can_reclaim
,
817 panfrost_resource_screen_deinit(struct panfrost_screen
*pscreen
)
819 pb_slabs_deinit(&pscreen
->slabs
);
823 panfrost_resource_context_init(struct pipe_context
*pctx
)
825 pctx
->transfer_map
= u_transfer_helper_transfer_map
;
826 pctx
->transfer_flush_region
= u_transfer_helper_transfer_flush_region
;
827 pctx
->transfer_unmap
= u_transfer_helper_transfer_unmap
;
828 pctx
->buffer_subdata
= u_default_buffer_subdata
;
829 pctx
->create_surface
= panfrost_create_surface
;
830 pctx
->surface_destroy
= panfrost_surface_destroy
;
831 pctx
->resource_copy_region
= util_resource_copy_region
;
832 pctx
->blit
= panfrost_blit
;
833 pctx
->generate_mipmap
= panfrost_generate_mipmap
;
834 pctx
->flush_resource
= panfrost_flush_resource
;
835 pctx
->invalidate_resource
= panfrost_invalidate_resource
;
836 pctx
->transfer_flush_region
= u_transfer_helper_transfer_flush_region
;
837 pctx
->buffer_subdata
= u_default_buffer_subdata
;
838 pctx
->texture_subdata
= u_default_texture_subdata
;