2 * Copyright (C) 2008 VMware, Inc.
3 * Copyright (C) 2014 Broadcom
4 * Copyright (C) 2018-2019 Alyssa Rosenzweig
5 * Copyright (C) 2019 Collabora
7 * Permission is hereby granted, free of charge, to any person obtaining a
8 * copy of this software and associated documentation files (the "Software"),
9 * to deal in the Software without restriction, including without limitation
10 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
11 * and/or sell copies of the Software, and to permit persons to whom the
12 * Software is furnished to do so, subject to the following conditions:
14 * The above copyright notice and this permission notice (including the next
15 * paragraph) shall be included in all copies or substantial portions of the
18 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
19 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
20 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
21 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
22 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
23 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
26 * Authors (Collabora):
27 * Tomeu Vizoso <tomeu.vizoso@collabora.com>
28 * Alyssa Rosenzweig <alyssa.rosenzweig@collabora.com>
34 #include "drm-uapi/drm_fourcc.h"
36 #include "state_tracker/winsys_handle.h"
37 #include "util/u_format.h"
38 #include "util/u_memory.h"
39 #include "util/u_surface.h"
40 #include "util/u_transfer.h"
41 #include "util/u_transfer_helper.h"
42 #include "util/u_gen_mipmap.h"
44 #include "pan_context.h"
45 #include "pan_screen.h"
46 #include "pan_resource.h"
48 #include "pan_tiling.h"
50 static struct pipe_resource
*
51 panfrost_resource_from_handle(struct pipe_screen
*pscreen
,
52 const struct pipe_resource
*templat
,
53 struct winsys_handle
*whandle
,
56 struct panfrost_screen
*screen
= pan_screen(pscreen
);
57 struct panfrost_resource
*rsc
;
58 struct pipe_resource
*prsc
;
60 assert(whandle
->type
== WINSYS_HANDLE_TYPE_FD
);
62 rsc
= rzalloc(pscreen
, struct panfrost_resource
);
70 pipe_reference_init(&prsc
->reference
, 1);
71 prsc
->screen
= pscreen
;
73 rsc
->bo
= panfrost_drm_import_bo(screen
, whandle
);
74 rsc
->bo
->slices
[0].stride
= whandle
->stride
;
75 rsc
->bo
->slices
[0].initialized
= true;
79 renderonly_create_gpu_import_for_resource(prsc
, screen
->ro
, NULL
);
80 /* failure is expected in some cases.. */
87 panfrost_resource_get_handle(struct pipe_screen
*pscreen
,
88 struct pipe_context
*ctx
,
89 struct pipe_resource
*pt
,
90 struct winsys_handle
*handle
,
93 struct panfrost_screen
*screen
= pan_screen(pscreen
);
94 struct panfrost_resource
*rsrc
= (struct panfrost_resource
*) pt
;
95 struct renderonly_scanout
*scanout
= rsrc
->scanout
;
97 handle
->modifier
= DRM_FORMAT_MOD_INVALID
;
99 if (handle
->type
== WINSYS_HANDLE_TYPE_SHARED
) {
101 } else if (handle
->type
== WINSYS_HANDLE_TYPE_KMS
) {
102 if (renderonly_get_handle(scanout
, handle
))
105 handle
->handle
= rsrc
->bo
->gem_handle
;
106 handle
->stride
= rsrc
->bo
->slices
[0].stride
;
108 } else if (handle
->type
== WINSYS_HANDLE_TYPE_FD
) {
110 struct drm_prime_handle args
= {
111 .handle
= scanout
->handle
,
112 .flags
= DRM_CLOEXEC
,
115 int ret
= drmIoctl(screen
->ro
->kms_fd
, DRM_IOCTL_PRIME_HANDLE_TO_FD
, &args
);
119 handle
->stride
= scanout
->stride
;
120 handle
->handle
= args
.fd
;
124 return panfrost_drm_export_bo(screen
, rsrc
->bo
->gem_handle
,
125 rsrc
->bo
->slices
[0].stride
,
133 panfrost_flush_resource(struct pipe_context
*pctx
, struct pipe_resource
*prsc
)
135 //DBG("TODO %s\n", __func__);
138 static struct pipe_surface
*
139 panfrost_create_surface(struct pipe_context
*pipe
,
140 struct pipe_resource
*pt
,
141 const struct pipe_surface
*surf_tmpl
)
143 struct pipe_surface
*ps
= NULL
;
145 ps
= rzalloc(pipe
, struct pipe_surface
);
148 pipe_reference_init(&ps
->reference
, 1);
149 pipe_resource_reference(&ps
->texture
, pt
);
151 ps
->format
= surf_tmpl
->format
;
153 if (pt
->target
!= PIPE_BUFFER
) {
154 assert(surf_tmpl
->u
.tex
.level
<= pt
->last_level
);
155 ps
->width
= u_minify(pt
->width0
, surf_tmpl
->u
.tex
.level
);
156 ps
->height
= u_minify(pt
->height0
, surf_tmpl
->u
.tex
.level
);
157 ps
->u
.tex
.level
= surf_tmpl
->u
.tex
.level
;
158 ps
->u
.tex
.first_layer
= surf_tmpl
->u
.tex
.first_layer
;
159 ps
->u
.tex
.last_layer
= surf_tmpl
->u
.tex
.last_layer
;
161 /* setting width as number of elements should get us correct renderbuffer width */
162 ps
->width
= surf_tmpl
->u
.buf
.last_element
- surf_tmpl
->u
.buf
.first_element
+ 1;
163 ps
->height
= pt
->height0
;
164 ps
->u
.buf
.first_element
= surf_tmpl
->u
.buf
.first_element
;
165 ps
->u
.buf
.last_element
= surf_tmpl
->u
.buf
.last_element
;
166 assert(ps
->u
.buf
.first_element
<= ps
->u
.buf
.last_element
);
167 assert(ps
->u
.buf
.last_element
< ps
->width
);
175 panfrost_surface_destroy(struct pipe_context
*pipe
,
176 struct pipe_surface
*surf
)
178 assert(surf
->texture
);
179 pipe_resource_reference(&surf
->texture
, NULL
);
184 panfrost_setup_slices(const struct pipe_resource
*tmpl
, struct panfrost_bo
*bo
)
186 unsigned width
= tmpl
->width0
;
187 unsigned height
= tmpl
->height0
;
188 unsigned depth
= tmpl
->depth0
;
189 unsigned bytes_per_pixel
= util_format_get_blocksize(tmpl
->format
);
193 /* Tiled operates blockwise; linear is packed. Also, anything
194 * we render to has to be tile-aligned. Maybe not strictly
195 * necessary, but we're not *that* pressed for memory and it
196 * makes code a lot simpler */
198 bool renderable
= tmpl
->bind
&
199 (PIPE_BIND_RENDER_TARGET
| PIPE_BIND_DEPTH_STENCIL
);
200 bool tiled
= bo
->layout
== PAN_TILED
;
201 bool should_align
= renderable
|| tiled
;
203 /* We don't know how to specify a 2D stride for 3D textures */
205 bool can_align_stride
=
206 tmpl
->target
!= PIPE_TEXTURE_3D
;
208 should_align
&= can_align_stride
;
211 unsigned size_2d
= 0;
213 for (unsigned l
= 0; l
<= tmpl
->last_level
; ++l
) {
214 struct panfrost_slice
*slice
= &bo
->slices
[l
];
216 unsigned effective_width
= width
;
217 unsigned effective_height
= height
;
218 unsigned effective_depth
= depth
;
221 effective_width
= ALIGN(effective_width
, 16);
222 effective_height
= ALIGN(effective_height
, 16);
224 /* We don't need to align depth */
227 slice
->offset
= offset
;
229 /* Compute the would-be stride */
230 unsigned stride
= bytes_per_pixel
* effective_width
;
232 /* ..but cache-line align it for performance */
233 if (can_align_stride
&& bo
->layout
== PAN_LINEAR
)
234 stride
= ALIGN(stride
, 64);
236 slice
->stride
= stride
;
238 unsigned slice_one_size
= slice
->stride
* effective_height
;
239 unsigned slice_full_size
= slice_one_size
* effective_depth
;
241 /* Report 2D size for 3D texturing */
244 size_2d
= slice_one_size
;
246 offset
+= slice_full_size
;
248 width
= u_minify(width
, 1);
249 height
= u_minify(height
, 1);
250 depth
= u_minify(depth
, 1);
253 assert(tmpl
->array_size
);
255 if (tmpl
->target
!= PIPE_TEXTURE_3D
) {
256 /* Arrays and cubemaps have the entire miptree duplicated */
258 bo
->cubemap_stride
= ALIGN(offset
, 64);
259 bo
->size
= ALIGN(bo
->cubemap_stride
* tmpl
->array_size
, 4096);
261 /* 3D strides across the 2D layers */
262 assert(tmpl
->array_size
== 1);
264 bo
->cubemap_stride
= size_2d
;
265 bo
->size
= ALIGN(offset
, 4096);
269 static struct panfrost_bo
*
270 panfrost_create_bo(struct panfrost_screen
*screen
, const struct pipe_resource
*template)
272 struct panfrost_bo
*bo
= rzalloc(screen
, struct panfrost_bo
);
273 pipe_reference_init(&bo
->reference
, 1);
275 /* Based on the usage, figure out what storing will be used. There are
278 * Linear: the basic format, bad for memory bandwidth, bad for cache
279 * use. Zero-copy, though. Renderable.
281 * Tiled: Not compressed, but cache-optimized. Expensive to write into
282 * (due to software tiling), but cheap to sample from. Ideal for most
285 * AFBC: Compressed and renderable (so always desirable for non-scanout
286 * rendertargets). Cheap to sample from. The format is black box, so we
287 * can't read/write from software.
290 /* Tiling textures is almost always faster, unless we only use it once */
292 bool is_texture
= (template->bind
& PIPE_BIND_SAMPLER_VIEW
);
293 bool is_2d
= template->depth0
== 1 && template->array_size
== 1;
294 bool is_streaming
= (template->usage
!= PIPE_USAGE_STREAM
);
296 bool should_tile
= is_streaming
&& is_texture
&& is_2d
;
298 /* Depth/stencil can't be tiled, only linear or AFBC */
299 should_tile
&= !(template->bind
& PIPE_BIND_DEPTH_STENCIL
);
301 /* Set the layout appropriately */
302 bo
->layout
= should_tile
? PAN_TILED
: PAN_LINEAR
;
304 panfrost_setup_slices(template, bo
);
306 if (bo
->layout
== PAN_TILED
|| bo
->layout
== PAN_LINEAR
) {
307 struct panfrost_memory mem
;
309 panfrost_drm_allocate_slab(screen
, &mem
, bo
->size
/ 4096, true, 0, 0, 0);
313 bo
->gem_handle
= mem
.gem_handle
;
319 static struct pipe_resource
*
320 panfrost_resource_create(struct pipe_screen
*screen
,
321 const struct pipe_resource
*template)
323 struct panfrost_resource
*so
= rzalloc(screen
, struct panfrost_resource
);
324 struct panfrost_screen
*pscreen
= (struct panfrost_screen
*) screen
;
326 so
->base
= *template;
327 so
->base
.screen
= screen
;
329 pipe_reference_init(&so
->base
.reference
, 1);
331 /* Make sure we're familiar */
332 switch (template->target
) {
334 case PIPE_TEXTURE_1D
:
335 case PIPE_TEXTURE_2D
:
336 case PIPE_TEXTURE_3D
:
337 case PIPE_TEXTURE_CUBE
:
338 case PIPE_TEXTURE_RECT
:
339 case PIPE_TEXTURE_2D_ARRAY
:
342 DBG("Unknown texture target %d\n", template->target
);
346 util_range_init(&so
->valid_buffer_range
);
348 if (template->bind
& PIPE_BIND_DISPLAY_TARGET
||
349 template->bind
& PIPE_BIND_SCANOUT
||
350 template->bind
& PIPE_BIND_SHARED
) {
351 struct pipe_resource scanout_templat
= *template;
352 struct renderonly_scanout
*scanout
;
353 struct winsys_handle handle
;
355 scanout
= renderonly_scanout_for_resource(&scanout_templat
,
356 pscreen
->ro
, &handle
);
360 assert(handle
.type
== WINSYS_HANDLE_TYPE_FD
);
361 /* TODO: handle modifiers? */
362 so
= pan_resource(screen
->resource_from_handle(screen
, template,
364 PIPE_HANDLE_USAGE_FRAMEBUFFER_WRITE
));
365 close(handle
.handle
);
369 so
->scanout
= scanout
;
370 pscreen
->display_target
= so
;
372 so
->bo
= panfrost_create_bo(pscreen
, template);
375 return (struct pipe_resource
*)so
;
379 panfrost_destroy_bo(struct panfrost_screen
*screen
, struct panfrost_bo
*bo
)
381 if ((bo
->layout
== PAN_LINEAR
|| bo
->layout
== PAN_TILED
) &&
383 struct panfrost_memory mem
= {
387 .gem_handle
= bo
->gem_handle
,
390 panfrost_drm_free_slab(screen
, &mem
);
393 if (bo
->layout
== PAN_AFBC
) {
395 DBG("--leaking afbc (%d bytes)--\n", bo
->afbc_metadata_size
);
398 if (bo
->has_checksum
) {
399 struct panfrost_memory mem
= {
400 .cpu
= bo
->checksum_slab
.cpu
,
401 .gpu
= bo
->checksum_slab
.gpu
,
402 .size
= bo
->checksum_slab
.size
,
403 .gem_handle
= bo
->checksum_slab
.gem_handle
,
406 panfrost_drm_free_slab(screen
, &mem
);
410 panfrost_drm_free_imported_bo(screen
, bo
);
417 panfrost_bo_reference(struct panfrost_bo
*bo
)
419 pipe_reference(NULL
, &bo
->reference
);
423 panfrost_bo_unreference(struct pipe_screen
*screen
, struct panfrost_bo
*bo
)
425 /* When the reference count goes to zero, we need to cleanup */
427 if (pipe_reference(&bo
->reference
, NULL
)) {
428 panfrost_destroy_bo(pan_screen(screen
), bo
);
433 panfrost_resource_destroy(struct pipe_screen
*screen
,
434 struct pipe_resource
*pt
)
436 struct panfrost_screen
*pscreen
= pan_screen(screen
);
437 struct panfrost_resource
*rsrc
= (struct panfrost_resource
*) pt
;
440 renderonly_scanout_destroy(rsrc
->scanout
, pscreen
->ro
);
443 panfrost_bo_unreference(screen
, rsrc
->bo
);
445 util_range_destroy(&rsrc
->valid_buffer_range
);
450 panfrost_transfer_map(struct pipe_context
*pctx
,
451 struct pipe_resource
*resource
,
453 unsigned usage
, /* a combination of PIPE_TRANSFER_x */
454 const struct pipe_box
*box
,
455 struct pipe_transfer
**out_transfer
)
457 int bytes_per_pixel
= util_format_get_blocksize(resource
->format
);
458 struct panfrost_resource
*rsrc
= pan_resource(resource
);
459 struct panfrost_bo
*bo
= rsrc
->bo
;
461 struct panfrost_gtransfer
*transfer
= rzalloc(pctx
, struct panfrost_gtransfer
);
462 transfer
->base
.level
= level
;
463 transfer
->base
.usage
= usage
;
464 transfer
->base
.box
= *box
;
466 pipe_resource_reference(&transfer
->base
.resource
, resource
);
468 *out_transfer
= &transfer
->base
;
470 /* Check if we're bound for rendering and this is a read pixels. If so,
471 * we need to flush */
473 struct panfrost_context
*ctx
= pan_context(pctx
);
474 struct pipe_framebuffer_state
*fb
= &ctx
->pipe_framebuffer
;
476 bool is_bound
= false;
478 for (unsigned c
= 0; c
< fb
->nr_cbufs
; ++c
) {
479 is_bound
|= fb
->cbufs
[c
]->texture
== resource
;
482 if (is_bound
&& (usage
& PIPE_TRANSFER_READ
)) {
484 panfrost_flush(pctx
, NULL
, PIPE_FLUSH_END_OF_FRAME
);
487 /* TODO: Respect usage flags */
489 if (usage
& PIPE_TRANSFER_DISCARD_WHOLE_RESOURCE
) {
490 /* TODO: reallocate */
491 //printf("debug: Missed reallocate\n");
492 } else if ((usage
& PIPE_TRANSFER_WRITE
)
493 && resource
->target
== PIPE_BUFFER
494 && !util_ranges_intersect(&rsrc
->valid_buffer_range
, box
->x
, box
->x
+ box
->width
)) {
495 /* No flush for writes to uninitialized */
496 } else if (!(usage
& PIPE_TRANSFER_UNSYNCHRONIZED
)) {
497 if (usage
& PIPE_TRANSFER_WRITE
) {
498 /* STUB: flush reading */
499 //printf("debug: missed reading flush %d\n", resource->target);
500 } else if (usage
& PIPE_TRANSFER_READ
) {
501 /* STUB: flush writing */
502 //printf("debug: missed writing flush %d (%d-%d)\n", resource->target, box->x, box->x + box->width);
504 /* Why are you even mapping?! */
508 if (bo
->layout
!= PAN_LINEAR
) {
509 /* Non-linear resources need to be indirectly mapped */
511 if (usage
& PIPE_TRANSFER_MAP_DIRECTLY
)
514 transfer
->base
.stride
= box
->width
* bytes_per_pixel
;
515 transfer
->base
.layer_stride
= transfer
->base
.stride
* box
->height
;
516 transfer
->map
= rzalloc_size(transfer
, transfer
->base
.layer_stride
* box
->depth
);
517 assert(box
->depth
== 1);
519 if ((usage
& PIPE_TRANSFER_READ
) && bo
->slices
[level
].initialized
) {
520 if (bo
->layout
== PAN_AFBC
) {
521 DBG("Unimplemented: reads from AFBC");
522 } else if (bo
->layout
== PAN_TILED
) {
523 panfrost_load_tiled_image(
525 bo
->cpu
+ bo
->slices
[level
].offset
,
527 transfer
->base
.stride
,
528 bo
->slices
[level
].stride
,
529 util_format_get_blocksize(resource
->format
));
533 return transfer
->map
;
535 transfer
->base
.stride
= bo
->slices
[level
].stride
;
536 transfer
->base
.layer_stride
= bo
->cubemap_stride
;
538 /* By mapping direct-write, we're implicitly already
539 * initialized (maybe), so be conservative */
541 if ((usage
& PIPE_TRANSFER_WRITE
) && (usage
& PIPE_TRANSFER_MAP_DIRECTLY
))
542 bo
->slices
[level
].initialized
= true;
545 + bo
->slices
[level
].offset
546 + transfer
->base
.box
.z
* bo
->cubemap_stride
547 + transfer
->base
.box
.y
* bo
->slices
[level
].stride
548 + transfer
->base
.box
.x
* bytes_per_pixel
;
553 panfrost_transfer_unmap(struct pipe_context
*pctx
,
554 struct pipe_transfer
*transfer
)
556 /* Gallium expects writeback here, so we tile */
558 struct panfrost_gtransfer
*trans
= pan_transfer(transfer
);
559 struct panfrost_resource
*prsrc
= (struct panfrost_resource
*) transfer
->resource
;
562 struct panfrost_bo
*bo
= prsrc
->bo
;
564 if (transfer
->usage
& PIPE_TRANSFER_WRITE
) {
565 unsigned level
= transfer
->level
;
566 bo
->slices
[level
].initialized
= true;
568 if (bo
->layout
== PAN_AFBC
) {
569 DBG("Unimplemented: writes to AFBC\n");
570 } else if (bo
->layout
== PAN_TILED
) {
571 assert(transfer
->box
.depth
== 1);
573 panfrost_store_tiled_image(
574 bo
->cpu
+ bo
->slices
[level
].offset
,
577 bo
->slices
[level
].stride
,
579 util_format_get_blocksize(prsrc
->base
.format
));
585 util_range_add(&prsrc
->valid_buffer_range
,
587 transfer
->box
.x
+ transfer
->box
.width
);
589 /* Derefence the resource */
590 pipe_resource_reference(&transfer
->resource
, NULL
);
592 /* Transfer itself is RALLOCed at the moment */
593 ralloc_free(transfer
);
597 panfrost_transfer_flush_region(struct pipe_context
*pctx
,
598 struct pipe_transfer
*transfer
,
599 const struct pipe_box
*box
)
601 struct panfrost_resource
*rsc
= pan_resource(transfer
->resource
);
603 if (transfer
->resource
->target
== PIPE_BUFFER
) {
604 util_range_add(&rsc
->valid_buffer_range
,
605 transfer
->box
.x
+ box
->x
,
606 transfer
->box
.x
+ box
->x
+ box
->width
);
610 static struct pb_slab
*
611 panfrost_slab_alloc(void *priv
, unsigned heap
, unsigned entry_size
, unsigned group_index
)
613 struct panfrost_screen
*screen
= (struct panfrost_screen
*) priv
;
614 struct panfrost_memory
*mem
= rzalloc(screen
, struct panfrost_memory
);
616 size_t slab_size
= (1 << (MAX_SLAB_ENTRY_SIZE
+ 1));
618 mem
->slab
.num_entries
= slab_size
/ entry_size
;
619 mem
->slab
.num_free
= mem
->slab
.num_entries
;
621 LIST_INITHEAD(&mem
->slab
.free
);
622 for (unsigned i
= 0; i
< mem
->slab
.num_entries
; ++i
) {
623 /* Create a slab entry */
624 struct panfrost_memory_entry
*entry
= rzalloc(mem
, struct panfrost_memory_entry
);
625 entry
->offset
= entry_size
* i
;
627 entry
->base
.slab
= &mem
->slab
;
628 entry
->base
.group_index
= group_index
;
630 LIST_ADDTAIL(&entry
->base
.head
, &mem
->slab
.free
);
633 /* Actually allocate the memory from kernel-space. Mapped, same_va, no
636 panfrost_drm_allocate_slab(screen
, mem
, slab_size
/ 4096, true, 0, 0, 0);
642 panfrost_slab_can_reclaim(void *priv
, struct pb_slab_entry
*entry
)
644 struct panfrost_memory_entry
*p_entry
= (struct panfrost_memory_entry
*) entry
;
645 return p_entry
->freed
;
649 panfrost_slab_free(void *priv
, struct pb_slab
*slab
)
651 struct panfrost_memory
*mem
= (struct panfrost_memory
*) slab
;
652 struct panfrost_screen
*screen
= (struct panfrost_screen
*) priv
;
654 panfrost_drm_free_slab(screen
, mem
);
659 panfrost_invalidate_resource(struct pipe_context
*pctx
, struct pipe_resource
*prsc
)
661 //DBG("TODO %s\n", __func__);
664 static enum pipe_format
665 panfrost_resource_get_internal_format(struct pipe_resource
*prsrc
)
667 return prsrc
->format
;
671 panfrost_generate_mipmap(
672 struct pipe_context
*pctx
,
673 struct pipe_resource
*prsrc
,
674 enum pipe_format format
,
677 unsigned first_layer
,
680 struct panfrost_context
*ctx
= pan_context(pctx
);
681 struct panfrost_resource
*rsrc
= pan_resource(prsrc
);
683 /* Generating a mipmap invalidates the written levels, so make that
684 * explicit so we don't try to wallpaper them back and end up with
685 * u_blitter recursion */
688 for (unsigned l
= base_level
+ 1; l
<= last_level
; ++l
)
689 rsrc
->bo
->slices
[l
].initialized
= false;
691 /* Beyond that, we just delegate the hard stuff. We're careful to
692 * include flushes on both ends to make sure the data is really valid.
693 * We could be doing a lot better perf-wise, especially once we have
694 * reorder-type optimizations in place. But for now prioritize
697 struct panfrost_job
*job
= panfrost_get_job_for_fbo(ctx
);
698 bool has_draws
= job
->last_job
.gpu
;
701 panfrost_flush(pctx
, NULL
, PIPE_FLUSH_END_OF_FRAME
);
703 /* We've flushed the original buffer if needed, now trigger a blit */
705 bool blit_res
= util_gen_mipmap(
707 base_level
, last_level
,
708 first_layer
, last_layer
,
709 PIPE_TEX_FILTER_LINEAR
);
711 /* If the blit was successful, flush once more. If it wasn't, well, let
712 * the state tracker deal with it. */
715 panfrost_flush(pctx
, NULL
, PIPE_FLUSH_END_OF_FRAME
);
721 panfrost_resource_set_stencil(struct pipe_resource
*prsrc
,
722 struct pipe_resource
*stencil
)
724 pan_resource(prsrc
)->separate_stencil
= pan_resource(stencil
);
727 static struct pipe_resource
*
728 panfrost_resource_get_stencil(struct pipe_resource
*prsrc
)
730 return &pan_resource(prsrc
)->separate_stencil
->base
;
733 static const struct u_transfer_vtbl transfer_vtbl
= {
734 .resource_create
= panfrost_resource_create
,
735 .resource_destroy
= panfrost_resource_destroy
,
736 .transfer_map
= panfrost_transfer_map
,
737 .transfer_unmap
= panfrost_transfer_unmap
,
738 .transfer_flush_region
= panfrost_transfer_flush_region
,
739 .get_internal_format
= panfrost_resource_get_internal_format
,
740 .set_stencil
= panfrost_resource_set_stencil
,
741 .get_stencil
= panfrost_resource_get_stencil
,
745 panfrost_resource_screen_init(struct panfrost_screen
*pscreen
)
747 //pscreen->base.resource_create_with_modifiers =
748 // panfrost_resource_create_with_modifiers;
749 pscreen
->base
.resource_create
= u_transfer_helper_resource_create
;
750 pscreen
->base
.resource_destroy
= u_transfer_helper_resource_destroy
;
751 pscreen
->base
.resource_from_handle
= panfrost_resource_from_handle
;
752 pscreen
->base
.resource_get_handle
= panfrost_resource_get_handle
;
753 pscreen
->base
.transfer_helper
= u_transfer_helper_create(&transfer_vtbl
,
757 pb_slabs_init(&pscreen
->slabs
,
761 3, /* Number of heaps */
765 panfrost_slab_can_reclaim
,
771 panfrost_resource_screen_deinit(struct panfrost_screen
*pscreen
)
773 pb_slabs_deinit(&pscreen
->slabs
);
777 panfrost_resource_context_init(struct pipe_context
*pctx
)
779 pctx
->transfer_map
= u_transfer_helper_transfer_map
;
780 pctx
->transfer_flush_region
= u_transfer_helper_transfer_flush_region
;
781 pctx
->transfer_unmap
= u_transfer_helper_transfer_unmap
;
782 pctx
->buffer_subdata
= u_default_buffer_subdata
;
783 pctx
->create_surface
= panfrost_create_surface
;
784 pctx
->surface_destroy
= panfrost_surface_destroy
;
785 pctx
->resource_copy_region
= util_resource_copy_region
;
786 pctx
->blit
= panfrost_blit
;
787 pctx
->generate_mipmap
= panfrost_generate_mipmap
;
788 pctx
->flush_resource
= panfrost_flush_resource
;
789 pctx
->invalidate_resource
= panfrost_invalidate_resource
;
790 pctx
->transfer_flush_region
= u_transfer_helper_transfer_flush_region
;
791 pctx
->buffer_subdata
= u_default_buffer_subdata
;
792 pctx
->texture_subdata
= u_default_texture_subdata
;