2 * Copyright (C) 2008 VMware, Inc.
3 * Copyright (C) 2014 Broadcom
4 * Copyright (C) 2018-2019 Alyssa Rosenzweig
5 * Copyright (C) 2019 Collabora, Ltd.
7 * Permission is hereby granted, free of charge, to any person obtaining a
8 * copy of this software and associated documentation files (the "Software"),
9 * to deal in the Software without restriction, including without limitation
10 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
11 * and/or sell copies of the Software, and to permit persons to whom the
12 * Software is furnished to do so, subject to the following conditions:
14 * The above copyright notice and this permission notice (including the next
15 * paragraph) shall be included in all copies or substantial portions of the
18 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
19 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
20 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
21 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
22 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
23 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
26 * Authors (Collabora):
27 * Tomeu Vizoso <tomeu.vizoso@collabora.com>
28 * Alyssa Rosenzweig <alyssa.rosenzweig@collabora.com>
34 #include "drm-uapi/drm_fourcc.h"
36 #include "state_tracker/winsys_handle.h"
37 #include "util/u_format.h"
38 #include "util/u_memory.h"
39 #include "util/u_surface.h"
40 #include "util/u_transfer.h"
41 #include "util/u_transfer_helper.h"
42 #include "util/u_gen_mipmap.h"
44 #include "pan_context.h"
45 #include "pan_screen.h"
46 #include "pan_resource.h"
48 #include "pan_tiling.h"
50 static struct pipe_resource
*
51 panfrost_resource_from_handle(struct pipe_screen
*pscreen
,
52 const struct pipe_resource
*templat
,
53 struct winsys_handle
*whandle
,
56 struct panfrost_screen
*screen
= pan_screen(pscreen
);
57 struct panfrost_resource
*rsc
;
58 struct pipe_resource
*prsc
;
60 assert(whandle
->type
== WINSYS_HANDLE_TYPE_FD
);
62 rsc
= rzalloc(pscreen
, struct panfrost_resource
);
70 pipe_reference_init(&prsc
->reference
, 1);
71 prsc
->screen
= pscreen
;
73 rsc
->bo
= panfrost_drm_import_bo(screen
, whandle
->handle
);
74 rsc
->slices
[0].stride
= whandle
->stride
;
75 rsc
->slices
[0].initialized
= true;
79 renderonly_create_gpu_import_for_resource(prsc
, screen
->ro
, NULL
);
80 /* failure is expected in some cases.. */
87 panfrost_resource_get_handle(struct pipe_screen
*pscreen
,
88 struct pipe_context
*ctx
,
89 struct pipe_resource
*pt
,
90 struct winsys_handle
*handle
,
93 struct panfrost_screen
*screen
= pan_screen(pscreen
);
94 struct panfrost_resource
*rsrc
= (struct panfrost_resource
*) pt
;
95 struct renderonly_scanout
*scanout
= rsrc
->scanout
;
97 handle
->modifier
= DRM_FORMAT_MOD_INVALID
;
99 if (handle
->type
== WINSYS_HANDLE_TYPE_SHARED
) {
101 } else if (handle
->type
== WINSYS_HANDLE_TYPE_KMS
) {
102 if (renderonly_get_handle(scanout
, handle
))
105 handle
->handle
= rsrc
->bo
->gem_handle
;
106 handle
->stride
= rsrc
->slices
[0].stride
;
108 } else if (handle
->type
== WINSYS_HANDLE_TYPE_FD
) {
110 struct drm_prime_handle args
= {
111 .handle
= scanout
->handle
,
112 .flags
= DRM_CLOEXEC
,
115 int ret
= drmIoctl(screen
->ro
->kms_fd
, DRM_IOCTL_PRIME_HANDLE_TO_FD
, &args
);
119 handle
->stride
= scanout
->stride
;
120 handle
->handle
= args
.fd
;
124 int fd
= panfrost_drm_export_bo(screen
, rsrc
->bo
);
130 handle
->stride
= rsrc
->slices
[0].stride
;
139 panfrost_flush_resource(struct pipe_context
*pctx
, struct pipe_resource
*prsc
)
141 //DBG("TODO %s\n", __func__);
144 static struct pipe_surface
*
145 panfrost_create_surface(struct pipe_context
*pipe
,
146 struct pipe_resource
*pt
,
147 const struct pipe_surface
*surf_tmpl
)
149 struct pipe_surface
*ps
= NULL
;
151 ps
= rzalloc(pipe
, struct pipe_surface
);
154 pipe_reference_init(&ps
->reference
, 1);
155 pipe_resource_reference(&ps
->texture
, pt
);
157 ps
->format
= surf_tmpl
->format
;
159 if (pt
->target
!= PIPE_BUFFER
) {
160 assert(surf_tmpl
->u
.tex
.level
<= pt
->last_level
);
161 ps
->width
= u_minify(pt
->width0
, surf_tmpl
->u
.tex
.level
);
162 ps
->height
= u_minify(pt
->height0
, surf_tmpl
->u
.tex
.level
);
163 ps
->u
.tex
.level
= surf_tmpl
->u
.tex
.level
;
164 ps
->u
.tex
.first_layer
= surf_tmpl
->u
.tex
.first_layer
;
165 ps
->u
.tex
.last_layer
= surf_tmpl
->u
.tex
.last_layer
;
167 /* setting width as number of elements should get us correct renderbuffer width */
168 ps
->width
= surf_tmpl
->u
.buf
.last_element
- surf_tmpl
->u
.buf
.first_element
+ 1;
169 ps
->height
= pt
->height0
;
170 ps
->u
.buf
.first_element
= surf_tmpl
->u
.buf
.first_element
;
171 ps
->u
.buf
.last_element
= surf_tmpl
->u
.buf
.last_element
;
172 assert(ps
->u
.buf
.first_element
<= ps
->u
.buf
.last_element
);
173 assert(ps
->u
.buf
.last_element
< ps
->width
);
181 panfrost_surface_destroy(struct pipe_context
*pipe
,
182 struct pipe_surface
*surf
)
184 assert(surf
->texture
);
185 pipe_resource_reference(&surf
->texture
, NULL
);
189 static struct pipe_resource
*
190 panfrost_create_scanout_res(struct pipe_screen
*screen
,
191 const struct pipe_resource
*template)
193 struct panfrost_screen
*pscreen
= pan_screen(screen
);
194 struct pipe_resource scanout_templat
= *template;
195 struct renderonly_scanout
*scanout
;
196 struct winsys_handle handle
;
197 struct pipe_resource
*res
;
199 scanout
= renderonly_scanout_for_resource(&scanout_templat
,
200 pscreen
->ro
, &handle
);
204 assert(handle
.type
== WINSYS_HANDLE_TYPE_FD
);
205 /* TODO: handle modifiers? */
206 res
= screen
->resource_from_handle(screen
, template, &handle
,
207 PIPE_HANDLE_USAGE_FRAMEBUFFER_WRITE
);
208 close(handle
.handle
);
212 struct panfrost_resource
*pres
= pan_resource(res
);
214 pres
->scanout
= scanout
;
219 /* Computes sizes for checksumming, which is 8 bytes per 16x16 tile */
221 #define CHECKSUM_TILE_WIDTH 16
222 #define CHECKSUM_TILE_HEIGHT 16
223 #define CHECKSUM_BYTES_PER_TILE 8
226 panfrost_compute_checksum_sizes(
227 struct panfrost_slice
*slice
,
231 unsigned aligned_width
= ALIGN_POT(width
, CHECKSUM_TILE_WIDTH
);
232 unsigned aligned_height
= ALIGN_POT(height
, CHECKSUM_TILE_HEIGHT
);
234 unsigned tile_count_x
= aligned_width
/ CHECKSUM_TILE_WIDTH
;
235 unsigned tile_count_y
= aligned_height
/ CHECKSUM_TILE_HEIGHT
;
237 slice
->checksum_stride
= tile_count_x
* CHECKSUM_BYTES_PER_TILE
;
239 return slice
->checksum_stride
* tile_count_y
;
242 /* Setup the mip tree given a particular layout, possibly with checksumming */
245 panfrost_setup_slices(struct panfrost_resource
*pres
, size_t *bo_size
)
247 struct pipe_resource
*res
= &pres
->base
;
248 unsigned width
= res
->width0
;
249 unsigned height
= res
->height0
;
250 unsigned depth
= res
->depth0
;
251 unsigned bytes_per_pixel
= util_format_get_blocksize(res
->format
);
255 /* Tiled operates blockwise; linear is packed. Also, anything
256 * we render to has to be tile-aligned. Maybe not strictly
257 * necessary, but we're not *that* pressed for memory and it
258 * makes code a lot simpler */
260 bool renderable
= res
->bind
&
261 (PIPE_BIND_RENDER_TARGET
| PIPE_BIND_DEPTH_STENCIL
);
262 bool afbc
= pres
->layout
== PAN_AFBC
;
263 bool tiled
= pres
->layout
== PAN_TILED
;
264 bool should_align
= renderable
|| tiled
;
266 /* We don't know how to specify a 2D stride for 3D textures */
268 bool can_align_stride
=
269 res
->target
!= PIPE_TEXTURE_3D
;
271 should_align
&= can_align_stride
;
274 unsigned size_2d
= 0;
276 for (unsigned l
= 0; l
<= res
->last_level
; ++l
) {
277 struct panfrost_slice
*slice
= &pres
->slices
[l
];
279 unsigned effective_width
= width
;
280 unsigned effective_height
= height
;
281 unsigned effective_depth
= depth
;
284 effective_width
= ALIGN_POT(effective_width
, 16);
285 effective_height
= ALIGN_POT(effective_height
, 16);
287 /* We don't need to align depth */
290 /* Align levels to cache-line as a performance improvement for
291 * linear/tiled and as a requirement for AFBC */
293 offset
= ALIGN_POT(offset
, 64);
295 slice
->offset
= offset
;
297 /* Compute the would-be stride */
298 unsigned stride
= bytes_per_pixel
* effective_width
;
300 /* ..but cache-line align it for performance */
301 if (can_align_stride
&& pres
->layout
== PAN_LINEAR
)
302 stride
= ALIGN_POT(stride
, 64);
304 slice
->stride
= stride
;
306 unsigned slice_one_size
= slice
->stride
* effective_height
;
307 unsigned slice_full_size
= slice_one_size
* effective_depth
;
309 /* Report 2D size for 3D texturing */
312 size_2d
= slice_one_size
;
314 /* Compute AFBC sizes if necessary */
317 panfrost_afbc_header_size(width
, height
);
319 offset
+= slice
->header_size
;
322 offset
+= slice_full_size
;
324 /* Add a checksum region if necessary */
325 if (pres
->checksummed
) {
326 slice
->checksum_offset
= offset
;
328 unsigned size
= panfrost_compute_checksum_sizes(
329 slice
, width
, height
);
334 width
= u_minify(width
, 1);
335 height
= u_minify(height
, 1);
336 depth
= u_minify(depth
, 1);
339 assert(res
->array_size
);
341 if (res
->target
!= PIPE_TEXTURE_3D
) {
342 /* Arrays and cubemaps have the entire miptree duplicated */
344 pres
->cubemap_stride
= ALIGN_POT(offset
, 64);
345 *bo_size
= ALIGN_POT(pres
->cubemap_stride
* res
->array_size
, 4096);
347 /* 3D strides across the 2D layers */
348 assert(res
->array_size
== 1);
350 pres
->cubemap_stride
= size_2d
;
351 *bo_size
= ALIGN_POT(offset
, 4096);
356 panfrost_resource_create_bo(struct panfrost_screen
*screen
, struct panfrost_resource
*pres
)
358 struct pipe_resource
*res
= &pres
->base
;
360 /* Based on the usage, figure out what storing will be used. There are
363 * Linear: the basic format, bad for memory bandwidth, bad for cache
364 * use. Zero-copy, though. Renderable.
366 * Tiled: Not compressed, but cache-optimized. Expensive to write into
367 * (due to software tiling), but cheap to sample from. Ideal for most
370 * AFBC: Compressed and renderable (so always desirable for non-scanout
371 * rendertargets). Cheap to sample from. The format is black box, so we
372 * can't read/write from software.
375 /* Tiling textures is almost always faster, unless we only use it once */
377 bool is_texture
= (res
->bind
& PIPE_BIND_SAMPLER_VIEW
);
378 bool is_2d
= res
->depth0
== 1 && res
->array_size
== 1;
379 bool is_streaming
= (res
->usage
!= PIPE_USAGE_STREAM
);
381 bool should_tile
= is_streaming
&& is_texture
&& is_2d
;
383 /* Depth/stencil can't be tiled, only linear or AFBC */
384 should_tile
&= !(res
->bind
& PIPE_BIND_DEPTH_STENCIL
);
386 /* FBOs we would like to checksum, if at all possible */
387 bool can_checksum
= !(res
->bind
& (PIPE_BIND_SCANOUT
| PIPE_BIND_SHARED
));
388 bool should_checksum
= res
->bind
& PIPE_BIND_RENDER_TARGET
;
390 pres
->checksummed
= can_checksum
&& should_checksum
;
392 /* Set the layout appropriately */
393 pres
->layout
= should_tile
? PAN_TILED
: PAN_LINEAR
;
397 panfrost_setup_slices(pres
, &bo_size
);
399 /* We create a BO immediately but don't bother mapping, since we don't
400 * care to map e.g. FBOs which the CPU probably won't touch */
401 pres
->bo
= panfrost_drm_create_bo(screen
, bo_size
, PAN_ALLOCATE_DELAY_MMAP
);
404 static struct pipe_resource
*
405 panfrost_resource_create(struct pipe_screen
*screen
,
406 const struct pipe_resource
*template)
408 /* Make sure we're familiar */
409 switch (template->target
) {
411 case PIPE_TEXTURE_1D
:
412 case PIPE_TEXTURE_2D
:
413 case PIPE_TEXTURE_3D
:
414 case PIPE_TEXTURE_CUBE
:
415 case PIPE_TEXTURE_RECT
:
416 case PIPE_TEXTURE_2D_ARRAY
:
419 DBG("Unknown texture target %d\n", template->target
);
424 (PIPE_BIND_DISPLAY_TARGET
| PIPE_BIND_SCANOUT
| PIPE_BIND_SHARED
))
425 return panfrost_create_scanout_res(screen
, template);
427 struct panfrost_resource
*so
= rzalloc(screen
, struct panfrost_resource
);
428 struct panfrost_screen
*pscreen
= (struct panfrost_screen
*) screen
;
430 so
->base
= *template;
431 so
->base
.screen
= screen
;
433 pipe_reference_init(&so
->base
.reference
, 1);
435 util_range_init(&so
->valid_buffer_range
);
437 panfrost_resource_create_bo(pscreen
, so
);
438 return (struct pipe_resource
*)so
;
442 panfrost_bo_reference(struct panfrost_bo
*bo
)
444 pipe_reference(NULL
, &bo
->reference
);
448 panfrost_bo_unreference(struct pipe_screen
*screen
, struct panfrost_bo
*bo
)
450 /* When the reference count goes to zero, we need to cleanup */
452 if (pipe_reference(&bo
->reference
, NULL
))
453 panfrost_drm_release_bo(pan_screen(screen
), bo
, true);
457 panfrost_resource_destroy(struct pipe_screen
*screen
,
458 struct pipe_resource
*pt
)
460 struct panfrost_screen
*pscreen
= pan_screen(screen
);
461 struct panfrost_resource
*rsrc
= (struct panfrost_resource
*) pt
;
464 renderonly_scanout_destroy(rsrc
->scanout
, pscreen
->ro
);
467 panfrost_bo_unreference(screen
, rsrc
->bo
);
469 util_range_destroy(&rsrc
->valid_buffer_range
);
474 panfrost_transfer_map(struct pipe_context
*pctx
,
475 struct pipe_resource
*resource
,
477 unsigned usage
, /* a combination of PIPE_TRANSFER_x */
478 const struct pipe_box
*box
,
479 struct pipe_transfer
**out_transfer
)
481 int bytes_per_pixel
= util_format_get_blocksize(resource
->format
);
482 struct panfrost_resource
*rsrc
= pan_resource(resource
);
483 struct panfrost_bo
*bo
= rsrc
->bo
;
485 struct panfrost_gtransfer
*transfer
= rzalloc(pctx
, struct panfrost_gtransfer
);
486 transfer
->base
.level
= level
;
487 transfer
->base
.usage
= usage
;
488 transfer
->base
.box
= *box
;
490 pipe_resource_reference(&transfer
->base
.resource
, resource
);
492 *out_transfer
= &transfer
->base
;
494 /* If we haven't already mmaped, now's the time */
497 struct panfrost_screen
*screen
= pan_screen(pctx
->screen
);
498 panfrost_drm_mmap_bo(screen
, bo
);
501 /* Check if we're bound for rendering and this is a read pixels. If so,
502 * we need to flush */
504 struct panfrost_context
*ctx
= pan_context(pctx
);
505 struct pipe_framebuffer_state
*fb
= &ctx
->pipe_framebuffer
;
507 bool is_bound
= false;
509 for (unsigned c
= 0; c
< fb
->nr_cbufs
; ++c
) {
510 is_bound
|= fb
->cbufs
[c
]->texture
== resource
;
513 if (is_bound
&& (usage
& PIPE_TRANSFER_READ
)) {
515 panfrost_flush(pctx
, NULL
, PIPE_FLUSH_END_OF_FRAME
);
518 /* TODO: Respect usage flags */
520 if (usage
& PIPE_TRANSFER_DISCARD_WHOLE_RESOURCE
) {
521 /* TODO: reallocate */
522 //printf("debug: Missed reallocate\n");
523 } else if ((usage
& PIPE_TRANSFER_WRITE
)
524 && resource
->target
== PIPE_BUFFER
525 && !util_ranges_intersect(&rsrc
->valid_buffer_range
, box
->x
, box
->x
+ box
->width
)) {
526 /* No flush for writes to uninitialized */
527 } else if (!(usage
& PIPE_TRANSFER_UNSYNCHRONIZED
)) {
528 if (usage
& PIPE_TRANSFER_WRITE
) {
529 /* STUB: flush reading */
530 //printf("debug: missed reading flush %d\n", resource->target);
531 } else if (usage
& PIPE_TRANSFER_READ
) {
532 /* STUB: flush writing */
533 //printf("debug: missed writing flush %d (%d-%d)\n", resource->target, box->x, box->x + box->width);
535 /* Why are you even mapping?! */
539 if (rsrc
->layout
!= PAN_LINEAR
) {
540 /* Non-linear resources need to be indirectly mapped */
542 if (usage
& PIPE_TRANSFER_MAP_DIRECTLY
)
545 transfer
->base
.stride
= box
->width
* bytes_per_pixel
;
546 transfer
->base
.layer_stride
= transfer
->base
.stride
* box
->height
;
547 transfer
->map
= rzalloc_size(transfer
, transfer
->base
.layer_stride
* box
->depth
);
548 assert(box
->depth
== 1);
550 if ((usage
& PIPE_TRANSFER_READ
) && rsrc
->slices
[level
].initialized
) {
551 if (rsrc
->layout
== PAN_AFBC
) {
552 DBG("Unimplemented: reads from AFBC");
553 } else if (rsrc
->layout
== PAN_TILED
) {
554 panfrost_load_tiled_image(
556 bo
->cpu
+ rsrc
->slices
[level
].offset
,
558 transfer
->base
.stride
,
559 rsrc
->slices
[level
].stride
,
560 util_format_get_blocksize(resource
->format
));
564 return transfer
->map
;
566 transfer
->base
.stride
= rsrc
->slices
[level
].stride
;
567 transfer
->base
.layer_stride
= rsrc
->cubemap_stride
;
569 /* By mapping direct-write, we're implicitly already
570 * initialized (maybe), so be conservative */
572 if ((usage
& PIPE_TRANSFER_WRITE
) && (usage
& PIPE_TRANSFER_MAP_DIRECTLY
))
573 rsrc
->slices
[level
].initialized
= true;
576 + rsrc
->slices
[level
].offset
577 + transfer
->base
.box
.z
* rsrc
->cubemap_stride
578 + transfer
->base
.box
.y
* rsrc
->slices
[level
].stride
579 + transfer
->base
.box
.x
* bytes_per_pixel
;
584 panfrost_transfer_unmap(struct pipe_context
*pctx
,
585 struct pipe_transfer
*transfer
)
587 /* Gallium expects writeback here, so we tile */
589 struct panfrost_gtransfer
*trans
= pan_transfer(transfer
);
590 struct panfrost_resource
*prsrc
= (struct panfrost_resource
*) transfer
->resource
;
593 struct panfrost_bo
*bo
= prsrc
->bo
;
595 if (transfer
->usage
& PIPE_TRANSFER_WRITE
) {
596 unsigned level
= transfer
->level
;
597 prsrc
->slices
[level
].initialized
= true;
599 if (prsrc
->layout
== PAN_AFBC
) {
600 DBG("Unimplemented: writes to AFBC\n");
601 } else if (prsrc
->layout
== PAN_TILED
) {
602 assert(transfer
->box
.depth
== 1);
604 panfrost_store_tiled_image(
605 bo
->cpu
+ prsrc
->slices
[level
].offset
,
608 prsrc
->slices
[level
].stride
,
610 util_format_get_blocksize(prsrc
->base
.format
));
616 util_range_add(&prsrc
->valid_buffer_range
,
618 transfer
->box
.x
+ transfer
->box
.width
);
620 /* Derefence the resource */
621 pipe_resource_reference(&transfer
->resource
, NULL
);
623 /* Transfer itself is RALLOCed at the moment */
624 ralloc_free(transfer
);
628 panfrost_transfer_flush_region(struct pipe_context
*pctx
,
629 struct pipe_transfer
*transfer
,
630 const struct pipe_box
*box
)
632 struct panfrost_resource
*rsc
= pan_resource(transfer
->resource
);
634 if (transfer
->resource
->target
== PIPE_BUFFER
) {
635 util_range_add(&rsc
->valid_buffer_range
,
636 transfer
->box
.x
+ box
->x
,
637 transfer
->box
.x
+ box
->x
+ box
->width
);
642 panfrost_invalidate_resource(struct pipe_context
*pctx
, struct pipe_resource
*prsc
)
644 //DBG("TODO %s\n", __func__);
647 static enum pipe_format
648 panfrost_resource_get_internal_format(struct pipe_resource
*prsrc
) {
649 return prsrc
->format
;
653 panfrost_generate_mipmap(
654 struct pipe_context
*pctx
,
655 struct pipe_resource
*prsrc
,
656 enum pipe_format format
,
659 unsigned first_layer
,
662 struct panfrost_context
*ctx
= pan_context(pctx
);
663 struct panfrost_resource
*rsrc
= pan_resource(prsrc
);
665 /* Generating a mipmap invalidates the written levels, so make that
666 * explicit so we don't try to wallpaper them back and end up with
667 * u_blitter recursion */
670 for (unsigned l
= base_level
+ 1; l
<= last_level
; ++l
)
671 rsrc
->slices
[l
].initialized
= false;
673 /* Beyond that, we just delegate the hard stuff. We're careful to
674 * include flushes on both ends to make sure the data is really valid.
675 * We could be doing a lot better perf-wise, especially once we have
676 * reorder-type optimizations in place. But for now prioritize
679 struct panfrost_job
*job
= panfrost_get_job_for_fbo(ctx
);
680 bool has_draws
= job
->last_job
.gpu
;
683 panfrost_flush(pctx
, NULL
, PIPE_FLUSH_END_OF_FRAME
);
685 /* We've flushed the original buffer if needed, now trigger a blit */
687 bool blit_res
= util_gen_mipmap(
689 base_level
, last_level
,
690 first_layer
, last_layer
,
691 PIPE_TEX_FILTER_LINEAR
);
693 /* If the blit was successful, flush once more. If it wasn't, well, let
694 * the state tracker deal with it. */
697 panfrost_flush(pctx
, NULL
, PIPE_FLUSH_END_OF_FRAME
);
702 /* Computes the address to a texture at a particular slice */
705 panfrost_get_texture_address(
706 struct panfrost_resource
*rsrc
,
707 unsigned level
, unsigned face
)
709 unsigned level_offset
= rsrc
->slices
[level
].offset
;
710 unsigned face_offset
= face
* rsrc
->cubemap_stride
;
712 return rsrc
->bo
->gpu
+ level_offset
+ face_offset
;
716 panfrost_resource_set_stencil(struct pipe_resource
*prsrc
,
717 struct pipe_resource
*stencil
)
719 pan_resource(prsrc
)->separate_stencil
= pan_resource(stencil
);
722 static struct pipe_resource
*
723 panfrost_resource_get_stencil(struct pipe_resource
*prsrc
)
725 return &pan_resource(prsrc
)->separate_stencil
->base
;
728 static const struct u_transfer_vtbl transfer_vtbl
= {
729 .resource_create
= panfrost_resource_create
,
730 .resource_destroy
= panfrost_resource_destroy
,
731 .transfer_map
= panfrost_transfer_map
,
732 .transfer_unmap
= panfrost_transfer_unmap
,
733 .transfer_flush_region
= panfrost_transfer_flush_region
,
734 .get_internal_format
= panfrost_resource_get_internal_format
,
735 .set_stencil
= panfrost_resource_set_stencil
,
736 .get_stencil
= panfrost_resource_get_stencil
,
740 panfrost_resource_screen_init(struct panfrost_screen
*pscreen
)
742 //pscreen->base.resource_create_with_modifiers =
743 // panfrost_resource_create_with_modifiers;
744 pscreen
->base
.resource_create
= u_transfer_helper_resource_create
;
745 pscreen
->base
.resource_destroy
= u_transfer_helper_resource_destroy
;
746 pscreen
->base
.resource_from_handle
= panfrost_resource_from_handle
;
747 pscreen
->base
.resource_get_handle
= panfrost_resource_get_handle
;
748 pscreen
->base
.transfer_helper
= u_transfer_helper_create(&transfer_vtbl
,
754 panfrost_resource_context_init(struct pipe_context
*pctx
)
756 pctx
->transfer_map
= u_transfer_helper_transfer_map
;
757 pctx
->transfer_flush_region
= u_transfer_helper_transfer_flush_region
;
758 pctx
->transfer_unmap
= u_transfer_helper_transfer_unmap
;
759 pctx
->buffer_subdata
= u_default_buffer_subdata
;
760 pctx
->create_surface
= panfrost_create_surface
;
761 pctx
->surface_destroy
= panfrost_surface_destroy
;
762 pctx
->resource_copy_region
= util_resource_copy_region
;
763 pctx
->blit
= panfrost_blit
;
764 pctx
->generate_mipmap
= panfrost_generate_mipmap
;
765 pctx
->flush_resource
= panfrost_flush_resource
;
766 pctx
->invalidate_resource
= panfrost_invalidate_resource
;
767 pctx
->transfer_flush_region
= u_transfer_helper_transfer_flush_region
;
768 pctx
->buffer_subdata
= u_default_buffer_subdata
;
769 pctx
->texture_subdata
= u_default_texture_subdata
;