2 * Copyright © 2013 Keith Packard
3 * Copyright © 2015 Boyan Ding
5 * Permission to use, copy, modify, distribute, and sell this software and its
6 * documentation for any purpose is hereby granted without fee, provided that
7 * the above copyright notice appear in all copies and that both that copyright
8 * notice and this permission notice appear in supporting documentation, and
9 * that the name of the copyright holders not be used in advertising or
10 * publicity pertaining to distribution of the software without specific,
11 * written prior permission. The copyright holders make no representations
12 * about the suitability of this software for any purpose. It is provided "as
13 * is" without express or implied warranty.
15 * THE COPYRIGHT HOLDERS DISCLAIM ALL WARRANTIES WITH REGARD TO THIS SOFTWARE,
16 * INCLUDING ALL IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS, IN NO
17 * EVENT SHALL THE COPYRIGHT HOLDERS BE LIABLE FOR ANY SPECIAL, INDIRECT OR
18 * CONSEQUENTIAL DAMAGES OR ANY DAMAGES WHATSOEVER RESULTING FROM LOSS OF USE,
19 * DATA OR PROFITS, WHETHER IN AN ACTION OF CONTRACT, NEGLIGENCE OR OTHER
20 * TORTIOUS ACTION, ARISING OUT OF OR IN CONNECTION WITH THE USE OR PERFORMANCE
29 #include <X11/xshmfence.h>
32 #include <xcb/present.h>
34 #include <X11/Xlib-xcb.h>
36 #include "loader_dri3_helper.h"
37 #include "util/macros.h"
38 #include "drm_fourcc.h"
40 /* From xmlpool/options.h, user exposed so should be stable */
41 #define DRI_CONF_VBLANK_NEVER 0
42 #define DRI_CONF_VBLANK_DEF_INTERVAL_0 1
43 #define DRI_CONF_VBLANK_DEF_INTERVAL_1 2
44 #define DRI_CONF_VBLANK_ALWAYS_SYNC 3
47 * A cached blit context.
49 struct loader_dri3_blit_context
{
52 __DRIscreen
*cur_screen
;
53 const __DRIcoreExtension
*core
;
56 /* For simplicity we maintain the cache only for a single screen at a time */
57 static struct loader_dri3_blit_context blit_context
= {
58 _MTX_INITIALIZER_NP
, NULL
62 dri3_flush_present_events(struct loader_dri3_drawable
*draw
);
64 static struct loader_dri3_buffer
*
65 dri3_find_back_alloc(struct loader_dri3_drawable
*draw
);
68 get_screen_for_root(xcb_connection_t
*conn
, xcb_window_t root
)
70 xcb_screen_iterator_t screen_iter
=
71 xcb_setup_roots_iterator(xcb_get_setup(conn
));
73 for (; screen_iter
.rem
; xcb_screen_next (&screen_iter
)) {
74 if (screen_iter
.data
->root
== root
)
75 return screen_iter
.data
;
81 static xcb_visualtype_t
*
82 get_xcb_visualtype_for_depth(struct loader_dri3_drawable
*draw
, int depth
)
84 xcb_visualtype_iterator_t visual_iter
;
85 xcb_screen_t
*screen
= draw
->screen
;
86 xcb_depth_iterator_t depth_iter
;
91 depth_iter
= xcb_screen_allowed_depths_iterator(screen
);
92 for (; depth_iter
.rem
; xcb_depth_next(&depth_iter
)) {
93 if (depth_iter
.data
->depth
!= depth
)
96 visual_iter
= xcb_depth_visuals_iterator(depth_iter
.data
);
98 return visual_iter
.data
;
104 /* Get red channel mask for given drawable at given depth. */
106 dri3_get_red_mask_for_depth(struct loader_dri3_drawable
*draw
, int depth
)
108 xcb_visualtype_t
*visual
= get_xcb_visualtype_for_depth(draw
, depth
);
111 return visual
->red_mask
;
117 * Do we have blit functionality in the image blit extension?
119 * \param draw[in] The drawable intended to blit from / to.
120 * \return true if we have blit functionality. false otherwise.
122 static bool loader_dri3_have_image_blit(const struct loader_dri3_drawable
*draw
)
124 return draw
->ext
->image
->base
.version
>= 9 &&
125 draw
->ext
->image
->blitImage
!= NULL
;
129 * Get and lock (for use with the current thread) a dri context associated
130 * with the drawable's dri screen. The context is intended to be used with
131 * the dri image extension's blitImage method.
133 * \param draw[in] Pointer to the drawable whose dri screen we want a
135 * \return A dri context or NULL if context creation failed.
137 * When the caller is done with the context (even if the context returned was
138 * NULL), the caller must call loader_dri3_blit_context_put.
140 static __DRIcontext
*
141 loader_dri3_blit_context_get(struct loader_dri3_drawable
*draw
)
143 mtx_lock(&blit_context
.mtx
);
145 if (blit_context
.ctx
&& blit_context
.cur_screen
!= draw
->dri_screen
) {
146 blit_context
.core
->destroyContext(blit_context
.ctx
);
147 blit_context
.ctx
= NULL
;
150 if (!blit_context
.ctx
) {
151 blit_context
.ctx
= draw
->ext
->core
->createNewContext(draw
->dri_screen
,
153 blit_context
.cur_screen
= draw
->dri_screen
;
154 blit_context
.core
= draw
->ext
->core
;
157 return blit_context
.ctx
;
161 * Release (for use with other threads) a dri context previously obtained using
162 * loader_dri3_blit_context_get.
165 loader_dri3_blit_context_put(void)
167 mtx_unlock(&blit_context
.mtx
);
171 * Blit (parts of) the contents of a DRI image to another dri image
173 * \param draw[in] The drawable which owns the images.
174 * \param dst[in] The destination image.
175 * \param src[in] The source image.
176 * \param dstx0[in] Start destination coordinate.
177 * \param dsty0[in] Start destination coordinate.
178 * \param width[in] Blit width.
179 * \param height[in] Blit height.
180 * \param srcx0[in] Start source coordinate.
181 * \param srcy0[in] Start source coordinate.
182 * \param flush_flag[in] Image blit flush flag.
183 * \return true iff successful.
186 loader_dri3_blit_image(struct loader_dri3_drawable
*draw
,
187 __DRIimage
*dst
, __DRIimage
*src
,
188 int dstx0
, int dsty0
, int width
, int height
,
189 int srcx0
, int srcy0
, int flush_flag
)
191 __DRIcontext
*dri_context
;
192 bool use_blit_context
= false;
194 if (!loader_dri3_have_image_blit(draw
))
197 dri_context
= draw
->vtable
->get_dri_context(draw
);
199 if (!dri_context
|| !draw
->vtable
->in_current_context(draw
)) {
200 dri_context
= loader_dri3_blit_context_get(draw
);
201 use_blit_context
= true;
202 flush_flag
|= __BLIT_FLAG_FLUSH
;
206 draw
->ext
->image
->blitImage(dri_context
, dst
, src
, dstx0
, dsty0
,
207 width
, height
, srcx0
, srcy0
,
208 width
, height
, flush_flag
);
210 if (use_blit_context
)
211 loader_dri3_blit_context_put();
213 return dri_context
!= NULL
;
217 dri3_fence_reset(xcb_connection_t
*c
, struct loader_dri3_buffer
*buffer
)
219 xshmfence_reset(buffer
->shm_fence
);
223 dri3_fence_set(struct loader_dri3_buffer
*buffer
)
225 xshmfence_trigger(buffer
->shm_fence
);
229 dri3_fence_trigger(xcb_connection_t
*c
, struct loader_dri3_buffer
*buffer
)
231 xcb_sync_trigger_fence(c
, buffer
->sync_fence
);
235 dri3_fence_await(xcb_connection_t
*c
, struct loader_dri3_drawable
*draw
,
236 struct loader_dri3_buffer
*buffer
)
239 xshmfence_await(buffer
->shm_fence
);
241 mtx_lock(&draw
->mtx
);
242 dri3_flush_present_events(draw
);
243 mtx_unlock(&draw
->mtx
);
248 dri3_update_num_back(struct loader_dri3_drawable
*draw
)
250 if (draw
->last_present_mode
== XCB_PRESENT_COMPLETE_MODE_FLIP
)
257 loader_dri3_set_swap_interval(struct loader_dri3_drawable
*draw
, int interval
)
259 draw
->swap_interval
= interval
;
262 /** dri3_free_render_buffer
264 * Free everything associated with one render buffer including pixmap, fence
265 * stuff and the driver image
268 dri3_free_render_buffer(struct loader_dri3_drawable
*draw
,
269 struct loader_dri3_buffer
*buffer
)
271 if (buffer
->own_pixmap
)
272 xcb_free_pixmap(draw
->conn
, buffer
->pixmap
);
273 xcb_sync_destroy_fence(draw
->conn
, buffer
->sync_fence
);
274 xshmfence_unmap_shm(buffer
->shm_fence
);
275 draw
->ext
->image
->destroyImage(buffer
->image
);
276 if (buffer
->linear_buffer
)
277 draw
->ext
->image
->destroyImage(buffer
->linear_buffer
);
282 loader_dri3_drawable_fini(struct loader_dri3_drawable
*draw
)
286 draw
->ext
->core
->destroyDrawable(draw
->dri_drawable
);
288 for (i
= 0; i
< ARRAY_SIZE(draw
->buffers
); i
++) {
289 if (draw
->buffers
[i
])
290 dri3_free_render_buffer(draw
, draw
->buffers
[i
]);
293 if (draw
->special_event
) {
294 xcb_void_cookie_t cookie
=
295 xcb_present_select_input_checked(draw
->conn
, draw
->eid
, draw
->drawable
,
296 XCB_PRESENT_EVENT_MASK_NO_EVENT
);
298 xcb_discard_reply(draw
->conn
, cookie
.sequence
);
299 xcb_unregister_for_special_event(draw
->conn
, draw
->special_event
);
302 cnd_destroy(&draw
->event_cnd
);
303 mtx_destroy(&draw
->mtx
);
307 loader_dri3_drawable_init(xcb_connection_t
*conn
,
308 xcb_drawable_t drawable
,
309 __DRIscreen
*dri_screen
,
310 bool is_different_gpu
,
311 bool multiplanes_available
,
312 const __DRIconfig
*dri_config
,
313 struct loader_dri3_extensions
*ext
,
314 const struct loader_dri3_vtable
*vtable
,
315 struct loader_dri3_drawable
*draw
)
317 xcb_get_geometry_cookie_t cookie
;
318 xcb_get_geometry_reply_t
*reply
;
319 xcb_generic_error_t
*error
;
320 GLint vblank_mode
= DRI_CONF_VBLANK_DEF_INTERVAL_1
;
325 draw
->vtable
= vtable
;
326 draw
->drawable
= drawable
;
327 draw
->dri_screen
= dri_screen
;
328 draw
->is_different_gpu
= is_different_gpu
;
329 draw
->multiplanes_available
= multiplanes_available
;
332 draw
->have_fake_front
= 0;
333 draw
->first_init
= true;
335 draw
->cur_blit_source
= -1;
336 draw
->back_format
= __DRI_IMAGE_FORMAT_NONE
;
337 mtx_init(&draw
->mtx
, mtx_plain
);
338 cnd_init(&draw
->event_cnd
);
340 if (draw
->ext
->config
)
341 draw
->ext
->config
->configQueryi(draw
->dri_screen
,
342 "vblank_mode", &vblank_mode
);
344 switch (vblank_mode
) {
345 case DRI_CONF_VBLANK_NEVER
:
346 case DRI_CONF_VBLANK_DEF_INTERVAL_0
:
349 case DRI_CONF_VBLANK_DEF_INTERVAL_1
:
350 case DRI_CONF_VBLANK_ALWAYS_SYNC
:
355 draw
->swap_interval
= swap_interval
;
357 dri3_update_num_back(draw
);
359 /* Create a new drawable */
361 draw
->ext
->image_driver
->createNewDrawable(dri_screen
,
365 if (!draw
->dri_drawable
)
368 cookie
= xcb_get_geometry(draw
->conn
, draw
->drawable
);
369 reply
= xcb_get_geometry_reply(draw
->conn
, cookie
, &error
);
370 if (reply
== NULL
|| error
!= NULL
) {
371 draw
->ext
->core
->destroyDrawable(draw
->dri_drawable
);
375 draw
->screen
= get_screen_for_root(draw
->conn
, reply
->root
);
376 draw
->width
= reply
->width
;
377 draw
->height
= reply
->height
;
378 draw
->depth
= reply
->depth
;
379 draw
->vtable
->set_drawable_size(draw
, draw
->width
, draw
->height
);
382 draw
->swap_method
= __DRI_ATTRIB_SWAP_UNDEFINED
;
383 if (draw
->ext
->core
->base
.version
>= 2) {
384 (void )draw
->ext
->core
->getConfigAttrib(dri_config
,
385 __DRI_ATTRIB_SWAP_METHOD
,
390 * Make sure server has the same swap interval we do for the new
393 loader_dri3_set_swap_interval(draw
, swap_interval
);
399 * Process one Present event
402 dri3_handle_present_event(struct loader_dri3_drawable
*draw
,
403 xcb_present_generic_event_t
*ge
)
405 switch (ge
->evtype
) {
406 case XCB_PRESENT_CONFIGURE_NOTIFY
: {
407 xcb_present_configure_notify_event_t
*ce
= (void *) ge
;
409 draw
->width
= ce
->width
;
410 draw
->height
= ce
->height
;
411 draw
->vtable
->set_drawable_size(draw
, draw
->width
, draw
->height
);
412 draw
->ext
->flush
->invalidate(draw
->dri_drawable
);
415 case XCB_PRESENT_COMPLETE_NOTIFY
: {
416 xcb_present_complete_notify_event_t
*ce
= (void *) ge
;
418 /* Compute the processed SBC number from the received 32-bit serial number
419 * merged with the upper 32-bits of the sent 64-bit serial number while
422 if (ce
->kind
== XCB_PRESENT_COMPLETE_KIND_PIXMAP
) {
423 uint64_t recv_sbc
= (draw
->send_sbc
& 0xffffffff00000000LL
) | ce
->serial
;
425 /* Only assume wraparound if that results in exactly the previous
426 * SBC + 1, otherwise ignore received SBC > sent SBC (those are
427 * probably from a previous loader_dri3_drawable instance) to avoid
428 * calculating bogus target MSC values in loader_dri3_swap_buffers_msc
430 if (recv_sbc
<= draw
->send_sbc
)
431 draw
->recv_sbc
= recv_sbc
;
432 else if (recv_sbc
== (draw
->recv_sbc
+ 0x100000001ULL
))
433 draw
->recv_sbc
= recv_sbc
- 0x100000000ULL
;
435 /* When moving from flip to copy, we assume that we can allocate in
436 * a more optimal way if we don't need to cater for the display
439 if (ce
->mode
== XCB_PRESENT_COMPLETE_MODE_COPY
&&
440 draw
->last_present_mode
== XCB_PRESENT_COMPLETE_MODE_FLIP
) {
441 for (int b
= 0; b
< ARRAY_SIZE(draw
->buffers
); b
++) {
442 if (draw
->buffers
[b
])
443 draw
->buffers
[b
]->reallocate
= true;
447 /* If the server tells us that our allocation is suboptimal, we
450 #ifdef HAVE_DRI3_MODIFIERS
451 if (ce
->mode
== XCB_PRESENT_COMPLETE_MODE_SUBOPTIMAL_COPY
&&
452 draw
->last_present_mode
!= ce
->mode
) {
453 for (int b
= 0; b
< ARRAY_SIZE(draw
->buffers
); b
++) {
454 if (draw
->buffers
[b
])
455 draw
->buffers
[b
]->reallocate
= true;
459 draw
->last_present_mode
= ce
->mode
;
461 if (draw
->vtable
->show_fps
)
462 draw
->vtable
->show_fps(draw
, ce
->ust
);
466 } else if (ce
->serial
== draw
->eid
) {
467 draw
->notify_ust
= ce
->ust
;
468 draw
->notify_msc
= ce
->msc
;
472 case XCB_PRESENT_EVENT_IDLE_NOTIFY
: {
473 xcb_present_idle_notify_event_t
*ie
= (void *) ge
;
476 for (b
= 0; b
< ARRAY_SIZE(draw
->buffers
); b
++) {
477 struct loader_dri3_buffer
*buf
= draw
->buffers
[b
];
479 if (buf
&& buf
->pixmap
== ie
->pixmap
)
489 dri3_wait_for_event_locked(struct loader_dri3_drawable
*draw
)
491 xcb_generic_event_t
*ev
;
492 xcb_present_generic_event_t
*ge
;
494 xcb_flush(draw
->conn
);
496 /* Only have one thread waiting for events at a time */
497 if (draw
->has_event_waiter
) {
498 cnd_wait(&draw
->event_cnd
, &draw
->mtx
);
499 /* Another thread has updated the protected info, so retest. */
502 draw
->has_event_waiter
= true;
503 /* Allow other threads access to the drawable while we're waiting. */
504 mtx_unlock(&draw
->mtx
);
505 ev
= xcb_wait_for_special_event(draw
->conn
, draw
->special_event
);
506 mtx_lock(&draw
->mtx
);
507 draw
->has_event_waiter
= false;
508 cnd_broadcast(&draw
->event_cnd
);
513 dri3_handle_present_event(draw
, ge
);
517 /** loader_dri3_wait_for_msc
519 * Get the X server to send an event when the target msc/divisor/remainder is
523 loader_dri3_wait_for_msc(struct loader_dri3_drawable
*draw
,
525 int64_t divisor
, int64_t remainder
,
526 int64_t *ust
, int64_t *msc
, int64_t *sbc
)
528 xcb_void_cookie_t cookie
= xcb_present_notify_msc(draw
->conn
,
534 xcb_generic_event_t
*ev
;
535 unsigned full_sequence
;
537 mtx_lock(&draw
->mtx
);
538 xcb_flush(draw
->conn
);
540 /* Wait for the event */
542 ev
= xcb_wait_for_special_event(draw
->conn
, draw
->special_event
);
544 mtx_unlock(&draw
->mtx
);
548 full_sequence
= ev
->full_sequence
;
549 dri3_handle_present_event(draw
, (void *) ev
);
550 } while (full_sequence
!= cookie
.sequence
|| draw
->notify_msc
< target_msc
);
552 *ust
= draw
->notify_ust
;
553 *msc
= draw
->notify_msc
;
554 *sbc
= draw
->recv_sbc
;
555 mtx_unlock(&draw
->mtx
);
560 /** loader_dri3_wait_for_sbc
562 * Wait for the completed swap buffer count to reach the specified
563 * target. Presumably the application knows that this will be reached with
564 * outstanding complete events, or we're going to be here awhile.
567 loader_dri3_wait_for_sbc(struct loader_dri3_drawable
*draw
,
568 int64_t target_sbc
, int64_t *ust
,
569 int64_t *msc
, int64_t *sbc
)
571 /* From the GLX_OML_sync_control spec:
573 * "If <target_sbc> = 0, the function will block until all previous
574 * swaps requested with glXSwapBuffersMscOML for that window have
577 mtx_lock(&draw
->mtx
);
579 target_sbc
= draw
->send_sbc
;
581 while (draw
->recv_sbc
< target_sbc
) {
582 if (!dri3_wait_for_event_locked(draw
)) {
583 mtx_unlock(&draw
->mtx
);
590 *sbc
= draw
->recv_sbc
;
591 mtx_unlock(&draw
->mtx
);
595 /** loader_dri3_find_back
597 * Find an idle back buffer. If there isn't one, then
598 * wait for a present idle notify event from the X server
601 dri3_find_back(struct loader_dri3_drawable
*draw
)
606 mtx_lock(&draw
->mtx
);
607 /* Increase the likelyhood of reusing current buffer */
608 dri3_flush_present_events(draw
);
610 /* Check whether we need to reuse the current back buffer as new back.
611 * In that case, wait until it's not busy anymore.
613 num_to_consider
= draw
->num_back
;
614 if (!loader_dri3_have_image_blit(draw
) && draw
->cur_blit_source
!= -1) {
616 draw
->cur_blit_source
= -1;
620 for (b
= 0; b
< num_to_consider
; b
++) {
621 int id
= LOADER_DRI3_BACK_ID((b
+ draw
->cur_back
) % draw
->num_back
);
622 struct loader_dri3_buffer
*buffer
= draw
->buffers
[id
];
624 if (!buffer
|| !buffer
->busy
) {
626 mtx_unlock(&draw
->mtx
);
630 if (!dri3_wait_for_event_locked(draw
)) {
631 mtx_unlock(&draw
->mtx
);
637 static xcb_gcontext_t
638 dri3_drawable_gc(struct loader_dri3_drawable
*draw
)
642 xcb_create_gc(draw
->conn
,
643 (draw
->gc
= xcb_generate_id(draw
->conn
)),
645 XCB_GC_GRAPHICS_EXPOSURES
,
652 static struct loader_dri3_buffer
*
653 dri3_back_buffer(struct loader_dri3_drawable
*draw
)
655 return draw
->buffers
[LOADER_DRI3_BACK_ID(draw
->cur_back
)];
658 static struct loader_dri3_buffer
*
659 dri3_fake_front_buffer(struct loader_dri3_drawable
*draw
)
661 return draw
->buffers
[LOADER_DRI3_FRONT_ID
];
665 dri3_copy_area(xcb_connection_t
*c
,
666 xcb_drawable_t src_drawable
,
667 xcb_drawable_t dst_drawable
,
676 xcb_void_cookie_t cookie
;
678 cookie
= xcb_copy_area_checked(c
,
688 xcb_discard_reply(c
, cookie
.sequence
);
692 * Asks the driver to flush any queued work necessary for serializing with the
693 * X command stream, and optionally the slightly more strict requirement of
694 * glFlush() equivalence (which would require flushing even if nothing had
695 * been drawn to a window system framebuffer, for example).
698 loader_dri3_flush(struct loader_dri3_drawable
*draw
,
700 enum __DRI2throttleReason throttle_reason
)
702 /* NEED TO CHECK WHETHER CONTEXT IS NULL */
703 __DRIcontext
*dri_context
= draw
->vtable
->get_dri_context(draw
);
706 draw
->ext
->flush
->flush_with_flags(dri_context
, draw
->dri_drawable
,
707 flags
, throttle_reason
);
712 loader_dri3_copy_sub_buffer(struct loader_dri3_drawable
*draw
,
714 int width
, int height
,
717 struct loader_dri3_buffer
*back
;
718 unsigned flags
= __DRI2_FLUSH_DRAWABLE
;
720 /* Check we have the right attachments */
721 if (!draw
->have_back
|| draw
->is_pixmap
)
725 flags
|= __DRI2_FLUSH_CONTEXT
;
726 loader_dri3_flush(draw
, flags
, __DRI2_THROTTLE_SWAPBUFFER
);
728 back
= dri3_find_back_alloc(draw
);
732 y
= draw
->height
- y
- height
;
734 if (draw
->is_different_gpu
) {
735 /* Update the linear buffer part of the back buffer
736 * for the dri3_copy_area operation
738 (void) loader_dri3_blit_image(draw
,
741 0, 0, back
->width
, back
->height
,
742 0, 0, __BLIT_FLAG_FLUSH
);
745 loader_dri3_swapbuffer_barrier(draw
);
746 dri3_fence_reset(draw
->conn
, back
);
747 dri3_copy_area(draw
->conn
,
750 dri3_drawable_gc(draw
),
751 x
, y
, x
, y
, width
, height
);
752 dri3_fence_trigger(draw
->conn
, back
);
753 /* Refresh the fake front (if present) after we just damaged the real
756 if (draw
->have_fake_front
&&
757 !loader_dri3_blit_image(draw
,
758 dri3_fake_front_buffer(draw
)->image
,
761 x
, y
, __BLIT_FLAG_FLUSH
) &&
762 !draw
->is_different_gpu
) {
763 dri3_fence_reset(draw
->conn
, dri3_fake_front_buffer(draw
));
764 dri3_copy_area(draw
->conn
,
766 dri3_fake_front_buffer(draw
)->pixmap
,
767 dri3_drawable_gc(draw
),
768 x
, y
, x
, y
, width
, height
);
769 dri3_fence_trigger(draw
->conn
, dri3_fake_front_buffer(draw
));
770 dri3_fence_await(draw
->conn
, NULL
, dri3_fake_front_buffer(draw
));
772 dri3_fence_await(draw
->conn
, draw
, back
);
776 loader_dri3_copy_drawable(struct loader_dri3_drawable
*draw
,
780 loader_dri3_flush(draw
, __DRI2_FLUSH_DRAWABLE
, 0);
782 dri3_fence_reset(draw
->conn
, dri3_fake_front_buffer(draw
));
783 dri3_copy_area(draw
->conn
,
785 dri3_drawable_gc(draw
),
786 0, 0, 0, 0, draw
->width
, draw
->height
);
787 dri3_fence_trigger(draw
->conn
, dri3_fake_front_buffer(draw
));
788 dri3_fence_await(draw
->conn
, draw
, dri3_fake_front_buffer(draw
));
792 loader_dri3_wait_x(struct loader_dri3_drawable
*draw
)
794 struct loader_dri3_buffer
*front
;
796 if (draw
== NULL
|| !draw
->have_fake_front
)
799 front
= dri3_fake_front_buffer(draw
);
801 loader_dri3_copy_drawable(draw
, front
->pixmap
, draw
->drawable
);
803 /* In the psc->is_different_gpu case, the linear buffer has been updated,
804 * but not yet the tiled buffer.
805 * Copy back to the tiled buffer we use for rendering.
806 * Note that we don't need flushing.
808 if (draw
->is_different_gpu
)
809 (void) loader_dri3_blit_image(draw
,
811 front
->linear_buffer
,
812 0, 0, front
->width
, front
->height
,
817 loader_dri3_wait_gl(struct loader_dri3_drawable
*draw
)
819 struct loader_dri3_buffer
*front
;
821 if (draw
== NULL
|| !draw
->have_fake_front
)
824 front
= dri3_fake_front_buffer(draw
);
826 /* In the psc->is_different_gpu case, we update the linear_buffer
827 * before updating the real front.
829 if (draw
->is_different_gpu
)
830 (void) loader_dri3_blit_image(draw
,
831 front
->linear_buffer
,
833 0, 0, front
->width
, front
->height
,
834 0, 0, __BLIT_FLAG_FLUSH
);
835 loader_dri3_swapbuffer_barrier(draw
);
836 loader_dri3_copy_drawable(draw
, draw
->drawable
, front
->pixmap
);
839 /** dri3_flush_present_events
841 * Process any present events that have been received from the X server
844 dri3_flush_present_events(struct loader_dri3_drawable
*draw
)
846 /* Check to see if any configuration changes have occurred
847 * since we were last invoked
849 if (draw
->has_event_waiter
)
852 if (draw
->special_event
) {
853 xcb_generic_event_t
*ev
;
855 while ((ev
= xcb_poll_for_special_event(draw
->conn
,
856 draw
->special_event
)) != NULL
) {
857 xcb_present_generic_event_t
*ge
= (void *) ev
;
858 dri3_handle_present_event(draw
, ge
);
863 /** loader_dri3_swap_buffers_msc
865 * Make the current back buffer visible using the present extension
868 loader_dri3_swap_buffers_msc(struct loader_dri3_drawable
*draw
,
869 int64_t target_msc
, int64_t divisor
,
870 int64_t remainder
, unsigned flush_flags
,
873 struct loader_dri3_buffer
*back
;
875 uint32_t options
= XCB_PRESENT_OPTION_NONE
;
877 draw
->vtable
->flush_drawable(draw
, flush_flags
);
879 back
= dri3_find_back_alloc(draw
);
881 mtx_lock(&draw
->mtx
);
882 if (draw
->is_different_gpu
&& back
) {
883 /* Update the linear buffer before presenting the pixmap */
884 (void) loader_dri3_blit_image(draw
,
887 0, 0, back
->width
, back
->height
,
888 0, 0, __BLIT_FLAG_FLUSH
);
891 /* If we need to preload the new back buffer, remember the source.
892 * The force_copy parameter is used by EGL to attempt to preserve
893 * the back buffer across a call to this function.
895 if (draw
->swap_method
!= __DRI_ATTRIB_SWAP_UNDEFINED
|| force_copy
)
896 draw
->cur_blit_source
= LOADER_DRI3_BACK_ID(draw
->cur_back
);
898 /* Exchange the back and fake front. Even though the server knows about these
899 * buffers, it has no notion of back and fake front.
901 if (back
&& draw
->have_fake_front
) {
902 struct loader_dri3_buffer
*tmp
;
904 tmp
= dri3_fake_front_buffer(draw
);
905 draw
->buffers
[LOADER_DRI3_FRONT_ID
] = back
;
906 draw
->buffers
[LOADER_DRI3_BACK_ID(draw
->cur_back
)] = tmp
;
908 if (draw
->swap_method
== __DRI_ATTRIB_SWAP_COPY
|| force_copy
)
909 draw
->cur_blit_source
= LOADER_DRI3_FRONT_ID
;
912 dri3_flush_present_events(draw
);
914 if (back
&& !draw
->is_pixmap
) {
915 dri3_fence_reset(draw
->conn
, back
);
917 /* Compute when we want the frame shown by taking the last known
918 * successful MSC and adding in a swap interval for each outstanding swap
919 * request. target_msc=divisor=remainder=0 means "Use glXSwapBuffers()
923 if (target_msc
== 0 && divisor
== 0 && remainder
== 0)
924 target_msc
= draw
->msc
+ draw
->swap_interval
*
925 (draw
->send_sbc
- draw
->recv_sbc
);
926 else if (divisor
== 0 && remainder
> 0) {
927 /* From the GLX_OML_sync_control spec:
928 * "If <divisor> = 0, the swap will occur when MSC becomes
929 * greater than or equal to <target_msc>."
931 * Note that there's no mention of the remainder. The Present
932 * extension throws BadValue for remainder != 0 with divisor == 0, so
933 * just drop the passed in value.
938 /* From the GLX_EXT_swap_control spec
939 * and the EGL 1.4 spec (page 53):
941 * "If <interval> is set to a value of 0, buffer swaps are not
942 * synchronized to a video frame."
944 * Implementation note: It is possible to enable triple buffering
945 * behaviour by not using XCB_PRESENT_OPTION_ASYNC, but this should not be
948 if (draw
->swap_interval
== 0)
949 options
|= XCB_PRESENT_OPTION_ASYNC
;
951 /* If we need to populate the new back, but need to reuse the back
952 * buffer slot due to lack of local blit capabilities, make sure
953 * the server doesn't flip and we deadlock.
955 if (!loader_dri3_have_image_blit(draw
) && draw
->cur_blit_source
!= -1)
956 options
|= XCB_PRESENT_OPTION_COPY
;
957 #ifdef HAVE_DRI3_MODIFIERS
958 if (draw
->multiplanes_available
)
959 options
|= XCB_PRESENT_OPTION_SUBOPTIMAL
;
962 back
->last_swap
= draw
->send_sbc
;
963 xcb_present_pixmap(draw
->conn
,
966 (uint32_t) draw
->send_sbc
,
971 None
, /* target_crtc */
978 ret
= (int64_t) draw
->send_sbc
;
980 /* Schedule a server-side back-preserving blit if necessary.
981 * This happens iff all conditions below are satisfied:
982 * a) We have a fake front,
983 * b) We need to preserve the back buffer,
984 * c) We don't have local blit capabilities.
986 if (!loader_dri3_have_image_blit(draw
) && draw
->cur_blit_source
!= -1 &&
987 draw
->cur_blit_source
!= LOADER_DRI3_BACK_ID(draw
->cur_back
)) {
988 struct loader_dri3_buffer
*new_back
= dri3_back_buffer(draw
);
989 struct loader_dri3_buffer
*src
= draw
->buffers
[draw
->cur_blit_source
];
991 dri3_fence_reset(draw
->conn
, new_back
);
992 dri3_copy_area(draw
->conn
, src
->pixmap
,
994 dri3_drawable_gc(draw
),
995 0, 0, 0, 0, draw
->width
, draw
->height
);
996 dri3_fence_trigger(draw
->conn
, new_back
);
997 new_back
->last_swap
= src
->last_swap
;
1000 xcb_flush(draw
->conn
);
1004 mtx_unlock(&draw
->mtx
);
1006 draw
->ext
->flush
->invalidate(draw
->dri_drawable
);
1012 loader_dri3_query_buffer_age(struct loader_dri3_drawable
*draw
)
1014 struct loader_dri3_buffer
*back
= dri3_find_back_alloc(draw
);
1017 mtx_lock(&draw
->mtx
);
1018 ret
= (!back
|| back
->last_swap
== 0) ? 0 :
1019 draw
->send_sbc
- back
->last_swap
+ 1;
1020 mtx_unlock(&draw
->mtx
);
1025 /** loader_dri3_open
1027 * Wrapper around xcb_dri3_open
1030 loader_dri3_open(xcb_connection_t
*conn
,
1034 xcb_dri3_open_cookie_t cookie
;
1035 xcb_dri3_open_reply_t
*reply
;
1038 cookie
= xcb_dri3_open(conn
,
1042 reply
= xcb_dri3_open_reply(conn
, cookie
, NULL
);
1046 if (reply
->nfd
!= 1) {
1051 fd
= xcb_dri3_open_reply_fds(conn
, reply
)[0];
1053 fcntl(fd
, F_SETFD
, fcntl(fd
, F_GETFD
) | FD_CLOEXEC
);
1059 dri3_cpp_for_format(uint32_t format
) {
1061 case __DRI_IMAGE_FORMAT_R8
:
1063 case __DRI_IMAGE_FORMAT_RGB565
:
1064 case __DRI_IMAGE_FORMAT_GR88
:
1066 case __DRI_IMAGE_FORMAT_XRGB8888
:
1067 case __DRI_IMAGE_FORMAT_ARGB8888
:
1068 case __DRI_IMAGE_FORMAT_ABGR8888
:
1069 case __DRI_IMAGE_FORMAT_XBGR8888
:
1070 case __DRI_IMAGE_FORMAT_XRGB2101010
:
1071 case __DRI_IMAGE_FORMAT_ARGB2101010
:
1072 case __DRI_IMAGE_FORMAT_XBGR2101010
:
1073 case __DRI_IMAGE_FORMAT_ABGR2101010
:
1074 case __DRI_IMAGE_FORMAT_SARGB8
:
1075 case __DRI_IMAGE_FORMAT_SABGR8
:
1077 case __DRI_IMAGE_FORMAT_NONE
:
1083 /* Map format of render buffer to corresponding format for the linear_buffer
1084 * used for sharing with the display gpu of a Prime setup (== is_different_gpu).
1085 * Usually linear_format == format, except for depth >= 30 formats, where
1086 * different gpu vendors have different preferences wrt. color channel ordering.
1089 dri3_linear_format_for_format(struct loader_dri3_drawable
*draw
, uint32_t format
)
1092 case __DRI_IMAGE_FORMAT_XRGB2101010
:
1093 case __DRI_IMAGE_FORMAT_XBGR2101010
:
1094 /* Different preferred formats for different hw */
1095 if (dri3_get_red_mask_for_depth(draw
, 30) == 0x3ff)
1096 return __DRI_IMAGE_FORMAT_XBGR2101010
;
1098 return __DRI_IMAGE_FORMAT_XRGB2101010
;
1100 case __DRI_IMAGE_FORMAT_ARGB2101010
:
1101 case __DRI_IMAGE_FORMAT_ABGR2101010
:
1102 /* Different preferred formats for different hw */
1103 if (dri3_get_red_mask_for_depth(draw
, 30) == 0x3ff)
1104 return __DRI_IMAGE_FORMAT_ABGR2101010
;
1106 return __DRI_IMAGE_FORMAT_ARGB2101010
;
1113 /* the DRIimage createImage function takes __DRI_IMAGE_FORMAT codes, while
1114 * the createImageFromFds call takes __DRI_IMAGE_FOURCC codes. To avoid
1115 * complete confusion, just deal in __DRI_IMAGE_FORMAT codes for now and
1116 * translate to __DRI_IMAGE_FOURCC codes in the call to createImageFromFds
1119 image_format_to_fourcc(int format
)
1122 /* Convert from __DRI_IMAGE_FORMAT to __DRI_IMAGE_FOURCC (sigh) */
1124 case __DRI_IMAGE_FORMAT_SARGB8
: return __DRI_IMAGE_FOURCC_SARGB8888
;
1125 case __DRI_IMAGE_FORMAT_SABGR8
: return __DRI_IMAGE_FOURCC_SABGR8888
;
1126 case __DRI_IMAGE_FORMAT_RGB565
: return __DRI_IMAGE_FOURCC_RGB565
;
1127 case __DRI_IMAGE_FORMAT_XRGB8888
: return __DRI_IMAGE_FOURCC_XRGB8888
;
1128 case __DRI_IMAGE_FORMAT_ARGB8888
: return __DRI_IMAGE_FOURCC_ARGB8888
;
1129 case __DRI_IMAGE_FORMAT_ABGR8888
: return __DRI_IMAGE_FOURCC_ABGR8888
;
1130 case __DRI_IMAGE_FORMAT_XBGR8888
: return __DRI_IMAGE_FOURCC_XBGR8888
;
1131 case __DRI_IMAGE_FORMAT_XRGB2101010
: return __DRI_IMAGE_FOURCC_XRGB2101010
;
1132 case __DRI_IMAGE_FORMAT_ARGB2101010
: return __DRI_IMAGE_FOURCC_ARGB2101010
;
1133 case __DRI_IMAGE_FORMAT_XBGR2101010
: return __DRI_IMAGE_FOURCC_XBGR2101010
;
1134 case __DRI_IMAGE_FORMAT_ABGR2101010
: return __DRI_IMAGE_FOURCC_ABGR2101010
;
1139 #ifdef HAVE_DRI3_MODIFIERS
1141 has_supported_modifier(struct loader_dri3_drawable
*draw
, unsigned int format
,
1142 uint64_t *modifiers
, uint32_t count
)
1144 uint64_t *supported_modifiers
;
1145 int32_t supported_modifiers_count
;
1149 if (!draw
->ext
->image
->queryDmaBufModifiers(draw
->dri_screen
,
1150 format
, 0, NULL
, NULL
,
1151 &supported_modifiers_count
) ||
1152 supported_modifiers_count
== 0)
1155 supported_modifiers
= malloc(supported_modifiers_count
* sizeof(uint64_t));
1156 if (!supported_modifiers
)
1159 draw
->ext
->image
->queryDmaBufModifiers(draw
->dri_screen
, format
,
1160 supported_modifiers_count
,
1161 supported_modifiers
, NULL
,
1162 &supported_modifiers_count
);
1164 for (i
= 0; !found
&& i
< supported_modifiers_count
; i
++) {
1165 for (j
= 0; !found
&& j
< count
; j
++) {
1166 if (supported_modifiers
[i
] == modifiers
[j
])
1171 free(supported_modifiers
);
1176 /** loader_dri3_alloc_render_buffer
1178 * Use the driver createImage function to construct a __DRIimage, then
1179 * get a file descriptor for that and create an X pixmap from that
1181 * Allocate an xshmfence for synchronization
1183 static struct loader_dri3_buffer
*
1184 dri3_alloc_render_buffer(struct loader_dri3_drawable
*draw
, unsigned int format
,
1185 int width
, int height
, int depth
)
1187 struct loader_dri3_buffer
*buffer
;
1188 __DRIimage
*pixmap_buffer
;
1189 xcb_pixmap_t pixmap
;
1190 xcb_sync_fence_t sync_fence
;
1191 struct xshmfence
*shm_fence
;
1192 int buffer_fds
[4], fence_fd
;
1197 /* Create an xshmfence object and
1198 * prepare to send that to the X server
1201 fence_fd
= xshmfence_alloc_shm();
1205 shm_fence
= xshmfence_map_shm(fence_fd
);
1206 if (shm_fence
== NULL
)
1209 /* Allocate the image from the driver
1211 buffer
= calloc(1, sizeof *buffer
);
1215 buffer
->cpp
= dri3_cpp_for_format(format
);
1219 if (!draw
->is_different_gpu
) {
1220 #ifdef HAVE_DRI3_MODIFIERS
1221 if (draw
->multiplanes_available
&&
1222 draw
->ext
->image
->base
.version
>= 15 &&
1223 draw
->ext
->image
->queryDmaBufModifiers
&&
1224 draw
->ext
->image
->createImageWithModifiers
) {
1225 xcb_dri3_get_supported_modifiers_cookie_t mod_cookie
;
1226 xcb_dri3_get_supported_modifiers_reply_t
*mod_reply
;
1227 xcb_generic_error_t
*error
= NULL
;
1228 uint64_t *modifiers
= NULL
;
1231 mod_cookie
= xcb_dri3_get_supported_modifiers(draw
->conn
,
1233 depth
, buffer
->cpp
* 8);
1234 mod_reply
= xcb_dri3_get_supported_modifiers_reply(draw
->conn
,
1240 if (mod_reply
->num_window_modifiers
) {
1241 count
= mod_reply
->num_window_modifiers
;
1242 modifiers
= malloc(count
* sizeof(uint64_t));
1249 xcb_dri3_get_supported_modifiers_window_modifiers(mod_reply
),
1250 count
* sizeof(uint64_t));
1252 if (!has_supported_modifier(draw
, image_format_to_fourcc(format
),
1253 modifiers
, count
)) {
1260 if (mod_reply
->num_screen_modifiers
&& modifiers
== NULL
) {
1261 count
= mod_reply
->num_screen_modifiers
;
1262 modifiers
= malloc(count
* sizeof(uint64_t));
1270 xcb_dri3_get_supported_modifiers_screen_modifiers(mod_reply
),
1271 count
* sizeof(uint64_t));
1276 buffer
->image
= draw
->ext
->image
->createImageWithModifiers(draw
->dri_screen
,
1286 buffer
->image
= draw
->ext
->image
->createImage(draw
->dri_screen
,
1289 __DRI_IMAGE_USE_SHARE
|
1290 __DRI_IMAGE_USE_SCANOUT
|
1291 __DRI_IMAGE_USE_BACKBUFFER
,
1294 pixmap_buffer
= buffer
->image
;
1299 buffer
->image
= draw
->ext
->image
->createImage(draw
->dri_screen
,
1308 buffer
->linear_buffer
=
1309 draw
->ext
->image
->createImage(draw
->dri_screen
,
1311 dri3_linear_format_for_format(draw
, format
),
1312 __DRI_IMAGE_USE_SHARE
|
1313 __DRI_IMAGE_USE_LINEAR
|
1314 __DRI_IMAGE_USE_BACKBUFFER
,
1316 pixmap_buffer
= buffer
->linear_buffer
;
1318 if (!buffer
->linear_buffer
)
1319 goto no_linear_buffer
;
1322 /* X want some information about the planes, so ask the image for it
1324 if (!draw
->ext
->image
->queryImage(pixmap_buffer
, __DRI_IMAGE_ATTRIB_NUM_PLANES
,
1328 for (i
= 0; i
< num_planes
; i
++) {
1329 __DRIimage
*image
= draw
->ext
->image
->fromPlanar(pixmap_buffer
, i
, NULL
);
1333 image
= pixmap_buffer
;
1336 ret
= draw
->ext
->image
->queryImage(image
, __DRI_IMAGE_ATTRIB_FD
,
1338 ret
&= draw
->ext
->image
->queryImage(image
, __DRI_IMAGE_ATTRIB_STRIDE
,
1339 &buffer
->strides
[i
]);
1340 ret
&= draw
->ext
->image
->queryImage(image
, __DRI_IMAGE_ATTRIB_OFFSET
,
1341 &buffer
->offsets
[i
]);
1342 if (image
!= pixmap_buffer
)
1343 draw
->ext
->image
->destroyImage(image
);
1346 goto no_buffer_attrib
;
1349 ret
= draw
->ext
->image
->queryImage(pixmap_buffer
,
1350 __DRI_IMAGE_ATTRIB_MODIFIER_UPPER
, &mod
);
1351 buffer
->modifier
= (uint64_t) mod
<< 32;
1352 ret
&= draw
->ext
->image
->queryImage(pixmap_buffer
,
1353 __DRI_IMAGE_ATTRIB_MODIFIER_LOWER
, &mod
);
1354 buffer
->modifier
|= (uint64_t)(mod
& 0xffffffff);
1357 buffer
->modifier
= DRM_FORMAT_MOD_INVALID
;
1359 pixmap
= xcb_generate_id(draw
->conn
);
1360 #ifdef HAVE_DRI3_MODIFIERS
1361 if (draw
->multiplanes_available
&&
1362 buffer
->modifier
!= DRM_FORMAT_MOD_INVALID
) {
1363 xcb_dri3_pixmap_from_buffers(draw
->conn
,
1368 buffer
->strides
[0], buffer
->offsets
[0],
1369 buffer
->strides
[1], buffer
->offsets
[1],
1370 buffer
->strides
[2], buffer
->offsets
[2],
1371 buffer
->strides
[3], buffer
->offsets
[3],
1372 depth
, buffer
->cpp
* 8,
1378 xcb_dri3_pixmap_from_buffer(draw
->conn
,
1382 width
, height
, buffer
->strides
[0],
1383 depth
, buffer
->cpp
* 8,
1387 xcb_dri3_fence_from_fd(draw
->conn
,
1389 (sync_fence
= xcb_generate_id(draw
->conn
)),
1393 buffer
->pixmap
= pixmap
;
1394 buffer
->own_pixmap
= true;
1395 buffer
->sync_fence
= sync_fence
;
1396 buffer
->shm_fence
= shm_fence
;
1397 buffer
->width
= width
;
1398 buffer
->height
= height
;
1400 /* Mark the buffer as idle
1402 dri3_fence_set(buffer
);
1408 close(buffer_fds
[i
]);
1410 draw
->ext
->image
->destroyImage(pixmap_buffer
);
1412 if (draw
->is_different_gpu
)
1413 draw
->ext
->image
->destroyImage(buffer
->image
);
1417 xshmfence_unmap_shm(shm_fence
);
1423 /** loader_dri3_update_drawable
1425 * Called the first time we use the drawable and then
1426 * after we receive present configure notify events to
1427 * track the geometry of the drawable
1430 dri3_update_drawable(__DRIdrawable
*driDrawable
,
1431 struct loader_dri3_drawable
*draw
)
1433 mtx_lock(&draw
->mtx
);
1434 if (draw
->first_init
) {
1435 xcb_get_geometry_cookie_t geom_cookie
;
1436 xcb_get_geometry_reply_t
*geom_reply
;
1437 xcb_void_cookie_t cookie
;
1438 xcb_generic_error_t
*error
;
1439 xcb_present_query_capabilities_cookie_t present_capabilities_cookie
;
1440 xcb_present_query_capabilities_reply_t
*present_capabilities_reply
;
1441 xcb_window_t root_win
;
1443 draw
->first_init
= false;
1445 /* Try to select for input on the window.
1447 * If the drawable is a window, this will get our events
1450 * Otherwise, we'll get a BadWindow error back from this request which
1451 * will let us know that the drawable is a pixmap instead.
1454 draw
->eid
= xcb_generate_id(draw
->conn
);
1456 xcb_present_select_input_checked(draw
->conn
, draw
->eid
, draw
->drawable
,
1457 XCB_PRESENT_EVENT_MASK_CONFIGURE_NOTIFY
|
1458 XCB_PRESENT_EVENT_MASK_COMPLETE_NOTIFY
|
1459 XCB_PRESENT_EVENT_MASK_IDLE_NOTIFY
);
1461 present_capabilities_cookie
=
1462 xcb_present_query_capabilities(draw
->conn
, draw
->drawable
);
1464 /* Create an XCB event queue to hold present events outside of the usual
1465 * application event queue
1467 draw
->special_event
= xcb_register_for_special_xge(draw
->conn
,
1471 geom_cookie
= xcb_get_geometry(draw
->conn
, draw
->drawable
);
1473 geom_reply
= xcb_get_geometry_reply(draw
->conn
, geom_cookie
, NULL
);
1476 mtx_unlock(&draw
->mtx
);
1479 draw
->width
= geom_reply
->width
;
1480 draw
->height
= geom_reply
->height
;
1481 draw
->depth
= geom_reply
->depth
;
1482 draw
->vtable
->set_drawable_size(draw
, draw
->width
, draw
->height
);
1483 root_win
= geom_reply
->root
;
1487 draw
->is_pixmap
= false;
1489 /* Check to see if our select input call failed. If it failed with a
1490 * BadWindow error, then assume the drawable is a pixmap. Destroy the
1491 * special event queue created above and mark the drawable as a pixmap
1494 error
= xcb_request_check(draw
->conn
, cookie
);
1496 present_capabilities_reply
=
1497 xcb_present_query_capabilities_reply(draw
->conn
,
1498 present_capabilities_cookie
,
1501 if (present_capabilities_reply
) {
1502 draw
->present_capabilities
= present_capabilities_reply
->capabilities
;
1503 free(present_capabilities_reply
);
1505 draw
->present_capabilities
= 0;
1508 if (error
->error_code
!= BadWindow
) {
1510 mtx_unlock(&draw
->mtx
);
1513 draw
->is_pixmap
= true;
1514 xcb_unregister_for_special_event(draw
->conn
, draw
->special_event
);
1515 draw
->special_event
= NULL
;
1518 if (draw
->is_pixmap
)
1519 draw
->window
= root_win
;
1521 draw
->window
= draw
->drawable
;
1523 dri3_flush_present_events(draw
);
1524 mtx_unlock(&draw
->mtx
);
1529 loader_dri3_create_image(xcb_connection_t
*c
,
1530 xcb_dri3_buffer_from_pixmap_reply_t
*bp_reply
,
1531 unsigned int format
,
1532 __DRIscreen
*dri_screen
,
1533 const __DRIimageExtension
*image
,
1534 void *loaderPrivate
)
1537 __DRIimage
*image_planar
, *ret
;
1540 /* Get an FD for the pixmap object
1542 fds
= xcb_dri3_buffer_from_pixmap_reply_fds(c
, bp_reply
);
1544 stride
= bp_reply
->stride
;
1547 /* createImageFromFds creates a wrapper __DRIimage structure which
1548 * can deal with multiple planes for things like Yuv images. So, once
1549 * we've gotten the planar wrapper, pull the single plane out of it and
1550 * discard the wrapper.
1552 image_planar
= image
->createImageFromFds(dri_screen
,
1555 image_format_to_fourcc(format
),
1557 &stride
, &offset
, loaderPrivate
);
1562 ret
= image
->fromPlanar(image_planar
, 0, loaderPrivate
);
1567 image
->destroyImage(image_planar
);
1572 #ifdef HAVE_DRI3_MODIFIERS
1574 loader_dri3_create_image_from_buffers(xcb_connection_t
*c
,
1575 xcb_dri3_buffers_from_pixmap_reply_t
*bp_reply
,
1576 unsigned int format
,
1577 __DRIscreen
*dri_screen
,
1578 const __DRIimageExtension
*image
,
1579 void *loaderPrivate
)
1583 uint32_t *strides_in
, *offsets_in
;
1584 int strides
[4], offsets
[4];
1588 if (bp_reply
->nfd
> 4)
1591 fds
= xcb_dri3_buffers_from_pixmap_reply_fds(c
, bp_reply
);
1592 strides_in
= xcb_dri3_buffers_from_pixmap_strides(bp_reply
);
1593 offsets_in
= xcb_dri3_buffers_from_pixmap_offsets(bp_reply
);
1594 for (i
= 0; i
< bp_reply
->nfd
; i
++) {
1595 strides
[i
] = strides_in
[i
];
1596 offsets
[i
] = offsets_in
[i
];
1599 ret
= image
->createImageFromDmaBufs2(dri_screen
,
1602 image_format_to_fourcc(format
),
1606 0, 0, 0, 0, /* UNDEFINED */
1607 &error
, loaderPrivate
);
1609 for (i
= 0; i
< bp_reply
->nfd
; i
++)
1616 /** dri3_get_pixmap_buffer
1618 * Get the DRM object for a pixmap from the X server and
1619 * wrap that with a __DRIimage structure using createImageFromFds
1621 static struct loader_dri3_buffer
*
1622 dri3_get_pixmap_buffer(__DRIdrawable
*driDrawable
, unsigned int format
,
1623 enum loader_dri3_buffer_type buffer_type
,
1624 struct loader_dri3_drawable
*draw
)
1626 int buf_id
= loader_dri3_pixmap_buf_id(buffer_type
);
1627 struct loader_dri3_buffer
*buffer
= draw
->buffers
[buf_id
];
1628 xcb_drawable_t pixmap
;
1629 xcb_sync_fence_t sync_fence
;
1630 struct xshmfence
*shm_fence
;
1634 __DRIscreen
*cur_screen
;
1639 pixmap
= draw
->drawable
;
1641 buffer
= calloc(1, sizeof *buffer
);
1645 fence_fd
= xshmfence_alloc_shm();
1648 shm_fence
= xshmfence_map_shm(fence_fd
);
1649 if (shm_fence
== NULL
) {
1654 /* Get the currently-bound screen or revert to using the drawable's screen if
1655 * no contexts are currently bound. The latter case is at least necessary for
1656 * obs-studio, when using Window Capture (Xcomposite) as a Source.
1658 cur_screen
= draw
->vtable
->get_dri_screen();
1660 cur_screen
= draw
->dri_screen
;
1663 xcb_dri3_fence_from_fd(draw
->conn
,
1665 (sync_fence
= xcb_generate_id(draw
->conn
)),
1668 #ifdef HAVE_DRI3_MODIFIERS
1669 if (draw
->multiplanes_available
&&
1670 draw
->ext
->image
->base
.version
>= 15 &&
1671 draw
->ext
->image
->createImageFromDmaBufs2
) {
1672 xcb_dri3_buffers_from_pixmap_cookie_t bps_cookie
;
1673 xcb_dri3_buffers_from_pixmap_reply_t
*bps_reply
;
1675 bps_cookie
= xcb_dri3_buffers_from_pixmap(draw
->conn
, pixmap
);
1676 bps_reply
= xcb_dri3_buffers_from_pixmap_reply(draw
->conn
, bps_cookie
,
1681 loader_dri3_create_image_from_buffers(draw
->conn
, bps_reply
, format
,
1682 cur_screen
, draw
->ext
->image
,
1684 width
= bps_reply
->width
;
1685 height
= bps_reply
->height
;
1690 xcb_dri3_buffer_from_pixmap_cookie_t bp_cookie
;
1691 xcb_dri3_buffer_from_pixmap_reply_t
*bp_reply
;
1693 bp_cookie
= xcb_dri3_buffer_from_pixmap(draw
->conn
, pixmap
);
1694 bp_reply
= xcb_dri3_buffer_from_pixmap_reply(draw
->conn
, bp_cookie
, NULL
);
1698 buffer
->image
= loader_dri3_create_image(draw
->conn
, bp_reply
, format
,
1699 cur_screen
, draw
->ext
->image
,
1701 width
= bp_reply
->width
;
1702 height
= bp_reply
->height
;
1709 buffer
->pixmap
= pixmap
;
1710 buffer
->own_pixmap
= false;
1711 buffer
->width
= width
;
1712 buffer
->height
= height
;
1713 buffer
->shm_fence
= shm_fence
;
1714 buffer
->sync_fence
= sync_fence
;
1716 draw
->buffers
[buf_id
] = buffer
;
1721 xcb_sync_destroy_fence(draw
->conn
, sync_fence
);
1722 xshmfence_unmap_shm(shm_fence
);
1731 * Find a front or back buffer, allocating new ones as necessary
1733 static struct loader_dri3_buffer
*
1734 dri3_get_buffer(__DRIdrawable
*driDrawable
,
1735 unsigned int format
,
1736 enum loader_dri3_buffer_type buffer_type
,
1737 struct loader_dri3_drawable
*draw
)
1739 struct loader_dri3_buffer
*buffer
;
1742 if (buffer_type
== loader_dri3_buffer_back
) {
1743 draw
->back_format
= format
;
1745 buf_id
= dri3_find_back(draw
);
1750 buf_id
= LOADER_DRI3_FRONT_ID
;
1753 buffer
= draw
->buffers
[buf_id
];
1755 /* Allocate a new buffer if there isn't an old one, if that
1756 * old one is the wrong size, or if it's suboptimal
1758 if (!buffer
|| buffer
->width
!= draw
->width
||
1759 buffer
->height
!= draw
->height
||
1760 buffer
->reallocate
) {
1761 struct loader_dri3_buffer
*new_buffer
;
1763 /* Allocate the new buffers
1765 new_buffer
= dri3_alloc_render_buffer(draw
,
1773 /* When resizing, copy the contents of the old buffer, waiting for that
1774 * copy to complete using our fences before proceeding
1776 if ((buffer_type
== loader_dri3_buffer_back
||
1777 (buffer_type
== loader_dri3_buffer_front
&& draw
->have_fake_front
))
1780 /* Fill the new buffer with data from an old buffer */
1781 dri3_fence_await(draw
->conn
, draw
, buffer
);
1782 if (!loader_dri3_blit_image(draw
,
1785 0, 0, draw
->width
, draw
->height
,
1787 !buffer
->linear_buffer
) {
1788 dri3_fence_reset(draw
->conn
, new_buffer
);
1789 dri3_copy_area(draw
->conn
,
1792 dri3_drawable_gc(draw
),
1794 draw
->width
, draw
->height
);
1795 dri3_fence_trigger(draw
->conn
, new_buffer
);
1797 dri3_free_render_buffer(draw
, buffer
);
1798 } else if (buffer_type
== loader_dri3_buffer_front
) {
1799 /* Fill the new fake front with data from a real front */
1800 loader_dri3_swapbuffer_barrier(draw
);
1801 dri3_fence_reset(draw
->conn
, new_buffer
);
1802 dri3_copy_area(draw
->conn
,
1805 dri3_drawable_gc(draw
),
1807 draw
->width
, draw
->height
);
1808 dri3_fence_trigger(draw
->conn
, new_buffer
);
1810 if (new_buffer
->linear_buffer
) {
1811 dri3_fence_await(draw
->conn
, draw
, new_buffer
);
1812 (void) loader_dri3_blit_image(draw
,
1814 new_buffer
->linear_buffer
,
1815 0, 0, draw
->width
, draw
->height
,
1819 buffer
= new_buffer
;
1820 draw
->buffers
[buf_id
] = buffer
;
1822 dri3_fence_await(draw
->conn
, draw
, buffer
);
1825 * Do we need to preserve the content of a previous buffer?
1827 * Note that this blit is needed only to avoid a wait for a buffer that
1828 * is currently in the flip chain or being scanned out from. That's really
1829 * a tradeoff. If we're ok with the wait we can reduce the number of back
1830 * buffers to 1 for SWAP_EXCHANGE, and 1 for SWAP_COPY,
1831 * but in the latter case we must disallow page-flipping.
1833 if (buffer_type
== loader_dri3_buffer_back
&&
1834 draw
->cur_blit_source
!= -1 &&
1835 draw
->buffers
[draw
->cur_blit_source
] &&
1836 buffer
!= draw
->buffers
[draw
->cur_blit_source
]) {
1838 struct loader_dri3_buffer
*source
= draw
->buffers
[draw
->cur_blit_source
];
1840 /* Avoid flushing here. Will propably do good for tiling hardware. */
1841 (void) loader_dri3_blit_image(draw
,
1844 0, 0, draw
->width
, draw
->height
,
1846 buffer
->last_swap
= source
->last_swap
;
1847 draw
->cur_blit_source
= -1;
1849 /* Return the requested buffer */
1853 /** dri3_free_buffers
1855 * Free the front bufffer or all of the back buffers. Used
1856 * when the application changes which buffers it needs
1859 dri3_free_buffers(__DRIdrawable
*driDrawable
,
1860 enum loader_dri3_buffer_type buffer_type
,
1861 struct loader_dri3_drawable
*draw
)
1863 struct loader_dri3_buffer
*buffer
;
1868 switch (buffer_type
) {
1869 case loader_dri3_buffer_back
:
1870 first_id
= LOADER_DRI3_BACK_ID(0);
1871 n_id
= LOADER_DRI3_MAX_BACK
;
1872 draw
->cur_blit_source
= -1;
1874 case loader_dri3_buffer_front
:
1875 first_id
= LOADER_DRI3_FRONT_ID
;
1876 /* Don't free a fake front holding new backbuffer content. */
1877 n_id
= (draw
->cur_blit_source
== LOADER_DRI3_FRONT_ID
) ? 0 : 1;
1880 for (buf_id
= first_id
; buf_id
< first_id
+ n_id
; buf_id
++) {
1881 buffer
= draw
->buffers
[buf_id
];
1883 dri3_free_render_buffer(draw
, buffer
);
1884 draw
->buffers
[buf_id
] = NULL
;
1889 /** loader_dri3_get_buffers
1891 * The published buffer allocation API.
1892 * Returns all of the necessary buffers, allocating
1896 loader_dri3_get_buffers(__DRIdrawable
*driDrawable
,
1897 unsigned int format
,
1899 void *loaderPrivate
,
1900 uint32_t buffer_mask
,
1901 struct __DRIimageList
*buffers
)
1903 struct loader_dri3_drawable
*draw
= loaderPrivate
;
1904 struct loader_dri3_buffer
*front
, *back
;
1907 buffers
->image_mask
= 0;
1908 buffers
->front
= NULL
;
1909 buffers
->back
= NULL
;
1914 if (!dri3_update_drawable(driDrawable
, draw
))
1917 dri3_update_num_back(draw
);
1919 /* Free no longer needed back buffers */
1920 for (buf_id
= draw
->num_back
; buf_id
< LOADER_DRI3_MAX_BACK
; buf_id
++) {
1921 if (draw
->cur_blit_source
!= buf_id
&& draw
->buffers
[buf_id
]) {
1922 dri3_free_render_buffer(draw
, draw
->buffers
[buf_id
]);
1923 draw
->buffers
[buf_id
] = NULL
;
1927 /* pixmaps always have front buffers.
1928 * Exchange swaps also mandate fake front buffers.
1930 if (draw
->is_pixmap
|| draw
->swap_method
== __DRI_ATTRIB_SWAP_EXCHANGE
)
1931 buffer_mask
|= __DRI_IMAGE_BUFFER_FRONT
;
1933 if (buffer_mask
& __DRI_IMAGE_BUFFER_FRONT
) {
1934 /* All pixmaps are owned by the server gpu.
1935 * When we use a different gpu, we can't use the pixmap
1936 * as buffer since it is potentially tiled a way
1937 * our device can't understand. In this case, use
1938 * a fake front buffer. Hopefully the pixmap
1939 * content will get synced with the fake front
1942 if (draw
->is_pixmap
&& !draw
->is_different_gpu
)
1943 front
= dri3_get_pixmap_buffer(driDrawable
,
1945 loader_dri3_buffer_front
,
1948 front
= dri3_get_buffer(driDrawable
,
1950 loader_dri3_buffer_front
,
1956 dri3_free_buffers(driDrawable
, loader_dri3_buffer_front
, draw
);
1957 draw
->have_fake_front
= 0;
1960 if (buffer_mask
& __DRI_IMAGE_BUFFER_BACK
) {
1961 back
= dri3_get_buffer(driDrawable
,
1963 loader_dri3_buffer_back
,
1967 draw
->have_back
= 1;
1969 dri3_free_buffers(driDrawable
, loader_dri3_buffer_back
, draw
);
1970 draw
->have_back
= 0;
1974 buffers
->image_mask
|= __DRI_IMAGE_BUFFER_FRONT
;
1975 buffers
->front
= front
->image
;
1976 draw
->have_fake_front
= draw
->is_different_gpu
|| !draw
->is_pixmap
;
1980 buffers
->image_mask
|= __DRI_IMAGE_BUFFER_BACK
;
1981 buffers
->back
= back
->image
;
1984 draw
->stamp
= stamp
;
1989 /** loader_dri3_update_drawable_geometry
1991 * Get the current drawable geometry.
1994 loader_dri3_update_drawable_geometry(struct loader_dri3_drawable
*draw
)
1996 xcb_get_geometry_cookie_t geom_cookie
;
1997 xcb_get_geometry_reply_t
*geom_reply
;
1999 geom_cookie
= xcb_get_geometry(draw
->conn
, draw
->drawable
);
2001 geom_reply
= xcb_get_geometry_reply(draw
->conn
, geom_cookie
, NULL
);
2004 draw
->width
= geom_reply
->width
;
2005 draw
->height
= geom_reply
->height
;
2006 draw
->vtable
->set_drawable_size(draw
, draw
->width
, draw
->height
);
2007 draw
->ext
->flush
->invalidate(draw
->dri_drawable
);
2015 * Make sure the server has flushed all pending swap buffers to hardware
2016 * for this drawable. Ideally we'd want to send an X protocol request to
2017 * have the server block our connection until the swaps are complete. That
2018 * would avoid the potential round-trip here.
2021 loader_dri3_swapbuffer_barrier(struct loader_dri3_drawable
*draw
)
2023 int64_t ust
, msc
, sbc
;
2025 (void) loader_dri3_wait_for_sbc(draw
, 0, &ust
, &msc
, &sbc
);
2029 * Perform any cleanup associated with a close screen operation.
2030 * \param dri_screen[in,out] Pointer to __DRIscreen about to be closed.
2032 * This function destroys the screen's cached swap context if any.
2035 loader_dri3_close_screen(__DRIscreen
*dri_screen
)
2037 mtx_lock(&blit_context
.mtx
);
2038 if (blit_context
.ctx
&& blit_context
.cur_screen
== dri_screen
) {
2039 blit_context
.core
->destroyContext(blit_context
.ctx
);
2040 blit_context
.ctx
= NULL
;
2042 mtx_unlock(&blit_context
.mtx
);
2046 * Find a backbuffer slot - potentially allocating a back buffer
2048 * \param draw[in,out] Pointer to the drawable for which to find back.
2049 * \return Pointer to a new back buffer or NULL if allocation failed or was
2052 * Find a potentially new back buffer, and if it's not been allocated yet and
2053 * in addition needs initializing, then try to allocate and initialize it.
2056 static struct loader_dri3_buffer
*
2057 dri3_find_back_alloc(struct loader_dri3_drawable
*draw
)
2059 struct loader_dri3_buffer
*back
;
2062 id
= dri3_find_back(draw
);
2066 back
= draw
->buffers
[id
];
2067 /* Allocate a new back if we haven't got one */
2068 if (!back
&& draw
->back_format
!= __DRI_IMAGE_FORMAT_NONE
&&
2069 dri3_update_drawable(draw
->dri_drawable
, draw
))
2070 back
= dri3_alloc_render_buffer(draw
, draw
->back_format
,
2071 draw
->width
, draw
->height
, draw
->depth
);
2076 draw
->buffers
[id
] = back
;
2078 /* If necessary, prefill the back with data according to swap_method mode. */
2079 if (draw
->cur_blit_source
!= -1 &&
2080 draw
->buffers
[draw
->cur_blit_source
] &&
2081 back
!= draw
->buffers
[draw
->cur_blit_source
]) {
2082 struct loader_dri3_buffer
*source
= draw
->buffers
[draw
->cur_blit_source
];
2084 dri3_fence_await(draw
->conn
, draw
, source
);
2085 dri3_fence_await(draw
->conn
, draw
, back
);
2086 (void) loader_dri3_blit_image(draw
,
2089 0, 0, draw
->width
, draw
->height
,
2091 back
->last_swap
= source
->last_swap
;
2092 draw
->cur_blit_source
= -1;