2 * Copyright 2010 Jerome Glisse <glisse@freedesktop.org>
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * on the rights to use, copy, modify, merge, publish, distribute, sub
8 * license, and/or sell copies of the Software, and to permit persons to whom
9 * the Software is furnished to do so, subject to the following conditions:
11 * The above copyright notice and this permission notice (including the next
12 * paragraph) shall be included in all copies or substantial portions of the
15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17 * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
18 * THE AUTHOR(S) AND/OR THEIR SUPPLIERS BE LIABLE FOR ANY CLAIM,
19 * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
20 * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
21 * USE OR OTHER DEALINGS IN THE SOFTWARE.
25 * Corbin Simpson <MostAwesomeDude@gmail.com>
27 #include "r600_pipe.h"
28 #include "util/u_upload_mgr.h"
29 #include "util/u_memory.h"
31 static void r600_buffer_destroy(struct pipe_screen
*screen
,
32 struct pipe_resource
*buf
)
34 struct r600_resource
*rbuffer
= r600_resource(buf
);
36 pb_reference(&rbuffer
->buf
, NULL
);
40 static struct pipe_transfer
*r600_get_transfer(struct pipe_context
*ctx
,
41 struct pipe_resource
*resource
,
44 const struct pipe_box
*box
)
46 struct r600_context
*rctx
= (struct r600_context
*)ctx
;
47 struct r600_transfer
*transfer
= util_slab_alloc(&rctx
->pool_transfers
);
49 assert(box
->x
+ box
->width
<= resource
->width0
);
51 transfer
->transfer
.resource
= resource
;
52 transfer
->transfer
.level
= level
;
53 transfer
->transfer
.usage
= usage
;
54 transfer
->transfer
.box
= *box
;
55 transfer
->transfer
.stride
= 0;
56 transfer
->transfer
.layer_stride
= 0;
57 transfer
->transfer
.data
= NULL
;
58 transfer
->staging
= NULL
;
61 /* Note strides are zero, this is ok for buffers, but not for
62 * textures 2d & higher at least.
64 return &transfer
->transfer
;
67 static void r600_set_constants_dirty_if_bound(struct r600_context
*rctx
,
68 struct r600_constbuf_state
*state
,
69 struct r600_resource
*rbuffer
)
72 uint32_t mask
= state
->enabled_mask
;
75 unsigned i
= u_bit_scan(&mask
);
76 if (state
->cb
[i
].buffer
== &rbuffer
->b
.b
) {
78 state
->dirty_mask
|= 1 << i
;
82 r600_constant_buffers_dirty(rctx
, state
);
86 static void *r600_buffer_transfer_map(struct pipe_context
*pipe
,
87 struct pipe_transfer
*transfer
)
89 struct r600_resource
*rbuffer
= r600_resource(transfer
->resource
);
90 struct r600_context
*rctx
= (struct r600_context
*)pipe
;
93 if (transfer
->usage
& PIPE_TRANSFER_DISCARD_WHOLE_RESOURCE
&&
94 !(transfer
->usage
& PIPE_TRANSFER_UNSYNCHRONIZED
)) {
95 assert(transfer
->usage
& PIPE_TRANSFER_WRITE
);
97 /* Check if mapping this buffer would cause waiting for the GPU. */
98 if (rctx
->ws
->cs_is_buffer_referenced(rctx
->cs
, rbuffer
->cs_buf
, RADEON_USAGE_READWRITE
) ||
99 rctx
->ws
->buffer_is_busy(rbuffer
->buf
, RADEON_USAGE_READWRITE
)) {
102 /* Discard the buffer. */
103 pb_reference(&rbuffer
->buf
, NULL
);
105 /* Create a new one in the same pipe_resource. */
106 /* XXX We probably want a different alignment for buffers and textures. */
107 r600_init_resource(rctx
->screen
, rbuffer
, rbuffer
->b
.b
.width0
, 4096,
108 rbuffer
->b
.b
.bind
, rbuffer
->b
.b
.usage
);
110 /* We changed the buffer, now we need to bind it where the old one was bound. */
111 /* Vertex buffers. */
112 mask
= rctx
->vertex_buffer_state
.enabled_mask
;
114 i
= u_bit_scan(&mask
);
115 if (rctx
->vertex_buffer_state
.vb
[i
].buffer
== &rbuffer
->b
.b
) {
116 rctx
->vertex_buffer_state
.dirty_mask
|= 1 << i
;
117 r600_vertex_buffers_dirty(rctx
);
120 /* Streamout buffers. */
121 for (i
= 0; i
< rctx
->num_so_targets
; i
++) {
122 if (rctx
->so_targets
[i
]->b
.buffer
== &rbuffer
->b
.b
) {
123 r600_context_streamout_end(rctx
);
124 rctx
->streamout_start
= TRUE
;
125 rctx
->streamout_append_bitmask
= ~0;
128 /* Constant buffers. */
129 r600_set_constants_dirty_if_bound(rctx
, &rctx
->vs_constbuf_state
, rbuffer
);
130 r600_set_constants_dirty_if_bound(rctx
, &rctx
->ps_constbuf_state
, rbuffer
);
133 else if ((transfer
->usage
& PIPE_TRANSFER_DISCARD_RANGE
) &&
134 !(transfer
->usage
& PIPE_TRANSFER_UNSYNCHRONIZED
) &&
135 rctx
->screen
->has_streamout
&&
136 /* The buffer range must be aligned to 4. */
137 transfer
->box
.x
% 4 == 0 && transfer
->box
.width
% 4 == 0) {
138 assert(transfer
->usage
& PIPE_TRANSFER_WRITE
);
140 /* Check if mapping this buffer would cause waiting for the GPU. */
141 if (rctx
->ws
->cs_is_buffer_referenced(rctx
->cs
, rbuffer
->cs_buf
, RADEON_USAGE_READWRITE
) ||
142 rctx
->ws
->buffer_is_busy(rbuffer
->buf
, RADEON_USAGE_READWRITE
)) {
143 /* Do a wait-free write-only transfer using a temporary buffer. */
144 struct r600_transfer
*rtransfer
= (struct r600_transfer
*)transfer
;
146 rtransfer
->staging
= (struct r600_resource
*)
147 pipe_buffer_create(pipe
->screen
, PIPE_BIND_VERTEX_BUFFER
,
148 PIPE_USAGE_STAGING
, transfer
->box
.width
);
149 return rctx
->ws
->buffer_map(rtransfer
->staging
->cs_buf
, rctx
->cs
, PIPE_TRANSFER_WRITE
);
153 data
= rctx
->ws
->buffer_map(rbuffer
->cs_buf
, rctx
->cs
, transfer
->usage
);
157 return (uint8_t*)data
+ transfer
->box
.x
;
160 static void r600_buffer_transfer_unmap(struct pipe_context
*pipe
,
161 struct pipe_transfer
*transfer
)
163 struct r600_transfer
*rtransfer
= (struct r600_transfer
*)transfer
;
165 if (rtransfer
->staging
) {
167 u_box_1d(0, transfer
->box
.width
, &box
);
169 /* Copy the staging buffer into the original one. */
170 r600_copy_buffer(pipe
, transfer
->resource
, transfer
->box
.x
,
171 &rtransfer
->staging
->b
.b
, &box
);
172 pipe_resource_reference((struct pipe_resource
**)&rtransfer
->staging
, NULL
);
176 static void r600_transfer_destroy(struct pipe_context
*ctx
,
177 struct pipe_transfer
*transfer
)
179 struct r600_context
*rctx
= (struct r600_context
*)ctx
;
180 util_slab_free(&rctx
->pool_transfers
, transfer
);
183 static const struct u_resource_vtbl r600_buffer_vtbl
=
185 u_default_resource_get_handle
, /* get_handle */
186 r600_buffer_destroy
, /* resource_destroy */
187 r600_get_transfer
, /* get_transfer */
188 r600_transfer_destroy
, /* transfer_destroy */
189 r600_buffer_transfer_map
, /* transfer_map */
190 NULL
, /* transfer_flush_region */
191 r600_buffer_transfer_unmap
, /* transfer_unmap */
192 NULL
/* transfer_inline_write */
195 bool r600_init_resource(struct r600_screen
*rscreen
,
196 struct r600_resource
*res
,
197 unsigned size
, unsigned alignment
,
198 unsigned bind
, unsigned usage
)
200 uint32_t initial_domain
, domains
;
202 /* Staging resources particpate in transfers and blits only
203 * and are used for uploads and downloads from regular
204 * resources. We generate them internally for some transfers.
206 if (usage
== PIPE_USAGE_STAGING
) {
207 domains
= RADEON_DOMAIN_GTT
;
208 initial_domain
= RADEON_DOMAIN_GTT
;
210 domains
= RADEON_DOMAIN_GTT
| RADEON_DOMAIN_VRAM
;
213 case PIPE_USAGE_DYNAMIC
:
214 case PIPE_USAGE_STREAM
:
215 case PIPE_USAGE_STAGING
:
216 initial_domain
= RADEON_DOMAIN_GTT
;
218 case PIPE_USAGE_DEFAULT
:
219 case PIPE_USAGE_STATIC
:
220 case PIPE_USAGE_IMMUTABLE
:
222 initial_domain
= RADEON_DOMAIN_VRAM
;
227 res
->buf
= rscreen
->ws
->buffer_create(rscreen
->ws
, size
, alignment
, bind
, initial_domain
);
232 res
->cs_buf
= rscreen
->ws
->buffer_get_cs_handle(res
->buf
);
233 res
->domains
= domains
;
237 struct pipe_resource
*r600_buffer_create(struct pipe_screen
*screen
,
238 const struct pipe_resource
*templ
)
240 struct r600_screen
*rscreen
= (struct r600_screen
*)screen
;
241 struct r600_resource
*rbuffer
;
242 /* XXX We probably want a different alignment for buffers and textures. */
243 unsigned alignment
= 4096;
245 rbuffer
= MALLOC_STRUCT(r600_resource
);
247 rbuffer
->b
.b
= *templ
;
248 pipe_reference_init(&rbuffer
->b
.b
.reference
, 1);
249 rbuffer
->b
.b
.screen
= screen
;
250 rbuffer
->b
.vtbl
= &r600_buffer_vtbl
;
252 if (!r600_init_resource(rscreen
, rbuffer
, templ
->width0
, alignment
, templ
->bind
, templ
->usage
)) {
256 return &rbuffer
->b
.b
;