X-Git-Url: https://git.libre-soc.org/?a=blobdiff_plain;f=src%2Fgallium%2Fdrivers%2Ffreedreno%2Ffreedreno_resource.h;h=9419097dcdfe82be30e148369c165eebab930718;hb=b54afde3ad631b33b42812c9ebf2225d9c53e38e;hp=7e10d7a6416419ffa8f5b0af148a303079db9e98;hpb=cb9e07aa84ff2dd8492fd8a7064ff2539d7d40f9;p=mesa.git diff --git a/src/gallium/drivers/freedreno/freedreno_resource.h b/src/gallium/drivers/freedreno/freedreno_resource.h index 7e10d7a6416..9419097dcdf 100644 --- a/src/gallium/drivers/freedreno/freedreno_resource.h +++ b/src/gallium/drivers/freedreno/freedreno_resource.h @@ -1,5 +1,3 @@ -/* -*- mode: C; c-file-style: "k&r"; tab-width 4; indent-tabs-mode: t; -*- */ - /* * Copyright (C) 2012 Rob Clark * @@ -29,45 +27,259 @@ #ifndef FREEDRENO_RESOURCE_H_ #define FREEDRENO_RESOURCE_H_ -#include "util/u_transfer.h" +#include "util/list.h" +#include "util/u_range.h" +#include "util/u_transfer_helper.h" +#include "util/simple_mtx.h" +#include "freedreno_batch.h" #include "freedreno_util.h" +#include "freedreno/fdl/freedreno_layout.h" -/* for mipmap, cubemap, etc, each level is represented by a slice. - * Currently all slices are part of same bo (just different offsets), - * this is at least how it needs to be for cubemaps, although mipmap - * can be different bo's (although, not sure if there is a strong - * advantage to doing that) - */ -struct fd_resource_slice { - uint32_t offset; /* offset of first layer in slice */ - uint32_t pitch; - uint32_t size0; /* size of first layer in slice */ +enum fd_lrz_direction { + FD_LRZ_UNKNOWN, + /* Depth func less/less-than: */ + FD_LRZ_LESS, + /* Depth func greater/greater-than: */ + FD_LRZ_GREATER, }; struct fd_resource { - struct u_resource base; + struct pipe_resource base; struct fd_bo *bo; - uint32_t cpp; - struct fd_resource_slice slices[MAX_MIP_LEVELS]; - uint32_t timestamp; - bool dirty; + enum pipe_format internal_format; + struct fdl_layout layout; + + /* buffer range that has been initialized */ + struct util_range valid_buffer_range; + bool valid; + struct renderonly_scanout *scanout; + + /* reference to the resource holding stencil data for a z32_s8 texture */ + /* TODO rename to secondary or auxiliary? */ + struct fd_resource *stencil; + + simple_mtx_t lock; + + /* bitmask of in-flight batches which reference this resource. Note + * that the batch doesn't hold reference to resources (but instead + * the fd_ringbuffer holds refs to the underlying fd_bo), but in case + * the resource is destroyed we need to clean up the batch's weak + * references to us. + */ + uint32_t batch_mask; + + /* reference to batch that writes this resource: */ + struct fd_batch *write_batch; + + /* Set of batches whose batch-cache key references this resource. + * We need to track this to know which batch-cache entries to + * invalidate if, for example, the resource is invalidated or + * shadowed. + */ + uint32_t bc_batch_mask; + + /* Sequence # incremented each time bo changes: */ + uint16_t seqno; + + /* bitmask of state this resource could potentially dirty when rebound, + * see rebind_resource() + */ + enum fd_dirty_3d_state dirty; + + /* + * LRZ + * + * TODO lrz width/height/pitch should probably also move to + * fdl_layout + */ + bool lrz_valid : 1; + enum fd_lrz_direction lrz_direction : 2; + uint16_t lrz_width; // for lrz clear, does this differ from lrz_pitch? + uint16_t lrz_height; + uint16_t lrz_pitch; + struct fd_bo *lrz; }; -static INLINE struct fd_resource * +static inline struct fd_resource * fd_resource(struct pipe_resource *ptex) { return (struct fd_resource *)ptex; } -static INLINE struct fd_resource_slice * +static inline const struct fd_resource * +fd_resource_const(const struct pipe_resource *ptex) +{ + return (const struct fd_resource *)ptex; +} + +static inline bool +pending(struct fd_resource *rsc, bool write) +{ + /* if we have a pending GPU write, we are busy in any case: */ + if (rsc->write_batch) + return true; + + /* if CPU wants to write, but we are pending a GPU read, we are busy: */ + if (write && rsc->batch_mask) + return true; + + if (rsc->stencil && pending(rsc->stencil, write)) + return true; + + return false; +} + +static inline bool +fd_resource_busy(struct fd_resource *rsc, unsigned op) +{ + return fd_bo_cpu_prep(rsc->bo, NULL, op | DRM_FREEDRENO_PREP_NOSYNC) != 0; +} + +static inline void +fd_resource_lock(struct fd_resource *rsc) +{ + simple_mtx_lock(&rsc->lock); +} + +static inline void +fd_resource_unlock(struct fd_resource *rsc) +{ + simple_mtx_unlock(&rsc->lock); +} + +static inline void +fd_resource_set_usage(struct pipe_resource *prsc, enum fd_dirty_3d_state usage) +{ + if (!prsc) + return; + struct fd_resource *rsc = fd_resource(prsc); + /* Bits are only ever ORed in, and we expect many set_usage() per + * resource, so do the quick check outside of the lock. + */ + if (likely(rsc->dirty & usage)) + return; + fd_resource_lock(rsc); + rsc->dirty |= usage; + fd_resource_unlock(rsc); +} + +static inline bool +has_depth(enum pipe_format format) +{ + const struct util_format_description *desc = + util_format_description(format); + return util_format_has_depth(desc); +} + +struct fd_transfer { + struct pipe_transfer base; + struct pipe_resource *staging_prsc; + struct pipe_box staging_box; +}; + +static inline struct fd_transfer * +fd_transfer(struct pipe_transfer *ptrans) +{ + return (struct fd_transfer *)ptrans; +} + +static inline struct fdl_slice * fd_resource_slice(struct fd_resource *rsc, unsigned level) { - assert(level <= rsc->base.b.last_level); - return &rsc->slices[level]; + assert(level <= rsc->base.last_level); + return &rsc->layout.slices[level]; +} + +static inline uint32_t +fd_resource_layer_stride(struct fd_resource *rsc, unsigned level) +{ + return fdl_layer_stride(&rsc->layout, level); +} + +/* get pitch (in bytes) for specified mipmap level */ +static inline uint32_t +fd_resource_pitch(struct fd_resource *rsc, unsigned level) +{ + if (is_a2xx(fd_screen(rsc->base.screen))) + return fdl2_pitch(&rsc->layout, level); + + return fdl_pitch(&rsc->layout, level); +} + +/* get offset for specified mipmap level and texture/array layer */ +static inline uint32_t +fd_resource_offset(struct fd_resource *rsc, unsigned level, unsigned layer) +{ + uint32_t offset = fdl_surface_offset(&rsc->layout, level, layer); + debug_assert(offset < fd_bo_size(rsc->bo)); + return offset; +} + +static inline uint32_t +fd_resource_ubwc_offset(struct fd_resource *rsc, unsigned level, unsigned layer) +{ + uint32_t offset = fdl_ubwc_offset(&rsc->layout, level, layer); + debug_assert(offset < fd_bo_size(rsc->bo)); + return offset; +} + +/* This might be a5xx specific, but higher mipmap levels are always linear: */ +static inline bool +fd_resource_level_linear(const struct pipe_resource *prsc, int level) +{ + struct fd_screen *screen = fd_screen(prsc->screen); + debug_assert(!is_a3xx(screen)); + + return fdl_level_linear(&fd_resource_const(prsc)->layout, level); +} + +static inline uint32_t +fd_resource_tile_mode(struct pipe_resource *prsc, int level) +{ + return fdl_tile_mode(&fd_resource(prsc)->layout, level); +} + +static inline bool +fd_resource_ubwc_enabled(struct fd_resource *rsc, int level) +{ + return fdl_ubwc_enabled(&rsc->layout, level); +} + +/* access # of samples, with 0 normalized to 1 (which is what we care about + * most of the time) + */ +static inline unsigned +fd_resource_nr_samples(struct pipe_resource *prsc) +{ + return MAX2(1, prsc->nr_samples); } void fd_resource_screen_init(struct pipe_screen *pscreen); void fd_resource_context_init(struct pipe_context *pctx); +uint32_t fd_setup_slices(struct fd_resource *rsc); +void fd_resource_resize(struct pipe_resource *prsc, uint32_t sz); +void fd_resource_uncompress(struct fd_context *ctx, struct fd_resource *rsc); + +bool fd_render_condition_check(struct pipe_context *pctx); + +static inline bool +fd_batch_references_resource(struct fd_batch *batch, struct fd_resource *rsc) +{ + return rsc->batch_mask & (1 << batch->idx); +} + +static inline void +fd_batch_resource_read(struct fd_batch *batch, + struct fd_resource *rsc) +{ + /* Fast path: if we hit this then we know we don't have anyone else + * writing to it (since both _write and _read flush other writers), and + * that we've already recursed for stencil. + */ + if (unlikely(!fd_batch_references_resource(batch, rsc))) + fd_batch_resource_read_slowpath(batch, rsc); +} + #endif /* FREEDRENO_RESOURCE_H_ */