nir: don't use GLenum16 in nir.h
authorMarek Olšák <marek.olsak@amd.com>
Tue, 12 Nov 2019 01:03:40 +0000 (20:03 -0500)
committerMarek Olšák <marek.olsak@amd.com>
Tue, 19 Nov 2019 23:20:12 +0000 (18:20 -0500)
Reviewed-by: Connor Abbott <cwabbott0@gmail.com>
src/compiler/nir/nir.h

index 29ef2ce0d7d9a9ef8881a08457f816c43fd08c0d..9bd1d321e032420a04f7b16f24e5ff789cff6e0d 100644 (file)
@@ -62,7 +62,6 @@ extern "C" {
 #define NIR_MAX_MATRIX_COLUMNS 4
 #define NIR_STREAM_PACKED (1 << 8)
 typedef uint8_t nir_component_mask_t;
-typedef unsigned short GLenum16;
 
 /** Defines a cast function
  *
@@ -488,7 +487,7 @@ typedef struct nir_variable {
       union {
          struct {
             /** Image internal format if specified explicitly, otherwise GL_NONE. */
-            GLenum16 format;
+            uint16_t format; /* GLenum */
          } image;
 
          struct {