|
|
//============ Copyright (c) Valve Corporation, All rights reserved. ============
//
// cglmtex.cpp
//
//===============================================================================
#include "glmgr/glmgr.h"
#include "glmgr/cglmtex.h"
#include "tier0/icommandline.h"
#include "../shaderapidx9/dxabstract.h"
// memdbgon -must- be the last include file in a .cpp file.
#include "tier0/memdbgon.h"
//===============================================================================
#define TEXSPACE_LOGGING 0
// encoding layout to an index where the bits read
// 4 : 1 if compressed
// 2 : 1 if not power of two
// 1 : 1 if mipmapped
bool pwroftwo (int val ) { return (val & (val-1)) == 0; }
int sEncodeLayoutAsIndex( GLMTexLayoutKey *key ) { int index = 0; if (key->m_texFlags & kGLMTexMipped) { index |= 1; }
if ( ! ( pwroftwo(key->m_xSize) && pwroftwo(key->m_ySize) && pwroftwo(key->m_zSize) ) ) { // if not all power of two
index |= 2; } if (GetFormatDesc( key->m_texFormat )->m_chunkSize >1 ) { index |= 4; }
return index; }
static unsigned long g_texGlobalBytes[8];
//===============================================================================
const GLMTexFormatDesc g_formatDescTable[] = { // not yet handled by this table:
// D3DFMT_INDEX16, D3DFMT_VERTEXDATA // D3DFMT_INDEX32,
// WTF { D3DFMT_R5G6R5 ???, GL_RGB, GL_RGB, GL_UNSIGNED_SHORT_5_6_5, 1, 2 },
// WTF { D3DFMT_A ???, GL_ALPHA8, GL_ALPHA, GL_UNSIGNED_BYTE, 1, 1 },
// ??? D3DFMT_V8U8,
// ??? D3DFMT_Q8W8V8U8,
// ??? D3DFMT_X8L8V8U8,
// ??? D3DFMT_R32F,
// ??? D3DFMT_D24X4S4 unsure how to handle or if it is ever used..
// ??? D3DFMT_D15S1 ever used ?
// ??? D3DFMT_D24X8 ever used?
// summ-name d3d-format gl-int-format gl-int-format-srgb gl-data-format gl-data-type chunksize, bytes-per-sqchunk
{ "_D16", D3DFMT_D16, GL_DEPTH_COMPONENT16, 0, GL_DEPTH_COMPONENT, GL_UNSIGNED_SHORT, 1, 2 }, { "_D24X8", D3DFMT_D24X8, GL_DEPTH_COMPONENT24, 0, GL_DEPTH_COMPONENT, GL_UNSIGNED_INT, 1, 4 }, // ??? unsure on this one
{ "_D24S8", D3DFMT_D24S8, GL_DEPTH24_STENCIL8_EXT, 0, GL_DEPTH_STENCIL_EXT, GL_UNSIGNED_INT_24_8_EXT, 1, 4 },
{ "_A8R8G8B8", D3DFMT_A8R8G8B8, GL_RGBA8, GL_SRGB8_ALPHA8_EXT, GL_BGRA, GL_UNSIGNED_INT_8_8_8_8_REV, 1, 4 }, { "_A4R4G4B4", D3DFMT_A4R4G4B4, GL_RGBA4, 0, GL_BGRA, GL_UNSIGNED_SHORT_4_4_4_4_REV, 1, 2 }, { "_X8R8G8B8", D3DFMT_X8R8G8B8, GL_RGB8, GL_SRGB8_EXT, GL_BGRA, GL_UNSIGNED_INT_8_8_8_8_REV, 1, 4 },
{ "_X1R5G5B5", D3DFMT_X1R5G5B5, GL_RGB5, 0, GL_BGRA, GL_UNSIGNED_SHORT_1_5_5_5_REV, 1, 2 }, { "_A1R5G5B5", D3DFMT_A1R5G5B5, GL_RGB5_A1, 0, GL_BGRA, GL_UNSIGNED_SHORT_1_5_5_5_REV, 1, 2 },
{ "_L8", D3DFMT_L8, GL_LUMINANCE8, GL_SLUMINANCE8_EXT, GL_LUMINANCE, GL_UNSIGNED_BYTE, 1, 1 }, { "_A8L8", D3DFMT_A8L8, GL_LUMINANCE8_ALPHA8, GL_SLUMINANCE8_ALPHA8_EXT, GL_LUMINANCE_ALPHA, GL_UNSIGNED_BYTE, 1, 2 },
{ "_DXT1", D3DFMT_DXT1, GL_COMPRESSED_RGB_S3TC_DXT1_EXT, GL_COMPRESSED_SRGB_S3TC_DXT1_EXT, GL_RGB, GL_UNSIGNED_BYTE, 4, 8 }, { "_DXT3", D3DFMT_DXT3, GL_COMPRESSED_RGBA_S3TC_DXT3_EXT, GL_COMPRESSED_SRGB_ALPHA_S3TC_DXT3_EXT, GL_RGBA, GL_UNSIGNED_BYTE, 4, 16 }, { "_DXT5", D3DFMT_DXT5, GL_COMPRESSED_RGBA_S3TC_DXT5_EXT, GL_COMPRESSED_SRGB_ALPHA_S3TC_DXT5_EXT, GL_RGBA, GL_UNSIGNED_BYTE, 4, 16 },
{ "_A16B16G16R16F", D3DFMT_A16B16G16R16F, GL_RGBA16F_ARB, 0, GL_RGBA, GL_HALF_FLOAT_ARB, 1, 8 }, { "_A16B16G16R16", D3DFMT_A16B16G16R16, GL_RGBA16, 0, GL_RGBA, GL_UNSIGNED_SHORT, 1, 8 }, // 16bpc integer tex
{ "_A32B32G32R32F", D3DFMT_A32B32G32R32F, GL_RGBA32F_ARB, 0, GL_RGBA, GL_FLOAT, 1, 16 },
{ "_R8G8B8", D3DFMT_R8G8B8, GL_RGB8, GL_SRGB8_EXT, GL_BGR, GL_UNSIGNED_BYTE, 1, 3 },
{ "_A8", D3DFMT_A8, GL_ALPHA8, 0, GL_ALPHA, GL_UNSIGNED_BYTE, 1, 1 }, { "_R5G6B5", D3DFMT_R5G6B5, GL_RGB, GL_SRGB_EXT, GL_RGB, GL_UNSIGNED_SHORT_5_6_5, 1, 2 },
// fakey tex formats: the stated GL format and the memory layout may not agree (U8V8 for example)
// _Q8W8V8U8 we just pass through as RGBA bytes. Shader does scale/bias fix
{ "_Q8W8V8U8", D3DFMT_Q8W8V8U8, GL_RGBA8, 0, GL_BGRA, GL_UNSIGNED_INT_8_8_8_8_REV, 1, 4 }, // straight ripoff of D3DFMT_A8R8G8B8
// U8V8 is exposed to the client as 2-bytes per texel, but we download it as 3-byte RGB.
// WriteTexels needs to do that conversion from rg8 to rgb8 in order to be able to download it correctly
{ "_V8U8", D3DFMT_V8U8, GL_RGB8, 0, GL_RG, GL_BYTE, 1, 2 },
// Test - this is oviously not correct:
{ "_R32F", D3DFMT_R32F, GL_RGBA8, GL_SRGB8_ALPHA8_EXT, GL_BGRA, GL_UNSIGNED_INT_8_8_8_8_REV, 1, 4 }, // { "_R32F", D3DFMT_R32F, GL_R32F, 0, GL_RED, GL_FLOAT, 1, 4 },
// Test - this is oviously not correct:
{ "_A2B10G10R10", D3DFMT_A2R10G10B10, GL_RGBA8, GL_SRGB8_ALPHA8_EXT, GL_BGRA, GL_UNSIGNED_INT_8_8_8_8_REV, 1, 4 }, // { "_A2B10G10R10", D3DFMT_A2R10G10B10, GL_UNSIGNED_INT_10_10_10_2, GL_UNSIGNED_INT_10_10_10_2, GL_RGBA, GL_UNSIGNED_INT, 1, 4 },
/*
// NV shadow depth tex
D3DFMT_NV_INTZ = 0x5a544e49, // MAKEFOURCC('I','N','T','Z')
D3DFMT_NV_RAWZ = 0x5a574152, // MAKEFOURCC('R','A','W','Z')
// NV null tex
D3DFMT_NV_NULL = 0x4c4c554e, // MAKEFOURCC('N','U','L','L')
// ATI shadow depth tex
D3DFMT_ATI_D16 = 0x36314644, // MAKEFOURCC('D','F','1','6')
D3DFMT_ATI_D24S8 = 0x34324644, // MAKEFOURCC('D','F','2','4')
// ATI 1N and 2N compressed tex
D3DFMT_ATI_2N = 0x32495441, // MAKEFOURCC('A', 'T', 'I', '2')
D3DFMT_ATI_1N = 0x31495441, // MAKEFOURCC('A', 'T', 'I', '1')
*/ };
int g_formatDescTableCount = sizeof(g_formatDescTable) / sizeof( g_formatDescTable[0] );
const GLMTexFormatDesc *GetFormatDesc( D3DFORMAT format ) { for( int i=0; i<g_formatDescTableCount; i++) { if (g_formatDescTable[i].m_d3dFormat == format) { return &g_formatDescTable[i]; } } return (const GLMTexFormatDesc *)NULL; // not found
}
//===============================================================================
void InsertTexelComponentFixed( float value, int width, unsigned long *valuebuf ) { unsigned long range = (1<<width); unsigned long scaled = (value * (float) range) * (range-1) / (range); if (scaled >= range) Debugger(); *valuebuf = (*valuebuf << width) | scaled; }
// return true if successful
bool GLMGenTexels( GLMGenTexelParams *params ) { unsigned char chunkbuf[256]; // can't think of any chunk this big..
const GLMTexFormatDesc *format = GetFormatDesc( params->m_format );
if (!format) { return FALSE; // fail
} // this section just generates one square chunk in the desired format
unsigned long *temp32 = (unsigned long*)chunkbuf; unsigned int chunksize = 0; // we can sanity check against the format table with this
switch( params->m_format ) { // comment shows byte order in RAM
// lowercase is bit arrangement in a byte
case D3DFMT_A8R8G8B8: // B G R A
InsertTexelComponentFixed( params->a, 8, temp32 ); // A is inserted first and winds up at most significant bits after insertions follow
InsertTexelComponentFixed( params->r, 8, temp32 ); InsertTexelComponentFixed( params->g, 8, temp32 ); InsertTexelComponentFixed( params->b, 8, temp32 ); chunksize = 4; break; case D3DFMT_A4R4G4B4: // [ggggbbbb] [aaaarrrr] RA (nibbles)
InsertTexelComponentFixed( params->a, 4, temp32 ); InsertTexelComponentFixed( params->r, 4, temp32 ); InsertTexelComponentFixed( params->g, 4, temp32 ); InsertTexelComponentFixed( params->b, 4, temp32 ); chunksize = 2; break; case D3DFMT_X8R8G8B8: // B G R X
InsertTexelComponentFixed( 0.0, 8, temp32 ); InsertTexelComponentFixed( params->r, 8, temp32 ); InsertTexelComponentFixed( params->g, 8, temp32 ); InsertTexelComponentFixed( params->b, 8, temp32 ); chunksize = 4; break; case D3DFMT_X1R5G5B5: // [gggbbbbb] [xrrrrrgg]
InsertTexelComponentFixed( 0.0, 1, temp32 ); InsertTexelComponentFixed( params->r, 5, temp32 ); InsertTexelComponentFixed( params->g, 5, temp32 ); InsertTexelComponentFixed( params->b, 5, temp32 ); chunksize = 2; break;
case D3DFMT_A1R5G5B5: // [gggbbbbb] [arrrrrgg]
InsertTexelComponentFixed( params->a, 1, temp32 ); InsertTexelComponentFixed( params->r, 5, temp32 ); InsertTexelComponentFixed( params->g, 5, temp32 ); InsertTexelComponentFixed( params->b, 5, temp32 ); chunksize = 2; break;
case D3DFMT_L8: // L // caller, use R for L
InsertTexelComponentFixed( params->r, 8, temp32 ); chunksize = 1; break; case D3DFMT_A8L8: // L A // caller, use R for L and A for A
InsertTexelComponentFixed( params->a, 8, temp32 ); InsertTexelComponentFixed( params->r, 8, temp32 ); chunksize = 2; break; case D3DFMT_R8G8B8: // B G R
InsertTexelComponentFixed( params->r, 8, temp32 ); InsertTexelComponentFixed( params->g, 8, temp32 ); InsertTexelComponentFixed( params->b, 8, temp32 ); chunksize = 3; break;
case D3DFMT_A8: // A
InsertTexelComponentFixed( params->a, 8, temp32 ); chunksize = 1; break; case D3DFMT_R5G6B5: // [gggbbbbb] [rrrrrggg]
InsertTexelComponentFixed( params->r, 5, temp32 ); InsertTexelComponentFixed( params->g, 6, temp32 ); InsertTexelComponentFixed( params->b, 5, temp32 ); chunksize = 2; break;
case D3DFMT_DXT1: { memset( temp32, 0, 8 ); // zap 8 bytes
// two 565 RGB words followed by 32 bits of 2-bit interp values for a 4x4 block
// we write the same color to both slots and all zeroes for the mask (one color total)
unsigned long dxt1_color = 0; // generate one such word and clone it
InsertTexelComponentFixed( params->r, 5, &dxt1_color ); InsertTexelComponentFixed( params->g, 6, &dxt1_color ); InsertTexelComponentFixed( params->b, 5, &dxt1_color ); // dupe
dxt1_color = dxt1_color | (dxt1_color<<16); // write into chunkbuf
*(unsigned long*)&chunkbuf[0] = dxt1_color; // color mask bits after that are already set to all zeroes. chunk is done.
chunksize = 8; } break; case D3DFMT_DXT3: { memset( temp32, 0, 16 ); // zap 16 bytes
// eight bytes of alpha (16 4-bit alpha nibbles)
// followed by a DXT1 block
unsigned long dxt3_alpha = 0; for( int i=0; i<8; i++) { // splat same alpha through block
InsertTexelComponentFixed( params->a, 4, &dxt3_alpha ); }
unsigned long dxt3_color = 0; // generate one such word and clone it
InsertTexelComponentFixed( params->r, 5, &dxt3_color ); InsertTexelComponentFixed( params->g, 6, &dxt3_color ); InsertTexelComponentFixed( params->b, 5, &dxt3_color ); // dupe
dxt3_color = dxt3_color | (dxt3_color<<16); // write into chunkbuf
*(unsigned long*)&chunkbuf[0] = dxt3_alpha; *(unsigned long*)&chunkbuf[4] = dxt3_alpha; *(unsigned long*)&chunkbuf[8] = dxt3_color; *(unsigned long*)&chunkbuf[12] = dxt3_color;
chunksize = 16; } break;
case D3DFMT_DXT5: { memset( temp32, 0, 16 ); // zap 16 bytes
// DXT5 has 8 bytes of compressed alpha, then 8 bytes of compressed RGB like DXT1.
// the 8 alpha bytes are 2 bytes of endpoint alpha values, then 16x3 bits of interpolants.
// so to write a single alpha value, just figure out the value, store it in both the first two bytes then store zeroes.
InsertTexelComponentFixed( params->a, 8, (unsigned long*)&chunkbuf[0] ); InsertTexelComponentFixed( params->a, 8, (unsigned long*)&chunkbuf[0] ); // rest of the alpha mask was already zeroed.
// now do colors
unsigned long dxt5_color = 0; // generate one such word and clone it
InsertTexelComponentFixed( params->r, 5, &dxt5_color ); InsertTexelComponentFixed( params->g, 6, &dxt5_color ); InsertTexelComponentFixed( params->b, 5, &dxt5_color ); // dupe
dxt5_color = dxt5_color | (dxt5_color<<16); // write into chunkbuf
*(unsigned long*)&chunkbuf[8] = dxt5_color; *(unsigned long*)&chunkbuf[12] = dxt5_color;
chunksize = 16; } break;
case D3DFMT_A32B32G32R32F: { *(float*)&chunkbuf[0] = params->r; *(float*)&chunkbuf[4] = params->g; *(float*)&chunkbuf[8] = params->b; *(float*)&chunkbuf[12] = params->a; chunksize = 16; } break;
case D3DFMT_A16B16G16R16: memset( chunkbuf, 0, 8 ); // R and G wind up in the first 32 bits
// B and A wind up in the second 32 bits
InsertTexelComponentFixed( params->a, 16, (unsigned long*)&chunkbuf[4] ); // winds up as MSW of second word (note [4]) - thus last in RAM
InsertTexelComponentFixed( params->b, 16, (unsigned long*)&chunkbuf[4] );
InsertTexelComponentFixed( params->g, 16, (unsigned long*)&chunkbuf[0] ); InsertTexelComponentFixed( params->r, 16, (unsigned long*)&chunkbuf[0] ); // winds up as LSW of first word, thus first in RAM
chunksize = 8; break; // not done yet
//case D3DFMT_D16:
//case D3DFMT_D24X8:
//case D3DFMT_D24S8:
//case D3DFMT_A16B16G16R16F:
default: return FALSE; // fail
break; } // once the chunk buffer is filled..
// sanity check the reported chunk size.
if (chunksize != format->m_bytesPerSquareChunk) { Debugger(); return FALSE; } // verify that the amount you want to write will not exceed the limit byte count
unsigned long destByteCount = chunksize * params->m_chunkCount; if (destByteCount > params->m_byteCountLimit) { Debugger(); return FALSE; } // write the bytes.
unsigned char *destP = (unsigned char*)params->m_dest; for( int chunk=0; chunk < params->m_chunkCount; chunk++) { for( int byteindex = 0; byteindex < chunksize; byteindex++) { *destP++ = chunkbuf[byteindex]; } } params->m_bytesWritten = destP - (unsigned char*)params->m_dest; return TRUE; }
//===============================================================================
bool LessFunc_GLMTexLayoutKey( const GLMTexLayoutKey &a, const GLMTexLayoutKey &b ) { #define DO_LESS(fff) if (a.fff != b.fff) { return (a.fff< b.fff); }
DO_LESS(m_texGLTarget); DO_LESS(m_texFormat); DO_LESS(m_texFlags); DO_LESS(m_xSize); DO_LESS(m_ySize) DO_LESS(m_zSize);
#undef DO_LESS
return false; // they are equal
}
CGLMTexLayoutTable::CGLMTexLayoutTable() { m_layoutMap.SetLessFunc( LessFunc_GLMTexLayoutKey ); }
GLMTexLayout *CGLMTexLayoutTable::NewLayoutRef( GLMTexLayoutKey *key ) { // look up 'key' in the map and see if it's a hit, if so, bump the refcount and return
// if not, generate a completed layout based on the key, add to map, set refcount to 1, return that
const GLMTexFormatDesc *formatDesc = GetFormatDesc( key->m_texFormat ); bool compression = (formatDesc->m_chunkSize > 1); if (!formatDesc) { GLMStop(); // bad news
} unsigned short index = m_layoutMap.Find( *key ); if (index != m_layoutMap.InvalidIndex()) { // found it
//printf(" -hit- ");
GLMTexLayout *layout = m_layoutMap[ index ]; // bump ref count
layout->m_refCount ++; return layout; } else { //printf(" -miss- ");
// need to make a new one
// to allocate it, we need to know how big to make it (slice count)
// figure out how many mip levels are in play
int mipCount = 1; if (key->m_texFlags & kGLMTexMipped) { int largestAxis = key->m_xSize; if (key->m_ySize > largestAxis) largestAxis = key->m_ySize; if (key->m_zSize > largestAxis) largestAxis = key->m_zSize; mipCount = 0; while( largestAxis > 0 ) { mipCount ++; largestAxis >>= 1; } }
int faceCount = 1; if (key->m_texGLTarget == GL_TEXTURE_CUBE_MAP) { faceCount = 6; } int sliceCount = mipCount * faceCount; if (key->m_texFlags & kGLMTexMultisampled) { Assert( (key->m_texGLTarget == GL_TEXTURE_2D) ); Assert( sliceCount == 1 ); // assume non mipped
Assert( (key->m_texFlags & kGLMTexMipped) == 0 ); Assert( (key->m_texFlags & kGLMTexMippedAuto) == 0 ); // assume renderable and srgb
Assert( (key->m_texFlags & kGLMTexRenderable) !=0 ); //Assert( (key->m_texFlags & kGLMTexSRGB) !=0 ); //FIXME don't assert on making depthstencil surfaces which are non srgb
// double check sample count (FIXME need real limit check here against device/driver)
Assert( (key->m_texSamples==2) || (key->m_texSamples==4) || (key->m_texSamples==6) || (key->m_texSamples==8) ); } // now we know enough to allocate and populate the new tex layout.
// malloc the new layout
int layoutSize = sizeof( GLMTexLayout ) + (sliceCount * sizeof( GLMTexLayoutSlice )); GLMTexLayout *layout = (GLMTexLayout *)malloc( layoutSize ); memset( layout, 0, layoutSize ); // clone the key in there
memset( &layout->m_key, 0x00, sizeof(layout->m_key) ); layout->m_key = *key;
// set refcount
layout->m_refCount = 1; // save the format desc
layout->m_format = (GLMTexFormatDesc *)formatDesc; // we know the mipcount from before
layout->m_mipCount = mipCount; // we know the face count too
layout->m_faceCount = faceCount; // slice count is the product
layout->m_sliceCount = mipCount * faceCount; // we can now fill in the slices.
GLMTexLayoutSlice *slicePtr = &layout->m_slices[0]; int storageOffset = 0; bool compressed = (formatDesc->m_chunkSize > 1); // true if DXT
for( int mip = 0; mip < mipCount; mip ++ ) { for( int face = 0; face < faceCount; face++ ) { // note application of chunk size which is 1 for uncompressed, and 4 for compressed tex (DXT)
// note also that the *dimensions* must scale down to 1
// but that the *storage* cannot go below 4x4.
// we introduce the "storage sizes" which are clamped, to compute the storage footprint.
int storage_x,storage_y,storage_z; slicePtr->m_xSize = layout->m_key.m_xSize >> mip; slicePtr->m_xSize = MAX( slicePtr->m_xSize, 1 ); // dimension can't go to zero
storage_x = MAX( slicePtr->m_xSize, formatDesc->m_chunkSize ); // storage extent can't go below chunk size
slicePtr->m_ySize = layout->m_key.m_ySize >> mip; slicePtr->m_ySize = MAX( slicePtr->m_ySize, 1 ); // dimension can't go to zero
storage_y = MAX( slicePtr->m_ySize, formatDesc->m_chunkSize ); // storage extent can't go below chunk size
slicePtr->m_zSize = layout->m_key.m_zSize >> mip; slicePtr->m_zSize = MAX( slicePtr->m_zSize, 1 ); // dimension can't go to zero
storage_z = MAX( slicePtr->m_zSize, 1); // storage extent for Z cannot go below '1'.
//if (compressed) NO NO NO do not lie about the dimensionality, just fudge the storage.
//{
// // round up to multiple of 4 in X and Y axes
// slicePtr->m_xSize = (slicePtr->m_xSize+3) & (~3);
// slicePtr->m_ySize = (slicePtr->m_ySize+3) & (~3);
//}
int xchunks = (storage_x / formatDesc->m_chunkSize ); int ychunks = (storage_y / formatDesc->m_chunkSize ); slicePtr->m_storageSize = (xchunks * ychunks * formatDesc->m_bytesPerSquareChunk) * storage_z; slicePtr->m_storageOffset = storageOffset; storageOffset += slicePtr->m_storageSize; storageOffset = ( (storageOffset+0x0F) & (~0x0F)); // keep each MIP starting on a 16 byte boundary.
slicePtr++; } } layout->m_storageTotalSize = storageOffset; //printf("\n size %08x for key (x=%d y=%d z=%d, fmt=%08x, bpsc=%d)", layout->m_storageTotalSize, key->m_xSize, key->m_ySize, key->m_zSize, key->m_texFormat, formatDesc->m_bytesPerSquareChunk );
// generate summary
// "target, format, +/- mips, base size"
char scratch[1024];
char *targetname; switch( key->m_texGLTarget ) { case GL_TEXTURE_2D: targetname = "2D "; break; case GL_TEXTURE_3D: targetname = "3D "; break; case GL_TEXTURE_CUBE_MAP: targetname = "CUBE"; break; } sprintf( scratch, "[%s %s %dx%dx%d mips=%d slices=%d flags=%02X%s]", targetname, formatDesc->m_formatSummary, layout->m_key.m_xSize, layout->m_key.m_ySize, layout->m_key.m_zSize, mipCount, sliceCount, layout->m_key.m_texFlags, (layout->m_key.m_texFlags & kGLMTexSRGB) ? " SRGB" : "" ); layout->m_layoutSummary = strdup( scratch ); //GLMPRINTF(("-D- new tex layout [ %s ]", scratch ));
// then insert into map. disregard returned index.
m_layoutMap.Insert( layout->m_key, layout ); return layout; } }
void CGLMTexLayoutTable::DelLayoutRef( GLMTexLayout *layout ) { // locate layout in hash, drop refcount
// (some GC step later on will harvest expired layouts - not like it's any big challenge to re-generate them)
unsigned short index = m_layoutMap.Find( layout->m_key ); if (index != m_layoutMap.InvalidIndex()) { // found it
GLMTexLayout *layout = m_layoutMap[ index ]; // drop ref count
layout->m_refCount --; assert( layout->m_refCount >= 0 ); } else { // that's bad
GLMStop(); } }
void CGLMTexLayoutTable::DumpStats( ) { for (int i=0; i<m_layoutMap.Count(); i++ ) { GLMTexLayout *layout = m_layoutMap[ i ]; // print it out
printf("\n%05d instances %08d bytes %08d totbytes %s", layout->m_refCount, layout->m_storageTotalSize, (layout->m_refCount*layout->m_storageTotalSize), layout->m_layoutSummary ); } }
ConVar gl_texclientstorage( "gl_texclientstorage", "1" ); // default 1 for L4D2
ConVar gl_texmsaalog ( "gl_texmsaalog", "0");
ConVar gl_rt_forcergba ( "gl_rt_forcergba", "1" ); // on teximage of a renderable tex, pass GL_RGBA in place of GL_BGRA
ConVar gl_minimize_rt_tex ( "gl_minimize_rt_tex", "0" ); // if 1, set the GL_TEXTURE_MINIMIZE_STORAGE_APPLE texture parameter to cut off mipmaps for RT's
ConVar gl_minimize_all_tex ( "gl_minimize_all_tex", "1" ); // if 1, set the GL_TEXTURE_MINIMIZE_STORAGE_APPLE texture parameter to cut off mipmaps for textures which are unmipped
ConVar gl_minimize_tex_log ( "gl_minimize_tex_log", "0" ); // if 1, printf the names of the tex that got minimized
CGLMTex::CGLMTex( GLMContext *ctx, GLMTexLayout *layout, GLMTexSamplingParams *sampling, const char *debugLabel ) { // caller has responsibility to make 'ctx' current, but we check to be sure.
ctx->CheckCurrent();
// note layout requested
m_layout = layout; m_maxActiveMip = -1; //index of highest mip that has been written - increase as each mip arrives
m_minActiveMip = 999; //index of lowest mip that has been written - lower it as each mip arrives
// note sampling (copy values)
m_sampling = *sampling; // note context owner
m_ctx = ctx; // clear the bind point flags
m_bindPoints.ClearAll(); // clear the RT attach count
m_rtAttachCount = 0; // come up with a GL name for this texture.
// for MTGL friendliness, we should generate our own names at some point..
glGenTextures( 1, &m_texName ); //sense whether to try and apply client storage upon teximage/subimage
m_texClientStorage = gl_texclientstorage.GetInt() != 0; // flag that we have not yet been explicitly kicked into VRAM..
m_texPreloaded = false; // clone the debug label if there is one.
m_debugLabel = debugLabel ? strdup(debugLabel) : NULL;
// if tex is MSAA renderable, make an RBO, else zero the RBO name and dirty bit
if (layout->m_key.m_texFlags & kGLMTexMultisampled) { glGenRenderbuffersEXT( 1, &m_rboName ); m_rboDirty = false; // so we have enough info to go ahead and bind the RBO and put storage on it?
// try it.
glBindRenderbufferEXT( GL_RENDERBUFFER_EXT, m_rboName ); GLMCheckError();
// quietly clamp if sample count exceeds known limit for the device
int sampleCount = layout->m_key.m_texSamples; if (sampleCount > ctx->Caps().m_maxSamples) { sampleCount = ctx->Caps().m_maxSamples; // clamp
} GLenum msaaFormat = (layout->m_key.m_texFlags & kGLMTexSRGB) ? layout->m_format->m_glIntFormatSRGB : layout->m_format->m_glIntFormat; glRenderbufferStorageMultisampleEXT( GL_RENDERBUFFER_EXT, sampleCount, // not "layout->m_key.m_texSamples"
msaaFormat, layout->m_key.m_xSize, layout->m_key.m_ySize ); GLMCheckError();
if (gl_texmsaalog.GetInt()) { printf( "\n == MSAA Tex %8x %s : MSAA RBO is intformat %s (%x)", this, m_debugLabel?m_debugLabel:"", GLMDecode( eGL_ENUM, msaaFormat ), msaaFormat ); }
glBindRenderbufferEXT( GL_RENDERBUFFER_EXT, 0 ); GLMCheckError(); } else { m_rboName = 0; m_rboDirty = false; }
// at this point we have the complete description of the texture, and a name for it, but no data and no actual GL object.
// we know this name has bever seen duty before, so we're going to hard-bind it to TMU 0, displacing any other tex that might have been bound there.
// any previously bound tex will be unbound and appropriately marked as a result.
// the active TMU will be set as a side effect.
ctx->BindTexToTMU( this, 0 ); // OK, our texture now exists and is bound on the active TMU. Not drawable yet though.
// impose the sampling params we were given, unconditionally
ApplySamplingParams( sampling, true );
// if not an RT, create backing storage and fill it
if ( !(layout->m_key.m_texFlags & kGLMTexRenderable) ) { m_backing = (char *)malloc( m_layout->m_storageTotalSize ); memset( m_backing, 0, m_layout->m_storageTotalSize ); // track bytes allocated for non-RT's
int formindex = sEncodeLayoutAsIndex( &layout->m_key ); g_texGlobalBytes[ formindex ] += m_layout->m_storageTotalSize; #if TEXSPACE_LOGGING
printf( "\n Tex %s added %d bytes in form %d which is now %d bytes", m_debugLabel ? m_debugLabel : "-", m_layout->m_storageTotalSize, formindex, g_texGlobalBytes[ formindex ] ); printf( "\n\t\t[ %d %d %d %d %d %d %d %d ]", g_texGlobalBytes[ 0 ],g_texGlobalBytes[ 1 ],g_texGlobalBytes[ 2 ],g_texGlobalBytes[ 3 ], g_texGlobalBytes[ 4 ],g_texGlobalBytes[ 5 ],g_texGlobalBytes[ 6 ],g_texGlobalBytes[ 7 ] ); #endif
} else { m_backing = NULL; m_texClientStorage = false; }
// init lock count
// lock reqs are tracked by the owning context
m_lockCount = 0;
m_sliceFlags.SetCount( m_layout->m_sliceCount ); for( int i=0; i< m_layout->m_sliceCount; i++) { m_sliceFlags[i] = 0; // kSliceValid = false (we have not teximaged each slice yet)
// kSliceStorageValid = false (the storage allocated does not reflect what is in the tex)
// kSliceLocked = false (the slices are not locked)
// kSliceFullyDirty = false (this does not come true til first lock)
} // texture minimize parameter keeps driver from allocing mips when it should not, by being explicit about the ones that have no mips.
bool setMinimizeParameter = false; bool minimize_rt = (gl_minimize_rt_tex.GetInt()!=0); bool minimize_all = (gl_minimize_all_tex.GetInt()!=0); if (layout->m_key.m_texFlags & kGLMTexRenderable) { // it's an RT. if mips were not explicitly requested, and "gl_minimize_rt_tex" is true, set the minimize parameter.
if ( (minimize_rt || minimize_all) && ( !(layout->m_key.m_texFlags & kGLMTexMipped) ) ) { setMinimizeParameter = true; } } else { // not an RT. if mips were not requested, and "gl_minimize_all_tex" is true, set the minimize parameter.
if ( minimize_all && ( !(layout->m_key.m_texFlags & kGLMTexMipped) ) ) { setMinimizeParameter = true; } }
if (setMinimizeParameter) { if (gl_minimize_tex_log.GetInt()) { printf("\n minimizing storage for tex '%s' [%s] ", m_debugLabel?m_debugLabel:"-", m_layout->m_layoutSummary ); } glTexParameteri( m_layout->m_key.m_texGLTarget, GL_TEXTURE_MINIMIZE_STORAGE_APPLE, 1 ); } // after a lot of pain with texture completeness...
// always push black into all slices of all newly created textures.
#if 0
bool pushRenderableSlices = (m_layout->m_key.m_texFlags & kGLMTexRenderable) != 0; bool pushTexSlices = true; // just do it everywhere (m_layout->m_mipCount>1) && (m_layout->m_format->m_chunkSize !=1) ;
if (pushTexSlices) { // fill storage with mostly-opaque purple
GLMGenTexelParams genp; memset( &genp, 0, sizeof(genp) ); genp.m_format = m_layout->m_format->m_d3dFormat; const GLMTexFormatDesc *format = GetFormatDesc( genp.m_format ); genp.m_dest = m_backing; // dest addr
genp.m_chunkCount = m_layout->m_storageTotalSize / format->m_bytesPerSquareChunk; // fill the whole slab
genp.m_byteCountLimit = m_layout->m_storageTotalSize; // limit writes to this amount
genp.r = 1.0; genp.g = 0.0; genp.b = 1.0; genp.a = 0.75; GLMGenTexels( &genp ); } #endif
//if (pushRenderableSlices || pushTexSlices)
if (1) { for( int face=0; face <m_layout->m_faceCount; face++) { for( int mip=0; mip <m_layout->m_mipCount; mip++) { // we're not really going to lock, we're just going to write the blank data from the backing store we just made
GLMTexLockDesc desc; desc.m_req.m_tex = this; desc.m_req.m_face = face; desc.m_req.m_mip = mip;
desc.m_sliceIndex = CalcSliceIndex( face, mip );
GLMTexLayoutSlice *slice = &m_layout->m_slices[ desc.m_sliceIndex ]; desc.m_req.m_region.xmin = desc.m_req.m_region.ymin = desc.m_req.m_region.zmin = 0; desc.m_req.m_region.xmax = slice->m_xSize; desc.m_req.m_region.ymax = slice->m_ySize; desc.m_req.m_region.zmax = slice->m_zSize;
desc.m_sliceBaseOffset = slice->m_storageOffset; // doesn't really matter... we're just pushing zeroes..
desc.m_sliceRegionOffset = 0;
this->WriteTexels( &desc, true, (layout->m_key.m_texFlags & kGLMTexRenderable)!=0 ); // write whole slice - but disable data source if it's an RT, as there's no backing
} } } GLMPRINTF(("-A- -**TEXNEW '%-60s' name=%06d size=%09d storage=%08x label=%s ", m_layout->m_layoutSummary, m_texName, m_layout->m_storageTotalSize, m_backing, m_debugLabel ? m_debugLabel : "-" )); }
CGLMTex::~CGLMTex( ) { if ( !(m_layout->m_key.m_texFlags & kGLMTexRenderable) ) { int formindex = sEncodeLayoutAsIndex( &m_layout->m_key );
g_texGlobalBytes[ formindex ] -= m_layout->m_storageTotalSize; #if TEXSPACE_LOGGING
printf( "\n Tex %s freed %d bytes in form %d which is now %d bytes", m_debugLabel ? m_debugLabel : "-", m_layout->m_storageTotalSize, formindex, g_texGlobalBytes[ formindex ] ); printf( "\n\t\t[ %d %d %d %d %d %d %d %d ]", g_texGlobalBytes[ 0 ],g_texGlobalBytes[ 1 ],g_texGlobalBytes[ 2 ],g_texGlobalBytes[ 3 ], g_texGlobalBytes[ 4 ],g_texGlobalBytes[ 5 ],g_texGlobalBytes[ 6 ],g_texGlobalBytes[ 7 ] ); #endif
} GLMPRINTF(("-A- -**TEXDEL '%-60s' name=%06d size=%09d storage=%08x label=%s ", m_layout->m_layoutSummary, m_texName, m_layout->m_storageTotalSize, m_backing, m_debugLabel ? m_debugLabel : "-" )); // check first to see if we were still bound anywhere or locked... these should be failures.
// if all that is OK, then delete the underlying tex
glDeleteTextures( 1, &m_texName ); GLMCheckError(); m_texName = 0;
if(m_rboName) { glDeleteRenderbuffersEXT( 1, &m_rboName ); GLMCheckError(); m_rboName = 0; m_rboDirty = false; } // release our usage of the layout
m_ctx->m_texLayoutTable->DelLayoutRef( m_layout ); m_layout = NULL; if (m_backing) { free( m_backing ); m_backing = NULL; } if (m_debugLabel) { free( m_debugLabel ); m_debugLabel = NULL; } m_ctx = NULL; }
int CGLMTex::CalcSliceIndex( int face, int mip ) { // faces of the same mip level are adjacent. "face major" storage
int index = (mip * m_layout->m_faceCount) + face;
return index; }
void CGLMTex::CalcTexelDataOffsetAndStrides( int sliceIndex, int x, int y, int z, int *offsetOut, int *yStrideOut, int *zStrideOut ) { int offset = 0; int yStride = 0; int zStride = 0; GLMTexFormatDesc *format = m_layout->m_format; if (format->m_chunkSize==1) { // figure out row stride and layer stride
yStride = format->m_bytesPerSquareChunk * m_layout->m_slices[sliceIndex].m_xSize; // bytes per texel row (y stride)
zStride = yStride * m_layout->m_slices[sliceIndex].m_ySize; // bytes per texel layer (if 3D tex)
offset = x * format->m_bytesPerSquareChunk; // lateral offset
offset += (y * yStride); // scanline offset
offset += (z * zStride); // should be zero for 2D tex
} else { yStride = format->m_bytesPerSquareChunk * (m_layout->m_slices[sliceIndex].m_xSize / format->m_chunkSize); zStride = yStride * (m_layout->m_slices[sliceIndex].m_ySize / format->m_chunkSize); // compressed format. scale the x,y,z values into chunks.
// assert if any of them are not multiples of a chunk.
int chunkx = x / format->m_chunkSize; int chunky = y / format->m_chunkSize; int chunkz = z / format->m_chunkSize; if ( (chunkx * format->m_chunkSize) != x) { GLMStop(); } if ( (chunky * format->m_chunkSize) != y) { GLMStop(); } if ( (chunkz * format->m_chunkSize) != z) { GLMStop(); } offset = chunkx * format->m_bytesPerSquareChunk; // lateral offset
offset += (chunky * yStride); // chunk row offset
offset += (chunkz * zStride); // should be zero for 2D tex
} *offsetOut = offset; *yStrideOut = yStride; *zStrideOut = zStride; }
void CGLMTex::ApplySamplingParams( GLMTexSamplingParams *params, bool noCheck ) { #define DIFF(fff) (noCheck || (params->fff != m_sampling.fff))
GLenum target = m_layout->m_key.m_texGLTarget;
// if the texture is compressed, and has a maxActiveMip that is >=0 but less than the mip count,
// (i.e. they supplied *some* but not *all* mips needed)...
// generate them, and fix the max mip count.
//if ( /*(m_layout->m_format->m_chunkSize !=1) &&*/ (m_layout->m_mipCount>3) )
//{
// m_maxActiveMip = m_layout->m_mipCount-3; // pull back three levels
// glTexParameteri( target, GL_TEXTURE_MAX_LEVEL, m_maxActiveMip);
// GLMCheckError();
//}
if (DIFF(m_addressModes[0])) { m_sampling.m_addressModes[0] = params->m_addressModes[0]; glTexParameteri( target, GL_TEXTURE_WRAP_S, m_sampling.m_addressModes[0]); GLMCheckError(); } if (DIFF(m_addressModes[1])) { m_sampling.m_addressModes[1] = params->m_addressModes[1]; glTexParameteri( target, GL_TEXTURE_WRAP_T, m_sampling.m_addressModes[1]); GLMCheckError(); } if (DIFF(m_addressModes[2])) { m_sampling.m_addressModes[2] = params->m_addressModes[2]; glTexParameteri( target, GL_TEXTURE_WRAP_R, m_sampling.m_addressModes[2]); GLMCheckError(); }
if ( noCheck || memcmp( params->m_borderColor, m_sampling.m_borderColor, sizeof(m_sampling.m_borderColor) ) ) { memcpy( m_sampling.m_borderColor, params->m_borderColor, sizeof(params->m_borderColor) ); glTexParameterfv( target, GL_TEXTURE_BORDER_COLOR, params->m_borderColor ); GLMCheckError(); } if (DIFF(m_magFilter)) { m_sampling.m_magFilter = params->m_magFilter; glTexParameteri( target, GL_TEXTURE_MAG_FILTER, params->m_magFilter); GLMCheckError(); } if (DIFF(m_minFilter)) { m_sampling.m_minFilter = params->m_minFilter; glTexParameteri( target, GL_TEXTURE_MIN_FILTER, params->m_minFilter); GLMCheckError(); } if (DIFF(m_mipmapBias)) { m_sampling.m_mipmapBias = params->m_mipmapBias; //glTexParameterf( target, GL_TEXTURE_LOD_BIAS, params->m_mipmapBias );
GLMCheckError(); } if (DIFF(m_minMipLevel)) { // don't let minmiplevel go below min active mip level
m_sampling.m_minMipLevel = MAX( m_minActiveMip, params->m_minMipLevel ); glTexParameteri( target, GL_TEXTURE_MIN_LOD, m_sampling.m_minMipLevel); GLMCheckError(); }
if (DIFF(m_maxMipLevel)) { // do not let max selectable LOD exceed the max submitted mip
m_sampling.m_maxMipLevel = MIN( m_maxActiveMip, params->m_maxMipLevel); glTexParameteri( target, GL_TEXTURE_MAX_LOD, m_sampling.m_maxMipLevel); GLMCheckError(); }
if (DIFF(m_maxAniso)) { m_sampling.m_maxAniso = MIN( m_ctx->m_caps.m_maxAniso, params->m_maxAniso ); glTexParameteri( target, GL_TEXTURE_MAX_ANISOTROPY_EXT, m_sampling.m_maxAniso ); GLMCheckError(); } if (DIFF(m_compareMode)) { m_sampling.m_compareMode = params->m_compareMode; glTexParameteri( target, GL_TEXTURE_COMPARE_MODE_ARB, params->m_compareMode ); GLMCheckError(); if (params->m_compareMode == GL_COMPARE_R_TO_TEXTURE_ARB) { glTexParameteri( target, GL_TEXTURE_COMPARE_FUNC_ARB, GL_LEQUAL ); GLMCheckError(); } } if (DIFF(m_srgb)) { m_sampling.m_srgb = params->m_srgb; // we might have to re-DL the tex if the SRGB read status changes..
} #undef DIFF
}
void CGLMTex::ReadTexels( GLMTexLockDesc *desc, bool readWholeSlice ) { GLMRegion readBox; if (readWholeSlice) { readBox.xmin = readBox.ymin = readBox.zmin = 0;
readBox.xmax = m_layout->m_slices[ desc->m_sliceIndex ].m_xSize; readBox.ymax = m_layout->m_slices[ desc->m_sliceIndex ].m_ySize; readBox.zmax = m_layout->m_slices[ desc->m_sliceIndex ].m_zSize; } else { readBox = desc->m_req.m_region; }
m_ctx->BindTexToTMU( this, 0, false ); // SelectTMU(n) is a side effect
if (readWholeSlice) { // make this work first.... then write the partial path
// (Hmmmm, I don't think we will ever actually need a partial path -
// since we have no notion of a partially valid slice of storage
GLMTexFormatDesc *format = m_layout->m_format; GLenum target = m_layout->m_key.m_texGLTarget; void *sliceAddress = m_backing + m_layout->m_slices[ desc->m_sliceIndex ].m_storageOffset; // this would change for PBO
int sliceSize = m_layout->m_slices[ desc->m_sliceIndex ].m_storageSize; // interestingly enough, we can use the same path for both 2D and 3D fetch
switch( target ) { case GL_TEXTURE_CUBE_MAP:
// adjust target to steer to the proper face, then fall through to the 2D texture path.
target = GL_TEXTURE_CUBE_MAP_POSITIVE_X + desc->m_req.m_face; case GL_TEXTURE_2D: case GL_TEXTURE_3D: { // check compressed or not
if (format->m_chunkSize != 1) { // compressed path
// http://www.opengl.org/sdk/docs/man/xhtml/glGetCompressedTexImage.xml
glGetCompressedTexImage( target, // target
desc->m_req.m_mip, // level
sliceAddress ); // destination
GLMCheckError(); } else { // uncompressed path
// http://www.opengl.org/sdk/docs/man/xhtml/glGetTexImage.xml
glGetTexImage( target, // target
desc->m_req.m_mip, // level
format->m_glDataFormat, // dataformat
format->m_glDataType, // datatype
sliceAddress ); // destination
GLMCheckError(); } } break; } } else { GLMStop(); } }
// defaulting the subimage support off, since it's breaking Ep2 at startup on some NV 9400 and friends
// defaulting it back to "1" for L4D2 and see if it flies
ConVar gl_enabletexsubimage( "gl_enabletexsubimage", "1" );
void CGLMTex::WriteTexels( GLMTexLockDesc *desc, bool writeWholeSlice, bool noDataWrite ) { GLMRegion writeBox;
bool needsExpand = false; char *expandTemp = NULL; switch( m_layout->m_format->m_d3dFormat) { case D3DFMT_V8U8: { needsExpand = true; writeWholeSlice = true; // shoot down client storage if we have to generate a new flavor of the data
m_texClientStorage = false; } break; } if (writeWholeSlice) { writeBox.xmin = writeBox.ymin = writeBox.zmin = 0;
writeBox.xmax = m_layout->m_slices[ desc->m_sliceIndex ].m_xSize; writeBox.ymax = m_layout->m_slices[ desc->m_sliceIndex ].m_ySize; writeBox.zmax = m_layout->m_slices[ desc->m_sliceIndex ].m_zSize; } else { writeBox = desc->m_req.m_region; }
// first thing is to get the GL texture bound to a TMU, or just select one if already bound
// to get this running we will just always slam TMU 0 and let the draw time code fix it back
// a later optimization would be to hoist the bind call to the caller, do it exactly once
m_ctx->BindTexToTMU( this, 0, false ); // SelectTMU(n) is a side effect
GLMTexFormatDesc *format = m_layout->m_format; GLenum target = m_layout->m_key.m_texGLTarget; GLenum glDataFormat = format->m_glDataFormat; // this could change if expansion kicks in
GLenum glDataType = format->m_glDataType; GLMTexLayoutSlice *slice = &m_layout->m_slices[ desc->m_sliceIndex ]; void *sliceAddress = m_backing ? (m_backing + slice->m_storageOffset) : NULL; // this would change for PBO
// allow use of subimage if the target is texture2D and it has already been teximage'd
bool mayUseSubImage = false; if ( (target==GL_TEXTURE_2D) && (m_sliceFlags[ desc->m_sliceIndex ] & kSliceValid) ) { mayUseSubImage = gl_enabletexsubimage.GetInt(); } // check flavor, 2D, 3D, or cube map
// we also have the choice to use subimage if this is a tex already created. (open question as to benefit)
// SRGB select. At this level (writetexels) we firmly obey the m_texFlags.
// (mechanism not policy)
GLenum intformat = (m_layout->m_key.m_texFlags & kGLMTexSRGB) ? format->m_glIntFormatSRGB : format->m_glIntFormat; if (CommandLine()->FindParm("-disable_srgbtex")) { // force non srgb flavor - experiment to make ATI r600 happy on 10.5.8 (maybe x1600 too!)
intformat = format->m_glIntFormat; } Assert( intformat != 0 ); if (m_layout->m_key.m_texFlags & kGLMTexSRGB) { Assert( m_layout->m_format->m_glDataFormat != GL_DEPTH_COMPONENT ); Assert( m_layout->m_format->m_glDataFormat != GL_DEPTH_STENCIL_EXT ); Assert( m_layout->m_format->m_glDataFormat != GL_ALPHA ); } // adjust min and max mip written
if (desc->m_req.m_mip > m_maxActiveMip) { m_maxActiveMip = desc->m_req.m_mip;
glTexParameteri( target, GL_TEXTURE_MAX_LEVEL, desc->m_req.m_mip); GLMCheckError(); } if (desc->m_req.m_mip < m_minActiveMip) { m_minActiveMip = desc->m_req.m_mip; glTexParameteri( target, GL_TEXTURE_BASE_LEVEL, desc->m_req.m_mip); GLMCheckError(); } if (needsExpand) { int expandSize = 0; switch( m_layout->m_format->m_d3dFormat) { case D3DFMT_V8U8: { // figure out new size based on 3byte RGB format
// easy, just take the two byte size and grow it by 50%
expandSize = (slice->m_storageSize * 3) / 2; expandTemp = (char*)malloc( expandSize ); char *src = (char*)sliceAddress; char *dst = expandTemp;
// transfer RG's to RGB's
while(expandSize>0) { *dst = *src++; // move first byte
*dst = *src++; // move second byte
*dst = 0xBB; // pad third byte
expandSize -= 3; } // move the slice pointer
sliceAddress = expandTemp; // change the data format we tell GL about
glDataFormat = GL_RGB; } break; default: Assert(!"Don't know how to expand that format.."); } }
// set up the client storage now, one way or another
glPixelStorei( GL_UNPACK_CLIENT_STORAGE_APPLE, m_texClientStorage ); GLMCheckError(); switch( target ) { case GL_TEXTURE_CUBE_MAP:
// adjust target to steer to the proper face, then fall through to the 2D texture path.
target = GL_TEXTURE_CUBE_MAP_POSITIVE_X + desc->m_req.m_face; case GL_TEXTURE_2D: { // check compressed or not
if (format->m_chunkSize != 1) { Assert( writeWholeSlice ); //subimage not implemented in this path yet
// compressed path
// http://www.opengl.org/sdk/docs/man/xhtml/glCompressedTexImage2D.xml
glCompressedTexImage2D( target, // target
desc->m_req.m_mip, // level
intformat, // internalformat - don't use format->m_glIntFormat because we have the SRGB select going on above
slice->m_xSize, // width
slice->m_ySize, // height
0, // border
slice->m_storageSize, // imageSize
sliceAddress ); // data
GLMCheckError(); } else { if (mayUseSubImage) { // go subimage2D if it's a replacement, not a creation
glPixelStorei( GL_UNPACK_ROW_LENGTH, slice->m_xSize ); // in pixels
glPixelStorei( GL_UNPACK_SKIP_PIXELS, writeBox.xmin ); // in pixels
glPixelStorei( GL_UNPACK_SKIP_ROWS, writeBox.ymin ); // in pixels
GLMCheckError();
glTexSubImage2D( target, desc->m_req.m_mip, // level
writeBox.xmin, // xoffset into dest
writeBox.ymin, // yoffset into dest
writeBox.xmax - writeBox.xmin, // width (was slice->m_xSize)
writeBox.ymax - writeBox.ymin, // height (was slice->m_ySize)
glDataFormat, // format
glDataType, // type
sliceAddress // data (will be offsetted by the SKIP_PIXELS and SKIP_ROWS - let GL do the math to find the first source texel)
); GLMCheckError();
glPixelStorei( GL_UNPACK_ROW_LENGTH, 0 ); glPixelStorei( GL_UNPACK_SKIP_PIXELS, 0 ); glPixelStorei( GL_UNPACK_SKIP_ROWS, 0 ); GLMCheckError();
/*
//http://www.opengl.org/sdk/docs/man/xhtml/glTexSubImage2D.xml
glTexSubImage2D( target, desc->m_req.m_mip, // level
0, // xoffset
0, // yoffset
slice->m_xSize, // width
slice->m_ySize, // height
glDataFormat, // format
glDataType, // type
sliceAddress // data
); GLMCheckError(); */ } else { if (m_layout->m_key.m_texFlags & kGLMTexRenderable) { if (gl_rt_forcergba.GetInt()) { if (glDataFormat == GL_BGRA) { // change it
glDataFormat = GL_RGBA; } } }
// uncompressed path
// http://www.opengl.org/documentation/specs/man_pages/hardcopy/GL/html/gl/teximage2d.html
glTexImage2D( target, // target
desc->m_req.m_mip, // level
intformat, // internalformat - don't use format->m_glIntFormat because we have the SRGB select going on above
slice->m_xSize, // width
slice->m_ySize, // height
0, // border
glDataFormat, // dataformat
glDataType, // datatype
noDataWrite ? NULL : sliceAddress ); // data (optionally suppressed in case ResetSRGB desires)
GLMCheckError();
if (m_layout->m_key.m_texFlags & kGLMTexMultisampled) { if (gl_texmsaalog.GetInt()) { printf( "\n == MSAA Tex %8x %s : glTexImage2D for flat tex using intformat %s (%x)", this, m_debugLabel?m_debugLabel:"", GLMDecode( eGL_ENUM, intformat ), intformat ); printf( "\n" ); } }
m_sliceFlags[ desc->m_sliceIndex ] |= kSliceValid; // for next time, we can subimage..
} } } break; case GL_TEXTURE_3D: { // check compressed or not
if (format->m_chunkSize != 1) { // compressed path
// http://www.opengl.org/sdk/docs/man/xhtml/glCompressedTexImage3D.xml
glCompressedTexImage3D( target, // target
desc->m_req.m_mip, // level
format->m_glIntFormat, // internalformat
slice->m_xSize, // width
slice->m_ySize, // height
slice->m_zSize, // depth
0, // border
slice->m_storageSize, // imageSize
sliceAddress ); // data
GLMCheckError(); } else { // uncompressed path
// http://www.opengl.org/sdk/docs/man/xhtml/glTexImage3D.xml
glTexImage3D( target, // target
desc->m_req.m_mip, // level
format->m_glIntFormat, // internalformat
slice->m_xSize, // width
slice->m_ySize, // height
slice->m_zSize, // depth
0, // border
glDataFormat, // dataformat
glDataType, // datatype
noDataWrite ? NULL : sliceAddress ); // data (optionally suppressed in case ResetSRGB desires)
GLMCheckError(); } } break; }
glPixelStorei( GL_UNPACK_CLIENT_STORAGE_APPLE, GL_FALSE ); GLMCheckError();
if ( expandTemp ) { free( expandTemp ); } }
void CGLMTex::Lock( GLMTexLockParams *params, char** addressOut, int* yStrideOut, int *zStrideOut ) { // locate appropriate slice in layout record
int sliceIndex = CalcSliceIndex( params->m_face, params->m_mip ); GLMTexLayoutSlice *slice = &m_layout->m_slices[sliceIndex];
// obtain offset
int sliceBaseOffset = slice->m_storageOffset; // cross check region req against slice bounds - figure out if it matches, exceeds, or is less than the whole slice.
char exceed = (params->m_region.xmin < 0) || (params->m_region.xmax > slice->m_xSize) || (params->m_region.ymin < 0) || (params->m_region.ymax > slice->m_ySize) || (params->m_region.zmin < 0) || (params->m_region.zmax > slice->m_zSize);
char partial = (params->m_region.xmin > 0) || (params->m_region.xmax < slice->m_xSize) || (params->m_region.ymin > 0) || (params->m_region.ymax < slice->m_ySize) || (params->m_region.zmin > 0) || (params->m_region.zmax < slice->m_zSize); bool copyout = false; // set if a readback of the texture slice from GL is needed
if (exceed) { // illegal rect, out of bounds
GLMStop(); } // on return, these things need to be true
// a - there needs to be storage allocated, which we will return an address within
// b - the region corresponding to the slice being locked, will have valid data there for the whole slice.
// c - the slice is marked as locked
// d - the params of the lock request have been saved in the lock table (in the context)
// so step 1 is unambiguous. If there's no backing storage, make some.
if (!m_backing) { m_backing = (char *)malloc( m_layout->m_storageTotalSize ); memset( m_backing, 0, m_layout->m_storageTotalSize ); // clear the kSliceStorageValid bit on all slices
for( int i=0; i<m_layout->m_sliceCount; i++) { m_sliceFlags[i] &= ~kSliceStorageValid; } } // work on this slice now
// storage is known to exist at this point, but we need to check if its contents are valid for this slice.
// this is tracked per-slice so we don't hoist all the texels back out of GL across all slices if caller only
// wanted to lock some of them.
// (i.e. if we just alloced it, it's blank)
// if storage is invalid, but the texture itself is valid, hoist the texels back to the storage and mark it valid.
// if storage is invalid, and texture itself is also invalid, go ahead and mark storage as valid and fully dirty... to force teximage.
// ???????????? we need to go over this more carefully re "slice valid" (it has been teximaged) vs "storage valid" (it has been copied out).
unsigned char *sliceFlags = &m_sliceFlags[ sliceIndex ]; if (params->m_readback) { // caller is letting us know that it wants to readback the real texels.
*sliceFlags |= kSliceStorageValid; *sliceFlags |= kSliceValid; *sliceFlags &= ~(kSliceFullyDirty); copyout = true; } else { // caller is pushing texels.
if (! (*sliceFlags & kSliceStorageValid) ) { // storage is invalid. check texture state
if ( *sliceFlags & kSliceValid ) { // kSliceValid set: the texture itself has a valid slice, but we don't have it in our backing copy, so copy it out.
copyout = true; } else { // kSliceValid not set: the texture does not have a valid slice to copy out - it hasn't been teximage'd yet.
// set the "full dirty" bit to make sure we teximage the whole thing on unlock.
*sliceFlags |= kSliceFullyDirty; // assert if they did not ask to lock the full slice size on this go-round
if (partial) { // choice here -
// 1 - stop cold, we don't know how to subimage yet.
// 2 - grin and bear it, mark whole slice dirty (ah, we already did... so, do nothing).
// choice 2: // GLMStop();
} } // one way or another, upon reaching here the slice storage is valid for read.
*sliceFlags |= kSliceStorageValid; } }
// when we arrive here, there is storage, and the content of the storage for this slice is valid
// (or zeroes if it's the first lock)
// log the lock request in the context.
int newdesc = m_ctx->m_texLocks.AddToTail(); GLMTexLockDesc *desc = &m_ctx->m_texLocks[newdesc]; desc->m_req = *params; desc->m_active = true; desc->m_sliceIndex = sliceIndex; desc->m_sliceBaseOffset = m_layout->m_slices[sliceIndex].m_storageOffset;
// to calculate the additional offset we need to look at the rect's min corner
// combined with the per-texel size and Y/Z stride
// also cross check it for 4x multiple if there is compression in play
int offsetInSlice = 0; int yStride = 0; int zStride = 0; CalcTexelDataOffsetAndStrides( sliceIndex, params->m_region.xmin, params->m_region.ymin, params->m_region.zmin, &offsetInSlice, &yStride, &zStride );
// for compressed case...
// since there is presently no way to texsubimage a DXT when the rect does not cover the whole width,
// we will probably need to inflate the dirty rect in the recorded lock req so that the entire span is
// pushed across at unlock time.
desc->m_sliceRegionOffset = offsetInSlice + desc->m_sliceBaseOffset;
if (copyout) { // read the whole slice
// (odds are we'll never request anything but a whole slice to be read..)
ReadTexels( desc, true ); } // this would be a good place to fill with scrub value if in debug...
*addressOut = m_backing + desc->m_sliceRegionOffset; *yStrideOut = yStride; *zStrideOut = zStride;
m_lockCount++; }
void CGLMTex::Unlock( GLMTexLockParams *params ) { // look for an active lock request on this face and mip (doesn't necessarily matter which one, if more than one)
// and mark it inactive.
// --> if you can't find one, fail. first line of defense against mismatched locks/unlocks..
int i=0; bool found = false; while( !found && (i<m_ctx->m_texLocks.Count()) ) { GLMTexLockDesc *desc = &m_ctx->m_texLocks[i]; // is lock at index 'i' targeted at the texture/face/mip in question?
if ( (desc->m_req.m_tex == this) && (desc->m_req.m_face == params->m_face) & (desc->m_req.m_mip == params->m_mip) && (desc->m_active) ) { // matched and active, so retire it
desc->m_active = false; // stop searching
found = true; } i++; }
if (!found) { GLMStop(); // bad news
} // found - so drop lock count
m_lockCount--; if (m_lockCount <0) { GLMStop(); // bad news
}
if (m_lockCount==0) { // there should not be any active locks remaining on this texture.
// motivation to defer all texel pushing til *all* open locks are closed out -
// if/when we back the texture with a PBO, we will need to unmap that PBO before teximaging from it;
// by waiting for all the locks to clear this gives us an unambiguous signal to act on.
// scan through all the retired locks for this texture and push the texels for each one.
// after each one is dispatched, remove it from the pile.
int j=0; while( j<m_ctx->m_texLocks.Count() ) { GLMTexLockDesc *desc = &m_ctx->m_texLocks[j]; if ( desc->m_req.m_tex == this ) { // if it's active, something is wrong
if (desc->m_active) { GLMStop(); } // write the texels
bool fullyDirty = false; fullyDirty |= ((m_sliceFlags[ desc->m_sliceIndex ] & kSliceFullyDirty) != 0);
// this is not optimal and will result in full downloads on any dirty.
// we're papering over the fact that subimage isn't done yet.
// but this is safe if the slice of storage is all valid.
// at some point we'll need to actually compare the lock box against the slice bounds.
// fullyDirty |= (m_sliceFlags[ desc->m_sliceIndex ] & kSliceStorageValid);
WriteTexels( desc, fullyDirty );
// logical place to trigger preloading
// only do it for an RT tex, if it is not yet attached to any FBO.
// also, only do it if the slice number is the last slice in the tex.
if ( desc->m_sliceIndex == (m_layout->m_sliceCount-1) ) { if ( !(m_layout->m_key.m_texFlags & kGLMTexRenderable) || (m_rtAttachCount==0) ) { m_ctx->PreloadTex( this ); // printf("( slice %d of %d )", desc->m_sliceIndex, m_layout->m_sliceCount );
} }
m_ctx->m_texLocks.FastRemove( j ); // remove from the pile, don't advance index
} else { j++; // move on to next one
} } // clear the locked and full-dirty flags for all slices
for( int slice=0; slice < m_layout->m_sliceCount; slice++) { m_sliceFlags[slice] &= ~( kSliceLocked | kSliceFullyDirty ); } } }
void CGLMTex::ResetSRGB( bool srgb, bool noDataWrite ) { // see if requested SRGB state differs from the known one
bool wasSRGB = (m_layout->m_key.m_texFlags & kGLMTexSRGB); GLMTexLayout *oldLayout = m_layout; // need to m_ctx->m_texLayoutTable->DelLayoutRef on this one if we flip
if (srgb != wasSRGB) { // we're going to need a new layout (though the storage size should be the same - check it)
GLMTexLayoutKey newKey = m_layout->m_key; newKey.m_texFlags &= (~kGLMTexSRGB); // turn off that bit
newKey.m_texFlags |= srgb ? kGLMTexSRGB : 0; // turn on that bit if it should be so
// get new layout
GLMTexLayout *newLayout = m_ctx->m_texLayoutTable->NewLayoutRef( &newKey );
// if SRGB requested, verify that the layout we just got can do it.
// if it can't, delete the new layout ref and bail.
if (srgb && (newLayout->m_format->m_glIntFormatSRGB == 0)) { Assert( !"Can't enable SRGB mode on this format" ); m_ctx->m_texLayoutTable->DelLayoutRef( newLayout ); return; }
// check sizes and fail if no match
if( newLayout->m_storageTotalSize != oldLayout->m_storageTotalSize ) { Assert( !"Bug: layout sizes don't match on SRGB change" ); m_ctx->m_texLayoutTable->DelLayoutRef( newLayout ); return; }
// commit to new layout
m_layout = newLayout; // check same size
Assert( m_layout->m_storageTotalSize == oldLayout->m_storageTotalSize ); // release old
m_ctx->m_texLayoutTable->DelLayoutRef( oldLayout ); oldLayout = NULL;
// force texel re-DL
// note this messes with TMU 0 as side effect of WriteTexels
// so we save and restore the TMU 0 binding first
// since we're likely to be called in dxabstract when it is syncing sampler state, we can't go trampling the bindings.
// a refinement would be to have each texture make a note of which TMU they're bound on, and just use that active TMU for DL instead of 0.
CGLMTex *tmu0save = m_ctx->m_samplers[0].m_drawTex; for( int face=0; face <m_layout->m_faceCount; face++) { for( int mip=0; mip <m_layout->m_mipCount; mip++) { // we're not really going to lock, we're just going to rewrite the orig data
GLMTexLockDesc desc; desc.m_req.m_tex = this; desc.m_req.m_face = face; desc.m_req.m_mip = mip;
desc.m_sliceIndex = CalcSliceIndex( face, mip );
GLMTexLayoutSlice *slice = &m_layout->m_slices[ desc.m_sliceIndex ]; desc.m_req.m_region.xmin = desc.m_req.m_region.ymin = desc.m_req.m_region.zmin = 0; desc.m_req.m_region.xmax = slice->m_xSize; desc.m_req.m_region.ymax = slice->m_ySize; desc.m_req.m_region.zmax = slice->m_zSize;
desc.m_sliceBaseOffset = slice->m_storageOffset; // doesn't really matter... we're just pushing zeroes..
desc.m_sliceRegionOffset = 0;
this->WriteTexels( &desc, true, noDataWrite ); // write whole slice. and avoid pushing real bits if the caller requests (RT's)
} } // put it back
m_ctx->BindTexToTMU( tmu0save, 0, true ); } }
|