|
|
#include "precomp.h"
//
// SBC.CPP
// Send Bitmap Cache
//
// Copyright(c) Microsoft 1997-
//
#define MLZ_FILE_ZONE ZONE_CORE
//
// SBC_HostStarting()
//
BOOL ASHost::SBC_HostStarting(void) { BITMAPINFO_ours bitmapInfo; int i; BOOL rc = FALSE;
DebugEntry(ASHost::SBC_HostStarting);
if (g_sbcEnabled) { //
// We create a DIB section for each tile size which we use during the
// conversion of a bitmap from the native (device) bpp to the protocol
// bpp. We create the DIB sections at the device bpp.
//
ZeroMemory(&bitmapInfo, sizeof(bitmapInfo)); m_pShare->USR_InitDIBitmapHeader((BITMAPINFOHEADER *)&bitmapInfo, g_usrCaptureBPP);
// We only capture at 8 or 24 for NT 5.0, otherwise the screen depth
if ((g_usrCaptureBPP > 8) && (g_usrCaptureBPP != 24)) { //
// If the device bpp is > 8 (but not 24), we have to set up the DIB
// section to use the same bitmasks as the device. This means
// setting the compression type to BI_BITFIELDS and setting the
// first 3 DWORDS of the bitmap info color table to be the bitmasks
// for R, G and B respectively.
//
// 24bpp does not use bitmasks - it must use
// regular BI_RGB format with 8 bits for each colour.
//
bitmapInfo.bmiHeader.biCompression = BI_BITFIELDS;
ASSERT(g_asbcBitMasks[0]); ASSERT(g_asbcBitMasks[1]); ASSERT(g_asbcBitMasks[2]);
bitmapInfo.bmiColors[0] = ((LPTSHR_RGBQUAD)g_asbcBitMasks)[0]; bitmapInfo.bmiColors[1] = ((LPTSHR_RGBQUAD)g_asbcBitMasks)[1]; bitmapInfo.bmiColors[2] = ((LPTSHR_RGBQUAD)g_asbcBitMasks)[2]; }
//
// Initialize m_asbcWorkInfo array which holds the info we use to
// convert from native bpp to protocol bpp.
//
//
// First, intialize all the fields to default values
//
for (i = 0; i < SBC_NUM_TILE_SIZES ; i++) { ASSERT(!m_asbcWorkInfo[i].pShuntBuffer); ASSERT(g_asbcShuntBuffers[i]);
m_asbcWorkInfo[i].pShuntBuffer = g_asbcShuntBuffers[i];
ASSERT(m_asbcWorkInfo[i].mruIndex == 0); ASSERT(m_asbcWorkInfo[i].workBitmap == 0); ASSERT(m_asbcWorkInfo[i].pWorkBitmapBits == NULL);
if (i == SBC_MEDIUM_TILE_INDEX) { m_asbcWorkInfo[i].tileWidth = MP_MEDIUM_TILE_WIDTH; m_asbcWorkInfo[i].tileHeight = MP_MEDIUM_TILE_HEIGHT; } else { m_asbcWorkInfo[i].tileWidth = MP_LARGE_TILE_WIDTH; m_asbcWorkInfo[i].tileHeight = MP_LARGE_TILE_HEIGHT; }
bitmapInfo.bmiHeader.biWidth = m_asbcWorkInfo[i].tileWidth; bitmapInfo.bmiHeader.biHeight = m_asbcWorkInfo[i].tileHeight;
m_asbcWorkInfo[i].workBitmap = CreateDIBSection(NULL, (BITMAPINFO*)&bitmapInfo, DIB_RGB_COLORS, (void **)&(m_asbcWorkInfo[i].pWorkBitmapBits), NULL, // File mapping object
0); // Offset into file
// mapping object
if (!m_asbcWorkInfo[i].workBitmap) { ERROR_OUT(("Failed to create SBC DIB section %d", i)); DC_QUIT; }
ASSERT(m_asbcWorkInfo[i].pWorkBitmapBits); TRACE_OUT(( "Created work DIB section %d, pBits = 0x%08x", i, m_asbcWorkInfo[i].pWorkBitmapBits)); }
//
// Initialize the fastpath
//
if (!SBCInitFastPath()) { TRACE_OUT(( "Failed to init fastpath")); DC_QUIT; }
if (!SBCInitInternalOrders()) { ERROR_OUT(( "Failed to init SBC internal order struct")); DC_QUIT; }
m_pShare->SBC_RecalcCaps(TRUE); }
rc = TRUE;
DC_EXIT_POINT: DebugExitBOOL(ASHost::SBC_HostStarting, rc); return(rc); }
//
// ASShare::SBC_HostEnded()
//
void ASHost::SBC_HostEnded(void) { int i;
DebugEntry(ASHost::SBC_HostEnded);
if (g_sbcEnabled) { //
// Free up the memory associated with sbcOrderInfo.
//
SBCFreeInternalOrders();
SBCInitCacheStructures();
//
// Free our fast path info
//
if (m_sbcFastPath) { delete m_sbcFastPath; m_sbcFastPath = NULL; }
//
// Clear our cache handles.
//
for (i = 0; i < NUM_BMP_CACHES; i++) { if (m_asbcBmpCaches[i].handle != 0) { TRACE_OUT(( "Clear cache %d", i)); CH_DestroyCache(m_asbcBmpCaches[i].handle); BMCFreeCacheData(&m_asbcBmpCaches[i]); } }
//
// Free our work DIB sections
//
//
// We just have to delete the DIB sections and reset our variables.
//
for (i = 0 ; i < SBC_NUM_TILE_SIZES ; i++) { m_asbcWorkInfo[i].pShuntBuffer = NULL;
if (m_asbcWorkInfo[i].workBitmap != NULL) { DeleteBitmap(m_asbcWorkInfo[i].workBitmap); m_asbcWorkInfo[i].workBitmap = NULL; m_asbcWorkInfo[i].pWorkBitmapBits = NULL; } } }
DebugExitVOID(ASHost::SBC_HostEnded); }
//
// SBC_SyncOutgoing()
// Called when we're already hosting and someone new joins the share.
// Resets the OUTGOING bitmap cache for bitblt orders.
//
void ASHost::SBC_SyncOutgoing(void) { int i;
DebugEntry(ASHost::SBC_SyncOutgoing);
//
// Only do anything if SBC is enabled
//
if (g_sbcEnabled) { //
// Discard all currently cached bitmaps and set the colour table to
// zero so that the next bitmap order which arrives will trigger the
// sending of a new colour table first. Note that if the colour table
// is then full of zeros(!) it will still be OK because the RBC zeros
// out its copy of the colour table when a new host joins the share.
//
TRACE_OUT(( "Clearing all send caches")); SBCInitCacheStructures();
//
// All we have to do here is to reset our MRU indices for each of the
// shunt buffers. Each of the entries in the shunt buffer will be
// marked as free down in the driver.
//
for (i = 0; i < SBC_NUM_TILE_SIZES; i++) { m_asbcWorkInfo[i].mruIndex = 0; } }
DebugExitVOID(ASHost::SBC_SyncOutgoing); }
//
//
// SBC_CopyPrivateOrderData()
//
//
UINT ASHost::SBC_CopyPrivateOrderData ( LPBYTE pDst, LPCOM_ORDER pOrder, UINT freeBytesInBuffer ) { UINT orderSize; LPBYTE pBitmapBits;
DebugEntry(ASHost::SBC_CopyPrivateOrderData);
//
// Copy the order header without the rectangle structure (which we
// do not use).
//
orderSize = sizeof(pOrder->OrderHeader) - sizeof(pOrder->OrderHeader.rcsDst); memcpy(pDst, pOrder, orderSize);
//
// Copy the basic order data.
//
memcpy(pDst + orderSize, pOrder->abOrderData, pOrder->OrderHeader.cbOrderDataLength); orderSize += pOrder->OrderHeader.cbOrderDataLength;
if (orderSize > freeBytesInBuffer) { ERROR_OUT(( "Overwritten end of buffer. (%u) > (%u)", orderSize, freeBytesInBuffer)); }
//
// Set the length field in the order header to be the total amount of
// data we have copied (including the partial header) minus the
// size of a full header. This is horrible! - but is needed because
// the OD2 code looks at the header (which it really should not know
// about) and uses the length field to calculate the total length of
// the order. The OD2 code does not know that we have omitted some
// of the header.
//
((LPCOM_ORDER)pDst)->OrderHeader.cbOrderDataLength = (WORD)(orderSize - sizeof(COM_ORDER_HEADER));
//
// Return the total number of bytes that we have copied.
//
DebugExitDWORD(ASHost::SBC_CopyPrivateOrderData, orderSize); return(orderSize); }
//
// Name: SBCInitCacheStructures()
//
// Purpose:
//
// Returns:
//
// Params:
//
// Operation:
//
//
void ASHost::SBCInitCacheStructures(void) { UINT i;
DebugEntry(ASHost::SBCInitCacheStructures);
ASSERT(g_sbcEnabled);
//
// Reset caches
//
for (i = 0; i < NUM_BMP_CACHES; i++) { if (m_asbcBmpCaches[i].handle) { CH_ClearCache(m_asbcBmpCaches[i].handle); } }
//
// Do any OS specific processing
//
SBC_CacheCleared();
DebugExitVOID(ASHost::SBCInitCacheStructures); }
//
// SBC_CacheCleared()
//
void ASHost::SBC_CacheCleared(void) { int i;
DebugEntry(ASHost::SBC_CacheCleared);
ASSERT(g_sbcEnabled); ASSERT(m_sbcFastPath);
//
// The cache has been cleared. Reset our fast path.
//
COM_BasedListInit(&m_sbcFastPath->usedList); COM_BasedListInit(&m_sbcFastPath->freeList);
for (i = 0; i < SBC_FASTPATH_ENTRIES; i++) { m_sbcFastPath->entry[i].list.next = 0; m_sbcFastPath->entry[i].list.prev = 0; COM_BasedListInsertBefore(&m_sbcFastPath->freeList, &m_sbcFastPath->entry[i].list); }
DebugExitVOID(ASHost::SBC_CacheCleared); }
//
//
// SBCSelectCache(..)
//
// Decides which cache a sub-bitmap from a source bitmap of the specified
// size should go in.
//
// To be cached, the sub-bitmap must:
// have a size, in compressed bytes, which fits in the cache
//
// The R1.1 cache selection is irrespective of the actual memory
// requirement for the cached data. This is wasteful of space, but is
// necessary for R1.1 compatibility. (The R1.1 cache paremeters mean that
// the total cache will be below about 128K in any case)
//
// For R2.0 the cache is selected by this function by comparing the
// post-compress size with the cell area of each of the caches. This gives
// us a much better space usage on both server and client.
//
// Returns:
// TRUE if the sub-bitmap can be cached.
// *pCache is updated with the index of the selected cache.
//
// FALSE if the sub-bitmap cannot be cached.
// *pCache is not updated.
//
//
BOOL ASHost::SBCSelectCache ( UINT cSize, UINT * pCache ) { BOOL fCacheSelected; BOOL fSelectedCacheIsFull; UINT i;
DebugEntry(ASHost::SBCSelectCache);
fCacheSelected = FALSE; fSelectedCacheIsFull = FALSE;
//
// This loop makes the assumption that cache 0 is the smallest. If
// abmcint.h changes this assumption it will need rewriting.
//
for (i = 0; i < NUM_BMP_CACHES; i++) { if (m_asbcBmpCaches[i].cEntries <= 0) { //
// No entries in this cache, so skip to the next one
//
continue; }
//
// R2 bitmap cache - only consider total cell size.
//
// Only consider this cache if
// - we haven't yet found a cache
// OR
// - we have found a cache, but it is full (i.e. will
// require an entry to be ejected) AND this one is not
// full
//
// (Note that a cache is full if freeEntry != NULL)
//
if (!fCacheSelected || (fSelectedCacheIsFull && ((m_asbcBmpCaches[i].freeEntry == NULL) || !m_asbcBmpCaches[i].freeEntry->inUse))) { if (cSize <= m_asbcBmpCaches[i].cSize) { if (fSelectedCacheIsFull) { TRACE_OUT(("Using cache %u because cache %u is full", *pCache, i)); }
*pCache = i; fCacheSelected = TRUE;
fSelectedCacheIsFull = ((m_asbcBmpCaches[i].freeEntry != NULL) && m_asbcBmpCaches[i].freeEntry->inUse);
if (!fSelectedCacheIsFull) { break; } } } }
DebugExitDWORD(ASHost::SBCSelectCache, fCacheSelected); return(fCacheSelected); }
//
// FUNCTION: SBC_RecreateSendCache
//
// DESCRIPTION:
//
// (Re)creates the send bitmap cache with a size suitable for the current
// capabilities.
//
// PARAMETERS:
// cache - index to the cache being recreated
// cOldEntries - the previous max number of entries in the cache
// oldCellSize - the previous cell size
//
// RETURNS: NONE
//
//
void ASHost::SBC_RecreateSendCache ( UINT cache, UINT newNumEntries, UINT newCellSize ) { PBMC_DIB_CACHE pCache = &(m_asbcBmpCaches[cache]);
DebugEntry(ASHost::SBC_RecreateSendCache);
//
// Allocate the memory for the new send cache
//
ASSERT((newCellSize != pCache->cCellSize) || (newNumEntries != pCache->cEntries));
//
// If the cache already exists then destroy it first
//
if (pCache->handle != 0) { TRACE_OUT(( "Destroy SBC cache %d", cache));
CH_DestroyCache(pCache->handle); pCache->handle = 0; }
//
// Now reallocate the cache data. This will free any memory previously
// allocated. If the entries/cellsize is zero, it will return success.
//
if (!BMCAllocateCacheData(newNumEntries, newCellSize, cache, pCache)) { ERROR_OUT(( "Bitmap caching disabled for cache %u", cache)); }
if (pCache->cEntries > 0) { //
// Allocate cache handler cache. Note that we force the cache
// handler to leave us with one entry in our hand at all times by
// decrementing its count of entries.
//
if (!CH_CreateCache(&(pCache->handle), pCache->cEntries - 1, SBC_NUM_CATEGORIES, BMC_DIB_NOT_HASHED, SBCCacheCallback )) { ERROR_OUT(( "Could not allocate SBC cache of (%u)", pCache->cEntries)); pCache->cEntries = 0; } }
TRACE_OUT(( "Created new cache: 0x%08x, size %u", pCache->handle, pCache->cEntries));
//
// Copy the relevant cache information into the shared memory buffer
//
m_asbcCacheInfo[cache].cEntries = (WORD)pCache->cEntries; m_asbcCacheInfo[cache].cCellSize = (WORD)pCache->cCellSize;
TRACE_OUT(("SBC cache %d: %d entries of size %d", cache, m_asbcCacheInfo[cache].cEntries, m_asbcCacheInfo[cache].cCellSize));
DebugExitVOID(ASHost::SBC_RecreateSendCache); }
//
// SBC_RecalcCaps()
//
// Enumerates all the people in the share and redetermines the size of the
// bitmap cache depending on their and the local receive capabilities.
//
//
// THIS CAN GO AWAY WHEN 2.X COMPAT DOES
//
void ASShare::SBC_RecalcCaps(BOOL fJoiner) { SBC_NEW_CAPABILITIES newCapabilities; UINT newSmallCellSize; UINT newSmallMaxEntries; UINT newMediumCellSize; UINT newMediumMaxEntries; UINT newLargeCellSize; UINT newLargeMaxEntries; PBMC_DIB_CACHE pSmall; PBMC_DIB_CACHE pMedium; PBMC_DIB_CACHE pLarge; BOOL cacheChanged = FALSE; ASPerson * pasT;
DebugEntry(ASShare::SBC_RecalcCaps);
if (!m_pHost || !g_sbcEnabled) { //
// Nothing to do -- we're not hosting, or there is no SBC. Note that
// 2.x always recalculated this stuff when somebody joined AND
// somebody left.
//
DC_QUIT; }
ValidatePerson(m_pasLocal);
pSmall = &(m_pHost->m_asbcBmpCaches[ID_SMALL_BMP_CACHE]); pMedium= &(m_pHost->m_asbcBmpCaches[ID_MEDIUM_BMP_CACHE]); pLarge = &(m_pHost->m_asbcBmpCaches[ID_LARGE_BMP_CACHE]);
//
// Enumerate all the bitmap cache receive capabilities of the parties
// in the share. The usable size of the send bitmap cache is then the
// minimum of all the remote receive caches and the local send cache
// size.
//
//
// Start by setting the size of the local send bitmap cache to the
// local default values.
//
newSmallCellSize = m_pasLocal->cpcCaps.bitmaps.sender.capsSmallCacheCellSize; newSmallMaxEntries = m_pasLocal->cpcCaps.bitmaps.sender.capsSmallCacheNumEntries;
newMediumCellSize = m_pasLocal->cpcCaps.bitmaps.sender.capsMediumCacheCellSize; newMediumMaxEntries = m_pasLocal->cpcCaps.bitmaps.sender.capsMediumCacheNumEntries;
newLargeCellSize = m_pasLocal->cpcCaps.bitmaps.sender.capsLargeCacheCellSize; newLargeMaxEntries = m_pasLocal->cpcCaps.bitmaps.sender.capsLargeCacheNumEntries;
if (m_scShareVersion < CAPS_VERSION_30) { TRACE_OUT(("In share with 2.x nodes, must recalc SBC caps"));
//
// Now enumerate all the REMOTE parties in the share and set our send bitmap
// size appropriately.
//
for (pasT = m_pasLocal->pasNext; pasT != NULL; pasT = pasT->pasNext) { //
// Set the size of the local send bitmap cache to the minimum of its
// current size and this party's receive bitmap cache size.
//
newSmallCellSize = min(newSmallCellSize, pasT->cpcCaps.bitmaps.receiver.capsSmallCacheCellSize); newSmallMaxEntries = min(newSmallMaxEntries, pasT->cpcCaps.bitmaps.receiver.capsSmallCacheNumEntries);
newMediumCellSize = min(newMediumCellSize, pasT->cpcCaps.bitmaps.receiver.capsMediumCacheCellSize); newMediumMaxEntries = min(newMediumMaxEntries, pasT->cpcCaps.bitmaps.receiver.capsMediumCacheNumEntries);
newLargeCellSize = min(newLargeCellSize, pasT->cpcCaps.bitmaps.receiver.capsLargeCacheCellSize); newLargeMaxEntries = min(newLargeMaxEntries, pasT->cpcCaps.bitmaps.receiver.capsLargeCacheNumEntries); } }
TRACE_OUT(("Recalced SBC caps: Small {%d of %d}, Medium {%d of %d}, Large {%d of %d}", newSmallMaxEntries, newSmallCellSize, newMediumMaxEntries, newMediumCellSize, newLargeMaxEntries, newLargeCellSize));
//
// If we've changed the size, reset the cache before continuing.
//
if ((pSmall->cCellSize != newSmallCellSize) || (pSmall->cEntries != newSmallMaxEntries)) { m_pHost->SBC_RecreateSendCache(ID_SMALL_BMP_CACHE, newSmallMaxEntries, newSmallCellSize); cacheChanged = TRUE; }
if ((pMedium->cCellSize != newMediumCellSize) || (pMedium->cEntries != newMediumMaxEntries)) { m_pHost->SBC_RecreateSendCache(ID_MEDIUM_BMP_CACHE, newMediumMaxEntries, newMediumCellSize); cacheChanged = TRUE; }
if ((pLarge->cCellSize != newLargeCellSize) || (pLarge->cEntries != newLargeMaxEntries)) { m_pHost->SBC_RecreateSendCache(ID_LARGE_BMP_CACHE, newLargeMaxEntries, newLargeCellSize); cacheChanged = TRUE; }
//
// If we had to recreate any of the send caches, make sure that we
// clear the fast path.
//
if (cacheChanged) { m_pHost->SBC_CacheCleared(); }
//
// Handle new capabilities
//
//
// Set up the new capabilities structure...
//
newCapabilities.sendingBpp = m_pHost->m_usrSendingBPP;
newCapabilities.cacheInfo = m_pHost->m_asbcCacheInfo;
//
// ... and pass it through to the driver.
//
if (! OSI_FunctionRequest(SBC_ESC_NEW_CAPABILITIES, (LPOSI_ESCAPE_HEADER)&newCapabilities, sizeof(newCapabilities))) { ERROR_OUT(("SBC_ESC_NEW_CAPABILITIES failed")); }
DC_EXIT_POINT: DebugExitVOID(ASShare::SBC_RecalcCaps); }
//
// FUNCTION: SBCCacheCallback
//
// DESCRIPTION:
//
// Send BMC Cache Manager callback function. Called whenever an entry is
// removed from the cache to allow us to free up the object.
//
// PARAMETERS:
//
// hCache - cache handle
//
// event - the cache event that has occured
//
// iCacheEntry - index of the cache entry that the event is affecting
//
// pData - pointer to the cache data associated with the given cache entry
//
// cbDataSize - size in bytes of the cached data
//
// RETURNS: Nothing
//
//
void SBCCacheCallback ( ASHost * pHost, PCHCACHE pCache, UINT iCacheEntry, LPBYTE pData ) { UINT cache;
DebugEntry(SBCCacheCallback);
//
// Simply release the cache entry for reuse. We must scan for
// the correct cache root
//
for (cache = 0; cache < NUM_BMP_CACHES; cache++) { if (pHost->m_asbcBmpCaches[cache].handle == pCache) { pHost->m_asbcBmpCaches[cache].freeEntry = (PBMC_DIB_ENTRY)pData; pHost->m_asbcBmpCaches[cache].freeEntry->inUse = FALSE;
TRACE_OUT(("0x%08x SBC cache entry 0x%08x now free", pCache, pData));
pHost->SBC_CacheEntryRemoved(cache, iCacheEntry); break; } }
DebugExitVOID(SBCCacheCallback); }
//
//
// SBC_ProcessMemBltOrder()
//
//
BOOL ASHost::SBC_ProcessMemBltOrder ( LPINT_ORDER pOrder, LPINT_ORDER * ppNextOrder ) { BOOL rc = FALSE; UINT orderType; UINT tileId; UINT tileType; LPSBC_TILE_DATA pTileData = NULL; UINT bitmapWidth; int bitmapHeight; LPINT_ORDER pBMCOrder = NULL; UINT colorCacheIndex; UINT bitsCache; UINT bitsCacheIndex; UINT numColors; LPLONG pXSrc; LPLONG pYSrc; BOOL isNewColorTableEntry; BOOL isNewBitsEntry; BOOL canFastPath = TRUE; LPMEMBLT_ORDER pMemBltOrder = (LPMEMBLT_ORDER)&(pOrder->abOrderData); LPMEM3BLT_ORDER pMem3BltOrder = (LPMEM3BLT_ORDER)pMemBltOrder; LPMEMBLT_R2_ORDER pMemBltR2Order = (LPMEMBLT_R2_ORDER)pMemBltOrder; LPMEM3BLT_R2_ORDER pMem3BltR2Order = (LPMEM3BLT_R2_ORDER)pMemBltOrder; BITMAPINFO_ours sbcBitmapInfo;
DebugEntry(ASHost::SBC_ProcessMemBltOrder);
*ppNextOrder = NULL;
//
// We may already have processed this MEMBLT order and have the color
// table and bitmap bits for it, ready to go across the wire. This
// would happen if the update packager called this function to process
// the MEMBLT, but then didn't have enough room in its current network
// packet to send the color table or the bitmap bits.
//
// So, if we've already processed this order, bail out now.
//
if (m_sbcOrderInfo.pOrder == pOrder) { //
// We've got a match ! Do we have valid data for it ? If we don't
// we must have failed last time, so we'll probably fail again (we
// don't do any memory allocation, so it's unlikely that the error
// condition has cleared up). In any case, we should not have been
// called again if we failed last time...
//
if (m_sbcOrderInfo.validData) { TRACE_OUT(( "Already have valid data for this MEMBLT")); rc = TRUE; } else { WARNING_OUT(( "Have invalid data for this MEMBLT")); } DC_QUIT; }
//
// Re-initialise m_sbcOrderInfo
//
m_sbcOrderInfo.pOrder = pOrder; m_sbcOrderInfo.validData = FALSE; m_sbcOrderInfo.sentColorTable = FALSE; m_sbcOrderInfo.sentBitmapBits = FALSE; m_sbcOrderInfo.sentMemBlt = FALSE;
//
// Here's on overview of what we do here...
//
// We've been given a MEMBLT order which references an entry in a shunt
// buffer containing the bits for the MEMBLT at the native bpp (the bpp
// of the display). We want to cache the bits and a color table at the
// protocol bpp. So, we
//
// - copy the bits from the shunt buffer into a work DIB section
// - call GetDIBits to get the data from the work DIB section at the
// protocol bpp
// - cache the bits and the color table
// - if we add new cache entries for the bits and / or the color table,
// we fill in m_sbcOrderInfo.pBitmapBits order and / or
// m_sbcOrderInfo.pColorTableInfo to hold the orders to be sent before
// the MEMBLT order.
//
//
// Make sure that we've been given the correct order type. Note that
// we will never be given the R2 versions of the MEMBLT orders.
//
orderType = pMemBltOrder->type; ASSERT(((orderType == ORD_MEMBLT_TYPE) || (orderType == ORD_MEM3BLT_TYPE)));
//
// Get a pointer to the entry in one of the shunt buffers which matches
// this order.
//
if (orderType == ORD_MEMBLT_TYPE) { tileId = pMemBltOrder->cacheId; } else { tileId = pMem3BltOrder->cacheId; }
if (!SBCGetTileData(tileId, &pTileData, &tileType)) { ERROR_OUT(( "Failed to find entry for tile %hx in shunt buffer", tileId)); DC_QUIT; }
bitmapWidth = pTileData->width; bitmapHeight = pTileData->height;
//
// Check if we should do any fast path operations on this bitmap
//
if (pTileData->majorCacheInfo == SBC_DONT_FASTPATH) { TRACE_OUT(( "Tile %x should not be fastpathed", tileId)); canFastPath = FALSE; } //
// Try to find an entry for this bitmap in the fast path (unless the
// bitmap is marked as being non-fastpathable).
//
if (canFastPath && SBCFindInFastPath(pTileData->majorCacheInfo, pTileData->minorCacheInfo, pTileData->majorPalette, pTileData->minorPalette, pTileData->srcX, pTileData->srcY, pTileData->tilingWidth, pTileData->tilingHeight, &bitsCache, &bitsCacheIndex, &colorCacheIndex)) { isNewBitsEntry = FALSE; isNewColorTableEntry = FALSE;
//
// Call the cache handler to get it to update its MRU entry for
// this cache entry
//
CH_TouchCacheEntry(m_asbcBmpCaches[bitsCache].handle, bitsCacheIndex); } else { //
// There is no entry in the fast path...
//
// Copy the data from the tile in the shunt buffer into the work
// DIB section. Note that this only works correctly because both
// our work DIB and the tile data are "top down" rather than the
// default of "bottom up". i.e the data for the first scanline is
// stored first in memory. If this wasn't the case, we'd have to
// work out an offset into the work DIB to start copying to.
//
memcpy(m_asbcWorkInfo[tileType].pWorkBitmapBits, pTileData->bitData, pTileData->bytesUsed);
//
// Now set up the destination for the GetDIBits call. First set up
// a bitmap info header to pass to GetDIBits. Only the header part
// of the structure will be sent across the network - the color
// table is sent via the palette packets.
//
// Note that we set the height in the bitmap info header to be
// negative. This forces a convertion from our "top down" DIB
// format to the default "bottom up" format which we want to cache
// and send over the wire.
//
ZeroMemory(&sbcBitmapInfo, sizeof(sbcBitmapInfo)); m_pShare->USR_InitDIBitmapHeader((BITMAPINFOHEADER *)&sbcBitmapInfo, m_usrSendingBPP); sbcBitmapInfo.bmiHeader.biWidth = m_asbcWorkInfo[tileType].tileWidth; sbcBitmapInfo.bmiHeader.biHeight = -(int)m_asbcWorkInfo[tileType].tileHeight;
//
// OK, we've set up the source and the destination, so now get the
// data at the protocol bpp. We get the bits into the usr general
// bitmap work buffer.
//
if (GetDIBits(m_usrWorkDC, m_asbcWorkInfo[tileType].workBitmap, 0, bitmapHeight, m_pShare->m_usrPBitmapBuffer, (BITMAPINFO *)&sbcBitmapInfo, DIB_RGB_COLORS) != (int)bitmapHeight) { ERROR_OUT(( "GetDIBits failed")); DC_QUIT; }
TRACE_OUT(( "%d x %d, (fixed %d) -> (%d, %d)", bitmapWidth, bitmapHeight, m_asbcWorkInfo[tileType].tileWidth, pMemBltOrder->nLeftRect, pMemBltOrder->nTopRect));
numColors = COLORS_FOR_BPP(m_usrSendingBPP);
//
// There is no color table to cache if there is no color table at
// all, which is the case when sending at 24BPP
//
if (numColors) { //
// Cache the color table. If this succeeds, colorCacheIndex will
// be set up to contain the details of the cache entry which the
// data is cached in. In addition, if isNewColorTableEntry is TRUE
// on return, psbcOrders.colorTableOrder will be fully initialized
// and ready to go across the wire.
//
if (!SBCCacheColorTable(m_sbcOrderInfo.pColorTableOrder, sbcBitmapInfo.bmiColors, numColors, &colorCacheIndex, &isNewColorTableEntry)) { TRACE_OUT(( "Failed to cache color table")); DC_QUIT; }
ASSERT(colorCacheIndex != COLORCACHEINDEX_NONE); } else { colorCacheIndex = COLORCACHEINDEX_NONE; isNewColorTableEntry = FALSE; }
//
// Cache the bits. If this succeeds, bitsCache and bitsCacheIndex
// will be set up to contain the details of the cache entry which
// the data is cached in. In addition, if isNewBitsEntry is TRUE
// on return, psbcOrders.bitmapBitsOrder will be fully initialized
// and ready to go across the wire.
//
// If this fails, the above values will be undefined.
//
if (!SBCCacheBits(m_sbcOrderInfo.pBitmapBitsOrder, m_sbcOrderInfo.bitmapBitsDataSize, m_pShare->m_usrPBitmapBuffer, bitmapWidth, m_asbcWorkInfo[tileType].tileWidth, bitmapHeight, BYTES_IN_BITMAP(m_asbcWorkInfo[tileType].tileWidth, bitmapHeight, sbcBitmapInfo.bmiHeader.biBitCount), &bitsCache, &bitsCacheIndex, &isNewBitsEntry)) { TRACE_OUT(( "Failed to cache bits")); DC_QUIT; }
//
// Add the newly cached item to the fast path (unless the bitmap is
// marked as being non-fastpathable).
//
if (canFastPath) { SBCAddToFastPath(pTileData->majorCacheInfo, pTileData->minorCacheInfo, pTileData->majorPalette, pTileData->minorPalette, pTileData->srcX, pTileData->srcY, pTileData->tilingWidth, pTileData->tilingHeight, bitsCache, bitsCacheIndex, colorCacheIndex); } }
//
// We've now got valid cache entries for the DIB bits and the color
// table, so we should now fill them into the MEMBLT order.
//
// Set up the source co-ordinates. For R1 protocols, the x-coordinate
// includes the offset which is required to get the right cell within
// the receive bitmap cache. For R2, we set up the cache entry in a
// separate field.
//
if (orderType == ORD_MEMBLT_TYPE) { pXSrc = &pMemBltOrder->nXSrc; pYSrc = &pMemBltOrder->nYSrc; } else { pXSrc = &pMem3BltOrder->nXSrc; pYSrc = &pMem3BltOrder->nYSrc; }
*pXSrc = *pXSrc % pTileData->tilingWidth; *pYSrc = *pYSrc % pTileData->tilingHeight;
//
// The sub-bitmap and color table are in the cache. Store a cache
// handle and color handle. Also store the cache index for R2
// protocols (see above).
//
if (orderType == ORD_MEMBLT_TYPE) { pMemBltOrder->cacheId = MEMBLT_COMBINEHANDLES(colorCacheIndex, bitsCache);
pMemBltR2Order->type = (TSHR_UINT16)ORD_MEMBLT_R2_TYPE; pMemBltR2Order->cacheIndex = (TSHR_UINT16)bitsCacheIndex;
TRACE_OUT(( "MEMBLT color %u bitmap %u:%u", colorCacheIndex, bitsCache, bitsCacheIndex)); } else { pMem3BltOrder->cacheId = MEMBLT_COMBINEHANDLES(colorCacheIndex, bitsCache);
pMem3BltR2Order->type = ORD_MEM3BLT_R2_TYPE; pMem3BltR2Order->cacheIndex = (TSHR_UINT16)bitsCacheIndex;
TRACE_OUT(( "MEM3BLT color %u bitmap %u:%u", colorCacheIndex, bitsCache, bitsCacheIndex)); }
//
// Must have successfully completed processing the order to get to
// here. Fill in the appropriate info in the m_sbcOrderInfo structure.
// If we got a cache hit on the color table or the bitmap bits then
// we've already sent the data for them.
//
m_sbcOrderInfo.validData = TRUE; m_sbcOrderInfo.sentColorTable = !isNewColorTableEntry; m_sbcOrderInfo.sentBitmapBits = !isNewBitsEntry; rc = TRUE;
DC_EXIT_POINT: if (rc) { //
// We've successfully processed the MEMBLT, so set up a pointer to
// the next order which should be sent by the caller.
//
// Note that if we have already sent these orders, then we return
// a NULL order.
//
if (!m_sbcOrderInfo.sentColorTable) { TRACE_OUT(( "Returning color table order")); *ppNextOrder = m_sbcOrderInfo.pColorTableOrder; } else if (!m_sbcOrderInfo.sentBitmapBits) { TRACE_OUT(( "Returning bitmap bits order")); *ppNextOrder = m_sbcOrderInfo.pBitmapBitsOrder; } else if (!m_sbcOrderInfo.sentMemBlt) { TRACE_OUT(( "Returning MemBlt order")); *ppNextOrder = pOrder; } else { TRACE_OUT(( "No order to return")); rc = FALSE; } }
//
// We've finished with the entry in the shunt buffer, so reset the
// inUse flag to allow the driver to re-use it.
//
if (pTileData != NULL) { pTileData->inUse = FALSE; }
DebugExitBOOL(ASHost::SBC_ProcessMemBltOrder, rc); return(rc); }
//
//
// SBC_OrderSentNotification()
//
//
void ASHost::SBC_OrderSentNotification(LPINT_ORDER pOrder) { DebugEntry(ASHost::SBC_OrderSentNotification);
//
// pOrder should be a pointer to either our internal bitmap bits order,
// or our color table order.
//
if (pOrder == m_sbcOrderInfo.pBitmapBitsOrder) { TRACE_OUT(( "Bitmap bits order has been sent")); m_sbcOrderInfo.sentBitmapBits = TRUE; } else if (pOrder == m_sbcOrderInfo.pColorTableOrder) { TRACE_OUT(( "Color table order has been sent")); m_sbcOrderInfo.sentColorTable = TRUE; } else if (pOrder == m_sbcOrderInfo.pOrder) { TRACE_OUT(( "Memblt order has been sent")); m_sbcOrderInfo.sentMemBlt = TRUE;
//
// All parts of the Memblt have been sent now, so reset our pointer
// to the order. This avoids a problem where
// SBC_ProcessMemBltOrder is called twice in a row with the same
// pOrder, but with different data (i.e. consecutive MemBlts
// ending up in the same point in the order heap). It can happen...
//
m_sbcOrderInfo.pOrder = NULL; } else { ERROR_OUT(( "Notification for unknown order %#.8lx", pOrder)); }
DebugExitVOID(ASHost::SBC_OrderSentNotification); }
//
//
// SBC_ProcessInternalOrder()
//
//
void ASHost::SBC_ProcessInternalOrder(LPINT_ORDER pOrder) { UINT orderType; LPINT_COLORTABLE_ORDER_1BPP pColorTableOrder; HBITMAP oldBitmap = 0; UINT numEntries; int i;
DebugEntry(ASHost::SBC_ProcessInternalOrder);
//
// Make sure that we've been given an order type which we recognise.
// Currently, the only internal order we support is a color table
// order.
//
pColorTableOrder = (LPINT_COLORTABLE_ORDER_1BPP)&(pOrder->abOrderData); orderType = pColorTableOrder->header.type;
ASSERT(orderType == INTORD_COLORTABLE_TYPE);
//
// Make sure that the color table order is the same bpp as the work DIB
// sections.
//
ASSERT(pColorTableOrder->header.bpp == g_usrCaptureBPP);
//
// All we have to do is to copy the color table from the order into our
// two work DIB sections. To do that, we have to select the DIB
// sections into a DC then set the color table for the DC - this sets
// the color table in the DIB section.
//
numEntries = COLORS_FOR_BPP(g_usrCaptureBPP); ASSERT(numEntries);
for (i = 0 ; i < SBC_NUM_TILE_SIZES; i++) { oldBitmap = SelectBitmap(m_usrWorkDC, m_asbcWorkInfo[i].workBitmap);
SetDIBColorTable(m_usrWorkDC, 0, // First index
numEntries, // Number of entries
(RGBQUAD*)pColorTableOrder->colorData); }
if (oldBitmap != NULL) { SelectBitmap(m_usrWorkDC, oldBitmap); }
DebugExitVOID(ASHost::SBC_ProcessInternalOrder); }
//
//
// SBC_PMCacheEntryRemoved()
//
//
void ASHost::SBC_PMCacheEntryRemoved(UINT cacheIndex) { LPSBC_FASTPATH_ENTRY pEntry; LPSBC_FASTPATH_ENTRY pNextEntry;
DebugEntry(ASHost::SBC_PMCacheEntryRemoved);
ASSERT(m_sbcFastPath);
//
// An entry has been removed from the color cache. We have to remove
// all entries from the fast path which reference this color table.
//
TRACE_OUT(( "Color table cache entry %d removed - removing references", cacheIndex));
pEntry = (LPSBC_FASTPATH_ENTRY)COM_BasedListFirst(&m_sbcFastPath->usedList, FIELD_OFFSET(SBC_FASTPATH_ENTRY, list)); while (pEntry != NULL) { pNextEntry = (LPSBC_FASTPATH_ENTRY)COM_BasedListNext(&m_sbcFastPath->usedList, pEntry, FIELD_OFFSET(SBC_FASTPATH_ENTRY, list));
if (pEntry->colorIndex == cacheIndex) { COM_BasedListRemove(&pEntry->list); COM_BasedListInsertAfter(&m_sbcFastPath->freeList, &pEntry->list); }
pEntry = pNextEntry; }
DebugExitVOID(ASHost::SBC_PMCacheEntryRemoved); }
//
//
// Name: SBCInitInternalOrders
//
// Purpose: Allocate memory for the internal orders used during MEMBLT
// order processing.
//
// Returns: TRUE if initialized OK, FALSE otherwise.
//
// Params: None
//
// Operation: If successful, this function initializes the following
//
// g_Share->sbcOrderInfo
//
//
BOOL ASHost::SBCInitInternalOrders(void) { BOOL initOK = FALSE; UINT orderSize; LPINT_ORDER_HEADER pOrderHeader;
DebugEntry(ASHost::SBCInitInternalOrders);
//
// Start with the bitmap bits order. Calculate the number of bytes
// required to store the bits for the largest bitmap bits order we will
// ever send. This includes room for the compression header which gets
// added before the bits if the data is compressed.
//
if (g_usrCaptureBPP >= 24) { // Can possibly send 24bpp TRUE COLOR data
m_sbcOrderInfo.bitmapBitsDataSize = BYTES_IN_BITMAP(MP_LARGE_TILE_WIDTH, MP_LARGE_TILE_HEIGHT, 24) + sizeof(CD_HEADER); } else { // Can't send 24bpp TRUE color data
m_sbcOrderInfo.bitmapBitsDataSize = BYTES_IN_BITMAP(MP_LARGE_TILE_WIDTH, MP_LARGE_TILE_WIDTH, 8) + sizeof(CD_HEADER); }
//
// Now allocate memory for the bitmap bits order. The size required
// is:
// The size of an INT_ORDER_HEADER (this is added in by OA when you
// call OA_AllocOrderMem)
// + the size of the largest BMC_BITMAP_BITS_ORDER structure
// + the number of bytes required for the bitmap bits
// + contingency for RLE compression overruns !
//
orderSize = sizeof(INT_ORDER_HEADER) + sizeof(BMC_BITMAP_BITS_ORDER_R2) + m_sbcOrderInfo.bitmapBitsDataSize + 4;
TRACE_OUT(( "Allocating %d bytes for SBC bitmap bits order (bits %d)", orderSize, m_sbcOrderInfo.bitmapBitsDataSize));
m_sbcOrderInfo.pBitmapBitsOrder = (LPINT_ORDER)new BYTE[orderSize]; if (!m_sbcOrderInfo.pBitmapBitsOrder) { ERROR_OUT(( "Failed to alloc %d bytes for SBC bitmap bits order (bits %d)", orderSize, m_sbcOrderInfo.bitmapBitsDataSize)); DC_QUIT; }
//
// Initialize the INT_ORDER_HEADER - this is normally done in
// OA_AllocOrderMem(). For the bitmap bits order, we can't fill in the
// orderLength because it is not a fixed size - this has to be done
// later when we fill in the bitmap bits. Note that the order length
// excludes the size of the INT_ORDER_HEADER.
//
pOrderHeader = &m_sbcOrderInfo.pBitmapBitsOrder->OrderHeader; pOrderHeader->additionalOrderData = 0; pOrderHeader->cbAdditionalOrderDataLength = 0;
//
// Now the color table order. The size required is:
// The size of an INT_ORDER_HEADER (this is added in by OA when you
// call OA_AllocOrderMem)
// + the size of a BMC_COLOR_TABLE_ORDER structure
// + the number of bytes required for the color table entries (note
// that the BMC_COLOR_TABLE_ORDER structure contains the first
// color table entry, so adjust the number of extra bytes required)
//
// Color tables are only for 8bpp and less.
orderSize = sizeof(INT_ORDER_HEADER) + sizeof(BMC_COLOR_TABLE_ORDER) + (COLORS_FOR_BPP(8) - 1) * sizeof(TSHR_RGBQUAD);
TRACE_OUT(( "Allocating %d bytes for SBC color table order", orderSize));
m_sbcOrderInfo.pColorTableOrder = (LPINT_ORDER)new BYTE[orderSize]; if (!m_sbcOrderInfo.pColorTableOrder) { ERROR_OUT(( "Failed to alloc %d bytes for SBC color table order", orderSize)); DC_QUIT; }
pOrderHeader = &m_sbcOrderInfo.pColorTableOrder->OrderHeader; pOrderHeader->additionalOrderData = 0; pOrderHeader->cbAdditionalOrderDataLength = 0; pOrderHeader->Common.cbOrderDataLength = (WORD)(orderSize - sizeof(INT_ORDER_HEADER));
//
// Fill in the remaining fields in m_sbcOrderInfo
//
m_sbcOrderInfo.pOrder = NULL; m_sbcOrderInfo.validData = FALSE; m_sbcOrderInfo.sentColorTable = FALSE; m_sbcOrderInfo.sentBitmapBits = FALSE; m_sbcOrderInfo.sentMemBlt = FALSE;
//
// Must be OK to get to here
//
initOK = TRUE;
DC_EXIT_POINT: DebugExitDWORD(ASHost::SBCInitInternalOrders, initOK); return(initOK); }
//
//
// Name: SBCFreeInternalOrders
//
// Purpose: Free up the internal orders used by SBC during MEMBLT order
// processing.
//
// Returns: Nothing
//
// Params: None
//
//
void ASHost::SBCFreeInternalOrders(void) { DebugEntry(ASHost::SBCFreeInternalOrders);
//
// First free up the memory.
//
if (m_sbcOrderInfo.pBitmapBitsOrder) { delete m_sbcOrderInfo.pBitmapBitsOrder; m_sbcOrderInfo.pBitmapBitsOrder = NULL; }
if (m_sbcOrderInfo.pColorTableOrder) { delete m_sbcOrderInfo.pColorTableOrder; m_sbcOrderInfo.pColorTableOrder = NULL; }
//
// Now reset the remaining fields in m_sbcOrderInfo
//
m_sbcOrderInfo.pOrder = NULL; m_sbcOrderInfo.validData = FALSE; m_sbcOrderInfo.sentColorTable = FALSE; m_sbcOrderInfo.sentBitmapBits = FALSE; m_sbcOrderInfo.bitmapBitsDataSize = 0;
DebugExitVOID(ASHost::SBCFreeInternalOrders); }
//
//
// Name: SBCInitFastPath
//
// Purpose: Initialize the SBC fast path
//
// Returns: TRUE if successful, FALSE otherwise
//
// Params: None
//
//
BOOL ASHost::SBCInitFastPath(void) { BOOL rc = FALSE;
DebugEntry(ASHost::SBCInitFastPath);
m_sbcFastPath = new SBC_FASTPATH; if (!m_sbcFastPath) { ERROR_OUT(("Failed to alloc m_sbcFastPath")); DC_QUIT; }
SET_STAMP(m_sbcFastPath, SBCFASTPATH);
//
// Initialize the structure.
//
SBC_CacheCleared();
rc = TRUE;
DC_EXIT_POINT: DebugExitBOOL(ASHost::SBCInitFastPath, rc); return(rc); }
//
//
// Name: SBCGetTileData
//
// Purpose: Given the ID of a tile data entry in one of the SBC shunt
// buffers, return a pointer to the entry with that ID.
//
// Returns: TRUE if the entry is found, FALSE otherwise
//
// Params: IN tileId - The ID of the shunt buffer entry to be
// found.
// OUT ppTileData - A pointer to the start of the shunt buffer
// entry (if found)
// OUT pTileType - The type of shunt buffer entry found. One
// of:
// SBC_MEDIUM_TILE
// SBC_LARGE_TILE
//
//
BOOL ASHost::SBCGetTileData ( UINT tileId, LPSBC_TILE_DATA * ppTileData, LPUINT pTileType ) { BOOL gotTileData = FALSE; UINT workTile; LPSBC_TILE_DATA pWorkTile;
DebugEntry(ASHost::SBCGetTileData);
TRACE_OUT(( "Looking for tile Id %x", tileId));
//
// Find out which of the shunt buffers the entry should be in.
//
*pTileType = SBC_TILE_TYPE(tileId);
//
// We implement the shunt buffers as circular FIFO queues, so in
// general, we are looking for the entry following the last one which
// we found. However, this wont always be the case because we do some
// out of order processing when we do spoiling.
//
// So, get the index of the last tile we accessed.
//
workTile = m_asbcWorkInfo[*pTileType].mruIndex;
//
// OK, so lets go for it ! Start at the tile following the last one we
// accessed, and loop through the circular buffer until we get a match,
// or have circled back to the beginning.
//
// Note that this has been coded as a "do while" loop, rather than just
// a "while" loop so that we don't miss mruTile.
//
do { //
// On to the next tile
//
workTile++; if (workTile == m_asbcWorkInfo[*pTileType].pShuntBuffer->numEntries) { workTile = 0; }
pWorkTile = SBCTilePtrFromIndex(m_asbcWorkInfo[*pTileType].pShuntBuffer, workTile);
if (pWorkTile->inUse) { if (pWorkTile->tileId == tileId) { //
// We've got a match.
//
TRACE_OUT(( "Matched tile Id %x at index %d", tileId, workTile)); *ppTileData = pWorkTile; gotTileData = TRUE; m_asbcWorkInfo[*pTileType].mruIndex = workTile; DC_QUIT; } } } while (workTile != m_asbcWorkInfo[*pTileType].mruIndex);
//
// If we get to here, we've not found a match.
//
TRACE_OUT(( "No match for tile Id %x", tileId));
DC_EXIT_POINT: DebugExitBOOL(ASHost::SBCGetTileData, gotTileData); return(gotTileData); }
//
//
// Name: SBCCacheColorTable
//
// Purpose: Ensure that the given color table is cached.
//
// Returns: TRUE if the color table is cached successfully, FALSE
// otherwise.
//
// Params: IN pOrder - A pointer to a color table order to be
// filled in.
// IN pColorTable - A pointer to the start of the color table
// to be cached.
// IN numColors - The number of colors in the color table.
// OUT pCacheIndex - The index of the cached color table.
// OUT pIsNewEntry - TRUE if we added a new cache entry,
// FALSE if we matched an existing entry.
//
// Operation: pOrder is only filled in if *pIsNewEntry is FALSE.
//
//
BOOL ASHost::SBCCacheColorTable ( LPINT_ORDER pOrder, LPTSHR_RGBQUAD pColorTable, UINT numColors, UINT * pCacheIndex, LPBOOL pIsNewEntry ) { BOOL cachedOK = FALSE; UINT cacheIndex; PBMC_COLOR_TABLE_ORDER pColorTableOrder;
DebugEntry(ASHost::SBCCacheColorTable);
//
// Call PM to do the caching.
//
if (!PM_CacheTxColorTable(&cacheIndex, pIsNewEntry, numColors, pColorTable)) { ERROR_OUT(( "Failed to cache color table")); DC_QUIT; }
//
// If the cache operation resulted in a cache update then we have to
// fill in the color table order.
//
if (*pIsNewEntry) { //
// The color table is new so we have to transmit it
//
TRACE_OUT(( "New color table"));
pOrder->OrderHeader.Common.fOrderFlags = OF_PRIVATE; pColorTableOrder = (PBMC_COLOR_TABLE_ORDER)(pOrder->abOrderData); pColorTableOrder->bmcPacketType = BMC_PT_COLOR_TABLE; pColorTableOrder->colorTableSize = (TSHR_UINT16)numColors; pColorTableOrder->index = (BYTE)cacheIndex;
//
// Copy the new color table into the Order Packet.
//
memcpy(pColorTableOrder->data, pColorTable, numColors * sizeof(TSHR_RGBQUAD)); } else { TRACE_OUT(( "Existing color table")); }
//
// Return the color table index to the caller
//
*pCacheIndex = cacheIndex; cachedOK = TRUE;
DC_EXIT_POINT: DebugExitBOOL(ASHost::SBCCacheColorTable, cachedOK); return(cachedOK); }
//
//
// Name: SBCCacheBits
//
// Purpose: This function adds the supplied bitmap bits to a bitmap
// cache. The cache selected depends on the bitmap size, but
// may be different for R1 and R2. SBCSelectCache handles the
// determination of the correct cache.
//
// Returns: TRUE if the bits have been cached OK, FALSE otherwise
//
// Params: IN pOrder - A pointer to a BMC order.
// IN destBitsSize - The number of bytes available in
// pOrder to store the bitmap data.
// IN pDIBits - A pointer to the bits to be cached.
// IN bitmapWidth - The "in use" width of the bitmap
// IN fixedBitmapWidth - The actual width of the bitmap
// IN bitmapHeight - The height of the bitmap
// IN numBytes - The number of bytes in the bitmap.
// OUT pCache - The cache that we put the bits into.
// OUT pCacheIndex - The cache index within *pCache at
// which we cached the data.
// OUT pIsNewEntry - TRUE if we added a new cache entry,
// FALSE if we matched an existing entry.
//
// Operation: pOrder is only filled in if *pIsNewEntry is FALSE.
//
//
BOOL ASHost::SBCCacheBits ( LPINT_ORDER pOrder, UINT destBitsSize, LPBYTE pDIBits, UINT bitmapWidth, UINT fixedBitmapWidth, UINT bitmapHeight, UINT numBytes, UINT * pCache, UINT * pCacheIndex, LPBOOL pIsNewEntry ) { BOOL cachedOK = FALSE; UINT cacheIndex; UINT i; LPBYTE pCompressed; UINT compressedSize; BOOL compressed; PBMC_DIB_ENTRY pEntry; PBMC_DIB_CACHE pCacheHdr; PBMC_BITMAP_BITS_ORDER_R2 pBitsOrderR2; PBMC_BITMAP_BITS_DATA pBmcData; LPBYTE pDestBits;
DebugEntry(ASHost::SBCCacheBits);
pBmcData = (PBMC_BITMAP_BITS_DATA)(pOrder->abOrderData); pBitsOrderR2 = (PBMC_BITMAP_BITS_ORDER_R2)pBmcData;
//
// Get a pointer to where the bitmap data starts in the order. This
// depends on whether it is an R1 or an R2 bitmap bits order.
//
pDestBits = pBitsOrderR2->data;
//
// Before we can select a cache entry we need to compress the bits.
// This therefore mandates a memcpy into the cache entry when we come
// to add it. The saving in memory by storing the bits compressed
// makes it all worthwhile.
//
// Compress the bitmap data. At this stage we don't know whether the
// bitmap will compress well or not, so allow cells that are larger
// than our maximum cell size. The largest we expect to see is 120*120*
// 24.
//
compressedSize = destBitsSize; if (m_pShare->BC_CompressBitmap(pDIBits, pDestBits, &compressedSize, fixedBitmapWidth, bitmapHeight, m_usrSendingBPP, NULL ) && (compressedSize < numBytes))
{ TRACE_OUT(( "Compressed bmp data from %u bytes to %u bytes", numBytes, compressedSize)); compressed = TRUE; pCompressed = pDestBits; } else { //
// The bitmap could not be compressed, or bitmap compression is not
// enabled. Send the bitmap uncompressed.
//
compressed = FALSE; compressedSize = numBytes; pCompressed = pDIBits; }
//
// Make sure that the data will fit into the order. Do this after
// compression since it is possible that the uncompressed data will not
// fit, but the compressed version will.
//
if (compressedSize > destBitsSize) { WARNING_OUT(( "Data (%d bytes) does not fit into order (%d bytes)", compressedSize, destBitsSize)); DC_QUIT; }
//
// Select the cache based on the compressed size - we pass in the
// sub-bitmap dimensions for R1 caching; R2 caching just uses the
// total size of the bits.
//
if (!SBCSelectCache(compressedSize + sizeof(BMC_DIB_ENTRY) - 1, pCache)) { TRACE_OUT(( "No cache selected")); DC_QUIT; } else { TRACE_OUT(( "Selected cache %d", *pCache)); }
//
// Find a free cache entry in our selected cache
//
// We arrange that our transmit cache is always one greater than the
// negotiated cache size so that we should never fail to find a free
// array entry. Once we have fully populated our Tx cache we will
// always find the free entry as the one last given back to us by CH.
// Note the scan to <= sbcTxCache[pmNumTxCacheEntries is NOT a mistake.
//
pCacheHdr = &(m_asbcBmpCaches[*pCache]); if (pCacheHdr->data == NULL) { ERROR_OUT(( "Asked to cache when no cache allocated")); DC_QUIT; }
//
// If the cache has returned an entry to us then use that without
// having to scan. This will be the default mode for adding entries
// to a fully populated cache.
//
if (pCacheHdr->freeEntry != NULL) { pEntry = pCacheHdr->freeEntry; pCacheHdr->freeEntry = NULL; TRACE_OUT(( "Cache fully populated - using entry 0x%08x", pEntry)); } else { //
// We are in the process of feeding the cache so we need to search
// for a free entry
//
pEntry = (PBMC_DIB_ENTRY)(pCacheHdr->data); for (i=0 ; i < pCacheHdr->cEntries ; i++) { if (!pEntry->inUse) { break; } pEntry = (PBMC_DIB_ENTRY)(((LPBYTE)pEntry) + pCacheHdr->cSize); }
//
// We should never run out of free entries, but cope with it
//
if (i == pCacheHdr->cEntries) { ERROR_OUT(( "All Tx DIB cache entries in use")); DC_QUIT; } }
//
// Set up the DIB entry for caching
//
pEntry->inUse = TRUE; pEntry->cx = (TSHR_UINT16)bitmapWidth; pEntry->cxFixed = (TSHR_UINT16)fixedBitmapWidth; pEntry->cy = (TSHR_UINT16)bitmapHeight; pEntry->bpp = (TSHR_UINT16)m_usrSendingBPP; pEntry->cBits = numBytes; pEntry->bCompressed = (BYTE)compressed; pEntry->cCompressed = compressedSize; memcpy(pEntry->bits, pCompressed, compressedSize);
//
// Now cache the data
//
if (CH_SearchAndCacheData(pCacheHdr->handle, (LPBYTE)pEntry, sizeof(BMC_DIB_ENTRY) + compressedSize - 1, 0, &cacheIndex)) { //
// The sub-bitmap is already in the cache
//
*pCacheIndex = cacheIndex; TRACE_OUT(( "Bitmap already cached %u:%u cx(%d) cy(%d)", *pCache, *pCacheIndex, bitmapWidth, bitmapHeight)); *pIsNewEntry = FALSE;
//
// Free up the entry we just created
//
pEntry->inUse = FALSE; } else { *pCacheIndex = cacheIndex; TRACE_OUT(( "Cache entry at 0x%08x now in use", pEntry)); TRACE_OUT(( "New cache entry %u:%u cx(%d) cy(%d)", *pCache, *pCacheIndex, bitmapWidth, bitmapHeight)); *pIsNewEntry = TRUE; pEntry->iCacheIndex = (TSHR_UINT16)*pCacheIndex; }
//
// We've got the bits into the cache. If the cache attempt added a
// cache entry we must fill in the bitmap cache order.
//
if (*pIsNewEntry) { //
// Fill in the order details.
//
// Remember that we have to fill in the order size into the
// INT_ORDER_HEADER as well as filling in the bitmap bits order
// header. When doing this, adjust for the number of bitmap bits
// which are included in the bitmap bits order header.
//
pOrder->OrderHeader.Common.fOrderFlags = OF_PRIVATE;
if (compressed) { pBmcData->bmcPacketType = BMC_PT_BITMAP_BITS_COMPRESSED; } else { pBmcData->bmcPacketType = BMC_PT_BITMAP_BITS_UNCOMPRESSED;
//
// The data is not compressed, so copy the uncompressed data
// into the order. In the case where we compressed the data
// successfully, we did so directly into the order, so the
// compressed bits are already there.
//
memcpy(pDestBits, pDIBits, compressedSize); }
pBmcData->cacheID = (BYTE)*pCache; pBmcData->cxSubBitmapWidth = (TSHR_UINT8)fixedBitmapWidth; pBmcData->cySubBitmapHeight = (TSHR_UINT8)bitmapHeight; pBmcData->bpp = (TSHR_UINT8)m_usrSendingBPP; pBmcData->cbBitmapBits = (TSHR_UINT16)compressedSize;
//
// The iCacheEntryR1 field is unused for R2 - we use
// iCacheEntryR2 instead.
//
pBmcData->iCacheEntryR1 = 0; pBitsOrderR2->iCacheEntryR2 = (TSHR_UINT16)*pCacheIndex;
pOrder->OrderHeader.Common.cbOrderDataLength = (compressedSize + sizeof(BMC_BITMAP_BITS_ORDER_R2) - sizeof(pBitsOrderR2->data)); }
cachedOK = TRUE;
DC_EXIT_POINT: DebugExitBOOL(ASHost::SBCCacheBits, cachedOK); return(cachedOK); }
//
//
// Name: SBCAddToFastPath
//
// Purpose: Add a bitmap to the fast path
//
// Returns: Nothing
//
// Params: IN majorInfo - The major caching info passed up from
// the driver (the bitmap ID)
// IN minorInfo - The minor caching info passed up from
// the driver (the bitmap revision number)
// IN majorPalette - The major palette info passed up from
// the driver (the XLATEOBJ)
// IN minorPalette - The minor palette info passed up from
// the driver (the XLATEOBJ iUniq)
// IN srcX - The x coord of the source of the Blt
// IN srcY - The y coord of the source of the Blt
// IN width - The width of the area being Blted
// IN height - The height of the area being Blted
// IN cache - The cache the bits were placed in
// IN cacheIndex - The index at which the bits were placed
// in the cache
// IN colorCacheIndex - The index in the color table cache of
// the color table associated with the bits
//
//
void ASHost::SBCAddToFastPath ( UINT_PTR majorInfo, UINT minorInfo, UINT_PTR majorPalette, UINT minorPalette, int srcX, int srcY, UINT width, UINT height, UINT cache, UINT cacheIndex, UINT colorCacheIndex ) { LPSBC_FASTPATH_ENTRY pEntry;
DebugEntry(ASHost::SBCAddToFastPath);
//
// First get a free entry
//
pEntry = (LPSBC_FASTPATH_ENTRY)COM_BasedListFirst(&m_sbcFastPath->freeList, FIELD_OFFSET(SBC_FASTPATH_ENTRY, list)); if (pEntry == NULL) { //
// There are no entries in the free list, so we have to use the
// oldest entry in the used list. The used list is stored in MRU
// order, so we just have to get the last item in the list.
//
pEntry = (LPSBC_FASTPATH_ENTRY)COM_BasedListLast(&m_sbcFastPath->usedList, FIELD_OFFSET(SBC_FASTPATH_ENTRY, list)); TRACE_OUT(( "Evicting fast path info for %x %x (%d, %d)", pEntry->majorInfo, pEntry->minorInfo, pEntry->srcX, pEntry->srcY)); }
//
// Remove the entry from its current list
//
COM_BasedListRemove(&pEntry->list);
//
// Now fill in the details
//
pEntry->majorInfo = majorInfo; pEntry->minorInfo = minorInfo; pEntry->majorPalette = majorPalette; pEntry->minorPalette = minorPalette; pEntry->srcX = srcX; pEntry->srcY = srcY; pEntry->width = width; pEntry->height = height; pEntry->cache = (WORD)cache; pEntry->cacheIndex = (WORD)cacheIndex; pEntry->colorIndex = (WORD)colorCacheIndex;
//
// Finally, add the entry to the front of the used list
//
TRACE_OUT(( "Adding fast path info for %x %x (%d, %d)", pEntry->majorInfo, pEntry->minorInfo, pEntry->srcX, pEntry->srcY)); COM_BasedListInsertAfter(&m_sbcFastPath->usedList, &pEntry->list);
DebugExitVOID(ASHost::SBCAddToFastPath); }
//
//
// Name: SBCFindInFastPath
//
// Purpose: Check to see if a bitmap with the given attributes is in the
// SBC fast path. If so, return the cache info for the bitmap.
//
// Returns: TRUE if the bitmap is in the fast path, FALSE if not.
//
// Params: IN majorInfo - The major caching info passed up from
// the driver (the bitmap ID)
// IN minorInfo - The minor caching info passed up from
// the driver (the bitmap revision
// number)
// IN majorPalette - The major palette info passed up from
// the driver (the XLATEOBJ)
// IN minorPalette - The minor palette info passed up from
// the driver (the XLATEOBJ iUniq)
// IN srcX - The x coord of the source of the Blt
// IN srcY - The y coord of the source of the Blt
// IN width - The width of the area being Blted
// IN height - The height of the area being Blted
// OUT pCache - The cache the bits were placed in
// OUT pCacheIndex - The index at which the bits were
// placed in the cache
// OUT pColorCacheIndex - The index in the color table cache of
// the color table associated with the
// bits
//
// Operation: The contents of pCache, pCacheIndex and pColorCacheIndex
// are only valid on return if the function returns TRUE.
//
//
BOOL ASHost::SBCFindInFastPath ( UINT_PTR majorInfo, UINT minorInfo, UINT_PTR majorPalette, UINT minorPalette, int srcX, int srcY, UINT width, UINT height, UINT * pCache, UINT * pCacheIndex, UINT * pColorCacheIndex ) { BOOL found = FALSE; LPSBC_FASTPATH_ENTRY pEntry; LPSBC_FASTPATH_ENTRY pNextEntry;
DebugEntry(ASHost::SBCFindInFastPath);
//
// Traverse the in use list looking for a match on the parameters
// passed in.
//
pEntry = (LPSBC_FASTPATH_ENTRY)COM_BasedListFirst(&m_sbcFastPath->usedList, FIELD_OFFSET(SBC_FASTPATH_ENTRY, list)); while (pEntry != NULL) { if ((pEntry->majorInfo == majorInfo) && (pEntry->minorInfo == minorInfo) && (pEntry->majorPalette == majorPalette) && (pEntry->minorPalette == minorPalette) && (pEntry->srcX == srcX) && (pEntry->srcY == srcY) && (pEntry->width == width) && (pEntry->height == height)) { //
// We've found a match - hurrah ! Fill in the return info.
//
TRACE_OUT(( "Hit for %x %x (%d, %d) cache %d", pEntry->majorInfo, pEntry->minorInfo, pEntry->srcX, pEntry->srcY, pEntry->cache, pEntry->cacheIndex));
found = TRUE; *pCache = pEntry->cache; *pCacheIndex = pEntry->cacheIndex; *pColorCacheIndex = pEntry->colorIndex;
//
// We order the used list in MRU order, so remove the entry
// from its current position and add it at the head of the used
// list.
//
COM_BasedListRemove(&pEntry->list); COM_BasedListInsertAfter(&m_sbcFastPath->usedList, &pEntry->list);
//
// Got a match, so we can break out of the while loop
//
break; } else if ((pEntry->majorInfo == majorInfo) && (pEntry->minorInfo != minorInfo)) { //
// We have been given a bitmap which we have seen before, but
// the revision number has changed i.e. the bitmap has been
// updated (majorInfo identifies the bitmap, and minorInfo
// identifies the revision number of that bitmap - it is
// incremented every time the bitmap is changed).
//
// We have to remove all entries from the used list which
// reference this bitmap. We can start from the current
// position since we know that we can't have an entry for this
// bitmap earlier in the list, but we have to be careful to get
// the next entry in the list before removing an entry.
//
TRACE_OUT(( "Bitmap %x updated - removing references", pEntry->majorInfo)); pNextEntry = pEntry;
while (pNextEntry != NULL) { pEntry = pNextEntry;
pNextEntry = (LPSBC_FASTPATH_ENTRY)COM_BasedListNext(&m_sbcFastPath->usedList, pNextEntry, FIELD_OFFSET(SBC_FASTPATH_ENTRY, list));
if (pEntry->majorInfo == majorInfo) { COM_BasedListRemove(&pEntry->list); COM_BasedListInsertAfter(&m_sbcFastPath->freeList, &pEntry->list); } }
//
// We know we wont find a match, so we can break out of the
// while loop
//
break; }
pEntry = (LPSBC_FASTPATH_ENTRY)COM_BasedListNext(&m_sbcFastPath->usedList, pEntry, FIELD_OFFSET(SBC_FASTPATH_ENTRY, list)); }
DebugExitBOOL(ASShare::SBCFindInFastPath, found); return(found); }
//
// SBC_CacheEntryRemoved()
//
void ASHost::SBC_CacheEntryRemoved ( UINT cache, UINT cacheIndex ) { LPSBC_FASTPATH_ENTRY pEntry; LPSBC_FASTPATH_ENTRY pNextEntry;
DebugEntry(ASHost::SBC_CacheEntryRemoved);
ASSERT(m_sbcFastPath);
//
// An entry has been removed from the cache. If we have this entry in
// our fast path, we have to remove it.
//
// Just traverse the used list looking for an entry with matching cache
// and cacheIndex. Note that there may be more than one entry - if the
// source bitmap has a repeating image, we will get a match on the bits
// when we cache different areas of the bitmap.
//
pNextEntry = (LPSBC_FASTPATH_ENTRY)COM_BasedListFirst(&m_sbcFastPath->usedList, FIELD_OFFSET(SBC_FASTPATH_ENTRY, list)); while (pNextEntry != NULL) { pEntry = pNextEntry;
pNextEntry = (LPSBC_FASTPATH_ENTRY)COM_BasedListNext(&m_sbcFastPath->usedList, pNextEntry, FIELD_OFFSET(SBC_FASTPATH_ENTRY, list));
if ((pEntry->cache == cache) && (pEntry->cacheIndex == cacheIndex)) { //
// Move the entry to the free list
//
TRACE_OUT(("Fast path entry %x %x (%d, %d) evicted from cache", pEntry->majorInfo, pEntry->minorInfo, pEntry->srcX, pEntry->srcY)); COM_BasedListRemove(&pEntry->list); COM_BasedListInsertAfter(&m_sbcFastPath->freeList, &pEntry->list); } }
DebugExitVOID(ASHost::SBC_CacheEntryRemoved); }
|