671 lines
17 KiB
C
671 lines
17 KiB
C
/*++
|
||
|
||
Copyright (c) 1989 Microsoft Corporation
|
||
|
||
Module Name:
|
||
|
||
heapmgr.c
|
||
|
||
Abstract:
|
||
|
||
This module contains initialization and termination routines for
|
||
server FSP heap, as well as debug routines for memory tracking.
|
||
|
||
--*/
|
||
|
||
#include "precomp.h"
|
||
#include "heapmgr.tmh"
|
||
#pragma hdrstop
|
||
|
||
// Make the retry time 15 milli-seconds
|
||
#define SRV_LOW_PRIORITY_RETRY_TIME -1*1000*10*15
|
||
|
||
#ifdef POOL_TAGGING
|
||
//
|
||
// Array correlating BlockType numbers to pool tags.
|
||
//
|
||
// *** This array must be maintained in concert with the BlockType
|
||
// definitions in srvblock.h!
|
||
//
|
||
|
||
ULONG SrvPoolTags[BlockTypeMax-1] = {
|
||
'fbSL', // BlockTypeBuffer
|
||
'ncSL', // BlockTypeConnection
|
||
'peSL', // BlockTypeEndpoint
|
||
'flSL', // BlockTypeLfcb
|
||
'fmSL', // BlockTypeMfcb
|
||
'frSL', // BlockTypeRfcb
|
||
'rsSL', // BlockTypeSearch
|
||
'csSL', // BlockTypeSearchCore
|
||
'lbSL', // BlockTypeByteRangeLock for persistent handles
|
||
'ssSL', // BlockTypeSession
|
||
'hsSL', // BlockTypeShare
|
||
'rtSL', // BlockTypeTransaction
|
||
'ctSL', // BlockTypeTreeConnect
|
||
'poSL', // BlockTypeOplockBreak
|
||
'dcSL', // BlockTypeCommDevice
|
||
'iwSL', // BlockTypeWorkContextInitial
|
||
'nwSL', // BlockTypeWorkContextNormal
|
||
'rwSL', // BlockTypeWorkContextRaw
|
||
'swSL', // BlockTypeWorkContextSpecial
|
||
'dcSL', // BlockTypeCachedDirectory
|
||
'bdSL', // BlockTypeDataBuffer
|
||
'btSL', // BlockTypeTable
|
||
'hnSL', // BlockTypeNonpagedHeader
|
||
'cpSL', // BlockTypePagedConnection
|
||
'rpSL', // BlockTypePagedRfcb
|
||
'mpSL', // BlockTypePagedMfcb
|
||
'itSL', // BlockTypeTimer
|
||
'caSL', // BlockTypeAdminCheck
|
||
'qwSL', // BlockTypeWorkQueue
|
||
'fsSL', // BlockTypeDfs
|
||
'rlSL', // BlockTypeLargeReadX
|
||
'saSL', // BlockTypeAdapterStatus
|
||
'rsSL', // BlockTypeShareRemark
|
||
'dsSL', // BlockTypeShareSecurityDescriptor
|
||
'ivSL', // BlockTypeVolumeInformation
|
||
'nfSL', // BlockTypeFSName
|
||
'inSL', // BlockTypeNameInfo
|
||
'idSL', // BlockTypeDirectoryInfo
|
||
'cdSL', // BlockTypeDirCache
|
||
'imSL', // BlockTypeMisc
|
||
'nsSL', // BlockTypeSnapShot
|
||
'esSL', // BlockTypeSecurityContext
|
||
};
|
||
|
||
#endif // def POOL_TAGGING
|
||
|
||
#ifdef ALLOC_PRAGMA
|
||
#pragma alloc_text( PAGE, SrvAllocatePagedPool )
|
||
#pragma alloc_text( PAGE, SrvFreePagedPool )
|
||
#pragma alloc_text( PAGE, SrvClearLookAsideList )
|
||
#endif
|
||
#if 0
|
||
NOT PAGEABLE -- SrvAllocateNonPagedPool
|
||
NOT PAGEABLE -- SrvFreeNonPagedPool
|
||
#endif
|
||
|
||
extern LONG SrvMemoryAllocationRetries;
|
||
extern LONG SrvMemoryAllocationRetriesSuccessful;
|
||
|
||
|
||
PVOID SRVFASTCALL
|
||
SrvInterlockedAllocate( PLOOK_ASIDE_LIST l, ULONG NumberOfBytes, PLONG statistics )
|
||
{
|
||
PPOOL_HEADER newPool;
|
||
PPOOL_HEADER *pentry = NumberOfBytes > LOOK_ASIDE_SWITCHOVER ?
|
||
l->LargeFreeList : l->SmallFreeList;
|
||
|
||
PPOOL_HEADER *pend = pentry + LOOK_ASIDE_MAX_ELEMENTS;
|
||
|
||
do {
|
||
//
|
||
// Exchange with the lookaside spot and see if we get anything
|
||
//
|
||
|
||
newPool = NULL;
|
||
newPool = (PPOOL_HEADER)InterlockedExchangePointer( pentry, newPool );
|
||
|
||
if( newPool == NULL ) {
|
||
continue;
|
||
}
|
||
|
||
if( newPool->RequestedSize >= NumberOfBytes ) {
|
||
//
|
||
// The one we got is big enough! Return it.
|
||
//
|
||
++(l->AllocHit);
|
||
return newPool + 1;
|
||
}
|
||
|
||
//
|
||
// It wasn't big enough, so put it back.
|
||
//
|
||
newPool = (PPOOL_HEADER)InterlockedExchangePointer( pentry, newPool );
|
||
if( newPool == NULL ) {
|
||
continue;
|
||
}
|
||
|
||
//
|
||
// Oops, somebody else freed some memory to this spot. Can we use it?
|
||
//
|
||
if( newPool->RequestedSize >= NumberOfBytes ) {
|
||
//
|
||
// We can use it!
|
||
//
|
||
++(l->AllocHit);
|
||
return newPool + 1;
|
||
}
|
||
|
||
//
|
||
// Can't use the memory -- so really free it and keep looking
|
||
//
|
||
if( statistics ) {
|
||
InterlockedExchangeAdd(
|
||
statistics,
|
||
-(LONG)newPool->RequestedSize
|
||
);
|
||
}
|
||
|
||
ExFreePool( newPool );
|
||
|
||
} while( ++pentry < pend );
|
||
|
||
++(l->AllocMiss);
|
||
return NULL;
|
||
}
|
||
|
||
PPOOL_HEADER SRVFASTCALL
|
||
SrvInterlockedFree( PPOOL_HEADER block )
|
||
{
|
||
PPOOL_HEADER *pentry = block->FreeList;
|
||
PPOOL_HEADER *pend = pentry + LOOK_ASIDE_MAX_ELEMENTS;
|
||
|
||
do {
|
||
|
||
block = (PPOOL_HEADER)InterlockedExchangePointer( pentry, block );
|
||
|
||
} while( block != NULL && ++pentry < pend );
|
||
|
||
return block;
|
||
}
|
||
|
||
VOID SRVFASTCALL
|
||
SrvClearLookAsideList( PLOOK_ASIDE_LIST l, VOID (SRVFASTCALL *FreeRoutine )( PVOID ) )
|
||
{
|
||
PPOOL_HEADER *pentry, *pend, block;
|
||
|
||
PAGED_CODE();
|
||
|
||
//
|
||
// Clear out the list of large chunks
|
||
//
|
||
pentry = l->LargeFreeList;
|
||
pend = pentry + LOOK_ASIDE_MAX_ELEMENTS;
|
||
|
||
do {
|
||
block = NULL;
|
||
block = (PPOOL_HEADER)InterlockedExchangePointer( pentry, block );
|
||
|
||
if( block != NULL ) {
|
||
block->FreeList = NULL;
|
||
FreeRoutine( block + 1 );
|
||
}
|
||
|
||
} while( ++pentry < pend );
|
||
|
||
//
|
||
// Clear out the list of small chunks
|
||
//
|
||
pentry = l->SmallFreeList;
|
||
pend = pentry + LOOK_ASIDE_MAX_ELEMENTS;
|
||
|
||
do {
|
||
block = NULL;
|
||
block = (PPOOL_HEADER)InterlockedExchangePointer( pentry, block );
|
||
|
||
if( block != NULL ) {
|
||
block->FreeList = NULL;
|
||
FreeRoutine( block + 1 );
|
||
}
|
||
|
||
} while( ++pentry < pend );
|
||
}
|
||
|
||
|
||
PVOID SRVFASTCALL
|
||
SrvAllocateNonPagedPool (
|
||
IN CLONG NumberOfBytes
|
||
#ifdef POOL_TAGGING
|
||
, IN CLONG BlockType
|
||
#endif
|
||
)
|
||
|
||
/*++
|
||
|
||
Routine Description:
|
||
|
||
This routine allocates nonpaged pool in the server. A check is
|
||
made to ensure that the server's total nonpaged pool usage is below
|
||
the configurable limit.
|
||
|
||
Arguments:
|
||
|
||
NumberOfBytes - the number of bytes to allocate.
|
||
|
||
BlockType - the type of block (used to pass pool tag to allocator)
|
||
|
||
Return Value:
|
||
|
||
PVOID - a pointer to the allocated memory or NULL if the memory could
|
||
not be allocated.
|
||
|
||
--*/
|
||
|
||
{
|
||
PPOOL_HEADER newPool;
|
||
PPOOL_HEADER *FreeList = NULL;
|
||
ULONG newUsage;
|
||
BOOLEAN IsLowPriority = FALSE;
|
||
LARGE_INTEGER interval;
|
||
|
||
#ifdef POOL_TAGGING
|
||
ASSERT( BlockType > 0 && BlockType < BlockTypeMax );
|
||
#endif
|
||
|
||
//
|
||
// Pull this allocation off the per-processor free list if we can
|
||
//
|
||
if( SrvWorkQueues ) {
|
||
|
||
PWORK_QUEUE queue = PROCESSOR_TO_QUEUE();
|
||
|
||
if( NumberOfBytes <= queue->NonPagedPoolLookAsideList.MaxSize ) {
|
||
|
||
newPool = SrvInterlockedAllocate(
|
||
&queue->NonPagedPoolLookAsideList,
|
||
NumberOfBytes,
|
||
(PLONG)&SrvStatistics.CurrentNonPagedPoolUsage
|
||
);
|
||
|
||
if( newPool != NULL ) {
|
||
return newPool;
|
||
}
|
||
|
||
FreeList = NumberOfBytes > LOOK_ASIDE_SWITCHOVER ?
|
||
queue->NonPagedPoolLookAsideList.LargeFreeList :
|
||
queue->NonPagedPoolLookAsideList.SmallFreeList ;
|
||
}
|
||
}
|
||
|
||
//
|
||
// Account for this allocation in the statistics database and make
|
||
// sure that this allocation will not put us over the limit of
|
||
// nonpaged pool that we can allocate.
|
||
//
|
||
|
||
newUsage = InterlockedExchangeAdd( (PLONG)&SrvStatistics.CurrentNonPagedPoolUsage,
|
||
(LONG)NumberOfBytes );
|
||
newUsage += NumberOfBytes;
|
||
|
||
if ( newUsage > SrvMaxNonPagedPoolUsage ) {
|
||
|
||
//
|
||
// Count the failure, but do NOT log an event. The scavenger
|
||
// will log an event when it next wakes up. This keeps us from
|
||
// flooding the event log.
|
||
//
|
||
|
||
SrvNonPagedPoolLimitHitCount++;
|
||
SrvStatistics.NonPagedPoolFailures++;
|
||
|
||
InterlockedExchangeAdd( (PLONG)&SrvStatistics.CurrentNonPagedPoolUsage,
|
||
-(LONG)NumberOfBytes );
|
||
|
||
return NULL;
|
||
|
||
}
|
||
|
||
if (SrvStatistics.CurrentNonPagedPoolUsage > SrvStatistics.PeakNonPagedPoolUsage) {
|
||
SrvStatistics.PeakNonPagedPoolUsage = SrvStatistics.CurrentNonPagedPoolUsage;
|
||
}
|
||
|
||
//
|
||
// Do the actual memory allocation. Allocate extra space so that we
|
||
// can store the size of the allocation for the free routine.
|
||
//
|
||
if( NumberOfBytes > 2*4096 )
|
||
{
|
||
IsLowPriority = TRUE;
|
||
}
|
||
|
||
newPool = ExAllocatePoolWithTagPriority(
|
||
NonPagedPool,
|
||
NumberOfBytes + sizeof(POOL_HEADER),
|
||
TAG_FROM_TYPE(BlockType),
|
||
IsLowPriority ? LowPoolPriority : NormalPoolPriority
|
||
);
|
||
|
||
if( (newPool == NULL) && IsLowPriority && (KeGetCurrentIrql() <= APC_LEVEL) )
|
||
{
|
||
interval.QuadPart = SRV_LOW_PRIORITY_RETRY_TIME;
|
||
InterlockedIncrement( &SrvMemoryAllocationRetries );
|
||
|
||
// Wait and try again
|
||
KeDelayExecutionThread( KernelMode, FALSE, &interval );
|
||
|
||
newPool = ExAllocatePoolWithTagPriority(
|
||
NonPagedPool,
|
||
NumberOfBytes + sizeof(POOL_HEADER),
|
||
TAG_FROM_TYPE(BlockType),
|
||
LowPoolPriority
|
||
);
|
||
|
||
if( newPool )
|
||
{
|
||
InterlockedIncrement( &SrvMemoryAllocationRetriesSuccessful );
|
||
}
|
||
}
|
||
|
||
|
||
//
|
||
// If the system couldn't satisfy the request, return NULL.
|
||
//
|
||
|
||
if ( newPool != NULL ) {
|
||
//
|
||
// Save the size of this block in the extra space we allocated.
|
||
//
|
||
|
||
newPool->RequestedSize = NumberOfBytes;
|
||
newPool->FreeList = FreeList;
|
||
|
||
//
|
||
// Return a pointer to the memory after the size longword.
|
||
//
|
||
|
||
return (PVOID)( newPool + 1 );
|
||
}
|
||
|
||
//
|
||
// Count the failure, but do NOT log an event. The scavenger
|
||
// will log an event when it next wakes up. This keeps us from
|
||
// flooding the event log.
|
||
//
|
||
|
||
SrvStatistics.NonPagedPoolFailures++;
|
||
|
||
|
||
InterlockedExchangeAdd( (PLONG)&SrvStatistics.CurrentNonPagedPoolUsage,
|
||
-(LONG)NumberOfBytes );
|
||
|
||
return NULL;
|
||
|
||
} // SrvAllocateNonPagedPool
|
||
|
||
VOID SRVFASTCALL
|
||
SrvFreeNonPagedPool (
|
||
IN PVOID Address
|
||
)
|
||
|
||
/*++
|
||
|
||
Routine Description:
|
||
|
||
Frees the memory allocated by a call to SrvAllocateNonPagedPool.
|
||
The statistics database is updated to reflect the current nonpaged
|
||
pool usage.
|
||
|
||
Arguments:
|
||
|
||
Address - the address of allocated memory returned by
|
||
SrvAllocateNonPagedPool.
|
||
|
||
Return Value:
|
||
|
||
None.
|
||
|
||
--*/
|
||
|
||
{
|
||
PPOOL_HEADER actualBlock = (PPOOL_HEADER)Address - 1;
|
||
|
||
//
|
||
// See if we can stash this bit of memory away in the NonPagedPoolFreeList
|
||
//
|
||
if( actualBlock->FreeList ) {
|
||
|
||
actualBlock = SrvInterlockedFree( actualBlock );
|
||
}
|
||
|
||
if( actualBlock != NULL ) {
|
||
|
||
//
|
||
// Update the nonpaged pool usage statistic.
|
||
//
|
||
InterlockedExchangeAdd(
|
||
(PLONG)&SrvStatistics.CurrentNonPagedPoolUsage,
|
||
-(LONG)actualBlock->RequestedSize
|
||
);
|
||
|
||
//
|
||
// Free the pool and return.
|
||
//
|
||
|
||
ExFreePool( actualBlock );
|
||
}
|
||
|
||
} // SrvFreeNonPagedPool
|
||
|
||
|
||
PVOID SRVFASTCALL
|
||
SrvAllocatePagedPool (
|
||
IN POOL_TYPE PoolType,
|
||
IN CLONG NumberOfBytes
|
||
#ifdef POOL_TAGGING
|
||
, IN CLONG BlockType
|
||
#endif
|
||
)
|
||
|
||
/*++
|
||
|
||
Routine Description:
|
||
|
||
This routine allocates Paged pool in the server. A check is
|
||
made to ensure that the server's total Paged pool usage is below
|
||
the configurable limit.
|
||
|
||
Arguments:
|
||
|
||
NumberOfBytes - the number of bytes to allocate.
|
||
|
||
BlockType - the type of block (used to pass pool tag to allocator)
|
||
|
||
Return Value:
|
||
|
||
PVOID - a pointer to the allocated memory or NULL if the memory could
|
||
not be allocated.
|
||
|
||
--*/
|
||
|
||
{
|
||
PPOOL_HEADER newPool;
|
||
PPOOL_HEADER *FreeList = NULL;
|
||
ULONG newUsage;
|
||
BOOLEAN IsLowPriority = FALSE;
|
||
LARGE_INTEGER interval;
|
||
|
||
PAGED_CODE();
|
||
|
||
#ifdef POOL_TAGGING
|
||
ASSERT( BlockType > 0 && BlockType < BlockTypeMax );
|
||
#endif
|
||
|
||
//
|
||
// Pull this allocation off the per-processor free list if we can
|
||
//
|
||
if( SrvWorkQueues ) {
|
||
|
||
PWORK_QUEUE queue = PROCESSOR_TO_QUEUE();
|
||
|
||
if( NumberOfBytes <= queue->PagedPoolLookAsideList.MaxSize ) {
|
||
|
||
newPool = SrvInterlockedAllocate(
|
||
&queue->PagedPoolLookAsideList,
|
||
NumberOfBytes,
|
||
(PLONG)&SrvStatistics.CurrentPagedPoolUsage
|
||
);
|
||
|
||
if( newPool != NULL ) {
|
||
return newPool;
|
||
}
|
||
|
||
FreeList = NumberOfBytes > LOOK_ASIDE_SWITCHOVER ?
|
||
queue->PagedPoolLookAsideList.LargeFreeList :
|
||
queue->PagedPoolLookAsideList.SmallFreeList ;
|
||
}
|
||
}
|
||
|
||
//
|
||
// Account for this allocation in the statistics database and make
|
||
// sure that this allocation will not put us over the limit of
|
||
// nonpaged pool that we can allocate.
|
||
//
|
||
|
||
|
||
newUsage = InterlockedExchangeAdd( (PLONG)&SrvStatistics.CurrentPagedPoolUsage,
|
||
(LONG)NumberOfBytes );
|
||
newUsage += NumberOfBytes;
|
||
|
||
if ( newUsage > SrvMaxPagedPoolUsage ) {
|
||
|
||
//
|
||
// Count the failure, but do NOT log an event. The scavenger
|
||
// will log an event when it next wakes up. This keeps us from
|
||
// flooding the event log.
|
||
//
|
||
|
||
SrvPagedPoolLimitHitCount++;
|
||
SrvStatistics.PagedPoolFailures++;
|
||
|
||
InterlockedExchangeAdd( (PLONG)&SrvStatistics.CurrentPagedPoolUsage,
|
||
-(LONG)NumberOfBytes );
|
||
|
||
return NULL;
|
||
}
|
||
|
||
if (SrvStatistics.CurrentPagedPoolUsage > SrvStatistics.PeakPagedPoolUsage ) {
|
||
SrvStatistics.PeakPagedPoolUsage = SrvStatistics.CurrentPagedPoolUsage;
|
||
}
|
||
|
||
//
|
||
// Do the actual memory allocation. Allocate extra space so that we
|
||
// can store the size of the allocation for the free routine.
|
||
//
|
||
if( NumberOfBytes > 2*4096 )
|
||
{
|
||
IsLowPriority = TRUE;
|
||
}
|
||
|
||
newPool = ExAllocatePoolWithTagPriority(
|
||
PoolType,
|
||
NumberOfBytes + sizeof(POOL_HEADER),
|
||
TAG_FROM_TYPE(BlockType),
|
||
IsLowPriority ? LowPoolPriority : NormalPoolPriority
|
||
);
|
||
|
||
if( (newPool == NULL) && IsLowPriority && (KeGetCurrentIrql() <= APC_LEVEL) )
|
||
{
|
||
interval.QuadPart = SRV_LOW_PRIORITY_RETRY_TIME;
|
||
InterlockedIncrement( &SrvMemoryAllocationRetries );
|
||
|
||
// Wait and try again
|
||
KeDelayExecutionThread( KernelMode, FALSE, &interval );
|
||
|
||
newPool = ExAllocatePoolWithTagPriority(
|
||
PoolType,
|
||
NumberOfBytes + sizeof(POOL_HEADER),
|
||
TAG_FROM_TYPE(BlockType),
|
||
LowPoolPriority
|
||
);
|
||
|
||
if( newPool )
|
||
{
|
||
InterlockedIncrement( &SrvMemoryAllocationRetriesSuccessful );
|
||
}
|
||
}
|
||
|
||
if( newPool != NULL ) {
|
||
|
||
newPool->FreeList = FreeList;
|
||
newPool->RequestedSize = NumberOfBytes;
|
||
|
||
//
|
||
// Return a pointer to the memory after the POOL_HEADER
|
||
//
|
||
|
||
return newPool + 1;
|
||
}
|
||
|
||
//
|
||
// If the system couldn't satisfy the request, return NULL.
|
||
//
|
||
// Count the failure, but do NOT log an event. The scavenger
|
||
// will log an event when it next wakes up. This keeps us from
|
||
// flooding the event log.
|
||
//
|
||
|
||
SrvStatistics.PagedPoolFailures++;
|
||
|
||
InterlockedExchangeAdd( (PLONG)&SrvStatistics.CurrentPagedPoolUsage,
|
||
-(LONG)NumberOfBytes );
|
||
|
||
|
||
return NULL;
|
||
|
||
|
||
} // SrvAllocatePagedPool
|
||
|
||
VOID SRVFASTCALL
|
||
SrvFreePagedPool (
|
||
IN PVOID Address
|
||
)
|
||
|
||
/*++
|
||
|
||
Routine Description:
|
||
|
||
Frees the memory allocated by a call to SrvAllocatePagedPool.
|
||
The statistics database is updated to reflect the current Paged
|
||
pool usage. If this routine is change, look at scavengr.c
|
||
|
||
Arguments:
|
||
|
||
Address - the address of allocated memory returned by
|
||
SrvAllocatePagedPool.
|
||
|
||
Return Value:
|
||
|
||
None.
|
||
|
||
--*/
|
||
|
||
{
|
||
PPOOL_HEADER actualBlock = (PPOOL_HEADER)Address - 1;
|
||
|
||
PAGED_CODE();
|
||
|
||
ASSERT( actualBlock != NULL );
|
||
|
||
//
|
||
// See if we can stash this bit of memory away in the PagedPoolFreeList
|
||
//
|
||
if( actualBlock->FreeList ) {
|
||
|
||
actualBlock = SrvInterlockedFree( actualBlock );
|
||
}
|
||
|
||
if( actualBlock != NULL ) {
|
||
|
||
//
|
||
// Update the Paged pool usage statistic.
|
||
//
|
||
|
||
ASSERT( SrvStatistics.CurrentPagedPoolUsage >= actualBlock->RequestedSize );
|
||
|
||
InterlockedExchangeAdd(
|
||
(PLONG)&SrvStatistics.CurrentPagedPoolUsage,
|
||
-(LONG)actualBlock->RequestedSize
|
||
);
|
||
|
||
ASSERT( (LONG)SrvStatistics.CurrentPagedPoolUsage >= 0 );
|
||
|
||
//
|
||
// Free the pool and return.
|
||
//
|
||
|
||
ExFreePool( actualBlock );
|
||
}
|
||
|
||
} // SrvFreePagedPool
|