Source code of Windows XP (NT5)
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
 
 
 
 
 
 

2815 lines
73 KiB

/*++
Copyright (c) 1990 Microsoft Corporation
Module Name:
wsmanage.c
Abstract:
This module contains routines which manage the set of active working
set lists.
Working set management is accomplished by a parallel group of actions
1. Writing modified pages.
2. Trimming working sets by :
a) Aging pages by turning off access bits and incrementing age
counts for pages which haven't been accessed.
b) Estimating the number of unused pages in a working set and
keeping a global count of that estimate.
c) When getting tight on memory, replacing rather than adding
pages in a working set when a fault occurs in a working set
that has a significant proportion of unused pages.
d) When memory is tight, reducing (trimming) working sets which
are above their maximum towards their minimum. This is done
especially if there are a large number of available pages
in it.
The metrics are set such that writing modified pages is typically
accomplished before trimming working sets, however, under certain cases
where modified pages are being generated at a very high rate, working
set trimming will be initiated to free up more pages.
When the first thread in a process is created, the memory management
system is notified that working set expansion is allowed. This
is noted by changing the FLINK field of the WorkingSetExpansionLink
entry in the process control block from MM_NO_WS_EXPANSION to
MM_ALLOW_WS_EXPANSION. As threads fault, the working set is eligible
for expansion if ample pages exist (MmAvailablePages is high enough).
Once a process has had its working set raised above the minimum
specified, the process is put on the Working Set Expanded list and
is now eligible for trimming. Note that at this time the FLINK field
in the WorkingSetExpansionLink has an address value.
When working set trimming is initiated, a process is removed from the
list (the expansion lock guards this list) and the FLINK field is set
to MM_NO_WS_EXPANSION, also, the BLINK field is set to
MM_WS_EXPANSION_IN_PROGRESS. The BLINK field value indicates to
the MmCleanUserAddressSpace function that working set trimming is
in progress for this process and it should wait until it completes.
This is accomplished by creating an event, putting the address of the
event in the BLINK field and then releasing the expansion lock and
waiting on the event atomically. When working set trimming is
complete, the BLINK field is no longer MM_EXPANSION_IN_PROGRESS
indicating that the event should be set.
Author:
Lou Perazzoli (loup) 10-Apr-1990
Landy Wang (landyw) 02-Jun-1997
Revision History:
--*/
#include "mi.h"
#ifdef ALLOC_PRAGMA
#pragma alloc_text(INIT, MiAdjustWorkingSetManagerParameters)
#pragma alloc_text(PAGE, MmIsMemoryAvailable)
#pragma alloc_text(PAGEVRFY, MmTrimSessionMemory)
#if !DBG
#pragma alloc_text(PAGEVRFY, MmTrimProcessMemory)
#endif
#endif
KEVENT MiWaitForEmptyEvent;
BOOLEAN MiWaitingForWorkingSetEmpty;
LOGICAL MiReplacing = FALSE;
extern ULONG MiStandbyRemoved;
PFN_NUMBER MmMoreThanEnoughFreePages = 1000;
#define MI_MAXIMUM_AGING_SHIFT 7
ULONG MiAgingShift = 4;
ULONG MiEstimationShift = 5;
PFN_NUMBER MmTotalClaim = 0;
PFN_NUMBER MmTotalEstimatedAvailable = 0;
LARGE_INTEGER MiLastAdjustmentOfClaimParams;
//
// Sixty seconds.
//
const LARGE_INTEGER MmClaimParameterAdjustUpTime = {60 * 1000 * 1000 * 10, 0};
//
// 2 seconds.
//
const LARGE_INTEGER MmClaimParameterAdjustDownTime = {2 * 1000 * 1000 * 10, 0};
ULONG MmPlentyFreePages = 400;
ULONG MmEnormousFreePages = 20000;
ULONG MmPlentyFreePagesValue;
LOGICAL MiHardTrim = FALSE;
WSLE_NUMBER MiMaximumWslesPerSweep = (1024 * 1024 * 1024) / PAGE_SIZE;
#define MI_MAXIMUM_SAMPLE 8192
#define MI_MINIMUM_SAMPLE 64
#define MI_MINIMUM_SAMPLE_SHIFT 7
#if DBG
PETHREAD MmWorkingSetThread;
#endif
//
// Number of times to retry when the target working set's mutex is not
// readily available.
//
ULONG MiWsRetryCount = 5;
typedef struct _MMWS_TRIM_CRITERIA {
ULONG NumPasses;
ULONG TrimAge;
PFN_NUMBER DesiredFreeGoal;
PFN_NUMBER NewTotalClaim;
PFN_NUMBER NewTotalEstimatedAvailable;
LOGICAL DoAging;
LOGICAL TrimAllPasses;
} MMWS_TRIM_CRITERIA, *PMMWS_TRIM_CRITERIA;
LOGICAL
MiCheckAndSetSystemTrimCriteria (
IN OUT PMMWS_TRIM_CRITERIA Criteria
);
LOGICAL
MiCheckSystemTrimEndCriteria (
IN OUT PMMWS_TRIM_CRITERIA Criteria,
IN KIRQL OldIrql
);
WSLE_NUMBER
MiDetermineWsTrimAmount (
IN PMMWS_TRIM_CRITERIA Criteria,
IN PMMSUPPORT VmSupport
);
VOID
MiAgePagesAndEstimateClaims (
LOGICAL EmptyIt
);
VOID
MiAdjustClaimParameters (
IN LOGICAL EnoughPages
);
VOID
MiRearrangeWorkingSetExpansionList (
VOID
);
VOID
MiAdjustWorkingSetManagerParameters (
IN LOGICAL WorkStation
)
/*++
Routine Description:
This function is called from MmInitSystem to adjust the working set manager
trim algorithms based on system type and size.
Arguments:
WorkStation - TRUE if this is a workstation, FALSE if not.
Return Value:
None.
Environment:
Kernel mode, INIT time only.
--*/
{
if (WorkStation && MmNumberOfPhysicalPages <= 257*1024*1024/PAGE_SIZE) {
MiAgingShift = 3;
MiEstimationShift = 4;
}
else {
MiAgingShift = 5;
MiEstimationShift = 6;
}
if (MmNumberOfPhysicalPages >= 63*1024*1024/PAGE_SIZE) {
MmPlentyFreePages *= 2;
}
MmPlentyFreePagesValue = MmPlentyFreePages;
MiWaitingForWorkingSetEmpty = FALSE;
KeInitializeEvent (&MiWaitForEmptyEvent, NotificationEvent, TRUE);
}
VOID
MiObtainFreePages (
VOID
)
/*++
Routine Description:
This function examines the size of the modified list and the
total number of pages in use because of working set increments
and obtains pages by writing modified pages and/or reducing
working sets.
Arguments:
None.
Return Value:
None.
Environment:
Kernel mode, APCs disabled, working set and PFN mutexes held.
--*/
{
//
// Check to see if there are enough modified pages to institute a
// write.
//
if (MmModifiedPageListHead.Total >= MmModifiedWriteClusterSize) {
//
// Start the modified page writer.
//
KeSetEvent (&MmModifiedPageWriterEvent, 0, FALSE);
}
//
// See if there are enough working sets above the minimum
// threshold to make working set trimming worthwhile.
//
if ((MmPagesAboveWsMinimum > MmPagesAboveWsThreshold) ||
(MmAvailablePages < 5)) {
//
// Start the working set manager to reduce working sets.
//
KeSetEvent (&MmWorkingSetManagerEvent, 0, FALSE);
}
}
LOGICAL
MmIsMemoryAvailable (
IN ULONG PagesDesired
)
/*++
Routine Description:
This function checks whether there are sufficient available pages based
on the caller's request. If currently active pages are needed to satisfy
this request and non-useful ones can be taken, then trimming is initiated
here to do so.
Arguments:
PagesRequested - Supplies the number of pages desired.
Return Value:
TRUE if sufficient pages exist to satisfy the request.
FALSE if not.
Environment:
Kernel mode, PASSIVE_LEVEL.
--*/
{
LOGICAL Status;
ULONG PageTarget;
ULONG PagePlentyTarget;
ULONG i;
ULONG CurrentAvailablePages;
PFN_NUMBER CurrentTotalClaim;
ASSERT (KeGetCurrentIrql () == PASSIVE_LEVEL);
CurrentAvailablePages = MmAvailablePages;
//
// If twice the pages that the caller asked for are available
// without trimming anything, return TRUE.
//
PageTarget = PagesDesired * 2;
if (CurrentAvailablePages >= PageTarget) {
return TRUE;
}
CurrentTotalClaim = MmTotalClaim;
//
// If there are few pages available or claimable, we adjust to do
// a hard trim.
//
if (CurrentAvailablePages + CurrentTotalClaim < PagesDesired) {
MiHardTrim = TRUE;
}
//
// Active pages must be trimmed to satisfy this request and it is believed
// that non-useful pages can be taken to accomplish this.
//
// Set the PagePlentyTarget to 125% of the readlist size and kick it off.
// Our actual trim goal will be 150% of the PagePlentyTarget.
//
PagePlentyTarget = PagesDesired + (PagesDesired >> 2);
MmPlentyFreePages = PagePlentyTarget;
KeSetEvent (&MmWorkingSetManagerEvent, 0, FALSE);
Status = FALSE;
for (i = 0; i < 10; i += 1) {
KeDelayExecutionThread (KernelMode, FALSE, (PLARGE_INTEGER)&Mm30Milliseconds);
if (MmAvailablePages >= PagesDesired) {
Status = TRUE;
break;
}
}
MmPlentyFreePages = MmPlentyFreePagesValue;
MiHardTrim = FALSE;
return Status;
}
LOGICAL
MiAttachAndLockWorkingSet (
IN PMMSUPPORT VmSupport,
IN PLOGICAL InformSessionOfRelease
)
/*++
Routine Description:
This function attaches to the proper address space and acquires the
relevant working set mutex for the address space being trimmed.
If successful, this routine returns with APCs blocked as well.
On failure, this routine returns without any APCs blocked, no working
set mutex acquired and no address space attached to.
Arguments:
VmSupport - Supplies the working set to attach to and lock.
Return Value:
TRUE on success.
FALSE if not.
Environment:
Kernel mode, PASSIVE_LEVEL.
--*/
{
ULONG count;
KIRQL OldIrql;
PEPROCESS ProcessToTrim;
LOGICAL Attached;
PMM_SESSION_SPACE SessionSpace;
ASSERT (KeGetCurrentIrql () == PASSIVE_LEVEL);
*InformSessionOfRelease = FALSE;
if (VmSupport == &MmSystemCacheWs) {
ASSERT (VmSupport->Flags.SessionSpace == 0);
ASSERT (VmSupport->Flags.TrimHard == 0);
//
// System cache,
//
KeRaiseIrql (APC_LEVEL, &OldIrql);
if (!ExTryToAcquireResourceExclusiveLite (&MmSystemWsLock)) {
//
// System working set mutex was not granted, don't trim
// the system cache.
//
KeLowerIrql (OldIrql);
return FALSE;
}
MmSystemLockOwner = PsGetCurrentThread ();
return TRUE;
}
if (VmSupport->Flags.SessionSpace == 0) {
ProcessToTrim = CONTAINING_RECORD (VmSupport, EPROCESS, Vm);
ASSERT ((ProcessToTrim->Flags & PS_PROCESS_FLAGS_VM_DELETED) == 0);
//
// Attach to the process in preparation for trimming.
//
Attached = 0;
if (ProcessToTrim != PsInitialSystemProcess) {
Attached = KeForceAttachProcess (&ProcessToTrim->Pcb);
if (Attached == 0) {
return FALSE;
}
if (ProcessToTrim->Flags & PS_PROCESS_FLAGS_OUTSWAP_ENABLED) {
ASSERT ((ProcessToTrim->Flags & PS_PROCESS_FLAGS_OUTSWAPPED) == 0);
if ((ProcessToTrim->Flags & PS_PROCESS_FLAGS_IN_SESSION ) &&
(VmSupport->Flags.SessionLeader == 0)) {
*InformSessionOfRelease = TRUE;
}
}
}
//
// Attempt to acquire the working set mutex. If the
// lock cannot be acquired, skip over this process.
//
count = 0;
do {
if (ExTryToAcquireFastMutex(&ProcessToTrim->WorkingSetLock) != FALSE) {
ASSERT (VmSupport->Flags.BeingTrimmed == 1);
return TRUE;
}
KeDelayExecutionThread (KernelMode, FALSE, (PLARGE_INTEGER)&MmShortTime);
count += 1;
} while (count < MiWsRetryCount);
//
// Could not get the lock, skip this process.
//
if (*InformSessionOfRelease == TRUE) {
LOCK_EXPANSION (OldIrql);
ASSERT (ProcessToTrim->Flags & PS_PROCESS_FLAGS_OUTSWAP_ENABLED);
PS_CLEAR_BITS (&ProcessToTrim->Flags, PS_PROCESS_FLAGS_OUTSWAP_ENABLED);
ASSERT (MmSessionSpace->ProcessOutSwapCount >= 1);
MmSessionSpace->ProcessOutSwapCount -= 1;
UNLOCK_EXPANSION (OldIrql);
*InformSessionOfRelease = FALSE;
}
if (Attached) {
KeDetachProcess ();
}
return FALSE;
}
SessionSpace = CONTAINING_RECORD (VmSupport, MM_SESSION_SPACE, Vm);
//
// Attach directly to the session space to be trimmed.
//
MiAttachSession (SessionSpace);
//
// Try for the session working set mutex.
//
KeRaiseIrql (APC_LEVEL, &OldIrql);
if (!ExTryToAcquireResourceExclusiveLite (&SessionSpace->WsLock)) {
//
// This session space's working set mutex was not
// granted, don't trim it.
//
KeLowerIrql (OldIrql);
MiDetachSession ();
return FALSE;
}
MM_SET_SESSION_RESOURCE_OWNER (PsGetCurrentThread ());
return TRUE;
}
VOID
MiDetachAndUnlockWorkingSet (
IN PMMSUPPORT VmSupport,
IN LOGICAL InformSessionOfRelease
)
/*++
Routine Description:
This function detaches from the target address space and releases the
relevant working set mutex for the address space that was trimmed.
Arguments:
VmSupport - Supplies the working set to detach from and unlock.
Return Value:
None.
Environment:
Kernel mode, APC_LEVEL.
--*/
{
KIRQL OldIrql;
PEPROCESS ProcessToTrim;
ASSERT (KeGetCurrentIrql () == APC_LEVEL);
if (VmSupport == &MmSystemCacheWs) {
ASSERT (VmSupport->Flags.SessionSpace == 0);
UNLOCK_SYSTEM_WS (PASSIVE_LEVEL);
return;
}
if (VmSupport->Flags.SessionSpace == 0) {
ProcessToTrim = CONTAINING_RECORD (VmSupport, EPROCESS, Vm);
UNLOCK_WS (ProcessToTrim);
ASSERT (KeGetCurrentIrql () == PASSIVE_LEVEL);
if (InformSessionOfRelease == TRUE) {
LOCK_EXPANSION (OldIrql);
ASSERT (ProcessToTrim->Flags & PS_PROCESS_FLAGS_OUTSWAP_ENABLED);
PS_CLEAR_BITS (&ProcessToTrim->Flags, PS_PROCESS_FLAGS_OUTSWAP_ENABLED);
ASSERT (MmSessionSpace->ProcessOutSwapCount >= 1);
MmSessionSpace->ProcessOutSwapCount -= 1;
UNLOCK_EXPANSION (OldIrql);
}
if (ProcessToTrim != PsInitialSystemProcess) {
KeDetachProcess ();
}
return;
}
UNLOCK_SESSION_SPACE_WS (PASSIVE_LEVEL);
MiDetachSession ();
return;
}
VOID
MmWorkingSetManager (
VOID
)
/*++
Routine Description:
Implements the NT working set manager thread. When the number
of free pages becomes critical and ample pages can be obtained by
reducing working sets, the working set manager's event is set, and
this thread becomes active.
Arguments:
None.
Return Value:
None.
Environment:
Kernel mode.
--*/
{
PLIST_ENTRY ListEntry;
WSLE_NUMBER Trim;
KIRQL OldIrql;
PMMSUPPORT VmSupport;
LARGE_INTEGER CurrentTime;
LOGICAL DoTrimming;
LOGICAL InformSessionOfRelease;
MMWS_TRIM_CRITERIA TrimCriteria;
static ULONG Initialized = 0;
PERFINFO_WSMANAGE_DECL();
if (Initialized == 0) {
PsGetCurrentThread()->MemoryMaker = 1;
Initialized = 1;
}
#if DBG
MmWorkingSetThread = PsGetCurrentThread ();
#endif
ASSERT (MmIsAddressValid (MmSessionSpace) == FALSE);
PERFINFO_WSMANAGE_CHECK();
//
// Set the trim criteria: If there are plenty of pages, the existing
// sets are aged and FALSE is returned to signify no trim is necessary.
// Otherwise, the working set expansion list is ordered so the best
// candidates for trimming are placed at the front and TRUE is returned.
//
DoTrimming = MiCheckAndSetSystemTrimCriteria (&TrimCriteria);
if (DoTrimming) {
//
// Clear the deferred entry list to free up some pages.
//
MiDeferredUnlockPages (0);
KeQuerySystemTime (&CurrentTime);
ASSERT (MmIsAddressValid (MmSessionSpace) == FALSE);
LOCK_EXPANSION (OldIrql);
while (!IsListEmpty (&MmWorkingSetExpansionHead.ListHead)) {
//
// Remove the entry at the head and trim it.
//
ListEntry = RemoveHeadList (&MmWorkingSetExpansionHead.ListHead);
VmSupport = CONTAINING_RECORD (ListEntry,
MMSUPPORT,
WorkingSetExpansionLinks);
//
// Note that other routines that set this bit must remove the
// entry from the expansion list first.
//
ASSERT (VmSupport->Flags.BeingTrimmed == 0);
//
// Check to see if we've been here before.
//
if ((*(PLARGE_INTEGER)&VmSupport->LastTrimTime).QuadPart ==
(*(PLARGE_INTEGER)&CurrentTime).QuadPart) {
InsertHeadList (&MmWorkingSetExpansionHead.ListHead,
&VmSupport->WorkingSetExpansionLinks);
//
// If we aren't finished we may sleep in this call.
//
if (MiCheckSystemTrimEndCriteria (&TrimCriteria, OldIrql)) {
//
// No more pages are needed so we're done.
//
break;
}
//
// Start a new round of trimming.
//
KeQuerySystemTime (&CurrentTime);
continue;
}
if ((VmSupport->WorkingSetSize > 3) ||
((VmSupport->Flags.TrimHard == 1) && (VmSupport->WorkingSetSize != 0))) {
//
// This working set is worth examining.
//
}
else {
InsertTailList (&MmWorkingSetExpansionHead.ListHead,
&VmSupport->WorkingSetExpansionLinks);
continue;
}
VmSupport->LastTrimTime = CurrentTime;
VmSupport->Flags.BeingTrimmed = 1;
VmSupport->WorkingSetExpansionLinks.Flink = MM_NO_WS_EXPANSION;
VmSupport->WorkingSetExpansionLinks.Blink =
MM_WS_EXPANSION_IN_PROGRESS;
UNLOCK_EXPANSION (OldIrql);
if (MiAttachAndLockWorkingSet (VmSupport, &InformSessionOfRelease) == FALSE) {
LOCK_EXPANSION (OldIrql);
VmSupport->Flags.AllowWorkingSetAdjustment = MM_FORCE_TRIM;
goto DoneWithWorkingSet;
}
//
// Determine how many pages to trim from this working set.
//
Trim = MiDetermineWsTrimAmount (&TrimCriteria, VmSupport);
//
// If there's something to trim...
//
if ((Trim != 0) &&
((TrimCriteria.TrimAllPasses > TrimCriteria.NumPasses) ||
(MmAvailablePages < TrimCriteria.DesiredFreeGoal))) {
//
// We haven't reached our goal, so trim now.
//
PERFINFO_WSMANAGE_TOTRIM(Trim);
Trim = MiTrimWorkingSet (Trim,
VmSupport,
TrimCriteria.TrimAge);
PERFINFO_WSMANAGE_ACTUALTRIM(Trim);
}
//
// Estimating the current claim is always done here by taking a
// sample of the working set. Aging is only done if the trim
// pass warrants it (ie: the first pass only).
//
MiAgeAndEstimateAvailableInWorkingSet (
VmSupport,
TrimCriteria.DoAging,
NULL,
&TrimCriteria.NewTotalClaim,
&TrimCriteria.NewTotalEstimatedAvailable);
MiDetachAndUnlockWorkingSet (VmSupport, InformSessionOfRelease);
LOCK_EXPANSION (OldIrql);
DoneWithWorkingSet:
ASSERT (VmSupport->Flags.BeingTrimmed == 1);
VmSupport->Flags.BeingTrimmed = 0;
ASSERT (VmSupport->WorkingSetExpansionLinks.Flink == MM_NO_WS_EXPANSION);
if (VmSupport->WorkingSetExpansionLinks.Blink ==
MM_WS_EXPANSION_IN_PROGRESS) {
//
// If the working set size is still above the minimum,
// add this back at the tail of the list.
//
InsertTailList (&MmWorkingSetExpansionHead.ListHead,
&VmSupport->WorkingSetExpansionLinks);
}
else {
//
// The value in the blink is the address of an event
// to set.
//
ASSERT (VmSupport != &MmSystemCacheWs);
KeSetEvent ((PKEVENT)VmSupport->WorkingSetExpansionLinks.Blink,
0,
FALSE);
}
}
MmTotalClaim = TrimCriteria.NewTotalClaim;
MmTotalEstimatedAvailable = TrimCriteria.NewTotalEstimatedAvailable;
PERFINFO_WSMANAGE_TRIMEND_CLAIMS(&TrimCriteria);
UNLOCK_EXPANSION (OldIrql);
}
MiStandbyRemoved = 0;
//
// If memory is critical and there are modified pages to be written
// (presumably because we've just trimmed them), then signal the
// modified page writer.
//
if ((MmAvailablePages < MmMinimumFreePages) ||
(MmModifiedPageListHead.Total >= MmModifiedPageMaximum)) {
KeSetEvent (&MmModifiedPageWriterEvent, 0, FALSE);
}
return;
}
LOGICAL
MiCheckAndSetSystemTrimCriteria (
IN PMMWS_TRIM_CRITERIA Criteria
)
/*++
Routine Description:
Decide whether to trim, age or adjust claim estimations at this time.
Arguments:
Criteria - Supplies a pointer to the trim criteria information. Various
fields in this structure are set as needed by this routine.
Return Value:
TRUE if the caller should initiate trimming, FALSE if not.
Environment:
Kernel mode. No locks held. APC level or below.
--*/
{
KIRQL OldIrql;
PFN_NUMBER Available;
ULONG StandbyRemoved;
ULONG WsRetryCount;
PERFINFO_WSMANAGE_DECL();
PERFINFO_WSMANAGE_CHECK();
//
// See if an empty-all-working-sets request has been queued to us.
//
WsRetryCount = MiWsRetryCount;
if (MiWaitingForWorkingSetEmpty == TRUE) {
MiWsRetryCount = 1;
MiAgePagesAndEstimateClaims (TRUE);
LOCK_EXPANSION (OldIrql);
KeSetEvent (&MiWaitForEmptyEvent, 0, FALSE);
MiWaitingForWorkingSetEmpty = FALSE;
UNLOCK_EXPANSION (OldIrql);
MiReplacing = FALSE;
MiWsRetryCount = WsRetryCount;
return FALSE;
}
//
// Check the number of pages available to see if any trimming (or aging)
// is really required.
//
Available = MmAvailablePages;
StandbyRemoved = MiStandbyRemoved;
if (StandbyRemoved != 0) {
//
// The value is nonzero, we need to synchronize so we get a coordinated
// snapshot of both values.
//
LOCK_PFN (OldIrql);
Available = MmAvailablePages;
StandbyRemoved = MiStandbyRemoved;
UNLOCK_PFN (OldIrql);
}
PERFINFO_WSMANAGE_STARTLOG_CLAIMS();
//
// If we've low on pages, or we've been replacing within a given
// working set, or we've been cannibalizing a large number of standby
// pages, then trim now.
//
if ((Available <= MmPlentyFreePages) ||
(MiReplacing == TRUE) ||
(StandbyRemoved >= (Available >> 2))) {
//
// Inform our caller to start trimming since we're below
// plenty pages - order the list so the bigger working sets are
// in front so our caller trims those first.
//
Criteria->NumPasses = 0;
Criteria->DesiredFreeGoal = MmPlentyFreePages + (MmPlentyFreePages / 2);
Criteria->NewTotalClaim = 0;
Criteria->NewTotalEstimatedAvailable = 0;
//
// If more than 25% of the available pages were recycled standby
// pages, then trim more aggresively in an attempt to get more of the
// cold pages into standby for the next pass.
//
if (StandbyRemoved >= (Available >> 2)) {
Criteria->TrimAllPasses = TRUE;
}
else {
Criteria->TrimAllPasses = FALSE;
}
//
// Start trimming the bigger working sets first.
//
MiRearrangeWorkingSetExpansionList ();
#if DBG
if (MmDebug & MM_DBG_WS_EXPANSION) {
DbgPrint("\nMM-wsmanage: Desired = %ld, Avail %ld\n",
Criteria->DesiredFreeGoal, MmAvailablePages);
}
#endif
PERFINFO_WSMANAGE_WILLTRIM_CLAIMS(Criteria);
//
// No need to lock synchronize the MiReplacing clearing as it
// gets set every time a page replacement happens anyway.
//
MiReplacing = FALSE;
return TRUE;
}
//
// If there is an overwhelming surplus of memory and this is a big
// server then don't even bother aging at this point.
//
if (Available > MmEnormousFreePages) {
//
// Note the claim and estimated available are not cleared so they
// may contain stale values, but at tihs level it doesn't really
// matter.
//
return FALSE;
}
//
// Don't trim but do age unused pages and estimate
// the amount available in working sets.
//
MiAgePagesAndEstimateClaims (FALSE);
MiAdjustClaimParameters (TRUE);
PERFINFO_WSMANAGE_TRIMACTION (PERFINFO_WS_ACTION_RESET_COUNTER);
PERFINFO_WSMANAGE_DUMPENTRIES_CLAIMS ();
return FALSE;
}
LOGICAL
MiCheckSystemTrimEndCriteria (
IN PMMWS_TRIM_CRITERIA Criteria,
IN KIRQL OldIrql
)
/*++
Routine Description:
Check the ending criteria. If we're not done, delay for a little
bit to let the modified writes catch up.
Arguments:
Criteria - Supplies the trim criteria information.
OldIrql - Supplies the old IRQL to lower to if the expansion lock needs
to be released.
Return Value:
TRUE if trimming can be stopped, FALSE otherwise.
Environment:
Kernel mode. Expansion lock held. APC level or below.
--*/
{
LOGICAL FinishedTrimming;
PERFINFO_WSMANAGE_DECL();
PERFINFO_WSMANAGE_CHECK();
if ((MmAvailablePages > Criteria->DesiredFreeGoal) ||
(Criteria->NumPasses >= MI_MAX_TRIM_PASSES)) {
//
// We have enough pages or we trimmed as many as we're going to get.
//
return TRUE;
}
//
// Update the global claim and estimate before we wait.
//
MmTotalClaim = Criteria->NewTotalClaim;
MmTotalEstimatedAvailable = Criteria->NewTotalEstimatedAvailable;
//
// We don't have enough pages - give the modified page writer
// 10 milliseconds to catch up. The wait is also important because a
// thread may have the system cache locked but has been preempted
// by the balance set manager due to its higher priority. We must
// give this thread a shot at running so it can release the system
// cache lock (all the trimmable pages may reside in the system cache).
//
UNLOCK_EXPANSION (OldIrql);
KeDelayExecutionThread (KernelMode,
FALSE,
(PLARGE_INTEGER)&MmShortTime);
PERFINFO_WSMANAGE_WAITFORWRITER_CLAIMS();
//
// Check again to see if we've met the criteria to stop trimming.
//
if (MmAvailablePages > Criteria->DesiredFreeGoal) {
//
// Now we have enough pages so break out.
//
FinishedTrimming = TRUE;
}
else {
//
// We don't have enough pages so let's do another pass.
// Go get the next working set list which is probably the
// one we put back before we gave up the processor.
//
FinishedTrimming = FALSE;
if (Criteria->NumPasses == 0) {
MiAdjustClaimParameters (FALSE);
}
Criteria->NumPasses += 1;
Criteria->NewTotalClaim = 0;
Criteria->NewTotalEstimatedAvailable = 0;
PERFINFO_WSMANAGE_TRIMACTION(PERFINFO_WS_ACTION_FORCE_TRIMMING_PROCESS);
}
LOCK_EXPANSION (OldIrql);
return FinishedTrimming;
}
WSLE_NUMBER
MiDetermineWsTrimAmount (
PMMWS_TRIM_CRITERIA Criteria,
PMMSUPPORT VmSupport
)
/*++
Routine Description:
Determine whether this process should be trimmed.
Arguments:
Criteria - Supplies the trim criteria information.
VmSupport - Supplies the working set information for the candidate.
Return Value:
TRUE if trimming should be done on this process, FALSE if not.
Environment:
Kernel mode. Expansion lock held. APC level or below.
--*/
{
PMMWSL WorkingSetList;
WSLE_NUMBER MaxTrim;
WSLE_NUMBER Trim;
LOGICAL OutswapEnabled;
PEPROCESS ProcessToTrim;
PMM_SESSION_SPACE SessionSpace;
WorkingSetList = VmSupport->VmWorkingSetList;
MaxTrim = VmSupport->WorkingSetSize;
if (MaxTrim <= WorkingSetList->FirstDynamic) {
return 0;
}
OutswapEnabled = FALSE;
if (VmSupport == &MmSystemCacheWs) {
PERFINFO_WSMANAGE_TRIMWS (NULL, NULL, VmSupport);
}
else if (VmSupport->Flags.SessionSpace == 0) {
ProcessToTrim = CONTAINING_RECORD (VmSupport, EPROCESS, Vm);
if (ProcessToTrim->Flags & PS_PROCESS_FLAGS_OUTSWAP_ENABLED) {
OutswapEnabled = TRUE;
}
PERFINFO_WSMANAGE_TRIMWS (ProcessToTrim, NULL, VmSupport);
}
else {
if (VmSupport->Flags.TrimHard == 1) {
OutswapEnabled = TRUE;
}
SessionSpace = CONTAINING_RECORD(VmSupport,
MM_SESSION_SPACE,
Vm);
PERFINFO_WSMANAGE_TRIMWS (NULL, SessionSpace, VmSupport);
}
if (OutswapEnabled == FALSE) {
//
// Don't trim the cache or non-swapped sessions or processes
// below their minimum.
//
MaxTrim -= VmSupport->MinimumWorkingSetSize;
}
switch (Criteria->NumPasses) {
case 0:
Trim = VmSupport->Claim >>
((VmSupport->Flags.MemoryPriority == MEMORY_PRIORITY_FOREGROUND)
? MI_FOREGROUND_CLAIM_AVAILABLE_SHIFT
: MI_BACKGROUND_CLAIM_AVAILABLE_SHIFT);
Criteria->TrimAge = MI_PASS0_TRIM_AGE;
Criteria->DoAging = TRUE;
break;
case 1:
Trim = VmSupport->Claim >>
((VmSupport->Flags.MemoryPriority == MEMORY_PRIORITY_FOREGROUND)
? MI_FOREGROUND_CLAIM_AVAILABLE_SHIFT
: MI_BACKGROUND_CLAIM_AVAILABLE_SHIFT);
Criteria->TrimAge = MI_PASS1_TRIM_AGE;
Criteria->DoAging = FALSE;
break;
case 2:
Trim = VmSupport->Claim;
Criteria->TrimAge = MI_PASS2_TRIM_AGE;
Criteria->DoAging = FALSE;
break;
case 3:
Trim = VmSupport->EstimatedAvailable;
Criteria->TrimAge = MI_PASS3_TRIM_AGE;
Criteria->DoAging = FALSE;
break;
default:
Trim = VmSupport->EstimatedAvailable;
Criteria->TrimAge = MI_PASS3_TRIM_AGE;
Criteria->DoAging = FALSE;
if (MiHardTrim == TRUE || MmAvailablePages < MM_HIGH_LIMIT + 64) {
if (VmSupport->WorkingSetSize > VmSupport->MinimumWorkingSetSize) {
Trim = (VmSupport->WorkingSetSize - VmSupport->MinimumWorkingSetSize) >> 2;
if (Trim == 0) {
Trim = VmSupport->WorkingSetSize - VmSupport->MinimumWorkingSetSize;
}
}
Criteria->TrimAge = MI_PASS4_TRIM_AGE;
Criteria->DoAging = TRUE;
}
break;
}
if (Trim > MaxTrim) {
Trim = MaxTrim;
}
#if DBG
if ((MmDebug & MM_DBG_WS_EXPANSION) && (Trim != 0)) {
if (VmSupport->Flags.SessionSpace == 0) {
ProcessToTrim = CONTAINING_RECORD (VmSupport, EPROCESS, Vm);
DbgPrint(" Trimming Process %16s, WS %6d, Trimming %5d ==> %5d\n",
ProcessToTrim ? ProcessToTrim->ImageFileName : (PUCHAR)"System Cache",
VmSupport->WorkingSetSize,
Trim,
VmSupport->WorkingSetSize-Trim);
}
else {
SessionSpace = CONTAINING_RECORD (VmSupport,
MM_SESSION_SPACE,
Vm);
DbgPrint(" Trimming Session 0x%x (id %d), WS %6d, Trimming %5d ==> %5d\n",
SessionSpace,
SessionSpace->SessionId,
VmSupport->WorkingSetSize,
Trim,
VmSupport->WorkingSetSize-Trim);
}
}
#endif
return Trim;
}
VOID
MiAgePagesAndEstimateClaims (
LOGICAL EmptyIt
)
/*++
Routine Description:
Walk through the sets on the working set expansion list.
Either age pages and estimate the claim (number of pages they aren't using),
or empty the working set.
Arguments:
EmptyIt - Supplies TRUE to empty the working set,
FALSE to just age and estimate it.
Return Value:
None.
Environment:
Kernel mode, APCs disabled. PFN lock NOT held.
--*/
{
WSLE_NUMBER WslesScanned;
PMMSUPPORT VmSupport;
PMMSUPPORT FirstSeen;
LOGICAL SystemCacheSeen;
KIRQL OldIrql;
PLIST_ENTRY ListEntry;
PFN_NUMBER NewTotalClaim;
PFN_NUMBER NewTotalEstimatedAvailable;
LOGICAL InformSessionOfRelease;
ULONG LoopCount;
FirstSeen = NULL;
SystemCacheSeen = FALSE;
NewTotalClaim = 0;
NewTotalEstimatedAvailable = 0;
LoopCount = 0;
WslesScanned = 0;
ASSERT (MmIsAddressValid (MmSessionSpace) == FALSE);
LOCK_EXPANSION (OldIrql);
while (!IsListEmpty (&MmWorkingSetExpansionHead.ListHead)) {
ASSERT (MmIsAddressValid (MmSessionSpace) == FALSE);
//
// Remove the entry at the head, try to lock it, if we can lock it
// then age some pages and estimate the number of available pages.
//
ListEntry = RemoveHeadList (&MmWorkingSetExpansionHead.ListHead);
VmSupport = CONTAINING_RECORD (ListEntry,
MMSUPPORT,
WorkingSetExpansionLinks);
if (VmSupport == &MmSystemCacheWs) {
if (SystemCacheSeen == TRUE) {
//
// Seen this one already.
//
FirstSeen = VmSupport;
}
SystemCacheSeen = TRUE;
}
ASSERT (VmSupport->Flags.BeingTrimmed == 0);
if (VmSupport == FirstSeen) {
InsertHeadList (&MmWorkingSetExpansionHead.ListHead,
&VmSupport->WorkingSetExpansionLinks);
break;
}
VmSupport->Flags.BeingTrimmed = 1;
VmSupport->WorkingSetExpansionLinks.Flink = MM_NO_WS_EXPANSION;
VmSupport->WorkingSetExpansionLinks.Blink =
MM_WS_EXPANSION_IN_PROGRESS;
UNLOCK_EXPANSION (OldIrql);
if (FirstSeen == NULL) {
FirstSeen = VmSupport;
}
if (MiAttachAndLockWorkingSet (VmSupport, &InformSessionOfRelease) == TRUE) {
if (EmptyIt == FALSE) {
MiAgeAndEstimateAvailableInWorkingSet (VmSupport,
TRUE,
&WslesScanned,
&NewTotalClaim,
&NewTotalEstimatedAvailable);
}
else {
MiEmptyWorkingSet (VmSupport, FALSE);
}
MiDetachAndUnlockWorkingSet (VmSupport, InformSessionOfRelease);
}
LOCK_EXPANSION (OldIrql);
ASSERT (VmSupport->Flags.BeingTrimmed == 1);
VmSupport->Flags.BeingTrimmed = 0;
ASSERT (VmSupport->WorkingSetExpansionLinks.Flink == MM_NO_WS_EXPANSION);
if (VmSupport->WorkingSetExpansionLinks.Blink ==
MM_WS_EXPANSION_IN_PROGRESS) {
//
// If the working set size is still above the minimum,
// add this back at the tail of the list.
//
InsertTailList (&MmWorkingSetExpansionHead.ListHead,
&VmSupport->WorkingSetExpansionLinks);
}
else {
//
// The value in the blink is the address of an event
// to set.
//
ASSERT (VmSupport != &MmSystemCacheWs);
KeSetEvent ((PKEVENT)VmSupport->WorkingSetExpansionLinks.Blink,
0,
FALSE);
}
//
// The initial working set that was chosen for FirstSeen may have
// been trimmed down under its minimum and been removed from the
// ExpansionHead links. It is possible that the system cache is not
// on the links either. This check detects this extremely rare
// situation so that the system does not spin forever.
//
LoopCount += 1;
if (LoopCount > 200) {
if (MmSystemCacheWs.WorkingSetExpansionLinks.Blink == MM_WS_EXPANSION_IN_PROGRESS) {
break;
}
}
}
UNLOCK_EXPANSION (OldIrql);
if (EmptyIt == FALSE) {
MmTotalClaim = NewTotalClaim;
MmTotalEstimatedAvailable = NewTotalEstimatedAvailable;
}
}
VOID
MiAgeAndEstimateAvailableInWorkingSet (
IN PMMSUPPORT VmSupport,
IN LOGICAL DoAging,
IN PWSLE_NUMBER WslesScanned,
IN OUT PPFN_NUMBER TotalClaim,
IN OUT PPFN_NUMBER TotalEstimatedAvailable
)
/*++
Routine Description:
Age pages (clear the access bit or if the page hasn't been
accessed, increment the age) for a portion of the working
set. Also, walk through a sample of the working set
building a set of counts of how old the pages are.
The counts are used to create a claim of the amount
the system can steal from this process if memory
becomes tight.
Arguments:
VmSupport - Supplies the VM support structure to age and estimate.
DoAging - TRUE if pages are to be aged. Regardless, the pages will be
added to the availability estimation.
WslesScanned - Total numbers of WSLEs scanned on this sweep, used as a
control to prevent excessive aging on large systems with
many processes.
TotalClaim - Supplies a pointer to system wide claim to update.
TotalEstimatedAvailable - Supplies a pointer to system wide estimate
to update.
Return Value:
None
Environment:
Kernel mode, APCs disabled, working set mutex. PFN lock NOT held.
--*/
{
LOGICAL RecalculateShift;
WSLE_NUMBER LastEntry;
WSLE_NUMBER StartEntry;
WSLE_NUMBER FirstDynamic;
WSLE_NUMBER CurrentEntry;
PMMWSL WorkingSetList;
PMMWSLE Wsle;
PMMPTE PointerPte;
WSLE_NUMBER NumberToExamine;
WSLE_NUMBER Claim;
ULONG Estimate;
ULONG SampledAgeCounts[MI_USE_AGE_COUNT] = {0};
MI_NEXT_ESTIMATION_SLOT_CONST NextConst;
WSLE_NUMBER SampleSize;
WSLE_NUMBER AgeSize;
ULONG CounterShift;
WSLE_NUMBER Temp;
ULONG i;
WorkingSetList = VmSupport->VmWorkingSetList;
Wsle = WorkingSetList->Wsle;
AgeSize = 0;
ASSERT ((VmSupport != &MmSystemCacheWs) || (PsGetCurrentThread() == MmSystemLockOwner));
LastEntry = WorkingSetList->LastEntry;
FirstDynamic = WorkingSetList->FirstDynamic;
if (DoAging == TRUE) {
//
// Clear the used bits or increment the age of a portion of the
// working set.
//
// Try to walk the entire working set every 2^MI_AGE_AGING_SHIFT
// seconds.
//
if (VmSupport->WorkingSetSize > WorkingSetList->FirstDynamic) {
NumberToExamine = (VmSupport->WorkingSetSize - WorkingSetList->FirstDynamic) >> MiAgingShift;
//
// Bigger machines can easily have working sets that span
// terabytes so limit the absolute walk.
//
if (NumberToExamine > MI_MAXIMUM_SAMPLE) {
NumberToExamine = MI_MAXIMUM_SAMPLE;
}
//
// In addition to large working sets, bigger machines may also
// have huge numbers of processes - checking the aggregate number
// of working set list entries scanned prevents this situation
// from triggering excessive scanning.
//
if ((WslesScanned != NULL) &&
(*WslesScanned >= MiMaximumWslesPerSweep)) {
NumberToExamine = 64;
}
AgeSize = NumberToExamine;
CurrentEntry = VmSupport->NextAgingSlot;
if (CurrentEntry > LastEntry || CurrentEntry < FirstDynamic) {
CurrentEntry = FirstDynamic;
}
if (Wsle[CurrentEntry].u1.e1.Valid == 0) {
MI_NEXT_VALID_AGING_SLOT(CurrentEntry, FirstDynamic, LastEntry, Wsle);
}
while (NumberToExamine != 0) {
PointerPte = MiGetPteAddress (Wsle[CurrentEntry].u1.VirtualAddress);
if (MI_GET_ACCESSED_IN_PTE(PointerPte) == 1) {
MI_SET_ACCESSED_IN_PTE(PointerPte, 0);
MI_RESET_WSLE_AGE(PointerPte, &Wsle[CurrentEntry]);
}
else {
MI_INC_WSLE_AGE(PointerPte, &Wsle[CurrentEntry]);
}
NumberToExamine -= 1;
MI_NEXT_VALID_AGING_SLOT(CurrentEntry, FirstDynamic, LastEntry, Wsle);
}
VmSupport->NextAgingSlot = CurrentEntry + 1; // Start here next time
}
}
//
// Estimate the number of unused pages in the working set.
//
// The working set may have shrunk or the non-paged portion may have
// grown since the last time. Put the next counter at the FirstDynamic
// if so.
//
CurrentEntry = VmSupport->NextEstimationSlot;
if (CurrentEntry > LastEntry || CurrentEntry < FirstDynamic) {
CurrentEntry = FirstDynamic;
}
//
// When aging, walk the entire working set every 2^MiEstimationShift
// seconds.
//
CounterShift = 0;
SampleSize = 0;
if (VmSupport->WorkingSetSize > WorkingSetList->FirstDynamic) {
RecalculateShift = FALSE;
SampleSize = VmSupport->WorkingSetSize - WorkingSetList->FirstDynamic;
NumberToExamine = SampleSize >> MiEstimationShift;
//
// Bigger machines may have huge numbers of processes - checking the
// aggregate number of working set list entries scanned prevents this
// situation from triggering excessive scanning.
//
if ((WslesScanned != NULL) &&
(*WslesScanned >= MiMaximumWslesPerSweep)) {
RecalculateShift = TRUE;
}
else if (NumberToExamine > MI_MAXIMUM_SAMPLE) {
//
// Bigger machines can easily have working sets that span
// terabytes so limit the absolute walk.
//
NumberToExamine = MI_MAXIMUM_SAMPLE;
Temp = SampleSize >> MI_MINIMUM_SAMPLE_SHIFT;
SampleSize = MI_MAXIMUM_SAMPLE;
//
// Calculate the necessary counter shift to estimate pages
// in use.
//
for ( ; Temp != 0; Temp = Temp >> 1) {
CounterShift += 1;
}
}
else if (NumberToExamine >= MI_MINIMUM_SAMPLE) {
//
// Ensure that NumberToExamine is at least the minimum size.
//
SampleSize = NumberToExamine;
CounterShift = MiEstimationShift;
}
else if (SampleSize > MI_MINIMUM_SAMPLE) {
RecalculateShift = TRUE;
}
if (RecalculateShift == TRUE) {
Temp = SampleSize >> MI_MINIMUM_SAMPLE_SHIFT;
SampleSize = MI_MINIMUM_SAMPLE;
//
// Calculate the necessary counter shift to estimate pages
// in use.
//
for ( ; Temp != 0; Temp = Temp >> 1) {
CounterShift += 1;
}
}
ASSERT (SampleSize != 0);
MI_CALC_NEXT_ESTIMATION_SLOT_CONST(NextConst, WorkingSetList);
StartEntry = FirstDynamic;
if (Wsle[CurrentEntry].u1.e1.Valid == 0) {
MI_NEXT_VALID_ESTIMATION_SLOT (CurrentEntry,
StartEntry,
FirstDynamic,
LastEntry,
NextConst,
Wsle);
}
for (i = 0; i < SampleSize; i += 1) {
PointerPte = MiGetPteAddress (Wsle[CurrentEntry].u1.VirtualAddress);
if (MI_GET_ACCESSED_IN_PTE(PointerPte) == 0) {
MI_UPDATE_USE_ESTIMATE (PointerPte,
&Wsle[CurrentEntry],
SampledAgeCounts);
}
if (i == NumberToExamine - 1) {
//
// Start estimation here next time.
//
VmSupport->NextEstimationSlot = CurrentEntry + 1;
}
MI_NEXT_VALID_ESTIMATION_SLOT (CurrentEntry,
StartEntry,
FirstDynamic,
LastEntry,
NextConst,
Wsle);
}
}
if (SampleSize < AgeSize) {
SampleSize = AgeSize;
}
if (WslesScanned != NULL) {
*WslesScanned += SampleSize;
}
Estimate = MI_CALCULATE_USAGE_ESTIMATE(SampledAgeCounts, CounterShift);
Claim = VmSupport->Claim + MI_CLAIM_INCR;
if (Claim > Estimate) {
Claim = Estimate;
}
VmSupport->Claim = Claim;
VmSupport->EstimatedAvailable = Estimate;
PERFINFO_WSMANAGE_DUMPWS(VmSupport, SampledAgeCounts);
VmSupport->GrowthSinceLastEstimate = 0;
*TotalClaim += Claim >> ((VmSupport->Flags.MemoryPriority == MEMORY_PRIORITY_FOREGROUND)
? MI_FOREGROUND_CLAIM_AVAILABLE_SHIFT
: MI_BACKGROUND_CLAIM_AVAILABLE_SHIFT);
*TotalEstimatedAvailable += Estimate;
return;
}
ULONG MiClaimAdjustmentThreshold[8] = { 0, 0, 4000, 8000, 12000, 24000, 32000, 32000};
VOID
MiAdjustClaimParameters (
IN LOGICAL EnoughPages
)
/*++
Routine Description:
Adjust the rate at which we walk through working sets. If we have
enough pages (we aren't trimming pages that aren't considered young),
then we check to see whether we should decrease the aging rate and
vice versa.
The limits for the aging rate are 1/8 and 1/128 of the working sets.
This means that the finest age granularities are 8 to 128 seconds in
these cases. With the current 2 bit counter, at the low end we would
start trimming pages > 16 seconds old and at the high end > 4 minutes.
Arguments:
EnoughPages - Supplies whether to increase the rate or decrease it.
Return Value:
None.
Environment:
Kernel mode.
--*/
{
LARGE_INTEGER CurrentTime;
KeQuerySystemTime (&CurrentTime);
if (EnoughPages == TRUE &&
((MmTotalClaim + MmAvailablePages) > MiClaimAdjustmentThreshold[MiAgingShift])) {
//
// Don't adjust the rate too frequently, don't go over the limit, and
// make sure there are enough claimed and/or available.
//
if (((CurrentTime.QuadPart - MiLastAdjustmentOfClaimParams.QuadPart) >
MmClaimParameterAdjustUpTime.QuadPart) &&
(MiAgingShift < MI_MAXIMUM_AGING_SHIFT ) ) {
//
// Set the time only when we change the rate.
//
MiLastAdjustmentOfClaimParams.QuadPart = CurrentTime.QuadPart;
MiAgingShift += 1;
MiEstimationShift += 1;
}
}
else if ((EnoughPages == FALSE) ||
(MmTotalClaim + MmAvailablePages) < MiClaimAdjustmentThreshold[MiAgingShift - 1]) {
//
// Don't adjust the rate down too frequently.
//
if ((CurrentTime.QuadPart - MiLastAdjustmentOfClaimParams.QuadPart) >
MmClaimParameterAdjustDownTime.QuadPart) {
//
// Always set the time so we don't adjust up too soon after
// a 2nd pass trim.
//
MiLastAdjustmentOfClaimParams.QuadPart = CurrentTime.QuadPart;
//
// Don't go under the limit.
//
if (MiAgingShift > 3) {
MiAgingShift -= 1;
MiEstimationShift -= 1;
}
}
}
}
#define MM_WS_REORG_BUCKETS_MAX 7
#if DBG
ULONG MiSessionIdleBuckets[MM_WS_REORG_BUCKETS_MAX];
#endif
VOID
MiRearrangeWorkingSetExpansionList (
VOID
)
/*++
Routine Description:
This function arranges the working set list into different
groups based upon the claim. This is done so the working set
trimming will take place on fat processes first.
The working sets are sorted into buckets and then linked back up.
Swapped out sessions and processes are put at the front.
Arguments:
None.
Return Value:
None.
Environment:
Kernel mode, no locks held.
--*/
{
KIRQL OldIrql;
PLIST_ENTRY ListEntry;
PMMSUPPORT VmSupport;
int Size;
int PreviousNonEmpty;
int NonEmpty;
LIST_ENTRY ListHead[MM_WS_REORG_BUCKETS_MAX];
LARGE_INTEGER CurrentTime;
LARGE_INTEGER SessionIdleTime;
ULONG IdleTime;
PMM_SESSION_SPACE SessionGlobal;
KeQuerySystemTime (&CurrentTime);
if (IsListEmpty (&MmWorkingSetExpansionHead.ListHead)) {
return;
}
for (Size = 0 ; Size < MM_WS_REORG_BUCKETS_MAX; Size++) {
InitializeListHead (&ListHead[Size]);
}
LOCK_EXPANSION (OldIrql);
while (!IsListEmpty (&MmWorkingSetExpansionHead.ListHead)) {
ListEntry = RemoveHeadList (&MmWorkingSetExpansionHead.ListHead);
VmSupport = CONTAINING_RECORD(ListEntry,
MMSUPPORT,
WorkingSetExpansionLinks);
if (VmSupport->Flags.TrimHard == 1) {
ASSERT (VmSupport->Flags.SessionSpace == 1);
SessionGlobal = CONTAINING_RECORD (VmSupport,
MM_SESSION_SPACE,
Vm);
SessionIdleTime.QuadPart = CurrentTime.QuadPart - SessionGlobal->LastProcessSwappedOutTime.QuadPart;
#if DBG
if (MmDebug & MM_DBG_SESSIONS) {
DbgPrint ("Mm: Session %d heavily trim/aged - all its processes (%d) swapped out %d seconds ago\n",
SessionGlobal->SessionId,
SessionGlobal->ReferenceCount,
(ULONG)(SessionIdleTime.QuadPart / 10000000));
}
#endif
if (SessionIdleTime.QuadPart < 0) {
//
// The administrator has moved the system time backwards.
// Give this session a fresh start.
//
SessionIdleTime.QuadPart = 0;
KeQuerySystemTime (&SessionGlobal->LastProcessSwappedOutTime);
}
IdleTime = (ULONG) (SessionIdleTime.QuadPart / 10000000);
}
else {
IdleTime = 0;
}
if (VmSupport->Flags.MemoryPriority == MEMORY_PRIORITY_FOREGROUND) {
//
// Put the foreground processes at the end of the list,
// to give them priority.
//
Size = 6;
}
else {
if (VmSupport->Claim > 400) {
Size = 0;
}
else if (IdleTime > 30) {
Size = 0;
#if DBG
MiSessionIdleBuckets[Size] += 1;
#endif
}
else if (VmSupport->Claim > 200) {
Size = 1;
}
else if (IdleTime > 20) {
Size = 1;
#if DBG
MiSessionIdleBuckets[Size] += 1;
#endif
}
else if (VmSupport->Claim > 100) {
Size = 2;
}
else if (IdleTime > 10) {
Size = 2;
#if DBG
MiSessionIdleBuckets[Size] += 1;
#endif
}
else if (VmSupport->Claim > 50) {
Size = 3;
}
else if (IdleTime) {
Size = 3;
#if DBG
MiSessionIdleBuckets[Size] += 1;
#endif
}
else if (VmSupport->Claim > 25) {
Size = 4;
}
else {
Size = 5;
#if DBG
if (VmSupport->Flags.SessionSpace == 1) {
MiSessionIdleBuckets[Size] += 1;
}
#endif
}
}
#if DBG
if (MmDebug & MM_DBG_WS_EXPANSION) {
DbgPrint("MM-rearrange: TrimHard = %d, WS Size = 0x%x, Claim 0x%x, Bucket %d\n",
VmSupport->Flags.TrimHard,
VmSupport->WorkingSetSize,
VmSupport->Claim,
Size);
}
#endif //DBG
//
// Note: this reverses the bucket order each time we
// reorganize the lists. This may be good or bad -
// if you change it you may want to think about it.
//
InsertHeadList (&ListHead[Size],
&VmSupport->WorkingSetExpansionLinks);
}
//
// Find the first non-empty list.
//
for (NonEmpty = 0 ; NonEmpty < MM_WS_REORG_BUCKETS_MAX ; NonEmpty += 1) {
if (!IsListEmpty (&ListHead[NonEmpty])) {
break;
}
}
//
// Put the head of first non-empty list at the beginning
// of the MmWorkingSetExpansion list.
//
MmWorkingSetExpansionHead.ListHead.Flink = ListHead[NonEmpty].Flink;
ListHead[NonEmpty].Flink->Blink = &MmWorkingSetExpansionHead.ListHead;
PreviousNonEmpty = NonEmpty;
//
// Link the rest of the lists together.
//
for (NonEmpty += 1; NonEmpty < MM_WS_REORG_BUCKETS_MAX; NonEmpty += 1) {
if (!IsListEmpty (&ListHead[NonEmpty])) {
ListHead[PreviousNonEmpty].Blink->Flink = ListHead[NonEmpty].Flink;
ListHead[NonEmpty].Flink->Blink = ListHead[PreviousNonEmpty].Blink;
PreviousNonEmpty = NonEmpty;
}
}
//
// Link the tail of last non-empty to the MmWorkingSetExpansion list.
//
MmWorkingSetExpansionHead.ListHead.Blink = ListHead[PreviousNonEmpty].Blink;
ListHead[PreviousNonEmpty].Blink->Flink = &MmWorkingSetExpansionHead.ListHead;
UNLOCK_EXPANSION (OldIrql);
return;
}
VOID
MmEmptyAllWorkingSets (
VOID
)
/*++
Routine Description:
This routine attempts to empty all the working sets on the
expansion list.
Arguments:
None.
Return Value:
None.
Environment:
Kernel mode. No locks held. APC level or below.
--*/
{
KIRQL OldIrql;
ASSERT (KeGetCurrentIrql () <= APC_LEVEL);
ASSERT (PsGetCurrentThread () != MmWorkingSetThread);
//
// For Hydra, we cannot attach directly to the session space to be
// trimmed because it would result in session space references by
// other threads in this process to the attached session instead
// of the (currently) correct one. In fact, we cannot even queue
// this to a worker thread because the working set manager
// (who shares the same page directory) may be attaching or
// detaching from a session (any session). So this must be queued
// to the working set manager.
//
LOCK_EXPANSION (OldIrql);
if (MiWaitingForWorkingSetEmpty == FALSE) {
MiWaitingForWorkingSetEmpty = TRUE;
KeClearEvent (&MiWaitForEmptyEvent);
}
UNLOCK_EXPANSION (OldIrql);
KeSetEvent (&MmWorkingSetManagerEvent, 0, FALSE);
KeWaitForSingleObject (&MiWaitForEmptyEvent,
WrVirtualMemory,
KernelMode,
FALSE,
(PLARGE_INTEGER)0);
return;
}
//
// This is deliberately initialized to 1 and only cleared when we have
// initialized enough of the system working set to support a trim.
//
LONG MiTrimInProgressCount = 1;
ULONG MiTrimAllPageFaultCount;
LOGICAL
MmTrimAllSystemPagableMemory (
IN LOGICAL PurgeTransition
)
/*++
Routine Description:
This routine unmaps all pagable system memory. This does not unmap user
memory or locked down kernel memory. Thus, the memory being unmapped
resides in paged pool, pagable kernel/driver code & data, special pool
and the system cache.
Note that pages with a reference count greater than 1 are skipped (ie:
they remain valid, as they are assumed to be locked down). This prevents
us from unmapping all of the system cache entries, etc.
Non-locked down kernel stacks must be outpaged by modifying the balance
set manager to operate in conjunction with a support routine. This is not
done here.
Arguments:
PurgeTransition - Supplies whether to purge all the clean pages from the
transition list.
Return Value:
TRUE if accomplished, FALSE if not.
Environment:
Kernel mode. APC_LEVEL or below.
--*/
{
KIRQL OldIrql;
KIRQL OldIrql2;
PLIST_ENTRY Next;
PMMSUPPORT VmSupport;
WSLE_NUMBER PagesInUse;
LOGICAL LockAvailable;
PETHREAD CurrentThread;
#if defined(_X86_)
ULONG flags;
#endif
//
// It's ok to check this without acquiring the system WS lock.
//
if (MiTrimAllPageFaultCount == MmSystemCacheWs.PageFaultCount) {
return FALSE;
}
//
// Working set mutexes will be acquired which require APC_LEVEL or below.
//
if (KeGetCurrentIrql() > APC_LEVEL) {
return FALSE;
}
//
// Just return if it's too early during system initialization or if
// another thread/processor is racing here to do the work for us.
//
if (InterlockedIncrement (&MiTrimInProgressCount) > 1) {
InterlockedDecrement (&MiTrimInProgressCount);
return FALSE;
}
#if defined(_X86_)
_asm {
pushfd
pop eax
mov flags, eax
}
if ((flags & EFLAGS_INTERRUPT_MASK) == 0) {
InterlockedDecrement (&MiTrimInProgressCount);
return FALSE;
}
#endif
LockAvailable = KeTryToAcquireSpinLock (&MmExpansionLock, &OldIrql);
if (LockAvailable == FALSE) {
InterlockedDecrement (&MiTrimInProgressCount);
return FALSE;
}
MM_SET_EXPANSION_OWNER ();
CurrentThread = PsGetCurrentThread();
//
// If the system cache resource is owned by this thread then don't bother
// trying to trim now. Note that checking the MmSystemLockOwner is not
// sufficient as this flag is cleared just before actually releasing it.
//
if ((CurrentThread == MmSystemLockOwner) ||
(ExTryToAcquireResourceExclusiveLite(&MmSystemWsLock) == FALSE)) {
UNLOCK_EXPANSION (OldIrql);
InterlockedDecrement (&MiTrimInProgressCount);
return FALSE;
}
Next = MmWorkingSetExpansionHead.ListHead.Flink;
while (Next != &MmWorkingSetExpansionHead.ListHead) {
if (Next == &MmSystemCacheWs.WorkingSetExpansionLinks) {
break;
}
Next = Next->Flink;
}
if (Next != &MmSystemCacheWs.WorkingSetExpansionLinks) {
ExReleaseResourceLite(&MmSystemWsLock);
UNLOCK_EXPANSION (OldIrql);
InterlockedDecrement (&MiTrimInProgressCount);
return FALSE;
}
RemoveEntryList (Next);
VmSupport = &MmSystemCacheWs;
VmSupport->WorkingSetExpansionLinks.Flink = MM_NO_WS_EXPANSION;
VmSupport->WorkingSetExpansionLinks.Blink = MM_WS_EXPANSION_IN_PROGRESS;
ASSERT (VmSupport->Flags.BeingTrimmed == 0);
VmSupport->Flags.BeingTrimmed = 1;
MiTrimAllPageFaultCount = VmSupport->PageFaultCount;
PagesInUse = VmSupport->WorkingSetSize;
//
// There are 2 issues here that are carefully dealt with :
//
// 1. APCs must be disabled while any resources are held to prevent
// suspend APCs from deadlocking the system.
// 2. Once the system cache has been marked MM_WS_EXPANSION_IN_PROGRESS,
// either the thread must not be preempted or the system cache working
// set mutex must be held throughout. Otherwise a high priority thread
// can fault on a system code and data address and the two pages will
// thrash forever (at high priority) because no system working set
// expansion is allowed while MM_WS_EXPANSION_IN_PROGRESS is set.
// The decision was to hold the system working set mutex throughout.
//
MmSystemLockOwner = PsGetCurrentThread ();
UNLOCK_EXPANSION (APC_LEVEL);
MiEmptyWorkingSet (VmSupport, FALSE);
LOCK_EXPANSION (OldIrql2);
ASSERT (OldIrql2 == APC_LEVEL);
ASSERT (VmSupport->WorkingSetExpansionLinks.Flink == MM_NO_WS_EXPANSION);
ASSERT (VmSupport->Flags.BeingTrimmed == 1);
VmSupport->Flags.BeingTrimmed = 0;
ASSERT (VmSupport->WorkingSetExpansionLinks.Blink ==
MM_WS_EXPANSION_IN_PROGRESS);
InsertTailList (&MmWorkingSetExpansionHead.ListHead,
&VmSupport->WorkingSetExpansionLinks);
UNLOCK_EXPANSION (APC_LEVEL);
//
// Since MiEmptyWorkingSet will attempt to recursively acquire and release
// the MmSystemWsLock, the MmSystemLockOwner field may get cleared.
// This means here the resource must be explicitly released instead of
// using UNLOCK_SYSTEM_WS.
//
MmSystemLockOwner = NULL;
ExReleaseResourceLite (&MmSystemWsLock);
KeLowerIrql (OldIrql);
ASSERT (KeGetCurrentIrql() <= APC_LEVEL);
if (PurgeTransition == TRUE) {
MiPurgeTransitionList ();
}
InterlockedDecrement (&MiTrimInProgressCount);
return TRUE;
}
LOGICAL
MmTrimProcessMemory (
IN LOGICAL PurgeTransition
)
/*++
Routine Description:
This routine unmaps all of the current process' user memory.
Arguments:
PurgeTransition - Supplies whether to purge all the clean pages from the
transition list.
Return Value:
TRUE if accomplished, FALSE if not.
Environment:
Kernel mode. APC_LEVEL or below.
--*/
{
WSLE_NUMBER Last;
KIRQL OldIrql;
PLIST_ENTRY Next;
PMMSUPPORT VmSupport;
LOGICAL LockAvailable;
PEPROCESS Process;
WSLE_NUMBER LastFreed;
PMMWSL WorkingSetList;
PMMWSLE Wsle;
PMMPTE PointerPte;
WSLE_NUMBER Entry;
#if defined(_X86_)
ULONG flags;
#endif
//
// Working set mutexes will be acquired which require APC_LEVEL or below.
//
if (KeGetCurrentIrql() > APC_LEVEL) {
return FALSE;
}
#if defined(_X86_)
_asm {
pushfd
pop eax
mov flags, eax
}
if ((flags & EFLAGS_INTERRUPT_MASK) == 0) {
return FALSE;
}
#endif
//
// Just return if it's too early during system initialization or if
// another thread/processor is racing here to do the work for us.
//
if (InterlockedIncrement (&MiTrimInProgressCount) > 1) {
InterlockedDecrement (&MiTrimInProgressCount);
return FALSE;
}
Process = PsGetCurrentProcess ();
VmSupport = &Process->Vm;
//
// If the WS mutex is not readily available then just return.
//
if (ExTryToAcquireFastMutex (&Process->WorkingSetLock) == FALSE) {
InterlockedDecrement (&MiTrimInProgressCount);
return FALSE;
}
//
// If the process is exiting then just return.
//
if (Process->Flags & PS_PROCESS_FLAGS_VM_DELETED) {
UNLOCK_WS (Process);
InterlockedDecrement (&MiTrimInProgressCount);
return FALSE;
}
ASSERT (!MI_IS_WS_UNSAFE(Process));
//
// If the expansion lock is not available then just return.
//
LockAvailable = KeTryToAcquireSpinLock (&MmExpansionLock, &OldIrql);
if (LockAvailable == FALSE) {
UNLOCK_WS (Process);
InterlockedDecrement (&MiTrimInProgressCount);
return FALSE;
}
ASSERT (OldIrql == APC_LEVEL);
MM_SET_EXPANSION_OWNER ();
Next = MmWorkingSetExpansionHead.ListHead.Flink;
while (Next != &MmWorkingSetExpansionHead.ListHead) {
if (Next == &VmSupport->WorkingSetExpansionLinks) {
break;
}
Next = Next->Flink;
}
if (Next != &VmSupport->WorkingSetExpansionLinks) {
UNLOCK_EXPANSION (OldIrql);
UNLOCK_WS (Process);
InterlockedDecrement (&MiTrimInProgressCount);
return FALSE;
}
RemoveEntryList (Next);
VmSupport->WorkingSetExpansionLinks.Flink = MM_NO_WS_EXPANSION;
VmSupport->WorkingSetExpansionLinks.Blink = MM_WS_EXPANSION_IN_PROGRESS;
ASSERT (VmSupport->Flags.BeingTrimmed == 0);
VmSupport->Flags.BeingTrimmed = 1;
UNLOCK_EXPANSION (APC_LEVEL);
//
// There are 2 issues here that are carefully dealt with :
//
// 1. APCs must be disabled while any resources are held to prevent
// suspend APCs from deadlocking the system.
// 2. Once the working set has been marked MM_WS_EXPANSION_IN_PROGRESS,
// the working set mutex must be held throughout, otherwise a high
// priority thread can fault on a code and data address and the two
// pages will thrash forever (at high priority) because no working set
// expansion is allowed while MM_WS_EXPANSION_IN_PROGRESS is set.
//
WorkingSetList = VmSupport->VmWorkingSetList;
Wsle = WorkingSetList->Wsle;
//
// Attempt to remove the pages starting at the bottom.
//
LastFreed = WorkingSetList->LastEntry;
for (Entry = WorkingSetList->FirstDynamic; Entry <= LastFreed; Entry += 1) {
if (Wsle[Entry].u1.e1.Valid != 0) {
PointerPte = MiGetPteAddress (Wsle[Entry].u1.VirtualAddress);
MiFreeWsle (Entry, VmSupport, PointerPte);
}
}
MiRemoveWorkingSetPages (WorkingSetList, VmSupport);
WorkingSetList->NextSlot = WorkingSetList->FirstDynamic;
//
// Attempt to remove the pages from the front to the end.
//
//
// Reorder the free list.
//
Last = 0;
Entry = WorkingSetList->FirstDynamic;
LastFreed = WorkingSetList->LastInitializedWsle;
while (Entry <= LastFreed) {
if (Wsle[Entry].u1.e1.Valid == 0) {
if (Last == 0) {
WorkingSetList->FirstFree = Entry;
}
else {
Wsle[Last].u1.Long = Entry << MM_FREE_WSLE_SHIFT;
}
Last = Entry;
}
Entry += 1;
}
if (Last != 0) {
Wsle[Last].u1.Long = WSLE_NULL_INDEX << MM_FREE_WSLE_SHIFT; // End of list.
}
LOCK_EXPANSION (OldIrql);
ASSERT (OldIrql == APC_LEVEL);
ASSERT (VmSupport->WorkingSetExpansionLinks.Flink == MM_NO_WS_EXPANSION);
ASSERT (VmSupport->Flags.BeingTrimmed == 1);
VmSupport->Flags.BeingTrimmed = 0;
ASSERT (VmSupport->WorkingSetExpansionLinks.Blink ==
MM_WS_EXPANSION_IN_PROGRESS);
InsertTailList (&MmWorkingSetExpansionHead.ListHead,
&VmSupport->WorkingSetExpansionLinks);
UNLOCK_EXPANSION (APC_LEVEL);
UNLOCK_WS (Process);
ASSERT (KeGetCurrentIrql() <= APC_LEVEL);
if (PurgeTransition == TRUE) {
MiPurgeTransitionList ();
}
InterlockedDecrement (&MiTrimInProgressCount);
return TRUE;
}
LOGICAL
MmTrimSessionMemory (
IN LOGICAL PurgeTransition
)
/*++
Routine Description:
This routine unmaps all of the current session's virtual addresses.
Arguments:
PurgeTransition - Supplies whether to purge all the clean pages from the
transition list.
Return Value:
TRUE if accomplished, FALSE if not.
Environment:
Kernel mode. APC_LEVEL or below.
--*/
{
KIRQL OldIrqlWs;
KIRQL OldIrql;
PLIST_ENTRY Next;
PMMSUPPORT VmSupport;
LOGICAL LockAvailable;
PEPROCESS Process;
PETHREAD Thread;
PMM_SESSION_SPACE SessionGlobal;
#if defined(_X86_)
ULONG flags;
#endif
//
// Working set mutexes will be acquired which require APC_LEVEL or below.
//
if (KeGetCurrentIrql() > APC_LEVEL) {
return FALSE;
}
#if defined(_X86_)
_asm {
pushfd
pop eax
mov flags, eax
}
if ((flags & EFLAGS_INTERRUPT_MASK) == 0) {
return FALSE;
}
#endif
//
// Just return if it's too early during system initialization or if
// another thread/processor is racing here to do the work for us.
//
if (InterlockedIncrement (&MiTrimInProgressCount) > 1) {
InterlockedDecrement (&MiTrimInProgressCount);
return FALSE;
}
Thread = PsGetCurrentThread ();
Process = PsGetCurrentProcessByThread (Thread);
if (((Process->Flags & PS_PROCESS_FLAGS_IN_SESSION) == 0) ||
(Process->Vm.Flags.SessionLeader == 1)) {
InterlockedDecrement (&MiTrimInProgressCount);
return FALSE;
}
//
// If the WS mutex is not readily available then just return.
//
SessionGlobal = SESSION_GLOBAL(MmSessionSpace);
KeRaiseIrql (APC_LEVEL, &OldIrqlWs);
//
// Check for the working set resource being owned by the current thread
// because the resource package allows recursive acquires.
//
if (MmSessionSpace->WorkingSetLockOwner == Thread) {
KeLowerIrql (OldIrqlWs);
InterlockedDecrement (&MiTrimInProgressCount);
return FALSE;
}
if (ExTryToAcquireResourceExclusiveLite (&MmSessionSpace->WsLock) == FALSE) {
KeLowerIrql (OldIrqlWs);
InterlockedDecrement (&MiTrimInProgressCount);
return FALSE;
}
MM_SET_SESSION_RESOURCE_OWNER (Thread);
VmSupport = &SessionGlobal->Vm;
//
// If the expansion lock is not available then just return.
//
LockAvailable = KeTryToAcquireSpinLock (&MmExpansionLock, &OldIrql);
if (LockAvailable == FALSE) {
UNLOCK_SESSION_SPACE_WS (OldIrqlWs);
InterlockedDecrement (&MiTrimInProgressCount);
return FALSE;
}
ASSERT (OldIrql == APC_LEVEL);
MM_SET_EXPANSION_OWNER ();
Next = MmWorkingSetExpansionHead.ListHead.Flink;
while (Next != &MmWorkingSetExpansionHead.ListHead) {
if (Next == &VmSupport->WorkingSetExpansionLinks) {
break;
}
Next = Next->Flink;
}
if (Next != &VmSupport->WorkingSetExpansionLinks) {
UNLOCK_EXPANSION (OldIrql);
UNLOCK_SESSION_SPACE_WS (OldIrqlWs);
InterlockedDecrement (&MiTrimInProgressCount);
return FALSE;
}
RemoveEntryList (Next);
VmSupport->WorkingSetExpansionLinks.Flink = MM_NO_WS_EXPANSION;
VmSupport->WorkingSetExpansionLinks.Blink = MM_WS_EXPANSION_IN_PROGRESS;
ASSERT (VmSupport->Flags.BeingTrimmed == 0);
VmSupport->Flags.BeingTrimmed = 1;
UNLOCK_EXPANSION (APC_LEVEL);
//
// There are 2 issues here that are carefully dealt with :
//
// 1. APCs must be disabled while any resources are held to prevent
// suspend APCs from deadlocking the system.
// 2. Once the working set has been marked MM_WS_EXPANSION_IN_PROGRESS,
// the working set mutex must be held throughout, otherwise a high
// priority thread can fault on a code and data address and the two
// pages will thrash forever (at high priority) because no working set
// expansion is allowed while MM_WS_EXPANSION_IN_PROGRESS is set.
//
MiEmptyWorkingSet (VmSupport, FALSE);
LOCK_EXPANSION (OldIrql);
ASSERT (OldIrql == APC_LEVEL);
ASSERT (VmSupport->WorkingSetExpansionLinks.Flink == MM_NO_WS_EXPANSION);
ASSERT (VmSupport->Flags.BeingTrimmed == 1);
VmSupport->Flags.BeingTrimmed = 0;
ASSERT (VmSupport->WorkingSetExpansionLinks.Blink ==
MM_WS_EXPANSION_IN_PROGRESS);
InsertTailList (&MmWorkingSetExpansionHead.ListHead,
&VmSupport->WorkingSetExpansionLinks);
UNLOCK_EXPANSION (APC_LEVEL);
//
// Since MiEmptyWorkingSet will attempt to recursively acquire and release
// the session space ws mutex, the owner field may get cleared.
// This means here the resource must be explicitly released instead of
// using UNLOCK_SESSION_SPACE_WS.
//
MmSessionSpace->WorkingSetLockOwner = NULL;
ExReleaseResourceLite (&MmSessionSpace->WsLock);
KeLowerIrql (OldIrqlWs);
ASSERT (KeGetCurrentIrql() <= APC_LEVEL);
if (PurgeTransition == TRUE) {
MiPurgeTransitionList ();
}
InterlockedDecrement (&MiTrimInProgressCount);
return TRUE;
}