windows-nt/Source/XPSP1/NT/base/ntos/cache/logsup.c

704 lines
19 KiB
C
Raw Normal View History

2020-09-26 03:20:57 -05:00
/*++
Copyright (c) 1990 Microsoft Corporation
Module Name:
logsup.c
Abstract:
This module implements the special cache manager support for logging
file systems.
Author:
Tom Miller [TomM] 30-Jul-1991
Revision History:
--*/
#include "cc.h"
//
// Define our debug constant
//
#define me 0x0000040
#ifdef ALLOC_PRAGMA
#pragma alloc_text(PAGE,CcSetLogHandleForFile)
#endif
VOID
CcSetAdditionalCacheAttributes (
IN PFILE_OBJECT FileObject,
IN BOOLEAN DisableReadAhead,
IN BOOLEAN DisableWriteBehind
)
/*++
Routine Description:
This routine supports the setting of disable read ahead or disable write
behind flags to control Cache Manager operation. This routine may be
called any time after calling CcInitializeCacheMap. Initially both
read ahead and write behind are enabled. Note that the state of both
of these flags must be specified on each call to this routine.
Arguments:
FileObject - File object for which the respective flags are to be set.
DisableReadAhead - FALSE to enable read ahead, TRUE to disable it.
DisableWriteBehind - FALSE to enable write behind, TRUE to disable it.
Return Value:
None.
--*/
{
PSHARED_CACHE_MAP SharedCacheMap;
KIRQL OldIrql;
//
// Get pointer to SharedCacheMap.
//
SharedCacheMap = FileObject->SectionObjectPointer->SharedCacheMap;
//
// Now set the flags and return.
//
CcAcquireMasterLock( &OldIrql );
if (DisableReadAhead) {
SetFlag(SharedCacheMap->Flags, DISABLE_READ_AHEAD);
} else {
ClearFlag(SharedCacheMap->Flags, DISABLE_READ_AHEAD);
}
if (DisableWriteBehind) {
SetFlag(SharedCacheMap->Flags, DISABLE_WRITE_BEHIND | MODIFIED_WRITE_DISABLED);
} else {
ClearFlag(SharedCacheMap->Flags, DISABLE_WRITE_BEHIND);
}
CcReleaseMasterLock( OldIrql );
}
NTKERNELAPI
BOOLEAN
CcSetPrivateWriteFile(
PFILE_OBJECT FileObject
)
/*++
Routine Description:
This routine will instruct the cache manager to treat the file as
a private-write stream, so that a caller can implement a private
logging mechanism for it. We will turn on both Mm's modify-no-write
and our disable-write-behind, and disallow non-aware flush/purge for
the file.
Caching must already be initiated on the file.
This routine is only exported to the kernel.
Arguments:
FileObject - File to make private-write.
Return Value:
None.
--*/
{
PSHARED_CACHE_MAP SharedCacheMap;
BOOLEAN Disabled;
KIRQL OldIrql;
PVACB Vacb;
ULONG ActivePage;
ULONG PageIsDirty;
//
// Pick up the file exclusive to synchronize against readahead and
// other purge/map activity.
//
FsRtlAcquireFileExclusive( FileObject );
//
// Get pointer to SharedCacheMap.
//
if( FileObject->SectionObjectPointer == NULL ) {
return FALSE;
}
SharedCacheMap = FileObject->SectionObjectPointer->SharedCacheMap;
if( !SharedCacheMap ) {
return FALSE;
}
//
// Unmap all the views in preparation for making the disable mw call.
//
//
// We still need to wait for any dangling cache read or writes.
//
// In fact we have to loop and wait because the lazy writer can
// sneak in and do an CcGetVirtualAddressIfMapped, and we are not
// synchronized.
//
// This is the same bit of code that our purge will do. We assume
// that a private writer has succesfully blocked out other activity.
//
//
// If there is an active Vacb, then nuke it now (before waiting!).
//
CcAcquireMasterLock( &OldIrql );
GetActiveVacbAtDpcLevel( SharedCacheMap, Vacb, ActivePage, PageIsDirty );
CcReleaseMasterLock( OldIrql );
if (Vacb != NULL) {
CcFreeActiveVacb( SharedCacheMap, Vacb, ActivePage, PageIsDirty );
}
while ((SharedCacheMap->Vacbs != NULL) &&
!CcUnmapVacbArray( SharedCacheMap, NULL, 0, FALSE )) {
CcWaitOnActiveCount( SharedCacheMap );
}
//
// Knock the file down.
//
CcFlushCache( FileObject->SectionObjectPointer, NULL, 0, NULL );
//
// Now the file is clean and unmapped. We can still have a racing
// lazy writer, though.
//
// We just wait for the lazy writer queue to drain before disabling
// modified write. There may be a better way to do this by having
// an event for the WRITE_QUEUED flag. ? This would also let us
// dispense with the pagingio pick/drop in the FS cache coherency
// paths, but there could be reasons why CcFlushCache shouldn't
// always do such a block. Investigate this.
//
// This wait takes on the order of ~.5s avg. case.
//
CcAcquireMasterLock( &OldIrql );
if (FlagOn( SharedCacheMap->Flags, WRITE_QUEUED )) {
CcReleaseMasterLock( OldIrql );
FsRtlReleaseFile( FileObject );
CcWaitForCurrentLazyWriterActivity();
FsRtlAcquireFileExclusive( FileObject );
} else {
CcReleaseMasterLock( OldIrql );
}
//
// Now set the flags and return. We do not set our MODIFIED_WRITE_DISABLED
// since we don't want to fully promote this cache map. Future?
//
Disabled = MmDisableModifiedWriteOfSection( FileObject->SectionObjectPointer );
if (Disabled) {
CcAcquireMasterLock( &OldIrql );
SetFlag(SharedCacheMap->Flags, DISABLE_WRITE_BEHIND | PRIVATE_WRITE);
CcReleaseMasterLock( OldIrql );
}
//
// Now release the file for regular operation.
//
FsRtlReleaseFile( FileObject );
return Disabled;
}
VOID
CcSetLogHandleForFile (
IN PFILE_OBJECT FileObject,
IN PVOID LogHandle,
IN PFLUSH_TO_LSN FlushToLsnRoutine
)
/*++
Routine Description:
This routine may be called to instruct the Cache Manager to store the
specified log handle with the shared cache map for a file, to support
subsequent calls to the other routines in this module which effectively
perform an associative search for files by log handle.
Arguments:
FileObject - File for which the log handle should be stored.
LogHandle - Log Handle to store.
FlushToLsnRoutine - A routine to call before flushing buffers for this
file, to insure a log file is flushed to the most
recent Lsn for any Bcb being flushed.
Return Value:
None.
--*/
{
PSHARED_CACHE_MAP SharedCacheMap;
//
// Get pointer to SharedCacheMap.
//
SharedCacheMap = FileObject->SectionObjectPointer->SharedCacheMap;
//
// Now set the log file handle and flush routine
//
SharedCacheMap->LogHandle = LogHandle;
SharedCacheMap->FlushToLsnRoutine = FlushToLsnRoutine;
}
LARGE_INTEGER
CcGetDirtyPages (
IN PVOID LogHandle,
IN PDIRTY_PAGE_ROUTINE DirtyPageRoutine,
IN PVOID Context1,
IN PVOID Context2
)
/*++
Routine Description:
This routine may be called to return all of the dirty pages in all files
for a given log handle. Each page is returned by an individual call to
the Dirty Page Routine. The Dirty Page Routine is defined by a prototype
in ntos\inc\cache.h.
Arguments:
LogHandle - Log Handle which must match the log handle previously stored
for all files which are to be returned.
DirtyPageRoutine -- The routine to call as each dirty page for this log
handle is found.
Context1 - First context parameter to be passed to the Dirty Page Routine.
Context2 - First context parameter to be passed to the Dirty Page Routine.
Return Value:
LARGE_INTEGER - Oldest Lsn found of all the dirty pages, or 0 if no dirty pages
--*/
{
PSHARED_CACHE_MAP SharedCacheMap;
PBCB Bcb, BcbToUnpin = NULL;
KLOCK_QUEUE_HANDLE LockHandle;
LARGE_INTEGER SavedFileOffset, SavedOldestLsn, SavedNewestLsn;
ULONG SavedByteLength;
LARGE_INTEGER OldestLsn = {0,0};
//
// Synchronize with changes to the SharedCacheMap list.
//
CcAcquireMasterLock( &LockHandle.OldIrql );
SharedCacheMap = CONTAINING_RECORD( CcDirtySharedCacheMapList.SharedCacheMapLinks.Flink,
SHARED_CACHE_MAP,
SharedCacheMapLinks );
//
// Use try/finally for cleanup. The only spot where we can raise is out of the
// filesystem callback, but we have the exception handler out here so we aren't
// constantly setting/unsetting it.
//
try {
while (&SharedCacheMap->SharedCacheMapLinks != &CcDirtySharedCacheMapList.SharedCacheMapLinks) {
//
// Skip over cursors, SharedCacheMaps for other LogHandles, and ones with
// no dirty pages
//
if (!FlagOn(SharedCacheMap->Flags, IS_CURSOR) && (SharedCacheMap->LogHandle == LogHandle) &&
(SharedCacheMap->DirtyPages != 0)) {
//
// This SharedCacheMap should stick around for a while in the dirty list.
//
CcIncrementOpenCount( SharedCacheMap, 'pdGS' );
SharedCacheMap->DirtyPages += 1;
CcReleaseMasterLock( LockHandle.OldIrql );
//
// Set our initial resume point and point to first Bcb in List.
//
KeAcquireInStackQueuedSpinLock( &SharedCacheMap->BcbSpinLock, &LockHandle );
Bcb = CONTAINING_RECORD( SharedCacheMap->BcbList.Flink, BCB, BcbLinks );
//
// Scan to the end of the Bcb list.
//
while (&Bcb->BcbLinks != &SharedCacheMap->BcbList) {
//
// If the Bcb is dirty, then capture the inputs for the
// callback routine so we can call without holding a spinlock.
//
if ((Bcb->NodeTypeCode == CACHE_NTC_BCB) && Bcb->Dirty) {
SavedFileOffset = Bcb->FileOffset;
SavedByteLength = Bcb->ByteLength;
SavedOldestLsn = Bcb->OldestLsn;
SavedNewestLsn = Bcb->NewestLsn;
//
// Increment PinCount so the Bcb sticks around
//
Bcb->PinCount += 1;
KeReleaseInStackQueuedSpinLock( &LockHandle );
//
// Any Bcb to unref from a previous loop?
//
if (BcbToUnpin != NULL) {
CcUnpinFileData( BcbToUnpin, TRUE, UNREF );
BcbToUnpin = NULL;
}
//
// Call the file system. This callback may raise status.
//
(*DirtyPageRoutine)( SharedCacheMap->FileObject,
&SavedFileOffset,
SavedByteLength,
&SavedOldestLsn,
&SavedNewestLsn,
Context1,
Context2 );
//
// Possibly update OldestLsn
//
if ((SavedOldestLsn.QuadPart != 0) &&
((OldestLsn.QuadPart == 0) || (SavedOldestLsn.QuadPart < OldestLsn.QuadPart ))) {
OldestLsn = SavedOldestLsn;
}
//
// Now reacquire the spinlock and scan from the resume point
// point to the next Bcb to return in the descending list.
//
KeAcquireInStackQueuedSpinLock( &SharedCacheMap->BcbSpinLock, &LockHandle );
//
// Normally the Bcb can stay around a while, but if not,
// we will just remember it for the next time we do not
// have the spin lock. We cannot unpin it now, because
// we would lose our place in the list.
//
// This is cheating, but it works and is sane since we're
// already traversing the bcb list - dropping the bcb count
// is OK, as long as we don't hit zero. Zero requires a
// slight bit more attention that shouldn't be replicated.
// (unmapping the view)
//
if (Bcb->PinCount > 1) {
Bcb->PinCount -= 1;
} else {
BcbToUnpin = Bcb;
}
}
Bcb = CONTAINING_RECORD( Bcb->BcbLinks.Flink, BCB, BcbLinks );
}
KeReleaseInStackQueuedSpinLock( &LockHandle );
//
// We need to unref any Bcb we are holding before moving on to
// the next SharedCacheMap, or else CcDeleteSharedCacheMap will
// also delete this Bcb.
//
if (BcbToUnpin != NULL) {
CcUnpinFileData( BcbToUnpin, TRUE, UNREF );
BcbToUnpin = NULL;
}
CcAcquireMasterLock( &LockHandle.OldIrql );
//
// Now release the SharedCacheMap, leaving it in the dirty list.
//
CcDecrementOpenCount( SharedCacheMap, 'pdGF' );
SharedCacheMap->DirtyPages -= 1;
}
//
// Now loop back for the next cache map.
//
SharedCacheMap =
CONTAINING_RECORD( SharedCacheMap->SharedCacheMapLinks.Flink,
SHARED_CACHE_MAP,
SharedCacheMapLinks );
}
CcReleaseMasterLock( LockHandle.OldIrql );
} finally {
//
// Drop the Bcb if we are being ejected. We are guaranteed that the
// only raise is from the callback, at which point we have an incremented
// pincount.
//
if (AbnormalTermination()) {
CcUnpinFileData( Bcb, TRUE, UNPIN );
}
}
return OldestLsn;
}
BOOLEAN
CcIsThereDirtyData (
IN PVPB Vpb
)
/*++
Routine Description:
This routine returns TRUE if the specified Vcb has any unwritten dirty
data in the cache.
Arguments:
Vpb - specifies Vpb to check for
Return Value:
FALSE - if the Vpb has no dirty data
TRUE - if the Vpb has dirty data
--*/
{
PSHARED_CACHE_MAP SharedCacheMap;
KIRQL OldIrql;
ULONG LoopsWithLockHeld = 0;
//
// Synchronize with changes to the SharedCacheMap list.
//
CcAcquireMasterLock( &OldIrql );
SharedCacheMap = CONTAINING_RECORD( CcDirtySharedCacheMapList.SharedCacheMapLinks.Flink,
SHARED_CACHE_MAP,
SharedCacheMapLinks );
while (&SharedCacheMap->SharedCacheMapLinks != &CcDirtySharedCacheMapList.SharedCacheMapLinks) {
//
// Look at this one if the Vpb matches and if there is dirty data.
// For what it's worth, don't worry about dirty data in temporary files,
// as that should not concern the caller if it wants to dismount.
//
if (!FlagOn(SharedCacheMap->Flags, IS_CURSOR) &&
(SharedCacheMap->FileObject->Vpb == Vpb) &&
(SharedCacheMap->DirtyPages != 0) &&
!FlagOn(SharedCacheMap->FileObject->Flags, FO_TEMPORARY_FILE)) {
CcReleaseMasterLock( OldIrql );
return TRUE;
}
//
// Make sure we occasionally drop the lock. Set WRITE_QUEUED
// to keep the guy from going away, and increment DirtyPages to
// keep it in this list.
//
if ((++LoopsWithLockHeld >= 20) &&
!FlagOn(SharedCacheMap->Flags, WRITE_QUEUED | IS_CURSOR)) {
SetFlag( *((ULONG volatile *)&SharedCacheMap->Flags), WRITE_QUEUED);
*((ULONG volatile *)&SharedCacheMap->DirtyPages) += 1;
CcReleaseMasterLock( OldIrql );
LoopsWithLockHeld = 0;
CcAcquireMasterLock( &OldIrql );
ClearFlag( *((ULONG volatile *)&SharedCacheMap->Flags), WRITE_QUEUED);
*((ULONG volatile *)&SharedCacheMap->DirtyPages) -= 1;
}
//
// Now loop back for the next cache map.
//
SharedCacheMap =
CONTAINING_RECORD( SharedCacheMap->SharedCacheMapLinks.Flink,
SHARED_CACHE_MAP,
SharedCacheMapLinks );
}
CcReleaseMasterLock( OldIrql );
return FALSE;
}
LARGE_INTEGER
CcGetLsnForFileObject(
IN PFILE_OBJECT FileObject,
OUT PLARGE_INTEGER OldestLsn OPTIONAL
)
/*++
Routine Description:
This routine returns the oldest and newest LSNs for a file object.
Arguments:
FileObject - File for which the log handle should be stored.
OldestLsn - pointer to location to store oldest LSN for file object.
Return Value:
The newest LSN for the file object.
--*/
{
PBCB Bcb;
KLOCK_QUEUE_HANDLE LockHandle;
LARGE_INTEGER Oldest, Newest;
PSHARED_CACHE_MAP SharedCacheMap = FileObject->SectionObjectPointer->SharedCacheMap;
//
// initialize lsn variables
//
Oldest.LowPart = 0;
Oldest.HighPart = 0;
Newest.LowPart = 0;
Newest.HighPart = 0;
if(SharedCacheMap == NULL) {
return Oldest;
}
KeAcquireInStackQueuedSpinLock(&SharedCacheMap->BcbSpinLock, &LockHandle);
//
// Now point to first Bcb in List, and loop through it.
//
Bcb = CONTAINING_RECORD( SharedCacheMap->BcbList.Flink, BCB, BcbLinks );
while (&Bcb->BcbLinks != &SharedCacheMap->BcbList) {
//
// If the Bcb is dirty then capture the oldest and newest lsn
//
if ((Bcb->NodeTypeCode == CACHE_NTC_BCB) && Bcb->Dirty) {
LARGE_INTEGER BcbLsn, BcbNewest;
BcbLsn = Bcb->OldestLsn;
BcbNewest = Bcb->NewestLsn;
if ((BcbLsn.QuadPart != 0) &&
((Oldest.QuadPart == 0) ||
(BcbLsn.QuadPart < Oldest.QuadPart))) {
Oldest = BcbLsn;
}
if ((BcbLsn.QuadPart != 0) && (BcbNewest.QuadPart > Newest.QuadPart)) {
Newest = BcbNewest;
}
}
Bcb = CONTAINING_RECORD( Bcb->BcbLinks.Flink, BCB, BcbLinks );
}
//
// Now release the spin lock for this Bcb list and generate a callback
// if we got something.
//
KeReleaseInStackQueuedSpinLock( &LockHandle );
if (ARGUMENT_PRESENT(OldestLsn)) {
*OldestLsn = Oldest;
}
return Newest;
}