2418 lines
70 KiB
C
2418 lines
70 KiB
C
/*++
|
|
|
|
Copyright (c) 1990 Microsoft Corporation
|
|
|
|
Module Name:
|
|
|
|
initia64.c
|
|
|
|
Abstract:
|
|
|
|
This module contains the machine dependent initialization for the
|
|
memory management component. It is specifically tailored to the
|
|
INTEL IA64 machine.
|
|
|
|
Author:
|
|
|
|
Lou Perazzoli (loup) 6-Jan-1990
|
|
|
|
Revision History:
|
|
|
|
--*/
|
|
|
|
#include "mi.h"
|
|
|
|
#ifdef ALLOC_PRAGMA
|
|
#pragma alloc_text(INIT,MiInitMachineDependent)
|
|
#endif
|
|
|
|
extern VOID
|
|
KeInitializeVhpt (
|
|
IN PVOID Virtual,
|
|
IN ULONG Size
|
|
);
|
|
|
|
extern VOID
|
|
KeFillLargePdePinned (
|
|
IN PMMPTE Pte,
|
|
IN PVOID Virtual
|
|
);
|
|
|
|
VOID
|
|
MiConvertToSuperPages(
|
|
PVOID StartVirtual,
|
|
PVOID EndVirtual,
|
|
SIZE_T PageSize,
|
|
ULONG PageShift
|
|
);
|
|
|
|
VOID
|
|
MiConvertBackToStandardPages(
|
|
IN PVOID StartVirtual,
|
|
IN PVOID EndVirtual
|
|
);
|
|
|
|
VOID
|
|
MiBuildPageTableForDrivers(
|
|
IN PLOADER_PARAMETER_BLOCK LoaderBlock
|
|
);
|
|
|
|
VOID
|
|
MiBuildPageTableForLoaderMemory(
|
|
IN PLOADER_PARAMETER_BLOCK LoaderBlock
|
|
);
|
|
|
|
VOID
|
|
MiRemoveLoaderSuperPages(
|
|
IN PLOADER_PARAMETER_BLOCK LoaderBlock
|
|
);
|
|
|
|
VOID
|
|
MiMakeKernelPagesPermanent(
|
|
IN PLOADER_PARAMETER_BLOCK LoaderBlock
|
|
);
|
|
|
|
VOID
|
|
MiCheckMemoryDescriptorList(
|
|
IN PLOADER_PARAMETER_BLOCK LoaderBlock
|
|
);
|
|
|
|
ULONG_PTR MmKseg3VhptBase;
|
|
ULONG MmKseg3VhptPs;
|
|
|
|
PFN_NUMBER MmSystemParentTablePage;
|
|
MMPTE MiSystemSelfMappedPte;
|
|
MMPTE MiUserSelfMappedPte;
|
|
PVOID MiKseg0Start = 0;
|
|
PVOID MiKseg0End = 0;
|
|
BOOLEAN MiKseg0Mapping = TRUE;
|
|
BOOLEAN MiPrintMemoryDescriptors = FALSE;
|
|
PFN_NUMBER MiNtoskrnlPhysicalBase;
|
|
ULONG_PTR MiNtoskrnlVirtualBase;
|
|
ULONG MiNtoskrnlPageShift;
|
|
PFN_NUMBER MiWasteStart = 0;
|
|
PFN_NUMBER MiWasteEnd = 0;
|
|
|
|
#define MiGetKSegPteAddress(Va) ((PMMPTE)__thash((ULONG_PTR)(Va)))
|
|
|
|
#define _x1mb (1024*1024)
|
|
#define _x1mbnp ((1024*1024) >> PAGE_SHIFT)
|
|
#define _x4mb (1024*1024*4)
|
|
#define _x4mbnp ((1024*1024*4) >> PAGE_SHIFT)
|
|
#define _x16mb (1024*1024*16)
|
|
#define _x16mbnp ((1024*1024*16) >> PAGE_SHIFT)
|
|
#define _x32mb (1024*1024*32)
|
|
#define _x32mbnp ((1024*1024*32) >> PAGE_SHIFT)
|
|
#define _x64mb (1024*1024*64)
|
|
#define _x64mbnp ((1024*1024*64) >> PAGE_SHIFT)
|
|
#define _x256mb (1024*1024*256)
|
|
#define _x256mbnp ((1024*1024*256) >> PAGE_SHIFT)
|
|
#define _x512mb (1024*1024*512)
|
|
#define _x512mbnp ((1024*1024*512) >> PAGE_SHIFT)
|
|
#define _x4gb (0x100000000UI64)
|
|
#define _x4gbnp (0x100000000UI64 >> PAGE_SHIFT)
|
|
|
|
#define ADD_HIGH_MEMORY 0
|
|
|
|
#if ADD_HIGH_MEMORY
|
|
MEMORY_ALLOCATION_DESCRIPTOR MiHighMemoryDescriptor;
|
|
#endif
|
|
|
|
PMEMORY_ALLOCATION_DESCRIPTOR MiFreeDescriptor;
|
|
PMEMORY_ALLOCATION_DESCRIPTOR MiFreeDescriptorLargest = NULL;
|
|
PMEMORY_ALLOCATION_DESCRIPTOR MiFreeDescriptorLowMem = NULL;
|
|
PMEMORY_ALLOCATION_DESCRIPTOR MiFreeDescriptorNonPaged = NULL;
|
|
|
|
PFN_NUMBER MiNextPhysicalPage;
|
|
PFN_NUMBER MiNumberOfPages;
|
|
PFN_NUMBER MiOldFreeDescriptorBase;
|
|
PFN_NUMBER MiOldFreeDescriptorCount;
|
|
|
|
|
|
// Processor Specific VM information
|
|
|
|
|
|
ULONG MiImplVirtualMsb = 50;
|
|
|
|
|
|
PFN_NUMBER
|
|
MiGetNextPhysicalPage (
|
|
VOID
|
|
)
|
|
|
|
/*++
|
|
|
|
Routine Description:
|
|
|
|
This function returns the next physical page number from either the
|
|
largest low memory descritor or the largest free descriptor. If there
|
|
are no physical pages left, then a bugcheck is executed since the
|
|
system cannot be initialized.
|
|
|
|
Arguments:
|
|
|
|
LoaderBlock - Supplies the address of the loader block.
|
|
|
|
Return Value:
|
|
|
|
None.
|
|
|
|
Environment:
|
|
|
|
Kernel mode.
|
|
|
|
--*/
|
|
|
|
{
|
|
|
|
|
|
// If there are free pages left in the current descriptor, then
|
|
// return the next physical page. Otherwise, attempt to switch
|
|
// descriptors.
|
|
|
|
|
|
if (MiNumberOfPages != 0) {
|
|
MiNumberOfPages -= 1;
|
|
return MiNextPhysicalPage++;
|
|
|
|
} else {
|
|
|
|
|
|
// If the current descriptor is not the largest free descriptor,
|
|
// then switch to the largest free descriptor. Otherwise, bugcheck.
|
|
|
|
|
|
if (MiFreeDescriptor == MiFreeDescriptorLargest) {
|
|
|
|
KeBugCheckEx(INSTALL_MORE_MEMORY,
|
|
MmNumberOfPhysicalPages,
|
|
MmLowestPhysicalPage,
|
|
MmHighestPhysicalPage,
|
|
0);
|
|
|
|
return 0;
|
|
|
|
} else if (MiFreeDescriptor == MiFreeDescriptorLowMem) {
|
|
|
|
MiOldFreeDescriptorCount = MiFreeDescriptorLargest->PageCount;
|
|
MiOldFreeDescriptorBase = MiFreeDescriptorLargest->BasePage;
|
|
|
|
MiFreeDescriptor = MiFreeDescriptorLargest;
|
|
MiNumberOfPages = MiFreeDescriptorLargest->PageCount - 1;
|
|
MiNextPhysicalPage = MiFreeDescriptorLargest->BasePage;
|
|
|
|
MiFreeDescriptorLowMem->PageCount = 0;
|
|
|
|
return MiNextPhysicalPage++;
|
|
|
|
} else {
|
|
|
|
MiOldFreeDescriptorCount = MiFreeDescriptorLowMem->PageCount;
|
|
MiOldFreeDescriptorBase = MiFreeDescriptorLowMem->BasePage;
|
|
|
|
MiFreeDescriptor = MiFreeDescriptorLowMem;
|
|
MiNumberOfPages = MiFreeDescriptorLowMem->PageCount - 1;
|
|
MiNextPhysicalPage = MiFreeDescriptorLowMem->BasePage;
|
|
|
|
MiFreeDescriptorNonPaged->PageCount = 0;
|
|
|
|
return MiNextPhysicalPage++;
|
|
}
|
|
}
|
|
}
|
|
|
|
BOOLEAN
|
|
MiEnsureAvailablePagesInFreeDescriptor(
|
|
IN PFN_NUMBER Pages,
|
|
IN PFN_NUMBER MaxPage
|
|
)
|
|
{
|
|
if ((MiNumberOfPages < Pages) &&
|
|
(MiFreeDescriptor == MiFreeDescriptorNonPaged)) {
|
|
|
|
MiFreeDescriptor->BasePage = (ULONG)MiNextPhysicalPage;
|
|
MiFreeDescriptor->PageCount = (PFN_COUNT)MiNumberOfPages;
|
|
|
|
if (MiFreeDescriptor != MiFreeDescriptorLowMem) {
|
|
|
|
MiOldFreeDescriptorCount = MiFreeDescriptorLowMem->PageCount;
|
|
MiOldFreeDescriptorBase = MiFreeDescriptorLowMem->BasePage;
|
|
|
|
MiFreeDescriptor = MiFreeDescriptorLowMem;
|
|
MiNumberOfPages = MiFreeDescriptorLowMem->PageCount;
|
|
MiNextPhysicalPage = MiFreeDescriptorLowMem->BasePage;
|
|
|
|
} else if (MiFreeDescriptor != MiFreeDescriptorLargest) {
|
|
|
|
MiOldFreeDescriptorCount = MiFreeDescriptorLargest->PageCount;
|
|
MiOldFreeDescriptorBase = MiFreeDescriptorLargest->BasePage;
|
|
|
|
MiFreeDescriptor = MiFreeDescriptorLargest;
|
|
MiNumberOfPages = MiFreeDescriptorLargest->PageCount;
|
|
MiNextPhysicalPage = MiFreeDescriptorLargest->BasePage;
|
|
|
|
} else {
|
|
|
|
KeBugCheckEx(INSTALL_MORE_MEMORY,
|
|
MmNumberOfPhysicalPages,
|
|
MmLowestPhysicalPage,
|
|
MmHighestPhysicalPage,
|
|
1);
|
|
}
|
|
}
|
|
|
|
if (MaxPage < (MiFreeDescriptor->BasePage + Pages)) {
|
|
|
|
return FALSE;
|
|
|
|
} else {
|
|
|
|
return TRUE;
|
|
|
|
}
|
|
}
|
|
|
|
|
|
VOID
|
|
MiInitMachineDependent (
|
|
IN PLOADER_PARAMETER_BLOCK LoaderBlock
|
|
)
|
|
|
|
/*++
|
|
|
|
Routine Description:
|
|
|
|
This routine performs the necessary operations to enable virtual
|
|
memory. This includes building the page directory page, building
|
|
page table pages to map the code section, the data section, the'
|
|
stack section and the trap handler.
|
|
|
|
It also initializes the PFN database and populates the free list.
|
|
|
|
|
|
Arguments:
|
|
|
|
LoaderBlock - Supplies a pointer to the firmware setup loader block.
|
|
|
|
Return Value:
|
|
|
|
None.
|
|
|
|
Environment:
|
|
|
|
Kernel mode.
|
|
|
|
--*/
|
|
|
|
{
|
|
PMMPFN BasePfn;
|
|
PMMPFN BottomPfn;
|
|
PMMPFN TopPfn;
|
|
SIZE_T Range;
|
|
PFN_NUMBER i;
|
|
ULONG j;
|
|
PFN_NUMBER PdePageNumber;
|
|
PFN_NUMBER PdePage;
|
|
PFN_NUMBER PageFrameIndex;
|
|
PFN_NUMBER NextPhysicalPage;
|
|
SPFN_NUMBER PfnAllocation;
|
|
SPFN_NUMBER NumberOfPages;
|
|
SIZE_T MaxPool;
|
|
PEPROCESS CurrentProcess;
|
|
PFN_NUMBER MostFreePage = 0;
|
|
PFN_NUMBER MostFreeLowMem = 0;
|
|
PFN_NUMBER MostFreeNonPaged = 0;
|
|
PLIST_ENTRY NextMd;
|
|
PMEMORY_ALLOCATION_DESCRIPTOR MemoryDescriptor;
|
|
MMPTE TempPte;
|
|
PMMPTE PointerPde;
|
|
PMMPTE PointerPte;
|
|
PMMPTE StartPte;
|
|
PMMPTE LastPte;
|
|
PMMPTE EndPte;
|
|
PMMPTE Pde;
|
|
PMMPTE StartPde;
|
|
PMMPTE StartPpe;
|
|
PMMPTE EndPde;
|
|
PMMPTE FirstPpe;
|
|
PMMPFN Pfn1;
|
|
PMMPFN Pfn2;
|
|
ULONG First;
|
|
KIRQL OldIrql;
|
|
ULONG_PTR va;
|
|
SIZE_T Kseg0Size = 0;
|
|
PVOID NonPagedPoolStartVirtual;
|
|
ULONG i1;
|
|
ULONG i2;
|
|
PFN_NUMBER PhysicalPages;
|
|
ULONG_PTR HighestPageAddress;
|
|
ULONG_PTR Kseg3VhptSize;
|
|
PFN_NUMBER Kseg3VhptPn;
|
|
ULONG_PTR size;
|
|
ULONG_PTR PageSize;
|
|
PFN_NUMBER KernelStart;
|
|
PFN_NUMBER KernelEnd;
|
|
|
|
if (InitializationPhase == 1) {
|
|
|
|
MiMakeKernelPagesPermanent(LoaderBlock);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
// Initialize the kernel mapping info
|
|
|
|
|
|
MiNtoskrnlPhysicalBase = LoaderBlock->u.Ia64.ItrInfo[ITR_KERNEL_INDEX].PhysicalAddress;
|
|
MiNtoskrnlVirtualBase = LoaderBlock->u.Ia64.ItrInfo[ITR_KERNEL_INDEX].VirtualAddress;
|
|
MiNtoskrnlPageShift = LoaderBlock->u.Ia64.ItrInfo[ITR_KERNEL_INDEX].PageSize;
|
|
|
|
KernelStart = MiNtoskrnlPhysicalBase >> PAGE_SHIFT;
|
|
PageSize = (ULONG_PTR)1 << MiNtoskrnlPageShift;
|
|
KernelEnd = KernelStart + (PageSize >> PAGE_SHIFT);
|
|
|
|
|
|
// Initialize MmDebugPte and MmCrashDumpPte
|
|
|
|
|
|
MmDebugPte = MiGetPteAddress(MM_DEBUG_VA);
|
|
|
|
MmCrashDumpPte = MiGetPteAddress(MM_CRASH_DUMP_VA);
|
|
|
|
|
|
// Set TempPte to ValidKernelPte for the future use
|
|
|
|
|
|
TempPte = ValidKernelPte;
|
|
|
|
|
|
// Check the memory descriptor list from NT loader
|
|
|
|
|
|
MiCheckMemoryDescriptorList(LoaderBlock);
|
|
|
|
|
|
|
|
// Get the lower bound of the free physical memory and the
|
|
// number of physical pages by walking the memory descriptor lists.
|
|
|
|
|
|
NextMd = LoaderBlock->MemoryDescriptorListHead.Flink;
|
|
|
|
#if ADD_HIGH_MEMORY
|
|
LoaderBlock->MemoryDescriptorListHead.Flink = &MiHighMemoryDescriptor.ListEntry;
|
|
MiHighMemoryDescriptor.ListEntry.Flink = NextMd;
|
|
MiHighMemoryDescriptor.MemoryType = 5;
|
|
MiHighMemoryDescriptor.BasePage = 0x80000;
|
|
MiHighMemoryDescriptor.PageCount = 0x20000;
|
|
NextMd = &MiHighMemoryDescriptor.ListEntry;
|
|
#endif
|
|
|
|
while (NextMd != &LoaderBlock->MemoryDescriptorListHead) {
|
|
|
|
MemoryDescriptor = CONTAINING_RECORD(NextMd,
|
|
MEMORY_ALLOCATION_DESCRIPTOR,
|
|
ListEntry);
|
|
|
|
|
|
// This check results in /BURNMEMORY chunks not being counted.
|
|
|
|
|
|
if (MemoryDescriptor->MemoryType != LoaderBad) {
|
|
MmNumberOfPhysicalPages += MemoryDescriptor->PageCount;
|
|
}
|
|
|
|
if (MemoryDescriptor->BasePage < MmLowestPhysicalPage) {
|
|
MmLowestPhysicalPage = MemoryDescriptor->BasePage;
|
|
}
|
|
if ((MemoryDescriptor->BasePage + MemoryDescriptor->PageCount) >
|
|
MmHighestPhysicalPage) {
|
|
MmHighestPhysicalPage =
|
|
MemoryDescriptor->BasePage + MemoryDescriptor->PageCount -1;
|
|
}
|
|
|
|
|
|
// We assume at this stage (Gambit Port) that the minimum memory requirement
|
|
// for the iA EM is 32mb. There will be no memory hole right below the 16m
|
|
// boundary on EM machines unlike some x86 of PC platforms.
|
|
|
|
|
|
if ((MemoryDescriptor->MemoryType == LoaderFree) ||
|
|
(MemoryDescriptor->MemoryType == LoaderLoadedProgram) ||
|
|
(MemoryDescriptor->MemoryType == LoaderFirmwareTemporary) ||
|
|
(MemoryDescriptor->MemoryType == LoaderOsloaderStack)) {
|
|
|
|
if (MemoryDescriptor->PageCount > MostFreePage) {
|
|
MostFreePage = MemoryDescriptor->PageCount;
|
|
MiFreeDescriptorLargest = MemoryDescriptor;
|
|
}
|
|
|
|
if (MemoryDescriptor->BasePage < _x256mbnp) {
|
|
|
|
|
|
// This memory descriptor is below 16mb.
|
|
|
|
|
|
if ((MostFreeLowMem < MemoryDescriptor->PageCount) &&
|
|
(MostFreeLowMem < _x256mbnp - MemoryDescriptor->BasePage)) {
|
|
|
|
MostFreeLowMem = _x256mbnp - MemoryDescriptor->BasePage;
|
|
if (MemoryDescriptor->PageCount < MostFreeLowMem) {
|
|
MostFreeLowMem = MemoryDescriptor->PageCount;
|
|
}
|
|
|
|
MiFreeDescriptorLowMem = MemoryDescriptor;
|
|
}
|
|
}
|
|
|
|
if ((MemoryDescriptor->BasePage >= KernelStart) &&
|
|
(MemoryDescriptor->PageCount > MostFreeNonPaged)) {
|
|
|
|
MostFreeNonPaged = MemoryDescriptor->PageCount;
|
|
MiFreeDescriptorNonPaged = MemoryDescriptor;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
#if DBG
|
|
if (MiPrintMemoryDescriptors) {
|
|
DbgPrint("MemoryType = %x\n", MemoryDescriptor->MemoryType);
|
|
DbgPrint("BasePage = %p\n", (PFN_NUMBER)MemoryDescriptor->BasePage << PAGE_SHIFT);
|
|
DbgPrint("PageCount = %x\n\n", MemoryDescriptor->PageCount << PAGE_SHIFT);
|
|
}
|
|
#endif
|
|
|
|
NextMd = MemoryDescriptor->ListEntry.Flink;
|
|
}
|
|
|
|
MmHighestPossiblePhysicalPage = MmHighestPhysicalPage;
|
|
|
|
|
|
// If the number of physical pages is less than 16mb, then
|
|
// bugcheck. There is not enough memory to run the system.
|
|
|
|
|
|
if (MmNumberOfPhysicalPages < _x16mbnp) {
|
|
KeBugCheckEx (INSTALL_MORE_MEMORY,
|
|
MmNumberOfPhysicalPages,
|
|
MmLowestPhysicalPage,
|
|
MmHighestPhysicalPage,
|
|
2);
|
|
}
|
|
|
|
|
|
// initialize the next page allocation variable & structures
|
|
|
|
|
|
|
|
if (MiFreeDescriptorNonPaged != NULL) {
|
|
|
|
MiFreeDescriptor = MiFreeDescriptorNonPaged;
|
|
|
|
} else if (MiFreeDescriptorLowMem != NULL) {
|
|
|
|
MiFreeDescriptor = MiFreeDescriptorLowMem;
|
|
|
|
} else {
|
|
|
|
MiFreeDescriptor = MiFreeDescriptorLargest;
|
|
}
|
|
|
|
if (MiFreeDescriptorLowMem == NULL) {
|
|
|
|
MiKseg0Mapping = FALSE;
|
|
|
|
}
|
|
|
|
MiNextPhysicalPage = MiFreeDescriptor->BasePage;
|
|
MiNumberOfPages = MiFreeDescriptor->PageCount;
|
|
MiOldFreeDescriptorCount = MiFreeDescriptor->PageCount;
|
|
MiOldFreeDescriptorBase = MiFreeDescriptor->BasePage;
|
|
|
|
|
|
// Produce the size of the Kseg 0 space
|
|
|
|
|
|
if (MiKseg0Mapping == TRUE) {
|
|
|
|
|
|
MiKseg0Start = KSEG0_ADDRESS(MiNextPhysicalPage);
|
|
|
|
Kseg0Size = MiNextPhysicalPage + MiNumberOfPages;
|
|
|
|
if (Kseg0Size > MM_PAGES_IN_KSEG0) {
|
|
Kseg0Size = MM_PAGES_IN_KSEG0 - (_x16mbnp * 3);
|
|
}
|
|
|
|
Kseg0Size = Kseg0Size << PAGE_SHIFT;
|
|
|
|
} else {
|
|
|
|
MiKseg0Mapping = FALSE;
|
|
Kseg0Size = 0;
|
|
|
|
}
|
|
|
|
|
|
// Initialize VHPT registers
|
|
|
|
|
|
KeInitializeVhpt ((PVOID)PTA_BASE, MiImplVirtualMsb - PAGE_SHIFT + PTE_SHIFT);
|
|
|
|
|
|
// Build 1 to 1 virtual to physical mapping space
|
|
|
|
|
|
#if defined(KSEG_VHPT)
|
|
HighestPageAddress = ((ULONG_PTR)MmHighestPhysicalPage << PAGE_SHIFT);
|
|
|
|
HighestPageAddress = (HighestPageAddress + (ULONG_PTR)X64K - 1) & ~((ULONG_PTR)X64K - 1);
|
|
|
|
Kseg3VhptSize = HighestPageAddress >> 16 << PTE_SHIFT;
|
|
|
|
size = (Kseg3VhptSize-1) >> 12; // make the base page size 4KB
|
|
PageSize = 12;
|
|
|
|
while (size != 0) {
|
|
size = size >> 2;
|
|
PageSize += 2;
|
|
}
|
|
|
|
Kseg3VhptSize = (ULONG_PTR)1 << PageSize;
|
|
|
|
Kseg3VhptPages = (PFN_NUMBER)(Kseg3VhptSize >> PAGE_SHIFT);
|
|
|
|
if ((MiFreeDescriptor->BasePage & (_x16mbnp - 1)) != 0) {
|
|
KeBugCheckEx (INSTALL_MORE_MEMORY,
|
|
MmNumberOfPhysicalPages,
|
|
MmLowestPhysicalPage,
|
|
MmHighestPhysicalPage,
|
|
3);
|
|
}
|
|
|
|
MmKseg3VhptBase = MiNextPhysicalPage;
|
|
MiNextPhysicalPage += Kseg3VhptPages;
|
|
MiNumberOfPages -= Kseg3VhptPages;
|
|
|
|
if (MiNumberOfPages < 0) {
|
|
|
|
KeBugCheckEx(INSTALL_MORE_MEMORY,
|
|
MmNumberOfPhysicalPages,
|
|
MmLowestPhysicalPage,
|
|
MmHighestPhysicalPage,
|
|
4);
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
// Build a parent directory page table for the kernel space
|
|
|
|
|
|
PdePageNumber = (PFN_NUMBER)LoaderBlock->u.Ia64.PdrPage;
|
|
|
|
MmSystemParentTablePage = MiGetNextPhysicalPage();
|
|
|
|
RtlZeroMemory(KSEG_ADDRESS(MmSystemParentTablePage), PAGE_SIZE);
|
|
|
|
TempPte.u.Hard.PageFrameNumber = MmSystemParentTablePage;
|
|
|
|
MiSystemSelfMappedPte = TempPte;
|
|
|
|
KeFillFixedEntryTb((PHARDWARE_PTE)&TempPte, (PVOID)PDE_KTBASE, DTR_KTBASE_INDEX_TMP);
|
|
|
|
|
|
// Install the self-mapped Ppe entry into the kernel parent directory table
|
|
|
|
|
|
PointerPte = MiGetPteAddress(PDE_KTBASE);
|
|
|
|
MI_WRITE_VALID_PTE(PointerPte, TempPte);
|
|
|
|
|
|
// Install the kernel image Ppe entry into the parent directory table
|
|
|
|
|
|
PointerPte = MiGetPteAddress(PDE_KBASE);
|
|
|
|
TempPte.u.Hard.PageFrameNumber = PdePageNumber;
|
|
|
|
MI_WRITE_VALID_PTE(PointerPte, TempPte);
|
|
|
|
|
|
// Build a parent directory page table for the user space
|
|
|
|
|
|
NextPhysicalPage = MiGetNextPhysicalPage();
|
|
|
|
RtlZeroMemory (KSEG_ADDRESS(NextPhysicalPage), PAGE_SIZE);
|
|
|
|
TempPte.u.Hard.PageFrameNumber = NextPhysicalPage;
|
|
|
|
INITIALIZE_DIRECTORY_TABLE_BASE
|
|
(&(PsGetCurrentProcess()->Pcb.DirectoryTableBase[0]), NextPhysicalPage);
|
|
|
|
MiUserSelfMappedPte = TempPte;
|
|
|
|
KeFillFixedEntryTb((PHARDWARE_PTE)&TempPte, (PVOID)PDE_UTBASE, DTR_UTBASE_INDEX_TMP);
|
|
|
|
|
|
// Install the self-mapped entry into the user parent directory table
|
|
|
|
|
|
PointerPte = MiGetPteAddress(PDE_UTBASE);
|
|
|
|
MI_WRITE_VALID_PTE(PointerPte, TempPte);
|
|
|
|
|
|
// Build a parent directory page table for the hydra space
|
|
|
|
|
|
NextPhysicalPage = MiGetNextPhysicalPage();
|
|
|
|
RtlZeroMemory (KSEG_ADDRESS(NextPhysicalPage), PAGE_SIZE);
|
|
|
|
TempPte.u.Hard.PageFrameNumber = NextPhysicalPage;
|
|
|
|
INITIALIZE_DIRECTORY_TABLE_BASE
|
|
(&(PsGetCurrentProcess()->Pcb.SessionParentBase), NextPhysicalPage);
|
|
|
|
KeFillFixedEntryTb((PHARDWARE_PTE)&TempPte, (PVOID)PDE_STBASE, DTR_STBASE_INDEX);
|
|
|
|
|
|
// Install the self-mapped entry into the hydra parent directory table
|
|
|
|
|
|
PointerPte = MiGetPteAddress(PDE_STBASE);
|
|
|
|
MI_WRITE_VALID_PTE(PointerPte, TempPte);
|
|
|
|
|
|
// Build Vhpt table for KSEG3
|
|
|
|
|
|
#if defined(KSEG_VHPT)
|
|
va = KSEG0_ADDRESS(MmKseg3VhptBase);
|
|
|
|
RtlZeroMemory((PVOID)va, Kseg3VhptSize);
|
|
|
|
TempPte.u.Hard.PageFrameNumber = MmKseg3VhptBase;
|
|
|
|
PointerPte = MiGetKSegPteAddress(KSEG3_BASE);
|
|
|
|
KeFillFixedLargeEntryTb((PHARDWARE_PTE)&TempPte,
|
|
(PVOID)PointerPte,
|
|
PageSize,
|
|
DTR_KSEG3_INDEX);
|
|
|
|
MmKseg3VhptPs = PageSize;
|
|
#endif
|
|
|
|
|
|
// Build non-paged pool using the physical pages following the
|
|
// data page in which to build the pool from. Non-page pool grows
|
|
// from the high range of the virtual address space and expands
|
|
// downward.
|
|
|
|
// At this time non-paged pool is constructed so virtual addresses
|
|
// are also physically contiguous.
|
|
|
|
|
|
if ((MmSizeOfNonPagedPoolInBytes >> PAGE_SHIFT) >
|
|
(7 * (MmNumberOfPhysicalPages << 3))) {
|
|
|
|
|
|
// More than 7/8 of memory allocated to nonpagedpool, reset to 0.
|
|
|
|
|
|
MmSizeOfNonPagedPoolInBytes = 0;
|
|
}
|
|
|
|
if (MmSizeOfNonPagedPoolInBytes < MmMinimumNonPagedPoolSize) {
|
|
|
|
|
|
// Calculate the size of nonpaged pool.
|
|
// Use the minimum size, then for every MB about 4mb add extra
|
|
// pages.
|
|
|
|
|
|
MmSizeOfNonPagedPoolInBytes = MmMinimumNonPagedPoolSize;
|
|
|
|
MmSizeOfNonPagedPoolInBytes +=
|
|
((MmNumberOfPhysicalPages - _x16mbnp)/_x1mbnp) *
|
|
MmMinAdditionNonPagedPoolPerMb;
|
|
}
|
|
|
|
if (MmSizeOfNonPagedPoolInBytes > MM_MAX_INITIAL_NONPAGED_POOL) {
|
|
MmSizeOfNonPagedPoolInBytes = MM_MAX_INITIAL_NONPAGED_POOL;
|
|
}
|
|
|
|
|
|
// Align to page size boundary.
|
|
|
|
|
|
MmSizeOfNonPagedPoolInBytes &= ~(PAGE_SIZE - 1);
|
|
|
|
|
|
// Calculate the maximum size of pool.
|
|
|
|
|
|
if (MmMaximumNonPagedPoolInBytes == 0) {
|
|
|
|
|
|
// Calculate the size of nonpaged pool. If 4mb of less use
|
|
// the minimum size, then for every MB about 4mb add extra
|
|
// pages.
|
|
|
|
|
|
MmMaximumNonPagedPoolInBytes = MmDefaultMaximumNonPagedPool;
|
|
|
|
|
|
// Make sure enough expansion for pfn database exists.
|
|
|
|
|
|
MmMaximumNonPagedPoolInBytes += (MmHighestPhysicalPage * sizeof(MMPFN)) & ~(PAGE_SIZE-1);
|
|
|
|
MmMaximumNonPagedPoolInBytes +=
|
|
((MmNumberOfPhysicalPages - _x16mbnp)/_x1mbnp) *
|
|
MmMaxAdditionNonPagedPoolPerMb;
|
|
|
|
}
|
|
|
|
MaxPool = MmSizeOfNonPagedPoolInBytes + PAGE_SIZE * 16 +
|
|
((MmHighestPhysicalPage * sizeof(MMPFN)) & ~(PAGE_SIZE -1));
|
|
|
|
if (MmMaximumNonPagedPoolInBytes < MaxPool) {
|
|
MmMaximumNonPagedPoolInBytes = MaxPool;
|
|
}
|
|
|
|
if (MmMaximumNonPagedPoolInBytes > MM_MAX_ADDITIONAL_NONPAGED_POOL) {
|
|
MmMaximumNonPagedPoolInBytes = MM_MAX_ADDITIONAL_NONPAGED_POOL;
|
|
}
|
|
|
|
MmNonPagedPoolStart = (PVOID)((ULONG_PTR)MmNonPagedPoolEnd
|
|
- MmMaximumNonPagedPoolInBytes);
|
|
|
|
MmNonPagedPoolStart = (PVOID)PAGE_ALIGN(MmNonPagedPoolStart);
|
|
|
|
MmPageAlignedPoolBase[NonPagedPool] = MmNonPagedPoolStart;
|
|
|
|
|
|
// Calculate the starting PDE for the system PTE pool which is
|
|
// right below the nonpaged pool.
|
|
|
|
|
|
MmNonPagedSystemStart = (PVOID)(((ULONG_PTR)MmNonPagedPoolStart -
|
|
((MmNumberOfSystemPtes + 1) * PAGE_SIZE)) &
|
|
(~PAGE_DIRECTORY2_MASK));
|
|
|
|
if (MmNonPagedSystemStart < MM_LOWEST_NONPAGED_SYSTEM_START) {
|
|
MmNonPagedSystemStart = MM_LOWEST_NONPAGED_SYSTEM_START;
|
|
MmNumberOfSystemPtes = (ULONG)(((ULONG_PTR)MmNonPagedPoolStart -
|
|
(ULONG_PTR)MmNonPagedSystemStart) >> PAGE_SHIFT)-1;
|
|
ASSERT (MmNumberOfSystemPtes > 1000);
|
|
}
|
|
|
|
|
|
|
|
// Allocate a page directory and a pair of page table pages.
|
|
// Map the hyper space page directory page into the top level parent
|
|
// directory & the hyper space page table page into the page directory
|
|
// and map an additional page that will eventually be used for the
|
|
// working set list. Page tables after the first two are set up later
|
|
// on during individual process working set initialization.
|
|
|
|
// The working set list page will eventually be a part of hyper space.
|
|
// It is mapped into the second level page directory page so it can be
|
|
// zeroed and so it will be accounted for in the PFN database. Later
|
|
// the page will be unmapped, and its page frame number captured in the
|
|
// system process object.
|
|
|
|
|
|
TempPte = ValidPdePde;
|
|
StartPde = MiGetPdeAddress(HYPER_SPACE);
|
|
StartPpe = MiGetPpeAddress(HYPER_SPACE);
|
|
EndPde = StartPde;
|
|
First = (StartPpe->u.Hard.Valid == 0) ? TRUE : FALSE;
|
|
|
|
while (StartPde <= EndPde) {
|
|
|
|
if (First == TRUE || MiIsPteOnPdeBoundary(StartPde)) {
|
|
First = FALSE;
|
|
StartPpe = MiGetPteAddress(StartPde);
|
|
if (StartPpe->u.Hard.Valid == 0) {
|
|
ASSERT (StartPpe->u.Long == 0);
|
|
NextPhysicalPage = MiGetNextPhysicalPage();
|
|
RtlZeroMemory(KSEG_ADDRESS(NextPhysicalPage), PAGE_SIZE);
|
|
TempPte.u.Hard.PageFrameNumber = NextPhysicalPage;
|
|
MI_WRITE_VALID_PTE(StartPpe, TempPte);
|
|
}
|
|
}
|
|
NextPhysicalPage = MiGetNextPhysicalPage();
|
|
RtlZeroMemory(KSEG_ADDRESS(NextPhysicalPage), PAGE_SIZE);
|
|
TempPte.u.Hard.PageFrameNumber = NextPhysicalPage;
|
|
MI_WRITE_VALID_PTE(StartPde, TempPte);
|
|
StartPde += 1;
|
|
}
|
|
|
|
|
|
// Allocate page directory and page table pages for
|
|
// system PTEs and nonpaged pool.
|
|
|
|
|
|
TempPte = ValidKernelPte;
|
|
StartPde = MiGetPdeAddress(MmNonPagedSystemStart);
|
|
StartPpe = MiGetPpeAddress(MmNonPagedSystemStart);
|
|
EndPde = MiGetPdeAddress((ULONG_PTR)MmNonPagedPoolEnd - 1);
|
|
First = (StartPpe->u.Hard.Valid == 0) ? TRUE : FALSE;
|
|
|
|
while (StartPde <= EndPde) {
|
|
|
|
if (First == TRUE || MiIsPteOnPdeBoundary(StartPde)) {
|
|
First = FALSE;
|
|
StartPpe = MiGetPteAddress(StartPde);
|
|
if (StartPpe->u.Hard.Valid == 0) {
|
|
NextPhysicalPage = MiGetNextPhysicalPage();
|
|
RtlZeroMemory(KSEG_ADDRESS(NextPhysicalPage), PAGE_SIZE);
|
|
TempPte.u.Hard.PageFrameNumber = NextPhysicalPage;
|
|
MI_WRITE_VALID_PTE(StartPpe, TempPte);
|
|
}
|
|
}
|
|
|
|
if (StartPde->u.Hard.Valid == 0) {
|
|
NextPhysicalPage = MiGetNextPhysicalPage();
|
|
RtlZeroMemory(KSEG_ADDRESS(NextPhysicalPage), PAGE_SIZE);
|
|
TempPte.u.Hard.PageFrameNumber = NextPhysicalPage;
|
|
MI_WRITE_VALID_PTE(StartPde, TempPte);
|
|
}
|
|
StartPde += 1;
|
|
}
|
|
|
|
NonPagedPoolStartVirtual = MmNonPagedPoolStart;
|
|
|
|
// MiBuildPageTableForDrivers(LoaderBlock);
|
|
|
|
|
|
MiBuildPageTableForLoaderMemory(LoaderBlock);
|
|
|
|
MiRemoveLoaderSuperPages(LoaderBlock);
|
|
|
|
|
|
// remove the temporary super pages for the root page table pages,
|
|
// and remap again with DTR_KTBASE_INDEX and DTR_UTBASE_INDEX.
|
|
|
|
|
|
KiFlushFixedDataTb(FALSE, (PVOID)PDE_KTBASE);
|
|
KiFlushFixedDataTb(FALSE, (PVOID)PDE_UTBASE);
|
|
KeFillFixedEntryTb((PHARDWARE_PTE)&MiSystemSelfMappedPte, (PVOID)PDE_KTBASE, DTR_KTBASE_INDEX);
|
|
KeFillFixedEntryTb((PHARDWARE_PTE)&MiUserSelfMappedPte, (PVOID)PDE_UTBASE, DTR_UTBASE_INDEX);
|
|
|
|
if (MiKseg0Mapping == TRUE) {
|
|
|
|
|
|
// Fill the PDEs for KSEG0 space
|
|
|
|
|
|
StartPde = MiGetPdeAddress(MiKseg0Start);
|
|
EndPde = MiGetPdeAddress((PCHAR)KSEG0_BASE + Kseg0Size);
|
|
|
|
while (StartPde <= EndPde) {
|
|
|
|
if (StartPde->u.Hard.Valid == 0) {
|
|
NextPhysicalPage = MiGetNextPhysicalPage();
|
|
RtlZeroMemory(KSEG_ADDRESS(NextPhysicalPage), PAGE_SIZE);
|
|
TempPte.u.Hard.PageFrameNumber = NextPhysicalPage;
|
|
MI_WRITE_VALID_PTE(StartPde, TempPte);
|
|
}
|
|
StartPde++;
|
|
}
|
|
}
|
|
|
|
|
|
// Initial NonPagedPool allocation priorities
|
|
|
|
// 1. allocate pages from 16mb kernel image maping space and use the 16mb super page
|
|
// addresses
|
|
// 2. allocate pages from low memory space ( < 256mb) and use the KSEG0 addresses
|
|
// 3. allocate pages from the largest free memory descriptor list and use
|
|
// MM_NON_PAGED_POOL addresses.
|
|
|
|
|
|
if (MiKseg0Mapping == TRUE) {
|
|
|
|
|
|
// When KiKseg0Mapping is enabled, check to see if
|
|
|
|
|
|
MiKseg0Mapping =
|
|
MiEnsureAvailablePagesInFreeDescriptor(BYTES_TO_PAGES(MmSizeOfNonPagedPoolInBytes),
|
|
MM_PAGES_IN_KSEG0);
|
|
|
|
}
|
|
|
|
|
|
// Fill in the PTEs to cover the initial nonpaged pool. The physical
|
|
// page frames to cover this range were reserved earlier from the
|
|
// largest low memory free descriptor. The initial allocation is both
|
|
// physically and virtually contiguous.
|
|
|
|
|
|
PointerPte = MiGetPteAddress(MmNonPagedPoolStart);
|
|
LastPte = MiGetPteAddress((PCHAR)MmNonPagedPoolStart +
|
|
MmSizeOfNonPagedPoolInBytes - 1);
|
|
|
|
while (PointerPte <= LastPte) {
|
|
NextPhysicalPage = MiGetNextPhysicalPage();
|
|
TempPte.u.Hard.PageFrameNumber = NextPhysicalPage;
|
|
MI_WRITE_VALID_PTE(PointerPte, TempPte);
|
|
PointerPte++;
|
|
}
|
|
|
|
|
|
// Zero the remaining PTEs (if any) for the initial nonpaged pool up to
|
|
// the end of the current page table page.
|
|
|
|
|
|
while (!MiIsPteOnPdeBoundary (PointerPte)) {
|
|
MI_WRITE_INVALID_PTE(PointerPte, ZeroKernelPte);
|
|
PointerPte += 1;
|
|
}
|
|
|
|
|
|
// Convert the starting nonpaged pool address to KSEG0 space
|
|
// address and set the address of the initial nonpaged pool allocation.
|
|
|
|
|
|
if (MiKseg0Mapping == TRUE) {
|
|
|
|
|
|
|
|
// No need to get page table pages for these as we can reference
|
|
// them via large pages.
|
|
|
|
|
|
PointerPte = MiGetPteAddress (MmNonPagedPoolStart);
|
|
MmNonPagedPoolStart = KSEG0_ADDRESS(PointerPte->u.Hard.PageFrameNumber);
|
|
MmSubsectionBase = KSEG0_BASE;
|
|
|
|
StartPte = MiGetPteAddress(MmNonPagedPoolStart);
|
|
LastPte = MiGetPteAddress((PCHAR)MmNonPagedPoolStart +
|
|
MmSizeOfNonPagedPoolInBytes - 1);
|
|
|
|
MiKseg0Start = MiGetVirtualAddressMappedByPte(StartPte);
|
|
|
|
while (StartPte <= LastPte) {
|
|
|
|
|
|
// duplicating PTEs to map non initial non paged pool
|
|
// in the KSEG0 space.
|
|
|
|
|
|
MI_WRITE_VALID_PTE(StartPte, *PointerPte);
|
|
StartPte++;
|
|
PointerPte++;
|
|
}
|
|
|
|
MiKseg0End = MiGetVirtualAddressMappedByPte(LastPte);
|
|
|
|
} else {
|
|
|
|
MiKseg0Mapping = FALSE;
|
|
MmSubsectionBase = 0;
|
|
|
|
}
|
|
|
|
|
|
// As only the initial non paged pool is mapped through superpages,
|
|
// MmSubsectionToPage is always set to zero.
|
|
|
|
|
|
MmSubsectionTopPage = 0;
|
|
|
|
|
|
MmNonPagedPoolExpansionStart = (PVOID)((PCHAR)NonPagedPoolStartVirtual +
|
|
MmSizeOfNonPagedPoolInBytes);
|
|
|
|
MmPageAlignedPoolBase[NonPagedPool] = MmNonPagedPoolStart;
|
|
|
|
|
|
|
|
// Non-paged pages now exist, build the pool structures.
|
|
|
|
|
|
MiInitializeNonPagedPool ();
|
|
|
|
|
|
// Before Non-paged pool can be used, the PFN database must
|
|
// be built. This is due to the fact that the start and end of
|
|
// allocation bits for nonpaged pool are maintained in the
|
|
// PFN elements for the corresponding pages.
|
|
|
|
|
|
|
|
// Calculate the number of pages required from page zero to
|
|
// the highest page.
|
|
|
|
// Get secondary color value from registry.
|
|
|
|
|
|
MmSecondaryColors = MmSecondaryColors >> PAGE_SHIFT;
|
|
|
|
if (MmSecondaryColors == 0) {
|
|
MmSecondaryColors = MM_SECONDARY_COLORS_DEFAULT;
|
|
} else {
|
|
|
|
|
|
// Make sure value is power of two and within limits.
|
|
|
|
|
|
if (((MmSecondaryColors & (MmSecondaryColors -1)) != 0) ||
|
|
(MmSecondaryColors < MM_SECONDARY_COLORS_MIN) ||
|
|
(MmSecondaryColors > MM_SECONDARY_COLORS_MAX)) {
|
|
MmSecondaryColors = MM_SECONDARY_COLORS_DEFAULT;
|
|
}
|
|
}
|
|
|
|
MmSecondaryColorMask = MmSecondaryColors - 1;
|
|
|
|
|
|
// Get the number of secondary colors and add the arrary for tracking
|
|
// secondary colors to the end of the PFN database.
|
|
|
|
|
|
|
|
PfnAllocation = 1 + ((((MmHighestPhysicalPage + 1) * sizeof(MMPFN)) +
|
|
(MmSecondaryColors * sizeof(MMCOLOR_TABLES)*2))
|
|
>> PAGE_SHIFT);
|
|
|
|
|
|
if ((MmHighestPhysicalPage < _x4gbnp) &&
|
|
(MiKseg0Mapping == TRUE) &&
|
|
(MiEnsureAvailablePagesInFreeDescriptor(PfnAllocation,
|
|
MM_PAGES_IN_KSEG0))) {
|
|
|
|
// Allocate the PFN database in the superpage space
|
|
|
|
// Compute the address of the PFN by allocating the appropriate
|
|
// number of pages from the end of the free descriptor.
|
|
|
|
|
|
MmPfnDatabase = (PMMPFN)KSEG0_ADDRESS(MiNextPhysicalPage);
|
|
|
|
StartPte = MiGetPteAddress(MmPfnDatabase);
|
|
LastPte = MiGetPteAddress((PCHAR)MmPfnDatabase + (PfnAllocation << PAGE_SHIFT) - 1);
|
|
|
|
while (StartPte <= LastPte) {
|
|
TempPte.u.Hard.PageFrameNumber = MiGetNextPhysicalPage();
|
|
MI_WRITE_VALID_PTE(StartPte, TempPte);
|
|
StartPte++;
|
|
}
|
|
|
|
RtlZeroMemory(MmPfnDatabase, PfnAllocation * PAGE_SIZE);
|
|
MiKseg0End = MiGetVirtualAddressMappedByPte(LastPte);
|
|
|
|
} else {
|
|
|
|
|
|
// Calculate the start of the Pfn Database (it starts a physical
|
|
// page zero, even if the Lowest physical page is not zero).
|
|
|
|
|
|
MmPfnDatabase = (PMMPFN)MM_PFN_DATABASE_START;
|
|
|
|
|
|
// Go through the memory descriptors and for each physical page
|
|
// make the PFN database has a valid PTE to map it. This allows
|
|
// machines with sparse physical memory to have a minimal PFN
|
|
// database.
|
|
|
|
|
|
NextMd = LoaderBlock->MemoryDescriptorListHead.Flink;
|
|
|
|
while (NextMd != &LoaderBlock->MemoryDescriptorListHead) {
|
|
|
|
MemoryDescriptor = CONTAINING_RECORD(NextMd,
|
|
MEMORY_ALLOCATION_DESCRIPTOR,
|
|
ListEntry);
|
|
|
|
PointerPte = MiGetPteAddress (MI_PFN_ELEMENT(
|
|
MemoryDescriptor->BasePage));
|
|
|
|
LastPte = MiGetPteAddress (((PCHAR)(MI_PFN_ELEMENT(
|
|
MemoryDescriptor->BasePage +
|
|
MemoryDescriptor->PageCount))) - 1);
|
|
|
|
First = TRUE;
|
|
|
|
while (PointerPte <= LastPte) {
|
|
|
|
if (First == TRUE || MiIsPteOnPpeBoundary(PointerPte)) {
|
|
StartPpe = MiGetPdeAddress(PointerPte);
|
|
if (StartPpe->u.Hard.Valid == 0) {
|
|
ASSERT (StartPpe->u.Long == 0);
|
|
NextPhysicalPage = MiGetNextPhysicalPage();
|
|
RtlZeroMemory(KSEG_ADDRESS(NextPhysicalPage), PAGE_SIZE);
|
|
TempPte.u.Hard.PageFrameNumber = NextPhysicalPage;
|
|
MI_WRITE_VALID_PTE(StartPpe, TempPte);
|
|
}
|
|
}
|
|
|
|
if ((First == TRUE) || MiIsPteOnPdeBoundary(PointerPte)) {
|
|
First = FALSE;
|
|
StartPde = MiGetPteAddress(PointerPte);
|
|
if (StartPde->u.Hard.Valid == 0) {
|
|
NextPhysicalPage = MiGetNextPhysicalPage();
|
|
RtlZeroMemory(KSEG_ADDRESS(NextPhysicalPage), PAGE_SIZE);
|
|
TempPte.u.Hard.PageFrameNumber = NextPhysicalPage;
|
|
MI_WRITE_VALID_PTE(StartPde, TempPte);
|
|
}
|
|
}
|
|
|
|
if (PointerPte->u.Hard.Valid == 0) {
|
|
NextPhysicalPage = MiGetNextPhysicalPage();
|
|
RtlZeroMemory(KSEG_ADDRESS(NextPhysicalPage), PAGE_SIZE);
|
|
TempPte.u.Hard.PageFrameNumber = NextPhysicalPage;
|
|
MI_WRITE_VALID_PTE(PointerPte, TempPte);
|
|
}
|
|
|
|
PointerPte++;
|
|
}
|
|
NextMd = MemoryDescriptor->ListEntry.Flink;
|
|
}
|
|
}
|
|
|
|
if (MiKseg0Mapping == TRUE) {
|
|
|
|
|
|
// Try to convert to superpages
|
|
|
|
|
|
MiConvertToSuperPages(MiKseg0Start,
|
|
MiKseg0End,
|
|
_x1mb,
|
|
MM_PTE_1MB_PAGE);
|
|
|
|
MiConvertToSuperPages(MiKseg0Start,
|
|
MiKseg0End,
|
|
_x4mb,
|
|
MM_PTE_4MB_PAGE);
|
|
|
|
MiConvertToSuperPages(MiKseg0Start,
|
|
MiKseg0End,
|
|
_x16mb,
|
|
MM_PTE_16MB_PAGE);
|
|
|
|
MiConvertToSuperPages(MiKseg0Start,
|
|
MiKseg0End,
|
|
_x64mb,
|
|
MM_PTE_64MB_PAGE);
|
|
|
|
MiConvertToSuperPages(MiKseg0Start,
|
|
MiKseg0End,
|
|
_x256mb,
|
|
MM_PTE_256MB_PAGE);
|
|
|
|
}
|
|
|
|
|
|
|
|
// Initialize support for colored pages.
|
|
|
|
|
|
MmFreePagesByColor[0] = (PMMCOLOR_TABLES)
|
|
&MmPfnDatabase[MmHighestPhysicalPage + 1];
|
|
|
|
MmFreePagesByColor[1] = &MmFreePagesByColor[0][MmSecondaryColors];
|
|
|
|
|
|
// Make sure the PTEs are mapped.
|
|
|
|
|
|
if (!MI_IS_PHYSICAL_ADDRESS(MmFreePagesByColor[0])) {
|
|
PointerPte = MiGetPteAddress (&MmFreePagesByColor[0][0]);
|
|
|
|
LastPte = MiGetPteAddress (
|
|
(PVOID)((PCHAR)&MmFreePagesByColor[1][MmSecondaryColors] - 1));
|
|
|
|
while (PointerPte <= LastPte) {
|
|
if (PointerPte->u.Hard.Valid == 0) {
|
|
NextPhysicalPage = MiGetNextPhysicalPage();
|
|
RtlZeroMemory(KSEG_ADDRESS(NextPhysicalPage), PAGE_SIZE);
|
|
TempPte.u.Hard.PageFrameNumber = NextPhysicalPage;
|
|
MI_WRITE_VALID_PTE(PointerPte, TempPte);
|
|
}
|
|
PointerPte++;
|
|
}
|
|
}
|
|
|
|
for (i = 0; i < MmSecondaryColors; i++) {
|
|
MmFreePagesByColor[ZeroedPageList][i].Flink = MM_EMPTY_LIST;
|
|
MmFreePagesByColor[FreePageList][i].Flink = MM_EMPTY_LIST;
|
|
}
|
|
|
|
#if MM_MAXIMUM_NUMBER_OF_COLORS > 1
|
|
for (i = 0; i < MM_MAXIMUM_NUMBER_OF_COLORS; i++) {
|
|
MmFreePagesByPrimaryColor[ZeroedPageList][i].ListName = ZeroedPageList;
|
|
MmFreePagesByPrimaryColor[FreePageList][i].ListName = FreePageList;
|
|
MmFreePagesByPrimaryColor[ZeroedPageList][i].Flink = MM_EMPTY_LIST;
|
|
MmFreePagesByPrimaryColor[FreePageList][i].Flink = MM_EMPTY_LIST;
|
|
MmFreePagesByPrimaryColor[ZeroedPageList][i].Blink = MM_EMPTY_LIST;
|
|
MmFreePagesByPrimaryColor[FreePageList][i].Blink = MM_EMPTY_LIST;
|
|
}
|
|
#endif
|
|
|
|
|
|
// Go through the page table entries and for any page which is
|
|
// valid, update the corresponding PFN database element.
|
|
|
|
|
|
StartPde = MiGetPdeAddress (HYPER_SPACE);
|
|
StartPpe = MiGetPpeAddress (HYPER_SPACE);
|
|
EndPde = MiGetPdeAddress(HYPER_SPACE_END);
|
|
First = (StartPpe->u.Hard.Valid == 0) ? TRUE : FALSE;
|
|
|
|
while (StartPde <= EndPde) {
|
|
|
|
if (First == TRUE || MiIsPteOnPdeBoundary(StartPde)) {
|
|
First = FALSE;
|
|
StartPpe = MiGetPteAddress(StartPde);
|
|
if (StartPpe->u.Hard.Valid == 0) {
|
|
StartPpe += 1;
|
|
StartPde = MiGetVirtualAddressMappedByPte (StartPpe);
|
|
continue;
|
|
}
|
|
|
|
PdePage = MI_GET_PAGE_FRAME_FROM_PTE(StartPpe);
|
|
|
|
Pfn1 = MI_PFN_ELEMENT(PdePage);
|
|
Pfn1->PteFrame = MmSystemParentTablePage;
|
|
Pfn1->PteAddress = StartPde;
|
|
Pfn1->u2.ShareCount += 1;
|
|
Pfn1->u3.e2.ReferenceCount = 1;
|
|
Pfn1->u3.e1.PageLocation = ActiveAndValid;
|
|
Pfn1->u3.e1.PageColor =
|
|
MI_GET_COLOR_FROM_SECONDARY(GET_PAGE_COLOR_FROM_PTE(StartPpe));
|
|
}
|
|
|
|
|
|
if (StartPde->u.Hard.Valid == 1) {
|
|
PdePage = MI_GET_PAGE_FRAME_FROM_PTE(StartPde);
|
|
Pfn1 = MI_PFN_ELEMENT(PdePage);
|
|
PointerPde = MiGetPteAddress(StartPde);
|
|
Pfn1->PteFrame = MI_GET_PAGE_FRAME_FROM_PTE(PointerPde);
|
|
Pfn1->PteAddress = StartPde;
|
|
Pfn1->u2.ShareCount += 1;
|
|
Pfn1->u3.e2.ReferenceCount = 1;
|
|
Pfn1->u3.e1.PageLocation = ActiveAndValid;
|
|
Pfn1->u3.e1.PageColor =
|
|
MI_GET_COLOR_FROM_SECONDARY(GET_PAGE_COLOR_FROM_PTE (StartPde));
|
|
|
|
PointerPte = MiGetVirtualAddressMappedByPte(StartPde);
|
|
for (j = 0 ; j < PTE_PER_PAGE; j++) {
|
|
if (PointerPte->u.Hard.Valid == 1) {
|
|
|
|
Pfn1->u2.ShareCount += 1;
|
|
|
|
if (PointerPte->u.Hard.PageFrameNumber <=
|
|
MmHighestPhysicalPage) {
|
|
Pfn2 = MI_PFN_ELEMENT(PointerPte->u.Hard.PageFrameNumber);
|
|
Pfn2->PteFrame = PdePage;
|
|
Pfn2->PteAddress = PointerPte;
|
|
Pfn2->u2.ShareCount += 1;
|
|
Pfn2->u3.e2.ReferenceCount = 1;
|
|
Pfn2->u3.e1.PageLocation = ActiveAndValid;
|
|
Pfn2->u3.e1.PageColor =
|
|
MI_GET_COLOR_FROM_SECONDARY(
|
|
MI_GET_PAGE_COLOR_FROM_PTE (
|
|
PointerPte));
|
|
}
|
|
}
|
|
PointerPte += 1;
|
|
}
|
|
}
|
|
|
|
StartPde++;
|
|
}
|
|
|
|
|
|
// do it for the kernel space
|
|
|
|
|
|
StartPde = MiGetPdeAddress (KADDRESS_BASE);
|
|
StartPpe = MiGetPpeAddress (KADDRESS_BASE);
|
|
EndPde = MiGetPdeAddress(MM_SYSTEM_SPACE_END);
|
|
First = (StartPpe->u.Hard.Valid == 0) ? TRUE : FALSE;
|
|
|
|
while (StartPde <= EndPde) {
|
|
|
|
if (First == TRUE || MiIsPteOnPdeBoundary(StartPde)) {
|
|
First = FALSE;
|
|
StartPpe = MiGetPteAddress(StartPde);
|
|
if (StartPpe->u.Hard.Valid == 0) {
|
|
StartPpe += 1;
|
|
StartPde = MiGetVirtualAddressMappedByPte (StartPpe);
|
|
continue;
|
|
}
|
|
|
|
PdePage = MI_GET_PAGE_FRAME_FROM_PTE(StartPpe);
|
|
|
|
Pfn1 = MI_PFN_ELEMENT(PdePage);
|
|
Pfn1->PteFrame = MmSystemParentTablePage;
|
|
Pfn1->PteAddress = StartPde;
|
|
Pfn1->u2.ShareCount += 1;
|
|
Pfn1->u3.e2.ReferenceCount = 1;
|
|
Pfn1->u3.e1.PageLocation = ActiveAndValid;
|
|
Pfn1->u3.e1.PageColor =
|
|
MI_GET_COLOR_FROM_SECONDARY(GET_PAGE_COLOR_FROM_PTE(StartPpe));
|
|
}
|
|
|
|
if (StartPde->u.Hard.Valid == 1) {
|
|
PdePage = MI_GET_PAGE_FRAME_FROM_PTE(StartPde);
|
|
Pfn1 = MI_PFN_ELEMENT(PdePage);
|
|
PointerPde = MiGetPteAddress(StartPde);
|
|
Pfn1->PteFrame = MI_GET_PAGE_FRAME_FROM_PTE(PointerPde);
|
|
Pfn1->PteAddress = StartPde;
|
|
Pfn1->u2.ShareCount += 1;
|
|
Pfn1->u3.e2.ReferenceCount = 1;
|
|
Pfn1->u3.e1.PageLocation = ActiveAndValid;
|
|
Pfn1->u3.e1.PageColor =
|
|
MI_GET_COLOR_FROM_SECONDARY(GET_PAGE_COLOR_FROM_PTE (StartPde));
|
|
|
|
PointerPte = MiGetVirtualAddressMappedByPte(StartPde);
|
|
for (j = 0 ; j < PTE_PER_PAGE; j++) {
|
|
if (PointerPte->u.Hard.Valid == 1) {
|
|
|
|
Pfn1->u2.ShareCount += 1;
|
|
|
|
if (PointerPte->u.Hard.PageFrameNumber <=
|
|
MmHighestPhysicalPage) {
|
|
Pfn2 = MI_PFN_ELEMENT(PointerPte->u.Hard.PageFrameNumber);
|
|
Pfn2->PteFrame = PdePage;
|
|
Pfn2->PteAddress = PointerPte;
|
|
Pfn2->u2.ShareCount += 1;
|
|
Pfn2->u3.e2.ReferenceCount = 1;
|
|
Pfn2->u3.e1.PageLocation = ActiveAndValid;
|
|
Pfn2->u3.e1.PageColor =
|
|
MI_GET_COLOR_FROM_SECONDARY(
|
|
MI_GET_PAGE_COLOR_FROM_PTE (
|
|
PointerPte));
|
|
}
|
|
}
|
|
PointerPte += 1;
|
|
}
|
|
}
|
|
|
|
StartPde++;
|
|
}
|
|
|
|
|
|
// If page zero is still unused, mark it as in use. This is
|
|
// temporary as we want to find bugs where a physical page
|
|
// is specified as zero.
|
|
|
|
|
|
Pfn1 = &MmPfnDatabase[MmLowestPhysicalPage];
|
|
if (Pfn1->u3.e2.ReferenceCount == 0) {
|
|
|
|
|
|
// Make the reference count non-zero and point it into a
|
|
// page directory.
|
|
|
|
|
|
Pde = MiGetPdeAddress (KADDRESS_BASE + 0xb0000000);
|
|
PdePage = MI_GET_PAGE_FRAME_FROM_PTE(Pde);
|
|
Pfn1->PteFrame = PdePageNumber;
|
|
Pfn1->PteAddress = Pde;
|
|
Pfn1->u2.ShareCount += 1;
|
|
Pfn1->u3.e2.ReferenceCount = 1;
|
|
Pfn1->u3.e1.PageLocation = ActiveAndValid;
|
|
Pfn1->u3.e1.PageColor = MI_GET_COLOR_FROM_SECONDARY(
|
|
MI_GET_PAGE_COLOR_FROM_PTE (Pde));
|
|
}
|
|
|
|
// end of temporary set to physical page zero.
|
|
|
|
|
|
|
|
// Walk through the memory descriptors and add pages to the
|
|
// free list in the PFN database.
|
|
|
|
|
|
MiFreeDescriptor->PageCount -=
|
|
(PFN_COUNT)(MiNextPhysicalPage - MiOldFreeDescriptorBase);
|
|
|
|
|
|
// Until BasePage (arc.h) is changed to PFN_NUMBER, NextPhysicalPage
|
|
// needs (ULONG) cast.
|
|
|
|
|
|
MiFreeDescriptor->BasePage = (ULONG)MiNextPhysicalPage;
|
|
|
|
|
|
// making unused pages inside the kernel super page mapping unusable
|
|
// so that no one is going to reclaim it for uncached pages.
|
|
|
|
|
|
if (MiFreeDescriptorNonPaged != NULL) {
|
|
|
|
if (MiFreeDescriptorNonPaged->BasePage > KernelEnd) {
|
|
|
|
MiWasteStart = KernelEnd;
|
|
MiWasteEnd = KernelEnd;
|
|
|
|
|
|
} else if ((MiFreeDescriptorNonPaged->BasePage +
|
|
MiFreeDescriptorNonPaged->PageCount) > KernelEnd) {
|
|
|
|
MiWasteStart = MiFreeDescriptorNonPaged->BasePage;
|
|
MiWasteEnd = KernelEnd;
|
|
|
|
MiFreeDescriptorNonPaged->PageCount -=
|
|
(PFN_COUNT) (KernelEnd - MiFreeDescriptorNonPaged->BasePage);
|
|
|
|
MiFreeDescriptorNonPaged->BasePage = (ULONG) KernelEnd;
|
|
|
|
} else if (MiFreeDescriptorNonPaged->PageCount != 0) {
|
|
|
|
MiWasteStart = MiFreeDescriptorNonPaged->BasePage;
|
|
MiWasteEnd = MiWasteStart + MiFreeDescriptorNonPaged->PageCount;
|
|
MiFreeDescriptorNonPaged->PageCount = 0;
|
|
|
|
}
|
|
}
|
|
|
|
NextMd = LoaderBlock->MemoryDescriptorListHead.Flink;
|
|
|
|
while (NextMd != &LoaderBlock->MemoryDescriptorListHead) {
|
|
|
|
MemoryDescriptor = CONTAINING_RECORD(NextMd,
|
|
MEMORY_ALLOCATION_DESCRIPTOR,
|
|
ListEntry);
|
|
|
|
i = MemoryDescriptor->PageCount;
|
|
NextPhysicalPage = MemoryDescriptor->BasePage;
|
|
|
|
switch (MemoryDescriptor->MemoryType) {
|
|
case LoaderBad:
|
|
while (i != 0) {
|
|
MiInsertPageInList (MmPageLocationList[BadPageList],
|
|
NextPhysicalPage);
|
|
i -= 1;
|
|
NextPhysicalPage += 1;
|
|
}
|
|
break;
|
|
|
|
case LoaderFree:
|
|
case LoaderLoadedProgram:
|
|
case LoaderFirmwareTemporary:
|
|
case LoaderOsloaderStack:
|
|
|
|
Pfn1 = MI_PFN_ELEMENT (NextPhysicalPage);
|
|
while (i != 0) {
|
|
if (Pfn1->u3.e2.ReferenceCount == 0) {
|
|
|
|
|
|
// Set the PTE address to the physical page for
|
|
// virtual address alignment checking.
|
|
|
|
|
|
Pfn1->PteAddress = KSEG_ADDRESS (NextPhysicalPage);
|
|
MiInsertPageInList (MmPageLocationList[FreePageList],
|
|
NextPhysicalPage);
|
|
}
|
|
Pfn1++;
|
|
i -= 1;
|
|
NextPhysicalPage += 1;
|
|
}
|
|
break;
|
|
|
|
default:
|
|
|
|
PointerPte = KSEG_ADDRESS(NextPhysicalPage);
|
|
Pfn1 = MI_PFN_ELEMENT (NextPhysicalPage);
|
|
while (i != 0) {
|
|
|
|
|
|
// Set page as in use.
|
|
|
|
|
|
if (Pfn1->u3.e2.ReferenceCount == 0) {
|
|
Pfn1->PteFrame = PdePageNumber;
|
|
Pfn1->PteAddress = PointerPte;
|
|
Pfn1->u2.ShareCount += 1;
|
|
Pfn1->u3.e2.ReferenceCount = 1;
|
|
Pfn1->u3.e1.PageLocation = ActiveAndValid;
|
|
Pfn1->u3.e1.PageColor = MI_GET_COLOR_FROM_SECONDARY(
|
|
MI_GET_PAGE_COLOR_FROM_PTE (
|
|
PointerPte));
|
|
}
|
|
Pfn1++;
|
|
i -= 1;
|
|
NextPhysicalPage += 1;
|
|
PointerPte += 1;
|
|
}
|
|
break;
|
|
}
|
|
|
|
NextMd = MemoryDescriptor->ListEntry.Flink;
|
|
}
|
|
|
|
if (MI_IS_PHYSICAL_ADDRESS(MmPfnDatabase) == FALSE) {
|
|
|
|
|
|
// Indicate that the PFN database is allocated in NonPaged pool.
|
|
|
|
|
|
PointerPte = MiGetPteAddress (&MmPfnDatabase[MmLowestPhysicalPage]);
|
|
Pfn1 = MI_PFN_ELEMENT(PointerPte->u.Hard.PageFrameNumber);
|
|
Pfn1->u3.e1.StartOfAllocation = 1;
|
|
|
|
|
|
// Set the end of the allocation.
|
|
|
|
|
|
PointerPte = MiGetPteAddress (&MmPfnDatabase[MmHighestPhysicalPage]);
|
|
Pfn1 = MI_PFN_ELEMENT(PointerPte->u.Hard.PageFrameNumber);
|
|
Pfn1->u3.e1.EndOfAllocation = 1;
|
|
|
|
} else {
|
|
|
|
// The PFN database is allocated in the superpage space
|
|
|
|
// Mark all pfn entries for the pfn pages in use.
|
|
|
|
|
|
PageFrameIndex = MI_CONVERT_PHYSICAL_TO_PFN (MmPfnDatabase);
|
|
Pfn1 = MI_PFN_ELEMENT(PageFrameIndex);
|
|
do {
|
|
Pfn1->PteAddress = KSEG_ADDRESS(PageFrameIndex);
|
|
Pfn1->u3.e1.PageColor = 0;
|
|
Pfn1->u3.e2.ReferenceCount += 1;
|
|
Pfn1->u3.e1.PageLocation = ActiveAndValid;
|
|
PageFrameIndex += 1;
|
|
Pfn1 += 1;
|
|
PfnAllocation -= 1;
|
|
} while (PfnAllocation != 0);
|
|
|
|
|
|
// To avoid creating WB/UC/WC aliasing problem, we should not scan
|
|
// and add free pages to the free list.
|
|
|
|
#if 0
|
|
// Scan the PFN database backward for pages that are completely zero.
|
|
// These pages are unused and can be added to the free list
|
|
|
|
|
|
BottomPfn = MI_PFN_ELEMENT(MmHighestPhysicalPage);
|
|
do {
|
|
|
|
|
|
// Compute the address of the start of the page that is next
|
|
// lower in memory and scan backwards until that page address
|
|
// is reached or just crossed.
|
|
|
|
|
|
if (((ULONG_PTR)BottomPfn & (PAGE_SIZE - 1)) != 0) {
|
|
BasePfn = (PMMPFN)((ULONG_PTR)BottomPfn & ~(PAGE_SIZE - 1));
|
|
TopPfn = BottomPfn + 1;
|
|
|
|
} else {
|
|
BasePfn = (PMMPFN)((ULONG_PTR)BottomPfn - PAGE_SIZE);
|
|
TopPfn = BottomPfn;
|
|
}
|
|
|
|
while (BottomPfn > BasePfn) {
|
|
BottomPfn -= 1;
|
|
}
|
|
|
|
|
|
// If the entire range over which the PFN entries span is
|
|
// completely zero and the PFN entry that maps the page is
|
|
// not in the range, then add the page to the appropriate
|
|
// free list.
|
|
|
|
|
|
Range = (ULONG_PTR)TopPfn - (ULONG_PTR)BottomPfn;
|
|
if (RtlCompareMemoryUlong((PVOID)BottomPfn, Range, 0) == Range) {
|
|
|
|
|
|
// Set the PTE address to the physical page for virtual
|
|
// address alignment checking.
|
|
|
|
|
|
PageFrameIndex = MI_CONVERT_PHYSICAL_TO_PFN (BasePfn);
|
|
Pfn1 = MI_PFN_ELEMENT(PageFrameIndex);
|
|
|
|
ASSERT (Pfn1->u3.e2.ReferenceCount == 1);
|
|
ASSERT (Pfn1->PteAddress == KSEG_ADDRESS(PageFrameIndex));
|
|
Pfn1->u3.e2.ReferenceCount = 0;
|
|
PfnAllocation += 1;
|
|
Pfn1->PteAddress = (PMMPTE)((ULONG_PTR)PageFrameIndex << PTE_SHIFT);
|
|
Pfn1->u3.e1.PageColor = 0;
|
|
MiInsertPageInList(MmPageLocationList[FreePageList],
|
|
PageFrameIndex);
|
|
}
|
|
|
|
} while (BottomPfn > MmPfnDatabase);
|
|
#endif
|
|
|
|
}
|
|
|
|
|
|
|
|
// Indicate that nonpaged pool must succeed is allocated in
|
|
// nonpaged pool.
|
|
|
|
|
|
if (MI_IS_PHYSICAL_ADDRESS(MmNonPagedMustSucceed)) {
|
|
Pfn1 = MI_PFN_ELEMENT(MI_CONVERT_PHYSICAL_TO_PFN (MmNonPagedMustSucceed));
|
|
} else {
|
|
PointerPte = MiGetPteAddress(MmNonPagedMustSucceed);
|
|
Pfn1 = MI_PFN_ELEMENT(PointerPte->u.Hard.PageFrameNumber);
|
|
}
|
|
|
|
i = MmSizeOfNonPagedMustSucceed;
|
|
while ((LONG)i > 0) {
|
|
Pfn1->u3.e1.StartOfAllocation = 1;
|
|
Pfn1->u3.e1.EndOfAllocation = 1;
|
|
i -= PAGE_SIZE;
|
|
Pfn1 += 1;
|
|
}
|
|
|
|
#if 0
|
|
|
|
// Adjust the memory descriptors to indicate that free pool has
|
|
// been used for nonpaged pool creation.
|
|
|
|
|
|
MiFreeDescriptor->PageCount = (ULONG)MiOldFreeDescriptorCount;
|
|
|
|
|
|
// Until PagePage is defined to PFN_NUMBER, we need (ULONG) cast to
|
|
// remove warning.
|
|
|
|
|
|
MiFreeDescriptor->BasePage = (ULONG)MiOldFreeDescriptorBase;
|
|
|
|
#endif
|
|
|
|
// moved from above for pool hack routines...
|
|
KeInitializeSpinLock (&MmSystemSpaceLock);
|
|
|
|
KeInitializeSpinLock (&MmPfnLock);
|
|
|
|
|
|
// Initialize the nonpaged available PTEs for mapping I/O space
|
|
// and kernel stacks.
|
|
|
|
|
|
PointerPte = MiGetPteAddress (MmNonPagedSystemStart);
|
|
ASSERT (((ULONG_PTR)PointerPte & (PAGE_SIZE - 1)) == 0);
|
|
|
|
MmNumberOfSystemPtes = (ULONG)(MiGetPteAddress(NonPagedPoolStartVirtual) - PointerPte - 1);
|
|
|
|
MiInitializeSystemPtes (PointerPte, MmNumberOfSystemPtes, SystemPteSpace);
|
|
|
|
|
|
// Initialize the nonpaged pool.
|
|
|
|
|
|
InitializePool(NonPagedPool,0);
|
|
|
|
|
|
// Initialize memory management structures for the system process.
|
|
|
|
// Set the address of the first and last reserved PTE in hyper space.
|
|
|
|
|
|
MmFirstReservedMappingPte = MiGetPteAddress (FIRST_MAPPING_PTE);
|
|
MmLastReservedMappingPte = MiGetPteAddress (LAST_MAPPING_PTE);
|
|
|
|
MmWorkingSetList = WORKING_SET_LIST;
|
|
MmWsle = (PMMWSLE)((PUCHAR)WORKING_SET_LIST + sizeof(MMWSL));
|
|
|
|
|
|
// The PFN element for the page directory parent will be initialized
|
|
// a second time when the process address space is initialized. Therefore,
|
|
// the share count and the reference count must be set to zero.
|
|
|
|
|
|
Pfn1 = MI_PFN_ELEMENT(MI_GET_PAGE_FRAME_FROM_PTE((PMMPTE)PDE_SELFMAP));
|
|
Pfn1->u2.ShareCount = 0;
|
|
Pfn1->u3.e2.ReferenceCount = 0;
|
|
|
|
|
|
// The PFN element for the hyper space page directory page will be
|
|
// initialized a second time when the process address space is initialized.
|
|
// Therefore, the share count and the reference count must be set to zero.
|
|
|
|
|
|
PointerPte = MiGetPpeAddress(HYPER_SPACE);
|
|
Pfn1 = MI_PFN_ELEMENT(MI_GET_PAGE_FRAME_FROM_PTE(PointerPte));
|
|
Pfn1->u2.ShareCount = 0;
|
|
Pfn1->u3.e2.ReferenceCount = 0;
|
|
|
|
|
|
// The PFN elements for the hyper space page table page and working set list
|
|
// page will be initialized a second time when the process address space
|
|
// is initialized. Therefore, the share count and the reference must be
|
|
// set to zero.
|
|
|
|
|
|
StartPde = MiGetPdeAddress(HYPER_SPACE);
|
|
|
|
Pfn1 = MI_PFN_ELEMENT(MI_GET_PAGE_FRAME_FROM_PTE(StartPde));
|
|
Pfn1->u2.ShareCount = 0;
|
|
Pfn1->u3.e2.ReferenceCount = 0;
|
|
|
|
|
|
// Initialize this process's memory management structures including
|
|
// the working set list.
|
|
|
|
|
|
|
|
// The pfn element for the page directory has already been initialized,
|
|
// zero the reference count and the share count so they won't be
|
|
// wrong.
|
|
|
|
|
|
Pfn1 = MI_PFN_ELEMENT (PdePageNumber);
|
|
Pfn1->u2.ShareCount = 0;
|
|
Pfn1->u3.e2.ReferenceCount = 0;
|
|
|
|
CurrentProcess = PsGetCurrentProcess ();
|
|
|
|
|
|
// Get a page for the working set list and map it into the Page
|
|
// directory at the page after hyperspace.
|
|
|
|
|
|
PageFrameIndex = MiRemoveAnyPage (0);
|
|
|
|
CurrentProcess->WorkingSetPage = PageFrameIndex;
|
|
TempPte.u.Hard.PageFrameNumber = PageFrameIndex;
|
|
|
|
RtlZeroMemory (KSEG_ADDRESS(PageFrameIndex), PAGE_SIZE);
|
|
|
|
CurrentProcess->Vm.MaximumWorkingSetSize = (ULONG)MmSystemProcessWorkingSetMax;
|
|
CurrentProcess->Vm.MinimumWorkingSetSize = (ULONG)MmSystemProcessWorkingSetMin;
|
|
|
|
MmInitializeProcessAddressSpace (CurrentProcess,
|
|
(PEPROCESS)NULL,
|
|
(PVOID)NULL,
|
|
(PVOID)NULL);
|
|
|
|
|
|
// *PointerPde = ZeroPte;
|
|
|
|
|
|
// page after hyperspace should be reclaimed for system cache structure
|
|
|
|
|
|
KeFlushCurrentTb();
|
|
|
|
|
|
// Check to see if moving the secondary page structures to the end
|
|
// of the PFN database is a waste of memory. And if so, copy it
|
|
// to paged pool.
|
|
|
|
// If the PFN datbase ends on a page aligned boundary and the
|
|
// size of the two arrays is less than a page, free the page
|
|
// and allocate nonpagedpool for this.
|
|
|
|
|
|
if ((((ULONG_PTR)MmFreePagesByColor[0] & (PAGE_SIZE - 1)) == 0) &&
|
|
((MmSecondaryColors * 2 * sizeof(MMCOLOR_TABLES)) < PAGE_SIZE)) {
|
|
|
|
PMMCOLOR_TABLES c;
|
|
|
|
c = MmFreePagesByColor[0];
|
|
|
|
MmFreePagesByColor[0] = ExAllocatePoolWithTag (NonPagedPoolMustSucceed,
|
|
MmSecondaryColors * 2 * sizeof(MMCOLOR_TABLES),
|
|
' mM');
|
|
|
|
MmFreePagesByColor[1] = &MmFreePagesByColor[0][MmSecondaryColors];
|
|
|
|
RtlMoveMemory (MmFreePagesByColor[0],
|
|
c,
|
|
MmSecondaryColors * 2 * sizeof(MMCOLOR_TABLES));
|
|
|
|
|
|
// Free the page.
|
|
|
|
|
|
if (!MI_IS_PHYSICAL_ADDRESS(c)) {
|
|
PointerPte = MiGetPteAddress(c);
|
|
PageFrameIndex = MI_GET_PAGE_FRAME_FROM_PTE(PointerPte);
|
|
MI_WRITE_INVALID_PTE(PointerPte, ZeroKernelPte);
|
|
} else {
|
|
PageFrameIndex = MI_CONVERT_PHYSICAL_TO_PFN (c);
|
|
}
|
|
|
|
Pfn1 = MI_PFN_ELEMENT (PageFrameIndex);
|
|
ASSERT ((Pfn1->u2.ShareCount <= 1) && (Pfn1->u3.e2.ReferenceCount <= 1));
|
|
Pfn1->u2.ShareCount = 0;
|
|
Pfn1->u3.e2.ReferenceCount = 1;
|
|
MI_SET_PFN_DELETED (Pfn1);
|
|
#if DBG
|
|
Pfn1->u3.e1.PageLocation = StandbyPageList;
|
|
#endif //DBG
|
|
MiDecrementReferenceCount (PageFrameIndex);
|
|
}
|
|
|
|
return;
|
|
}
|
|
|
|
PVOID
|
|
MiGetKSegAddress (
|
|
PFN_NUMBER FrameNumber
|
|
)
|
|
/*++
|
|
|
|
Routine Description:
|
|
|
|
This function returns the KSEG3 address which maps the given physical page.
|
|
|
|
Arguments:
|
|
|
|
FrameNumber - Supplies the physical page number to get the KSEG3 address for
|
|
|
|
Return Value:
|
|
|
|
Virtual address mapped in KSEG3 space
|
|
|
|
TBS
|
|
|
|
--*/
|
|
{
|
|
PVOID Virtual;
|
|
PMMPTE PointerPte;
|
|
MMPTE TempPte;
|
|
|
|
ASSERT (FrameNumber <= MmHighestPhysicalPage);
|
|
|
|
Virtual = ((PVOID)(KSEG3_BASE | ((ULONG_PTR)(FrameNumber) << PAGE_SHIFT)));
|
|
|
|
#if defined(KSEG_VHPT)
|
|
PointerPte = MiGetKSegPteAddress (Virtual);
|
|
|
|
if (PointerPte->u.Long == 0) {
|
|
|
|
|
|
// if the VHPT entry for the KSEG3 address is still zero, build it here.
|
|
|
|
|
|
TempPte = ValidKernelPte;
|
|
TempPte.u.Hard.PageFrameNumber = FrameNumber;
|
|
MI_WRITE_VALID_PTE(PointerPte, TempPte);
|
|
|
|
__mf();
|
|
}
|
|
#endif
|
|
|
|
return (Virtual);
|
|
}
|
|
|
|
VOID
|
|
MiConvertToSuperPages(
|
|
IN PVOID StartVirtual,
|
|
IN PVOID EndVirtual,
|
|
IN SIZE_T PageSize,
|
|
IN ULONG PageShift
|
|
)
|
|
/*++
|
|
|
|
Routine Description:
|
|
|
|
This function makes contiguous non-paged memory use super pages rather than
|
|
using page tables.
|
|
|
|
Arguments:
|
|
|
|
StartVirtual - the start address of the region of pages to be mapped by
|
|
super pages.
|
|
|
|
EndVirtual - the end address of the region of pages to be mapped by super
|
|
pages.
|
|
|
|
Page Size - the page size to be used by the super page.
|
|
|
|
Page Shift - the page shift count to be used by the super page.
|
|
|
|
Return Value:
|
|
|
|
None.
|
|
|
|
--*/
|
|
{
|
|
ULONG_PTR VirtualAddress;
|
|
ULONG_PTR i;
|
|
ULONG_PTR NumberOfPte;
|
|
PMMPTE StartPte;
|
|
|
|
VirtualAddress = (ULONG_PTR) PAGE_ALIGN(StartVirtual);
|
|
i = VirtualAddress & (PageSize - 1);
|
|
|
|
if (i != 0) {
|
|
|
|
VirtualAddress = (VirtualAddress + PageSize - 1) & ~(PageSize - 1);
|
|
|
|
}
|
|
|
|
StartPte = MiGetPteAddress(VirtualAddress);
|
|
NumberOfPte = PageSize >> PAGE_SHIFT;
|
|
|
|
i = 0;
|
|
|
|
while (VirtualAddress <= (ULONG_PTR)EndVirtual) {
|
|
|
|
if (i == NumberOfPte) {
|
|
|
|
StartPte -= NumberOfPte;
|
|
|
|
for (i = 0; i < NumberOfPte; i++) {
|
|
|
|
StartPte->u.Hard.Valid = 0;
|
|
StartPte->u.Large.LargePage = 1;
|
|
StartPte->u.Large.PageSize = PageShift;
|
|
StartPte += 1;
|
|
}
|
|
|
|
i = 0;
|
|
}
|
|
|
|
i += 1;
|
|
StartPte += 1;
|
|
VirtualAddress += PAGE_SIZE;
|
|
}
|
|
}
|
|
|
|
VOID
|
|
MiConvertBackToStandardPages(
|
|
IN PVOID StartVirtual,
|
|
IN PVOID EndVirtual
|
|
)
|
|
/*++
|
|
|
|
Routine Description:
|
|
|
|
This function disables the use of the super pages.
|
|
|
|
Arguments:
|
|
|
|
StartVirtual - the start address of the region of pages to disable super pages.
|
|
super pages.
|
|
|
|
EndVirtual - the end address of the region of pages to disable super pages.
|
|
|
|
Return Value:
|
|
|
|
None.
|
|
|
|
--*/
|
|
{
|
|
ULONG_PTR VirtualAddress;
|
|
ULONG_PTR i;
|
|
ULONG_PTR NumberOfPte;
|
|
PMMPTE StartPte;
|
|
MMPTE TempPte;
|
|
|
|
VirtualAddress = (ULONG_PTR) PAGE_ALIGN(StartVirtual);
|
|
|
|
StartPte = MiGetPteAddress(VirtualAddress);
|
|
|
|
while (VirtualAddress <= (ULONG_PTR)EndVirtual) {
|
|
|
|
TempPte = *StartPte;
|
|
TempPte.u.Large.PageSize = 0;
|
|
TempPte.u.Large.PageSize = 0;
|
|
TempPte.u.Hard.Valid = 1;
|
|
MI_WRITE_VALID_PTE (StartPte, TempPte);
|
|
|
|
StartPte += 1;
|
|
VirtualAddress += PAGE_SIZE;
|
|
}
|
|
}
|
|
|
|
VOID
|
|
MiSweepCacheMachineDependent(
|
|
IN PVOID VirtualAddress,
|
|
IN SIZE_T Size,
|
|
IN MEMORY_CACHING_TYPE CacheType
|
|
)
|
|
/*++
|
|
|
|
Routine Description:
|
|
|
|
This function checks and perform appropriate cache flushing operations.
|
|
|
|
Arguments:
|
|
|
|
StartVirtual - the start address of the region of pages to be examined.
|
|
|
|
Size - the size of the region of pages
|
|
|
|
Cache - the new cache type
|
|
|
|
Return Value:
|
|
|
|
None.
|
|
|
|
--*/
|
|
{
|
|
PFN_NUMBER j;
|
|
PFN_NUMBER NumberOfPages;
|
|
KIRQL OldIrql;
|
|
PMMPTE PointerPte;
|
|
MMPTE TempPte;
|
|
|
|
NumberOfPages = COMPUTE_PAGES_SPANNED (VirtualAddress, Size);
|
|
VirtualAddress = PAGE_ALIGN(VirtualAddress);
|
|
Size = NumberOfPages * PAGE_SIZE;
|
|
|
|
KeSweepCacheRangeWithDrain(TRUE, VirtualAddress, (ULONG)Size);
|
|
|
|
if (CacheType == MmWriteCombined) {
|
|
PointerPte = MiGetPteAddress(VirtualAddress);
|
|
for (j = 0; j < NumberOfPages; j += 1) {
|
|
TempPte = *PointerPte;
|
|
MI_SET_PTE_WRITE_COMBINE (TempPte);
|
|
MI_WRITE_VALID_PTE (PointerPte, TempPte);
|
|
PointerPte += 1;
|
|
}
|
|
}
|
|
}
|
|
|
|
VOID
|
|
MiBuildPageTableForDrivers(
|
|
IN PLOADER_PARAMETER_BLOCK LoaderBlock
|
|
)
|
|
/*++
|
|
|
|
Routine Description:
|
|
|
|
This function builds page ables for loader loaded drivers.
|
|
|
|
Arguments:
|
|
|
|
LoaderBlock - Supplies the address of the loader block.
|
|
|
|
Return Value:
|
|
|
|
None.
|
|
|
|
--*/
|
|
{
|
|
PMMPTE StartPte;
|
|
PMMPTE EndPte;
|
|
PMMPTE StartPde;
|
|
PMMPTE StartPpe;
|
|
MMPTE TempPte;
|
|
ULONG First;
|
|
ULONG_PTR i;
|
|
PLIST_ENTRY NextEntry;
|
|
ULONG NumberOfLoaderPtes;
|
|
PFN_NUMBER NextPhysicalPage;
|
|
PVOID Va;
|
|
PLDR_DATA_TABLE_ENTRY DataTableEntry;
|
|
|
|
TempPte = ValidKernelPte;
|
|
TempPte.u.Long |= MM_PTE_EXECUTE;
|
|
|
|
i = 0;
|
|
NextEntry = LoaderBlock->LoadOrderListHead.Flink;
|
|
|
|
for ( ; NextEntry != &LoaderBlock->LoadOrderListHead; NextEntry = NextEntry->Flink) {
|
|
|
|
|
|
// As it is mapped through the translation registers, skip the kernel.
|
|
|
|
|
|
i += 1;
|
|
if (i <= 1) {
|
|
continue;
|
|
}
|
|
|
|
DataTableEntry = CONTAINING_RECORD(NextEntry,
|
|
LDR_DATA_TABLE_ENTRY,
|
|
InLoadOrderLinks);
|
|
|
|
NumberOfLoaderPtes = (ULONG)((ROUND_TO_PAGES(DataTableEntry->SizeOfImage)) >> PAGE_SHIFT);
|
|
|
|
Va = DataTableEntry->DllBase;
|
|
StartPte = MiGetPteAddress(Va);
|
|
EndPte = StartPte + NumberOfLoaderPtes;
|
|
|
|
First = TRUE;
|
|
|
|
while (StartPte <= EndPte) {
|
|
|
|
if (First == TRUE || MiIsPteOnPpeBoundary(StartPte)) {
|
|
StartPpe = MiGetPdeAddress(StartPte);
|
|
if (StartPpe->u.Hard.Valid == 0) {
|
|
ASSERT (StartPpe->u.Long == 0);
|
|
NextPhysicalPage = MiGetNextPhysicalPage();
|
|
RtlZeroMemory(KSEG_ADDRESS(NextPhysicalPage), PAGE_SIZE);
|
|
TempPte.u.Hard.PageFrameNumber = NextPhysicalPage;
|
|
MI_WRITE_VALID_PTE(StartPpe, TempPte);
|
|
}
|
|
}
|
|
|
|
if ((First == TRUE) || MiIsPteOnPdeBoundary(StartPte)) {
|
|
First = FALSE;
|
|
StartPde = MiGetPteAddress(StartPte);
|
|
if (StartPde->u.Hard.Valid == 0) {
|
|
NextPhysicalPage = MiGetNextPhysicalPage();
|
|
RtlZeroMemory(KSEG_ADDRESS(NextPhysicalPage), PAGE_SIZE);
|
|
TempPte.u.Hard.PageFrameNumber = NextPhysicalPage;
|
|
MI_WRITE_VALID_PTE(StartPde, TempPte);
|
|
}
|
|
}
|
|
|
|
TempPte.u.Hard.PageFrameNumber = MI_CONVERT_PHYSICAL_TO_PFN(Va);
|
|
MI_WRITE_VALID_PTE (StartPte, TempPte);
|
|
StartPte += 1;
|
|
Va = (PVOID)((ULONG_PTR)Va + PAGE_SIZE);
|
|
}
|
|
}
|
|
}
|
|
|
|
PVOID
|
|
MiConvertToLoaderVirtual(
|
|
IN PFN_NUMBER Page,
|
|
IN PLOADER_PARAMETER_BLOCK LoaderBlock
|
|
)
|
|
{
|
|
ULONG_PTR PageAddress = Page << PAGE_SHIFT;
|
|
PTR_INFO ItrInfo = &LoaderBlock->u.Ia64.ItrInfo[0];
|
|
|
|
|
|
if ((PageAddress >= ItrInfo[ITR_KERNEL_INDEX].PhysicalAddress) &&
|
|
(PageAddress <= ItrInfo[ITR_KERNEL_INDEX].PhysicalAddress +
|
|
((ULONG_PTR)1 << ItrInfo[ITR_KERNEL_INDEX].PageSize))) {
|
|
|
|
return (PVOID)(ItrInfo[ITR_KERNEL_INDEX].VirtualAddress +
|
|
(PageAddress - ItrInfo[ITR_KERNEL_INDEX].PhysicalAddress));
|
|
|
|
} else if ((PageAddress >= ItrInfo[ITR_DRIVER0_INDEX].PhysicalAddress) &&
|
|
(PageAddress <= ItrInfo[ITR_DRIVER0_INDEX].PhysicalAddress +
|
|
((ULONG_PTR)1 << ItrInfo[ITR_DRIVER0_INDEX].PageSize))) {
|
|
|
|
return (PVOID)(ItrInfo[ITR_DRIVER0_INDEX].VirtualAddress +
|
|
(PageAddress - ItrInfo[ITR_DRIVER0_INDEX].PhysicalAddress));
|
|
|
|
} else if ((PageAddress >= ItrInfo[ITR_DRIVER1_INDEX].PhysicalAddress) &&
|
|
(PageAddress <= ItrInfo[ITR_DRIVER1_INDEX].PhysicalAddress +
|
|
((ULONG_PTR)1 << ItrInfo[ITR_DRIVER1_INDEX].PageSize))) {
|
|
|
|
return (PVOID)(ItrInfo[ITR_DRIVER1_INDEX].VirtualAddress +
|
|
(PageAddress - ItrInfo[ITR_DRIVER1_INDEX].PhysicalAddress));
|
|
|
|
} else {
|
|
|
|
KeBugCheckEx (MEMORY_MANAGEMENT,
|
|
0x01010101,
|
|
PageAddress,
|
|
(ULONG_PTR)&ItrInfo[0],
|
|
(ULONG_PTR)LoaderBlock);
|
|
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
|
|
VOID
|
|
MiBuildPageTableForLoaderMemory(
|
|
IN PLOADER_PARAMETER_BLOCK LoaderBlock
|
|
)
|
|
/*++
|
|
|
|
Routine Description:
|
|
|
|
This function builds page ables for loader loaded drivers and loader
|
|
allocated memory.
|
|
|
|
Arguments:
|
|
|
|
LoaderBlock - Supplies the address of the loader block.
|
|
|
|
Return Value:
|
|
|
|
None.
|
|
|
|
--*/
|
|
{
|
|
PMMPTE StartPte;
|
|
PMMPTE EndPte;
|
|
PMMPTE StartPde;
|
|
PMMPTE StartPpe;
|
|
MMPTE TempPte;
|
|
ULONG First;
|
|
PLIST_ENTRY NextEntry;
|
|
PFN_NUMBER NextPhysicalPage;
|
|
PVOID Va;
|
|
PFN_NUMBER PfnNumber;
|
|
PMEMORY_ALLOCATION_DESCRIPTOR MemoryDescriptor;
|
|
|
|
TempPte = ValidKernelPte;
|
|
NextEntry = LoaderBlock->MemoryDescriptorListHead.Flink;
|
|
|
|
for ( ; NextEntry != &LoaderBlock->MemoryDescriptorListHead; NextEntry = NextEntry->Flink) {
|
|
|
|
MemoryDescriptor = CONTAINING_RECORD(NextEntry,
|
|
MEMORY_ALLOCATION_DESCRIPTOR,
|
|
ListEntry);
|
|
|
|
if ((MemoryDescriptor->MemoryType == LoaderOsloaderHeap) ||
|
|
(MemoryDescriptor->MemoryType == LoaderRegistryData) ||
|
|
(MemoryDescriptor->MemoryType == LoaderNlsData) ||
|
|
(MemoryDescriptor->MemoryType == LoaderStartupDpcStack) ||
|
|
(MemoryDescriptor->MemoryType == LoaderStartupKernelStack) ||
|
|
(MemoryDescriptor->MemoryType == LoaderStartupPanicStack) ||
|
|
(MemoryDescriptor->MemoryType == LoaderStartupPdrPage) ||
|
|
(MemoryDescriptor->MemoryType == LoaderMemoryData)) {
|
|
|
|
TempPte.u.Hard.Execute = 0;
|
|
|
|
} else if ((MemoryDescriptor->MemoryType == LoaderSystemCode) ||
|
|
(MemoryDescriptor->MemoryType == LoaderHalCode) ||
|
|
(MemoryDescriptor->MemoryType == LoaderBootDriver) ||
|
|
(MemoryDescriptor->MemoryType == LoaderStartupDpcStack)) {
|
|
|
|
TempPte.u.Hard.Execute = 1;
|
|
|
|
} else {
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
PfnNumber = MemoryDescriptor->BasePage;
|
|
Va = MiConvertToLoaderVirtual(MemoryDescriptor->BasePage, LoaderBlock);
|
|
|
|
StartPte = MiGetPteAddress(Va);
|
|
EndPte = StartPte + MemoryDescriptor->PageCount;
|
|
|
|
First = TRUE;
|
|
|
|
while (StartPte <= EndPte) {
|
|
|
|
if (First == TRUE || MiIsPteOnPpeBoundary(StartPte)) {
|
|
StartPpe = MiGetPdeAddress(StartPte);
|
|
if (StartPpe->u.Hard.Valid == 0) {
|
|
ASSERT (StartPpe->u.Long == 0);
|
|
NextPhysicalPage = MiGetNextPhysicalPage();
|
|
RtlZeroMemory(KSEG_ADDRESS(NextPhysicalPage), PAGE_SIZE);
|
|
TempPte.u.Hard.PageFrameNumber = NextPhysicalPage;
|
|
MI_WRITE_VALID_PTE(StartPpe, TempPte);
|
|
}
|
|
}
|
|
|
|
if ((First == TRUE) || MiIsPteOnPdeBoundary(StartPte)) {
|
|
First = FALSE;
|
|
StartPde = MiGetPteAddress(StartPte);
|
|
if (StartPde->u.Hard.Valid == 0) {
|
|
NextPhysicalPage = MiGetNextPhysicalPage();
|
|
RtlZeroMemory(KSEG_ADDRESS(NextPhysicalPage), PAGE_SIZE);
|
|
TempPte.u.Hard.PageFrameNumber = NextPhysicalPage;
|
|
MI_WRITE_VALID_PTE(StartPde, TempPte);
|
|
}
|
|
}
|
|
|
|
TempPte.u.Hard.PageFrameNumber = PfnNumber;
|
|
MI_WRITE_VALID_PTE (StartPte, TempPte);
|
|
StartPte += 1;
|
|
PfnNumber += 1;
|
|
Va = (PVOID)((ULONG_PTR)Va + PAGE_SIZE);
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
VOID MiRemoveLoaderSuperPages(IN PLOADER_PARAMETER_BLOCK LoaderBlock)
|
|
{
|
|
// remove the super pages for the boot drivers
|
|
KiFlushFixedInstTb(FALSE, LoaderBlock->u.Ia64.ItrInfo[ITR_DRIVER0_INDEX].VirtualAddress);
|
|
KiFlushFixedInstTb(FALSE, LoaderBlock->u.Ia64.ItrInfo[ITR_DRIVER1_INDEX].VirtualAddress);
|
|
KiFlushFixedDataTb(FALSE, LoaderBlock->u.Ia64.DtrInfo[DTR_DRIVER0_INDEX].VirtualAddress);
|
|
KiFlushFixedDataTb(FALSE, LoaderBlock->u.Ia64.DtrInfo[DTR_DRIVER1_INDEX].VirtualAddress);
|
|
}
|
|
|
|
|
|
VOID MiMakeKernelPagesPermanent(IN PLOADER_PARAMETER_BLOCK LoaderBlock)
|
|
{
|
|
PFN_NUMBER KernelStart;
|
|
PFN_NUMBER KernelEnd;
|
|
ULONG_PTR PageSize;
|
|
PLIST_ENTRY NextEntry;
|
|
PMEMORY_ALLOCATION_DESCRIPTOR MemoryDescriptor;
|
|
|
|
KernelStart = MiNtoskrnlPhysicalBase >> PAGE_SHIFT;
|
|
PageSize = (ULONG_PTR)1 << MiNtoskrnlPageShift;
|
|
KernelEnd = KernelStart + (PageSize >> PAGE_SHIFT);
|
|
|
|
NextEntry = LoaderBlock->MemoryDescriptorListHead.Flink;
|
|
|
|
for ( ; NextEntry != &LoaderBlock->MemoryDescriptorListHead; NextEntry = NextEntry->Flink) {
|
|
MemoryDescriptor = CONTAINING_RECORD(NextEntry, MEMORY_ALLOCATION_DESCRIPTOR, ListEntry);
|
|
if (((MemoryDescriptor->BasePage >= KernelStart) && (MemoryDescriptor->BasePage < KernelEnd)) &&
|
|
((MemoryDescriptor->MemoryType == LoaderOsloaderHeap) || (MemoryDescriptor->MemoryType == LoaderRegistryData) || (MemoryDescriptor->MemoryType == LoaderNlsData))) {
|
|
// prevent these pages from being reclaimed later
|
|
MemoryDescriptor->MemoryType = LoaderSpecialMemory;
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
VOID MiCheckMemoryDescriptorList(IN PLOADER_PARAMETER_BLOCK LoaderBlock)
|
|
{
|
|
PFN_NUMBER KernelStart;
|
|
PFN_NUMBER KernelEnd;
|
|
ULONG_PTR PageSize;
|
|
PLIST_ENTRY NextEntry;
|
|
PLIST_ENTRY PreviousEntry;
|
|
PMEMORY_ALLOCATION_DESCRIPTOR MemoryDescriptor;
|
|
PMEMORY_ALLOCATION_DESCRIPTOR PreviousMemoryDescriptor;
|
|
|
|
KernelStart = MiNtoskrnlPhysicalBase >> PAGE_SHIFT;
|
|
PageSize = (ULONG_PTR)1 << MiNtoskrnlPageShift;
|
|
KernelEnd = KernelStart + (PageSize >> PAGE_SHIFT);
|
|
|
|
PreviousMemoryDescriptor = NULL;
|
|
PreviousEntry = NULL;
|
|
|
|
NextEntry = LoaderBlock->MemoryDescriptorListHead.Flink;
|
|
|
|
for ( ; NextEntry != &LoaderBlock->MemoryDescriptorListHead; NextEntry = NextEntry->Flink) {
|
|
MemoryDescriptor = CONTAINING_RECORD(NextEntry, MEMORY_ALLOCATION_DESCRIPTOR, ListEntry);
|
|
|
|
#if DBG
|
|
if (MiPrintMemoryDescriptors) {
|
|
DbgPrint("MemoryType = %x\n", MemoryDescriptor->MemoryType);
|
|
DbgPrint("BasePage = %p\n", (PFN_NUMBER)MemoryDescriptor->BasePage << PAGE_SHIFT);
|
|
DbgPrint("PageCount = %x\n\n", MemoryDescriptor->PageCount << PAGE_SHIFT);
|
|
}
|
|
#endif
|
|
if ((MemoryDescriptor->BasePage >= KernelStart) && (MemoryDescriptor->BasePage + MemoryDescriptor->PageCount <= KernelEnd)) {
|
|
if (MemoryDescriptor->MemoryType == LoaderSystemBlock) {
|
|
MemoryDescriptor->MemoryType = LoaderFirmwareTemporary;
|
|
} else if (MemoryDescriptor->MemoryType == LoaderSpecialMemory) {
|
|
MemoryDescriptor->MemoryType = LoaderFirmwareTemporary;
|
|
}
|
|
}
|
|
|
|
if ((PreviousMemoryDescriptor != NULL) &&
|
|
(MemoryDescriptor->MemoryType == PreviousMemoryDescriptor->MemoryType) &&
|
|
(MemoryDescriptor->BasePage == (PreviousMemoryDescriptor->BasePage + PreviousMemoryDescriptor->PageCount))) {
|
|
PreviousMemoryDescriptor->PageCount += MemoryDescriptor->PageCount;
|
|
PreviousEntry->Flink = NextEntry->Flink;
|
|
} else {
|
|
PreviousMemoryDescriptor = MemoryDescriptor;
|
|
PreviousEntry = NextEntry;
|
|
}
|
|
}
|
|
} |