mirror of
https://git.naxdy.org/Mirror/Ryujinx.git
synced 2025-03-25 11:40:19 +00:00
* Add host tracked memory manager mode * Skipping flush is no longer needed * Formatting + revert unrelated change * LightningJit: Ensure that dest register is saved for load ops that do partial updates * avoid allocations when doing address space lookup Add missing improvement * IsRmwMemory -> IsPartialRegisterUpdateMemory * Ensure we iterate all private allocations in range * PR feedback and potential fixes * Simplified bridges a lot * Skip calling SignalMappingChanged if Guest is true * Late map bridge too * Force address masking for prefetch instructions * Reprotection for bridges * Move partition list validation to separate debug method * Move host tracked related classes to HostTracked folder * New HostTracked namespace * Move host tracked modes to the end of enum to avoid PPTC invalidation --------- Co-authored-by: riperiperi <rhy3756547@hotmail.com>
256 lines
8 KiB
C#
256 lines
8 KiB
C#
using Ryujinx.Horizon.Common;
|
|
using Ryujinx.Memory;
|
|
using Ryujinx.Memory.Range;
|
|
using System;
|
|
using System.Collections.Generic;
|
|
using System.Diagnostics;
|
|
|
|
namespace Ryujinx.HLE.HOS.Kernel.Memory
|
|
{
|
|
class KPageTable : KPageTableBase
|
|
{
|
|
private readonly IVirtualMemoryManager _cpuMemory;
|
|
|
|
protected override bool Supports4KBPages => _cpuMemory.Supports4KBPages;
|
|
|
|
public KPageTable(KernelContext context, IVirtualMemoryManager cpuMemory, ulong reservedAddressSpaceSize) : base(context, reservedAddressSpaceSize)
|
|
{
|
|
_cpuMemory = cpuMemory;
|
|
}
|
|
|
|
/// <inheritdoc/>
|
|
protected override IEnumerable<HostMemoryRange> GetHostRegions(ulong va, ulong size)
|
|
{
|
|
return _cpuMemory.GetHostRegions(va, size);
|
|
}
|
|
|
|
/// <inheritdoc/>
|
|
protected override void GetPhysicalRegions(ulong va, ulong size, KPageList pageList)
|
|
{
|
|
var ranges = _cpuMemory.GetPhysicalRegions(va, size);
|
|
foreach (var range in ranges)
|
|
{
|
|
pageList.AddRange(range.Address + DramMemoryMap.DramBase, range.Size / PageSize);
|
|
}
|
|
}
|
|
|
|
/// <inheritdoc/>
|
|
protected override ReadOnlySpan<byte> GetSpan(ulong va, int size)
|
|
{
|
|
return _cpuMemory.GetSpan(va, size);
|
|
}
|
|
|
|
/// <inheritdoc/>
|
|
protected override Result MapMemory(ulong src, ulong dst, ulong pagesCount, KMemoryPermission oldSrcPermission, KMemoryPermission newDstPermission)
|
|
{
|
|
KPageList pageList = new();
|
|
GetPhysicalRegions(src, pagesCount * PageSize, pageList);
|
|
|
|
Result result = Reprotect(src, pagesCount, KMemoryPermission.None);
|
|
|
|
if (result != Result.Success)
|
|
{
|
|
return result;
|
|
}
|
|
|
|
result = MapPages(dst, pageList, newDstPermission, MemoryMapFlags.Private, false, 0);
|
|
|
|
if (result != Result.Success)
|
|
{
|
|
Result reprotectResult = Reprotect(src, pagesCount, oldSrcPermission);
|
|
Debug.Assert(reprotectResult == Result.Success);
|
|
}
|
|
|
|
return result;
|
|
}
|
|
|
|
/// <inheritdoc/>
|
|
protected override Result UnmapMemory(ulong dst, ulong src, ulong pagesCount, KMemoryPermission oldDstPermission, KMemoryPermission newSrcPermission)
|
|
{
|
|
ulong size = pagesCount * PageSize;
|
|
|
|
KPageList srcPageList = new();
|
|
KPageList dstPageList = new();
|
|
|
|
GetPhysicalRegions(src, size, srcPageList);
|
|
GetPhysicalRegions(dst, size, dstPageList);
|
|
|
|
if (!dstPageList.IsEqual(srcPageList))
|
|
{
|
|
return KernelResult.InvalidMemRange;
|
|
}
|
|
|
|
Result result = Unmap(dst, pagesCount);
|
|
|
|
if (result != Result.Success)
|
|
{
|
|
return result;
|
|
}
|
|
|
|
result = Reprotect(src, pagesCount, newSrcPermission);
|
|
|
|
if (result != Result.Success)
|
|
{
|
|
Result mapResult = MapPages(dst, dstPageList, oldDstPermission, MemoryMapFlags.Private, false, 0);
|
|
Debug.Assert(mapResult == Result.Success);
|
|
}
|
|
|
|
return result;
|
|
}
|
|
|
|
/// <inheritdoc/>
|
|
protected override Result MapPages(
|
|
ulong dstVa,
|
|
ulong pagesCount,
|
|
ulong srcPa,
|
|
KMemoryPermission permission,
|
|
MemoryMapFlags flags,
|
|
bool shouldFillPages,
|
|
byte fillValue)
|
|
{
|
|
ulong size = pagesCount * PageSize;
|
|
|
|
Context.CommitMemory(srcPa - DramMemoryMap.DramBase, size);
|
|
|
|
_cpuMemory.Map(dstVa, srcPa - DramMemoryMap.DramBase, size, flags);
|
|
|
|
if (DramMemoryMap.IsHeapPhysicalAddress(srcPa))
|
|
{
|
|
Context.MemoryManager.IncrementPagesReferenceCount(srcPa, pagesCount);
|
|
}
|
|
|
|
if (shouldFillPages)
|
|
{
|
|
_cpuMemory.Fill(dstVa, size, fillValue);
|
|
}
|
|
|
|
return Result.Success;
|
|
}
|
|
|
|
/// <inheritdoc/>
|
|
protected override Result MapPages(
|
|
ulong address,
|
|
KPageList pageList,
|
|
KMemoryPermission permission,
|
|
MemoryMapFlags flags,
|
|
bool shouldFillPages,
|
|
byte fillValue)
|
|
{
|
|
using var scopedPageList = new KScopedPageList(Context.MemoryManager, pageList);
|
|
|
|
ulong currentVa = address;
|
|
|
|
foreach (var pageNode in pageList)
|
|
{
|
|
ulong addr = pageNode.Address - DramMemoryMap.DramBase;
|
|
ulong size = pageNode.PagesCount * PageSize;
|
|
|
|
Context.CommitMemory(addr, size);
|
|
|
|
_cpuMemory.Map(currentVa, addr, size, flags);
|
|
|
|
if (shouldFillPages)
|
|
{
|
|
_cpuMemory.Fill(currentVa, size, fillValue);
|
|
}
|
|
|
|
currentVa += size;
|
|
}
|
|
|
|
scopedPageList.SignalSuccess();
|
|
|
|
return Result.Success;
|
|
}
|
|
|
|
/// <inheritdoc/>
|
|
protected override Result MapForeign(IEnumerable<HostMemoryRange> regions, ulong va, ulong size)
|
|
{
|
|
ulong backingStart = (ulong)Context.Memory.Pointer;
|
|
ulong backingEnd = backingStart + Context.Memory.Size;
|
|
|
|
KPageList pageList = new();
|
|
|
|
foreach (HostMemoryRange region in regions)
|
|
{
|
|
// If the range is inside the physical memory, it is shared and we should increment the page count,
|
|
// otherwise it is private and we don't need to increment the page count.
|
|
|
|
if (region.Address >= backingStart && region.Address < backingEnd)
|
|
{
|
|
pageList.AddRange(region.Address - backingStart + DramMemoryMap.DramBase, region.Size / PageSize);
|
|
}
|
|
}
|
|
|
|
using var scopedPageList = new KScopedPageList(Context.MemoryManager, pageList);
|
|
|
|
foreach (var pageNode in pageList)
|
|
{
|
|
Context.CommitMemory(pageNode.Address - DramMemoryMap.DramBase, pageNode.PagesCount * PageSize);
|
|
}
|
|
|
|
ulong offset = 0;
|
|
|
|
foreach (var region in regions)
|
|
{
|
|
_cpuMemory.MapForeign(va + offset, region.Address, region.Size);
|
|
|
|
offset += region.Size;
|
|
}
|
|
|
|
scopedPageList.SignalSuccess();
|
|
|
|
return Result.Success;
|
|
}
|
|
|
|
/// <inheritdoc/>
|
|
protected override Result Unmap(ulong address, ulong pagesCount)
|
|
{
|
|
KPageList pagesToClose = new();
|
|
|
|
var regions = _cpuMemory.GetPhysicalRegions(address, pagesCount * PageSize);
|
|
|
|
foreach (var region in regions)
|
|
{
|
|
ulong pa = region.Address + DramMemoryMap.DramBase;
|
|
if (DramMemoryMap.IsHeapPhysicalAddress(pa))
|
|
{
|
|
pagesToClose.AddRange(pa, region.Size / PageSize);
|
|
}
|
|
}
|
|
|
|
_cpuMemory.Unmap(address, pagesCount * PageSize);
|
|
|
|
pagesToClose.DecrementPagesReferenceCount(Context.MemoryManager);
|
|
|
|
return Result.Success;
|
|
}
|
|
|
|
/// <inheritdoc/>
|
|
protected override Result Reprotect(ulong address, ulong pagesCount, KMemoryPermission permission)
|
|
{
|
|
_cpuMemory.Reprotect(address, pagesCount * PageSize, permission.Convert());
|
|
|
|
return Result.Success;
|
|
}
|
|
|
|
/// <inheritdoc/>
|
|
protected override Result ReprotectAndFlush(ulong address, ulong pagesCount, KMemoryPermission permission)
|
|
{
|
|
// TODO: Flush JIT cache.
|
|
|
|
return Reprotect(address, pagesCount, permission);
|
|
}
|
|
|
|
/// <inheritdoc/>
|
|
protected override void SignalMemoryTracking(ulong va, ulong size, bool write)
|
|
{
|
|
_cpuMemory.SignalMemoryTracking(va, size, write);
|
|
}
|
|
|
|
/// <inheritdoc/>
|
|
protected override void Write(ulong va, ReadOnlySpan<byte> data)
|
|
{
|
|
_cpuMemory.Write(va, data);
|
|
}
|
|
}
|
|
}
|