// SPDX-FileCopyrightText: Copyright 2023 yuzu Emulator Project // SPDX-License-Identifier: GPL-2.0-or-later #include #include #include #include #include "common/address_space.h" #include "common/address_space.inc" #include "common/alignment.h" #include "common/assert.h" #include "common/div_ceil.h" #include "common/scope_exit.h" #include "common/settings.h" #include "core/device_memory.h" #include "core/device_memory_manager.h" #include "core/memory.h" namespace Core { namespace { class MultiAddressContainer { public: MultiAddressContainer() = default; ~MultiAddressContainer() = default; void GatherValues(u32 start_entry, Common::ScratchBuffer& buffer) { buffer.resize(8); buffer.resize(0); size_t index = 0; const auto add_value = [&](u32 value) { buffer.resize(index + 1); buffer[index++] = value; }; u32 iter_entry = start_entry; Entry* current = &storage[iter_entry - 1]; add_value(current->value); while (current->next_entry != 0) { iter_entry = current->next_entry; current = &storage[iter_entry - 1]; add_value(current->value); } } u32 Register(u32 value) { return RegisterImplementation(value); } void Register(u32 value, u32 start_entry) { auto entry_id = RegisterImplementation(value); u32 iter_entry = start_entry; Entry* current = &storage[iter_entry - 1]; while (current->next_entry != 0) { iter_entry = current->next_entry; current = &storage[iter_entry - 1]; } current->next_entry = entry_id; } std::pair Unregister(u32 value, u32 start_entry) { u32 iter_entry = start_entry; Entry* previous{}; Entry* current = &storage[iter_entry - 1]; Entry* next{}; bool more_than_one_remaining = false; u32 result_start{start_entry}; size_t count = 0; while (current->value != value) { count++; previous = current; iter_entry = current->next_entry; current = &storage[iter_entry - 1]; } // Find next u32 next_entry = current->next_entry; if (next_entry != 0) { next = &storage[next_entry - 1]; more_than_one_remaining = next->next_entry != 0 || previous != nullptr; } if (previous) { previous->next_entry = next_entry; } else { result_start = next_entry; } free_entries.emplace_back(iter_entry); return std::make_pair(more_than_one_remaining || count > 1, result_start); } u32 ReleaseEntry(u32 start_entry) { Entry* current = &storage[start_entry - 1]; free_entries.emplace_back(start_entry); return current->value; } private: u32 RegisterImplementation(u32 value) { auto entry_id = GetNewEntry(); auto& entry = storage[entry_id - 1]; entry.next_entry = 0; entry.value = value; return entry_id; } u32 GetNewEntry() { if (!free_entries.empty()) { u32 result = free_entries.front(); free_entries.pop_front(); return result; } storage.emplace_back(); u32 new_entry = static_cast(storage.size()); return new_entry; } struct Entry { u32 next_entry{}; u32 value{}; }; std::deque storage; std::deque free_entries; }; struct EmptyAllocator { EmptyAllocator([[maybe_unused]] DAddr address) {} }; } // namespace template struct DeviceMemoryManagerAllocator { static constexpr size_t device_virtual_bits = DTraits::device_virtual_bits; static constexpr DAddr first_address = 1ULL << Memory::YUZU_PAGEBITS; static constexpr DAddr max_device_area = 1ULL << device_virtual_bits; DeviceMemoryManagerAllocator() : main_allocator(first_address) {} Common::FlatAllocator main_allocator; MultiAddressContainer multi_dev_address; /// Returns true when vaddr -> vaddr+size is fully contained in the buffer template [[nodiscard]] bool IsInBounds(VAddr addr, u64 size) const noexcept { return addr >= 0 && addr + size <= max_device_area; } DAddr Allocate(size_t size) { return main_allocator.Allocate(size); } void AllocateFixed(DAddr b_address, size_t b_size) { main_allocator.AllocateFixed(b_address, b_size); } void Free(DAddr b_address, size_t b_size) { main_allocator.Free(b_address, b_size); } }; template DeviceMemoryManager::DeviceMemoryManager(const DeviceMemory& device_memory_) : physical_base{reinterpret_cast(device_memory_.buffer.BackingBasePointer())}, device_inter{nullptr}, compressed_physical_ptr(device_as_size >> Memory::YUZU_PAGEBITS), compressed_device_addr(1ULL << ((Settings::values.memory_layout_mode.GetValue() == Settings::MemoryLayout::Memory_4Gb ? physical_min_bits : physical_max_bits) - Memory::YUZU_PAGEBITS)), continuity_tracker(device_as_size >> Memory::YUZU_PAGEBITS), cpu_backing_address(device_as_size >> Memory::YUZU_PAGEBITS) { impl = std::make_unique>(); cached_pages = std::make_unique(); const size_t total_virtual = device_as_size >> Memory::YUZU_PAGEBITS; for (size_t i = 0; i < total_virtual; i++) { compressed_physical_ptr[i] = 0; continuity_tracker[i] = 1; cpu_backing_address[i] = 0; } const size_t total_phys = 1ULL << ((Settings::values.memory_layout_mode.GetValue() == Settings::MemoryLayout::Memory_4Gb ? physical_min_bits : physical_max_bits) - Memory::YUZU_PAGEBITS); for (size_t i = 0; i < total_phys; i++) { compressed_device_addr[i] = 0; } } template DeviceMemoryManager::~DeviceMemoryManager() = default; template void DeviceMemoryManager::BindInterface(DeviceInterface* device_inter_) { device_inter = device_inter_; } template DAddr DeviceMemoryManager::Allocate(size_t size) { return impl->Allocate(size); } template void DeviceMemoryManager::AllocateFixed(DAddr start, size_t size) { return impl->AllocateFixed(start, size); } template void DeviceMemoryManager::Free(DAddr start, size_t size) { impl->Free(start, size); } template void DeviceMemoryManager::Map(DAddr address, VAddr virtual_address, size_t size, Asid asid, bool track) { Core::Memory::Memory* process_memory = registered_processes[asid.id]; size_t start_page_d = address >> Memory::YUZU_PAGEBITS; size_t num_pages = Common::AlignUp(size, Memory::YUZU_PAGESIZE) >> Memory::YUZU_PAGEBITS; std::scoped_lock lk(mapping_guard); for (size_t i = 0; i < num_pages; i++) { const VAddr new_vaddress = virtual_address + i * Memory::YUZU_PAGESIZE; auto* ptr = process_memory->GetPointerSilent(Common::ProcessAddress(new_vaddress)); if (ptr == nullptr) [[unlikely]] { compressed_physical_ptr[start_page_d + i] = 0; continue; } auto phys_addr = static_cast(GetRawPhysicalAddr(ptr) >> Memory::YUZU_PAGEBITS) + 1U; compressed_physical_ptr[start_page_d + i] = phys_addr; InsertCPUBacking(start_page_d + i, new_vaddress, asid); const u32 base_dev = compressed_device_addr[phys_addr - 1U]; const u32 new_dev = static_cast(start_page_d + i); if (base_dev == 0) [[likely]] { compressed_device_addr[phys_addr - 1U] = new_dev; continue; } u32 start_id = base_dev & MULTI_MASK; if ((base_dev >> MULTI_FLAG_BITS) == 0) { start_id = impl->multi_dev_address.Register(base_dev); compressed_device_addr[phys_addr - 1U] = MULTI_FLAG | start_id; } impl->multi_dev_address.Register(new_dev, start_id); } if (track) { TrackContinuityImpl(address, virtual_address, size, asid); } } template void DeviceMemoryManager::Unmap(DAddr address, size_t size) { size_t start_page_d = address >> Memory::YUZU_PAGEBITS; size_t num_pages = Common::AlignUp(size, Memory::YUZU_PAGESIZE) >> Memory::YUZU_PAGEBITS; device_inter->InvalidateRegion(address, size); std::scoped_lock lk(mapping_guard); for (size_t i = 0; i < num_pages; i++) { auto phys_addr = compressed_physical_ptr[start_page_d + i]; compressed_physical_ptr[start_page_d + i] = 0; cpu_backing_address[start_page_d + i] = 0; if (phys_addr != 0) [[likely]] { const u32 base_dev = compressed_device_addr[phys_addr - 1U]; if ((base_dev >> MULTI_FLAG_BITS) == 0) [[likely]] { compressed_device_addr[phys_addr - 1] = 0; continue; } const auto [more_entries, new_start] = impl->multi_dev_address.Unregister( static_cast(start_page_d + i), base_dev & MULTI_MASK); if (!more_entries) { compressed_device_addr[phys_addr - 1] = impl->multi_dev_address.ReleaseEntry(new_start); continue; } compressed_device_addr[phys_addr - 1] = new_start | MULTI_FLAG; } } } template void DeviceMemoryManager::TrackContinuityImpl(DAddr address, VAddr virtual_address, size_t size, Asid asid) { Core::Memory::Memory* process_memory = registered_processes[asid.id]; size_t start_page_d = address >> Memory::YUZU_PAGEBITS; size_t num_pages = Common::AlignUp(size, Memory::YUZU_PAGESIZE) >> Memory::YUZU_PAGEBITS; uintptr_t last_ptr = 0; size_t page_count = 1; for (size_t i = num_pages; i > 0; i--) { size_t index = i - 1; const VAddr new_vaddress = virtual_address + index * Memory::YUZU_PAGESIZE; const uintptr_t new_ptr = reinterpret_cast( process_memory->GetPointerSilent(Common::ProcessAddress(new_vaddress))); if (new_ptr + page_size == last_ptr) { page_count++; } else { page_count = 1; } last_ptr = new_ptr; continuity_tracker[start_page_d + index] = static_cast(page_count); } } template u8* DeviceMemoryManager::GetSpan(const DAddr src_addr, const std::size_t size) { size_t page_index = src_addr >> page_bits; size_t subbits = src_addr & page_mask; if ((static_cast(continuity_tracker[page_index]) << page_bits) >= size + subbits) { return GetPointer(src_addr); } return nullptr; } template const u8* DeviceMemoryManager::GetSpan(const DAddr src_addr, const std::size_t size) const { size_t page_index = src_addr >> page_bits; size_t subbits = src_addr & page_mask; if ((static_cast(continuity_tracker[page_index]) << page_bits) >= size + subbits) { return GetPointer(src_addr); } return nullptr; } template void DeviceMemoryManager::InnerGatherDeviceAddresses(Common::ScratchBuffer& buffer, PAddr address) { size_t phys_addr = address >> page_bits; std::scoped_lock lk(mapping_guard); u32 backing = compressed_device_addr[phys_addr]; if ((backing >> MULTI_FLAG_BITS) != 0) { impl->multi_dev_address.GatherValues(backing & MULTI_MASK, buffer); return; } buffer.resize(1); buffer[0] = backing; } template template T* DeviceMemoryManager::GetPointer(DAddr address) { const size_t index = address >> Memory::YUZU_PAGEBITS; const size_t offset = address & Memory::YUZU_PAGEMASK; auto phys_addr = compressed_physical_ptr[index]; if (phys_addr == 0) [[unlikely]] { return nullptr; } return GetPointerFromRaw((static_cast(phys_addr - 1) << Memory::YUZU_PAGEBITS) + offset); } template template const T* DeviceMemoryManager::GetPointer(DAddr address) const { const size_t index = address >> Memory::YUZU_PAGEBITS; const size_t offset = address & Memory::YUZU_PAGEMASK; auto phys_addr = compressed_physical_ptr[index]; if (phys_addr == 0) [[unlikely]] { return nullptr; } return GetPointerFromRaw((static_cast(phys_addr - 1) << Memory::YUZU_PAGEBITS) + offset); } template template void DeviceMemoryManager::Write(DAddr address, T value) { T* ptr = GetPointer(address); if (!ptr) [[unlikely]] { return; } std::memcpy(ptr, &value, sizeof(T)); } template template T DeviceMemoryManager::Read(DAddr address) const { const T* ptr = GetPointer(address); T result{}; if (!ptr) [[unlikely]] { return result; } std::memcpy(&result, ptr, sizeof(T)); return result; } template void DeviceMemoryManager::WalkBlock(DAddr addr, std::size_t size, auto on_unmapped, auto on_memory, auto increment) { std::size_t remaining_size = size; std::size_t page_index = addr >> Memory::YUZU_PAGEBITS; std::size_t page_offset = addr & Memory::YUZU_PAGEMASK; while (remaining_size) { const size_t next_pages = static_cast(continuity_tracker[page_index]); const std::size_t copy_amount = std::min((next_pages << Memory::YUZU_PAGEBITS) - page_offset, remaining_size); const auto current_vaddr = static_cast((page_index << Memory::YUZU_PAGEBITS) + page_offset); SCOPE_EXIT({ page_index += next_pages; page_offset = 0; increment(copy_amount); remaining_size -= copy_amount; }); auto phys_addr = compressed_physical_ptr[page_index]; if (phys_addr == 0) { on_unmapped(copy_amount, current_vaddr); continue; } auto* mem_ptr = GetPointerFromRaw( (static_cast(phys_addr - 1) << Memory::YUZU_PAGEBITS) + page_offset); on_memory(copy_amount, mem_ptr); } } template void DeviceMemoryManager::ReadBlock(DAddr address, void* dest_pointer, size_t size) { device_inter->FlushRegion(address, size); WalkBlock( address, size, [&](size_t copy_amount, DAddr current_vaddr) { LOG_ERROR( HW_Memory, "Unmapped Device ReadBlock @ 0x{:016X} (start address = 0x{:016X}, size = {})", current_vaddr, address, size); std::memset(dest_pointer, 0, copy_amount); }, [&](size_t copy_amount, const u8* const src_ptr) { std::memcpy(dest_pointer, src_ptr, copy_amount); }, [&](const std::size_t copy_amount) { dest_pointer = static_cast(dest_pointer) + copy_amount; }); } template void DeviceMemoryManager::WriteBlock(DAddr address, const void* src_pointer, size_t size) { WalkBlock( address, size, [&](size_t copy_amount, DAddr current_vaddr) { LOG_ERROR( HW_Memory, "Unmapped Device WriteBlock @ 0x{:016X} (start address = 0x{:016X}, size = {})", current_vaddr, address, size); }, [&](size_t copy_amount, u8* const dst_ptr) { std::memcpy(dst_ptr, src_pointer, copy_amount); }, [&](const std::size_t copy_amount) { src_pointer = static_cast(src_pointer) + copy_amount; }); device_inter->InvalidateRegion(address, size); } template void DeviceMemoryManager::ReadBlockUnsafe(DAddr address, void* dest_pointer, size_t size) { WalkBlock( address, size, [&](size_t copy_amount, DAddr current_vaddr) { LOG_ERROR( HW_Memory, "Unmapped Device ReadBlock @ 0x{:016X} (start address = 0x{:016X}, size = {})", current_vaddr, address, size); std::memset(dest_pointer, 0, copy_amount); }, [&](size_t copy_amount, const u8* const src_ptr) { std::memcpy(dest_pointer, src_ptr, copy_amount); }, [&](const std::size_t copy_amount) { dest_pointer = static_cast(dest_pointer) + copy_amount; }); } template void DeviceMemoryManager::WriteBlockUnsafe(DAddr address, const void* src_pointer, size_t size) { WalkBlock( address, size, [&](size_t copy_amount, DAddr current_vaddr) { LOG_ERROR( HW_Memory, "Unmapped Device WriteBlock @ 0x{:016X} (start address = 0x{:016X}, size = {})", current_vaddr, address, size); }, [&](size_t copy_amount, u8* const dst_ptr) { std::memcpy(dst_ptr, src_pointer, copy_amount); }, [&](const std::size_t copy_amount) { src_pointer = static_cast(src_pointer) + copy_amount; }); } template Asid DeviceMemoryManager::RegisterProcess(Memory::Memory* memory_device_inter) { size_t new_id{}; if (!id_pool.empty()) { new_id = id_pool.front(); id_pool.pop_front(); registered_processes[new_id] = memory_device_inter; } else { registered_processes.emplace_back(memory_device_inter); new_id = registered_processes.size() - 1U; } return Asid{new_id}; } template void DeviceMemoryManager::UnregisterProcess(Asid asid) { registered_processes[asid.id] = nullptr; id_pool.push_front(asid.id); } template void DeviceMemoryManager::UpdatePagesCachedCount(DAddr addr, size_t size, s32 delta) { std::unique_lock lk(counter_guard, std::defer_lock); const auto Lock = [&] { if (!lk) { lk.lock(); } }; u64 uncache_begin = 0; u64 cache_begin = 0; u64 uncache_bytes = 0; u64 cache_bytes = 0; const auto MarkRegionCaching = &DeviceMemoryManager::DeviceMethods::MarkRegionCaching; std::atomic_thread_fence(std::memory_order_acquire); const size_t page_end = Common::DivCeil(addr + size, Memory::YUZU_PAGESIZE); size_t page = addr >> Memory::YUZU_PAGEBITS; auto [asid, base_vaddress] = ExtractCPUBacking(page); size_t vpage = base_vaddress >> Memory::YUZU_PAGEBITS; auto* memory_device_inter = registered_processes[asid.id]; for (; page != page_end; ++page) { std::atomic_uint8_t& count = cached_pages->at(page >> 3).Count(page); if (delta > 0) { ASSERT_MSG(count.load(std::memory_order::relaxed) < std::numeric_limits::max(), "Count may overflow!"); } else if (delta < 0) { ASSERT_MSG(count.load(std::memory_order::relaxed) > 0, "Count may underflow!"); } else { ASSERT_MSG(false, "Delta must be non-zero!"); } // Adds or subtracts 1, as count is a unsigned 8-bit value count.fetch_add(static_cast(delta), std::memory_order_release); // Assume delta is either -1 or 1 if (count.load(std::memory_order::relaxed) == 0) { if (uncache_bytes == 0) { uncache_begin = vpage; } uncache_bytes += Memory::YUZU_PAGESIZE; } else if (uncache_bytes > 0) { Lock(); MarkRegionCaching(memory_device_inter, uncache_begin << Memory::YUZU_PAGEBITS, uncache_bytes, false); uncache_bytes = 0; } if (count.load(std::memory_order::relaxed) == 1 && delta > 0) { if (cache_bytes == 0) { cache_begin = vpage; } cache_bytes += Memory::YUZU_PAGESIZE; } else if (cache_bytes > 0) { Lock(); MarkRegionCaching(memory_device_inter, cache_begin << Memory::YUZU_PAGEBITS, cache_bytes, true); cache_bytes = 0; } vpage++; } if (uncache_bytes > 0) { Lock(); MarkRegionCaching(memory_device_inter, uncache_begin << Memory::YUZU_PAGEBITS, uncache_bytes, false); } if (cache_bytes > 0) { Lock(); MarkRegionCaching(memory_device_inter, cache_begin << Memory::YUZU_PAGEBITS, cache_bytes, true); } } } // namespace Core