// SPDX-FileCopyrightText: 2023 yuzu Emulator Project // SPDX-License-Identifier: GPL-3.0-or-later #pragma once #include #include #include "common/alignment.h" #include "common/settings.h" #include "video_core/control/channel_state.h" #include "video_core/dirty_flags.h" #include "video_core/engines/kepler_compute.h" #include "video_core/texture_cache/image_view_base.h" #include "video_core/texture_cache/samples_helper.h" #include "video_core/texture_cache/texture_cache_base.h" #include "video_core/texture_cache/util.h" namespace VideoCommon { using Tegra::Texture::TICEntry; using Tegra::Texture::TSCEntry; using VideoCore::Surface::GetFormatType; using VideoCore::Surface::PixelFormat; using VideoCore::Surface::SurfaceType; using namespace Common::Literals; template TextureCache

::TextureCache(Runtime& runtime_, VideoCore::RasterizerInterface& rasterizer_) : runtime{runtime_}, rasterizer{rasterizer_} { // Configure null sampler TSCEntry sampler_descriptor{}; sampler_descriptor.min_filter.Assign(Tegra::Texture::TextureFilter::Linear); sampler_descriptor.mag_filter.Assign(Tegra::Texture::TextureFilter::Linear); sampler_descriptor.mipmap_filter.Assign(Tegra::Texture::TextureMipmapFilter::Linear); sampler_descriptor.cubemap_anisotropy.Assign(1); // These values were chosen based on typical peak swizzle data sizes seen in some titles static constexpr size_t SWIZZLE_DATA_BUFFER_INITIAL_CAPACITY = 8_MiB; static constexpr size_t UNSWIZZLE_DATA_BUFFER_INITIAL_CAPACITY = 1_MiB; swizzle_data_buffer.resize_destructive(SWIZZLE_DATA_BUFFER_INITIAL_CAPACITY); unswizzle_data_buffer.resize_destructive(UNSWIZZLE_DATA_BUFFER_INITIAL_CAPACITY); // Make sure the first index is reserved for the null resources // This way the null resource becomes a compile time constant void(slot_images.insert(NullImageParams{})); void(slot_image_views.insert(runtime, NullImageViewParams{})); void(slot_samplers.insert(runtime, sampler_descriptor)); if constexpr (HAS_DEVICE_MEMORY_INFO) { const s64 device_memory = static_cast(runtime.GetDeviceLocalMemory()); const s64 min_spacing_expected = device_memory - 1_GiB - 512_MiB; const s64 min_spacing_critical = device_memory - 1_GiB; const s64 mem_threshold = std::min(device_memory, TARGET_THRESHOLD); const s64 min_vacancy_expected = (6 * mem_threshold) / 10; const s64 min_vacancy_critical = (3 * mem_threshold) / 10; expected_memory = static_cast( std::max(std::min(device_memory - min_vacancy_expected, min_spacing_expected), DEFAULT_EXPECTED_MEMORY)); critical_memory = static_cast( std::max(std::min(device_memory - min_vacancy_critical, min_spacing_critical), DEFAULT_CRITICAL_MEMORY)); minimum_memory = static_cast((device_memory - mem_threshold) / 2); } else { expected_memory = DEFAULT_EXPECTED_MEMORY + 512_MiB; critical_memory = DEFAULT_CRITICAL_MEMORY + 1_GiB; minimum_memory = 0; } } template void TextureCache

::RunGarbageCollector() { bool high_priority_mode = total_used_memory >= expected_memory; bool aggressive_mode = total_used_memory >= critical_memory; const u64 ticks_to_destroy = aggressive_mode ? 10ULL : high_priority_mode ? 25ULL : 50ULL; size_t num_iterations = aggressive_mode ? 40 : (high_priority_mode ? 20 : 10); const auto clean_up = [this, &num_iterations, &high_priority_mode, &aggressive_mode](ImageId image_id) { if (num_iterations == 0) { return true; } --num_iterations; auto& image = slot_images[image_id]; if (True(image.flags & ImageFlagBits::IsDecoding)) { // This image is still being decoded, deleting it will invalidate the slot // used by the async decoder thread. return false; } const bool must_download = image.IsSafeDownload() && False(image.flags & ImageFlagBits::BadOverlap); if (!high_priority_mode && (must_download || True(image.flags & ImageFlagBits::CostlyLoad))) { return false; } if (must_download) { auto map = runtime.DownloadStagingBuffer(image.unswizzled_size_bytes); const auto copies = FullDownloadCopies(image.info); image.DownloadMemory(map, copies); runtime.Finish(); SwizzleImage(*gpu_memory, image.gpu_addr, image.info, copies, map.mapped_span, swizzle_data_buffer); } if (True(image.flags & ImageFlagBits::Tracked)) { UntrackImage(image, image_id); } UnregisterImage(image_id); DeleteImage(image_id, image.scale_tick > frame_tick + 5); if (total_used_memory < critical_memory) { if (aggressive_mode) { // Sink the aggresiveness. num_iterations >>= 2; aggressive_mode = false; return false; } if (high_priority_mode && total_used_memory < expected_memory) { num_iterations >>= 1; high_priority_mode = false; } } return false; }; lru_cache.ForEachItemBelow(frame_tick - ticks_to_destroy, clean_up); } template void TextureCache

::TickFrame() { // If we can obtain the memory info, use it instead of the estimate. if (runtime.CanReportMemoryUsage()) { total_used_memory = runtime.GetDeviceMemoryUsage(); } if (total_used_memory > minimum_memory) { RunGarbageCollector(); } sentenced_images.Tick(); sentenced_framebuffers.Tick(); sentenced_image_view.Tick(); TickAsyncDecode(); runtime.TickFrame(); critical_gc = 0; ++frame_tick; if constexpr (IMPLEMENTS_ASYNC_DOWNLOADS) { for (auto& buffer : async_buffers_death_ring) { runtime.FreeDeferredStagingBuffer(buffer); } async_buffers_death_ring.clear(); } } template const typename P::ImageView& TextureCache

::GetImageView(ImageViewId id) const noexcept { return slot_image_views[id]; } template typename P::ImageView& TextureCache

::GetImageView(ImageViewId id) noexcept { return slot_image_views[id]; } template typename P::ImageView& TextureCache

::GetImageView(u32 index) noexcept { const auto image_view_id = VisitImageView(channel_state->graphics_image_table, channel_state->graphics_image_view_ids, index); return slot_image_views[image_view_id]; } template void TextureCache

::MarkModification(ImageId id) noexcept { MarkModification(slot_images[id]); } template template void TextureCache

::FillGraphicsImageViews(std::span views) { FillImageViews(channel_state->graphics_image_table, channel_state->graphics_image_view_ids, views); } template void TextureCache

::FillComputeImageViews(std::span views) { FillImageViews(channel_state->compute_image_table, channel_state->compute_image_view_ids, views); } template typename P::Sampler* TextureCache

::GetGraphicsSampler(u32 index) { if (index > channel_state->graphics_sampler_table.Limit()) { LOG_DEBUG(HW_GPU, "Invalid sampler index={}", index); return &slot_samplers[NULL_SAMPLER_ID]; } const auto [descriptor, is_new] = channel_state->graphics_sampler_table.Read(index); SamplerId& id = channel_state->graphics_sampler_ids[index]; if (is_new) { id = FindSampler(descriptor); } return &slot_samplers[id]; } template typename P::Sampler* TextureCache

::GetComputeSampler(u32 index) { if (index > channel_state->compute_sampler_table.Limit()) { LOG_DEBUG(HW_GPU, "Invalid sampler index={}", index); return &slot_samplers[NULL_SAMPLER_ID]; } const auto [descriptor, is_new] = channel_state->compute_sampler_table.Read(index); SamplerId& id = channel_state->compute_sampler_ids[index]; if (is_new) { id = FindSampler(descriptor); } return &slot_samplers[id]; } template void TextureCache

::SynchronizeGraphicsDescriptors() { using SamplerBinding = Tegra::Engines::Maxwell3D::Regs::SamplerBinding; const bool linked_tsc = maxwell3d->regs.sampler_binding == SamplerBinding::ViaHeaderBinding; const u32 tic_limit = maxwell3d->regs.tex_header.limit; const u32 tsc_limit = linked_tsc ? tic_limit : maxwell3d->regs.tex_sampler.limit; if (channel_state->graphics_sampler_table.Synchronize(maxwell3d->regs.tex_sampler.Address(), tsc_limit)) { channel_state->graphics_sampler_ids.resize(tsc_limit + 1, CORRUPT_ID); } if (channel_state->graphics_image_table.Synchronize(maxwell3d->regs.tex_header.Address(), tic_limit)) { channel_state->graphics_image_view_ids.resize(tic_limit + 1, CORRUPT_ID); } } template void TextureCache

::SynchronizeComputeDescriptors() { const bool linked_tsc = kepler_compute->launch_description.linked_tsc; const u32 tic_limit = kepler_compute->regs.tic.limit; const u32 tsc_limit = linked_tsc ? tic_limit : kepler_compute->regs.tsc.limit; const GPUVAddr tsc_gpu_addr = kepler_compute->regs.tsc.Address(); if (channel_state->compute_sampler_table.Synchronize(tsc_gpu_addr, tsc_limit)) { channel_state->compute_sampler_ids.resize(tsc_limit + 1, CORRUPT_ID); } if (channel_state->compute_image_table.Synchronize(kepler_compute->regs.tic.Address(), tic_limit)) { channel_state->compute_image_view_ids.resize(tic_limit + 1, CORRUPT_ID); } } template bool TextureCache

::RescaleRenderTargets(bool is_clear) { auto& flags = maxwell3d->dirty.flags; u32 scale_rating = 0; bool rescaled = false; std::array tmp_color_images{}; ImageId tmp_depth_image{}; do { flags[Dirty::RenderTargets] = false; has_deleted_images = false; // Render target control is used on all render targets, so force look ups when this one is // up const bool force = flags[Dirty::RenderTargetControl]; flags[Dirty::RenderTargetControl] = false; scale_rating = 0; bool any_rescaled = false; bool can_rescale = true; const auto check_rescale = [&](ImageViewId view_id, ImageId& id_save) { if (view_id != NULL_IMAGE_VIEW_ID && view_id != ImageViewId{}) { const auto& view = slot_image_views[view_id]; const auto image_id = view.image_id; id_save = image_id; auto& image = slot_images[image_id]; can_rescale &= ImageCanRescale(image); any_rescaled |= True(image.flags & ImageFlagBits::Rescaled) || GetFormatType(image.info.format) != SurfaceType::ColorTexture; scale_rating = std::max(scale_rating, image.scale_tick <= frame_tick ? image.scale_rating + 1U : image.scale_rating); } else { id_save = CORRUPT_ID; } }; for (size_t index = 0; index < NUM_RT; ++index) { ImageViewId& color_buffer_id = render_targets.color_buffer_ids[index]; if (flags[Dirty::ColorBuffer0 + index] || force) { flags[Dirty::ColorBuffer0 + index] = false; BindRenderTarget(&color_buffer_id, FindColorBuffer(index, is_clear)); } check_rescale(color_buffer_id, tmp_color_images[index]); } if (flags[Dirty::ZetaBuffer] || force) { flags[Dirty::ZetaBuffer] = false; BindRenderTarget(&render_targets.depth_buffer_id, FindDepthBuffer(is_clear)); } check_rescale(render_targets.depth_buffer_id, tmp_depth_image); if (can_rescale) { rescaled = any_rescaled || scale_rating >= 2; const auto scale_up = [this](ImageId image_id) { if (image_id != CORRUPT_ID) { Image& image = slot_images[image_id]; ScaleUp(image); } }; if (rescaled) { for (size_t index = 0; index < NUM_RT; ++index) { scale_up(tmp_color_images[index]); } scale_up(tmp_depth_image); scale_rating = 2; } } else { rescaled = false; const auto scale_down = [this](ImageId image_id) { if (image_id != CORRUPT_ID) { Image& image = slot_images[image_id]; ScaleDown(image); } }; for (size_t index = 0; index < NUM_RT; ++index) { scale_down(tmp_color_images[index]); } scale_down(tmp_depth_image); scale_rating = 1; } } while (has_deleted_images); const auto set_rating = [this, scale_rating](ImageId image_id) { if (image_id != CORRUPT_ID) { Image& image = slot_images[image_id]; image.scale_rating = scale_rating; if (image.scale_tick <= frame_tick) { image.scale_tick = frame_tick + 1; } } }; for (size_t index = 0; index < NUM_RT; ++index) { set_rating(tmp_color_images[index]); } set_rating(tmp_depth_image); return rescaled; } template void TextureCache

::UpdateRenderTargets(bool is_clear) { using namespace VideoCommon::Dirty; auto& flags = maxwell3d->dirty.flags; if (!flags[Dirty::RenderTargets]) { for (size_t index = 0; index < NUM_RT; ++index) { ImageViewId& color_buffer_id = render_targets.color_buffer_ids[index]; PrepareImageView(color_buffer_id, true, is_clear && IsFullClear(color_buffer_id)); } const ImageViewId depth_buffer_id = render_targets.depth_buffer_id; PrepareImageView(depth_buffer_id, true, is_clear && IsFullClear(depth_buffer_id)); return; } const bool rescaled = RescaleRenderTargets(is_clear); if (is_rescaling != rescaled) { flags[Dirty::RescaleViewports] = true; flags[Dirty::RescaleScissors] = true; is_rescaling = rescaled; } for (size_t index = 0; index < NUM_RT; ++index) { ImageViewId& color_buffer_id = render_targets.color_buffer_ids[index]; PrepareImageView(color_buffer_id, true, is_clear && IsFullClear(color_buffer_id)); } const ImageViewId depth_buffer_id = render_targets.depth_buffer_id; PrepareImageView(depth_buffer_id, true, is_clear && IsFullClear(depth_buffer_id)); for (size_t index = 0; index < NUM_RT; ++index) { render_targets.draw_buffers[index] = static_cast(maxwell3d->regs.rt_control.Map(index)); } u32 up_scale = 1; u32 down_shift = 0; if (is_rescaling) { up_scale = Settings::values.resolution_info.up_scale; down_shift = Settings::values.resolution_info.down_shift; } render_targets.size = Extent2D{ (maxwell3d->regs.surface_clip.width * up_scale) >> down_shift, (maxwell3d->regs.surface_clip.height * up_scale) >> down_shift, }; render_targets.is_rescaled = is_rescaling; flags[Dirty::DepthBiasGlobal] = true; } template typename P::Framebuffer* TextureCache

::GetFramebuffer() { return &slot_framebuffers[GetFramebufferId(render_targets)]; } template template void TextureCache

::FillImageViews(DescriptorTable& table, std::span cached_image_view_ids, std::span views) { bool has_blacklisted = false; do { has_deleted_images = false; if constexpr (has_blacklists) { has_blacklisted = false; } for (ImageViewInOut& view : views) { view.id = VisitImageView(table, cached_image_view_ids, view.index); if constexpr (has_blacklists) { if (view.blacklist && view.id != NULL_IMAGE_VIEW_ID) { const ImageViewBase& image_view{slot_image_views[view.id]}; auto& image = slot_images[image_view.image_id]; has_blacklisted |= ScaleDown(image); image.scale_rating = 0; } } } } while (has_deleted_images || (has_blacklists && has_blacklisted)); } template ImageViewId TextureCache

::VisitImageView(DescriptorTable& table, std::span cached_image_view_ids, u32 index) { if (index > table.Limit()) { LOG_DEBUG(HW_GPU, "Invalid image view index={}", index); return NULL_IMAGE_VIEW_ID; } const auto [descriptor, is_new] = table.Read(index); ImageViewId& image_view_id = cached_image_view_ids[index]; if (is_new) { image_view_id = FindImageView(descriptor); } if (image_view_id != NULL_IMAGE_VIEW_ID) { PrepareImageView(image_view_id, false, false); } return image_view_id; } template FramebufferId TextureCache

::GetFramebufferId(const RenderTargets& key) { const auto [pair, is_new] = framebuffers.try_emplace(key); FramebufferId& framebuffer_id = pair->second; if (!is_new) { return framebuffer_id; } std::array color_buffers; std::ranges::transform(key.color_buffer_ids, color_buffers.begin(), [this](ImageViewId id) { return id ? &slot_image_views[id] : nullptr; }); ImageView* const depth_buffer = key.depth_buffer_id ? &slot_image_views[key.depth_buffer_id] : nullptr; framebuffer_id = slot_framebuffers.insert(runtime, color_buffers, depth_buffer, key); return framebuffer_id; } template void TextureCache

::WriteMemory(VAddr cpu_addr, size_t size) { ForEachImageInRegion(cpu_addr, size, [this](ImageId image_id, Image& image) { if (True(image.flags & ImageFlagBits::CpuModified)) { return; } image.flags |= ImageFlagBits::CpuModified; if (True(image.flags & ImageFlagBits::Tracked)) { UntrackImage(image, image_id); } }); } template void TextureCache

::DownloadMemory(VAddr cpu_addr, size_t size) { std::vector images; ForEachImageInRegion(cpu_addr, size, [&images](ImageId image_id, ImageBase& image) { if (!image.IsSafeDownload()) { return; } image.flags &= ~ImageFlagBits::GpuModified; images.push_back(image_id); }); if (images.empty()) { return; } std::ranges::sort(images, [this](ImageId lhs, ImageId rhs) { return slot_images[lhs].modification_tick < slot_images[rhs].modification_tick; }); for (const ImageId image_id : images) { Image& image = slot_images[image_id]; auto map = runtime.DownloadStagingBuffer(image.unswizzled_size_bytes); const auto copies = FullDownloadCopies(image.info); image.DownloadMemory(map, copies); runtime.Finish(); SwizzleImage(*gpu_memory, image.gpu_addr, image.info, copies, map.mapped_span, swizzle_data_buffer); } } template std::optional TextureCache

::GetFlushArea(VAddr cpu_addr, u64 size) { std::optional area{}; ForEachImageInRegion(cpu_addr, size, [&](ImageId, ImageBase& image) { if (False(image.flags & ImageFlagBits::GpuModified)) { return; } if (!area) { area.emplace(); area->start_address = cpu_addr; area->end_address = cpu_addr + size; area->preemtive = true; } area->start_address = std::min(area->start_address, image.cpu_addr); area->end_address = std::max(area->end_address, image.cpu_addr_end); for (auto image_view_id : image.image_view_ids) { auto& image_view = slot_image_views[image_view_id]; image_view.flags |= ImageViewFlagBits::PreemtiveDownload; } area->preemtive &= image.info.forced_flushed; image.info.forced_flushed = true; }); return area; } template void TextureCache

::UnmapMemory(VAddr cpu_addr, size_t size) { std::vector deleted_images; ForEachImageInRegion(cpu_addr, size, [&](ImageId id, Image&) { deleted_images.push_back(id); }); for (const ImageId id : deleted_images) { Image& image = slot_images[id]; if (True(image.flags & ImageFlagBits::Tracked)) { UntrackImage(image, id); } UnregisterImage(id); DeleteImage(id); } } template void TextureCache

::UnmapGPUMemory(size_t as_id, GPUVAddr gpu_addr, size_t size) { std::vector deleted_images; ForEachImageInRegionGPU(as_id, gpu_addr, size, [&](ImageId id, Image&) { deleted_images.push_back(id); }); for (const ImageId id : deleted_images) { Image& image = slot_images[id]; if (True(image.flags & ImageFlagBits::CpuModified)) { return; } image.flags |= ImageFlagBits::CpuModified; if (True(image.flags & ImageFlagBits::Tracked)) { UntrackImage(image, id); } /* if (True(image.flags & ImageFlagBits::Remapped)) { continue; } image.flags |= ImageFlagBits::Remapped; if (True(image.flags & ImageFlagBits::Tracked)) { UntrackImage(image, id); } */ } } template bool TextureCache

::BlitImage(const Tegra::Engines::Fermi2D::Surface& dst, const Tegra::Engines::Fermi2D::Surface& src, const Tegra::Engines::Fermi2D::Config& copy) { const auto result = GetBlitImages(dst, src, copy); if (!result) { return false; } const BlitImages images = *result; const ImageId dst_id = images.dst_id; const ImageId src_id = images.src_id; PrepareImage(src_id, false, false); PrepareImage(dst_id, true, false); Image& dst_image = slot_images[dst_id]; Image& src_image = slot_images[src_id]; bool is_src_rescaled = True(src_image.flags & ImageFlagBits::Rescaled); bool is_dst_rescaled = True(dst_image.flags & ImageFlagBits::Rescaled); const bool is_resolve = src_image.info.num_samples != 1 && dst_image.info.num_samples == 1; if (is_src_rescaled != is_dst_rescaled) { if (ImageCanRescale(src_image)) { ScaleUp(src_image); is_src_rescaled = True(src_image.flags & ImageFlagBits::Rescaled); if (is_resolve) { dst_image.info.rescaleable = true; for (const auto& alias : dst_image.aliased_images) { Image& other_image = slot_images[alias.id]; other_image.info.rescaleable = true; } } } if (ImageCanRescale(dst_image)) { ScaleUp(dst_image); is_dst_rescaled = True(dst_image.flags & ImageFlagBits::Rescaled); } } if (is_resolve && (is_src_rescaled != is_dst_rescaled)) { // A resolve requires both images to be the same dimensions. Resize down if needed. ScaleDown(src_image); ScaleDown(dst_image); is_src_rescaled = True(src_image.flags & ImageFlagBits::Rescaled); is_dst_rescaled = True(dst_image.flags & ImageFlagBits::Rescaled); } const auto& resolution = Settings::values.resolution_info; const auto scale_region = [&](Region2D& region) { region.start.x = resolution.ScaleUp(region.start.x); region.start.y = resolution.ScaleUp(region.start.y); region.end.x = resolution.ScaleUp(region.end.x); region.end.y = resolution.ScaleUp(region.end.y); }; // TODO: Deduplicate const std::optional src_base = src_image.TryFindBase(src.Address()); const SubresourceRange src_range{.base = src_base.value(), .extent = {1, 1}}; const ImageViewInfo src_view_info(ImageViewType::e2D, images.src_format, src_range); const auto [src_framebuffer_id, src_view_id] = RenderTargetFromImage(src_id, src_view_info); const auto [src_samples_x, src_samples_y] = SamplesLog2(src_image.info.num_samples); Region2D src_region{ Offset2D{.x = copy.src_x0 >> src_samples_x, .y = copy.src_y0 >> src_samples_y}, Offset2D{.x = copy.src_x1 >> src_samples_x, .y = copy.src_y1 >> src_samples_y}, }; if (is_src_rescaled) { scale_region(src_region); } const std::optional dst_base = dst_image.TryFindBase(dst.Address()); const SubresourceRange dst_range{.base = dst_base.value(), .extent = {1, 1}}; const ImageViewInfo dst_view_info(ImageViewType::e2D, images.dst_format, dst_range); const auto [dst_framebuffer_id, dst_view_id] = RenderTargetFromImage(dst_id, dst_view_info); const auto [dst_samples_x, dst_samples_y] = SamplesLog2(dst_image.info.num_samples); Region2D dst_region{ Offset2D{.x = copy.dst_x0 >> dst_samples_x, .y = copy.dst_y0 >> dst_samples_y}, Offset2D{.x = copy.dst_x1 >> dst_samples_x, .y = copy.dst_y1 >> dst_samples_y}, }; if (is_dst_rescaled) { scale_region(dst_region); } // Always call this after src_framebuffer_id was queried, as the address might be invalidated. Framebuffer* const dst_framebuffer = &slot_framebuffers[dst_framebuffer_id]; if constexpr (FRAMEBUFFER_BLITS) { // OpenGL blits from framebuffers, not images Framebuffer* const src_framebuffer = &slot_framebuffers[src_framebuffer_id]; runtime.BlitFramebuffer(dst_framebuffer, src_framebuffer, dst_region, src_region, copy.filter, copy.operation); } else { // Vulkan can blit images, but it lacks format reinterpretations // Provide a framebuffer in case it's necessary ImageView& dst_view = slot_image_views[dst_view_id]; ImageView& src_view = slot_image_views[src_view_id]; runtime.BlitImage(dst_framebuffer, dst_view, src_view, dst_region, src_region, copy.filter, copy.operation); } return true; } template typename P::ImageView* TextureCache

::TryFindFramebufferImageView(VAddr cpu_addr) { // TODO: Properly implement this const auto it = page_table.find(cpu_addr >> YUZU_PAGEBITS); if (it == page_table.end()) { return nullptr; } const auto& image_map_ids = it->second; for (const ImageMapId map_id : image_map_ids) { const ImageMapView& map = slot_map_views[map_id]; const ImageBase& image = slot_images[map.image_id]; if (image.cpu_addr != cpu_addr) { continue; } if (image.image_view_ids.empty()) { continue; } return &slot_image_views[image.image_view_ids.at(0)]; } return nullptr; } template bool TextureCache

::HasUncommittedFlushes() const noexcept { return !uncommitted_downloads.empty(); } template bool TextureCache

::ShouldWaitAsyncFlushes() const noexcept { return !committed_downloads.empty() && !committed_downloads.front().empty(); } template void TextureCache

::CommitAsyncFlushes() { // This is intentionally passing the value by copy if constexpr (IMPLEMENTS_ASYNC_DOWNLOADS) { auto& download_ids = uncommitted_downloads; if (download_ids.empty()) { committed_downloads.emplace_back(std::move(uncommitted_downloads)); uncommitted_downloads.clear(); async_buffers.emplace_back(std::move(uncommitted_async_buffers)); uncommitted_async_buffers.clear(); return; } size_t total_size_bytes = 0; size_t last_async_buffer_id = uncommitted_async_buffers.size(); bool any_none_dma = false; for (PendingDownload& download_info : download_ids) { if (download_info.is_swizzle) { total_size_bytes += Common::AlignUp(slot_images[download_info.object_id].unswizzled_size_bytes, 64); any_none_dma = true; download_info.async_buffer_id = last_async_buffer_id; } } if (any_none_dma) { auto download_map = runtime.DownloadStagingBuffer(total_size_bytes, true); for (const PendingDownload& download_info : download_ids) { if (download_info.is_swizzle) { Image& image = slot_images[download_info.object_id]; const auto copies = FullDownloadCopies(image.info); image.DownloadMemory(download_map, copies); download_map.offset += Common::AlignUp(image.unswizzled_size_bytes, 64); } } uncommitted_async_buffers.emplace_back(download_map); } async_buffers.emplace_back(std::move(uncommitted_async_buffers)); uncommitted_async_buffers.clear(); } committed_downloads.emplace_back(std::move(uncommitted_downloads)); uncommitted_downloads.clear(); } template void TextureCache

::PopAsyncFlushes() { if (committed_downloads.empty()) { return; } if constexpr (IMPLEMENTS_ASYNC_DOWNLOADS) { const auto& download_ids = committed_downloads.front(); if (download_ids.empty()) { committed_downloads.pop_front(); async_buffers.pop_front(); return; } auto download_map = std::move(async_buffers.front()); for (size_t i = download_ids.size(); i > 0; i--) { auto& download_info = download_ids[i - 1]; auto& download_buffer = download_map[download_info.async_buffer_id]; if (download_info.is_swizzle) { const ImageBase& image = slot_images[download_info.object_id]; const auto copies = FullDownloadCopies(image.info); download_buffer.offset -= Common::AlignUp(image.unswizzled_size_bytes, 64); std::span download_span = download_buffer.mapped_span.subspan(download_buffer.offset); SwizzleImage(*gpu_memory, image.gpu_addr, image.info, copies, download_span, swizzle_data_buffer); } else { const BufferDownload& buffer_info = slot_buffer_downloads[download_info.object_id]; std::span download_span = download_buffer.mapped_span.subspan(download_buffer.offset); gpu_memory->WriteBlockUnsafe(buffer_info.address, download_span.data(), buffer_info.size); slot_buffer_downloads.erase(download_info.object_id); } } for (auto& download_buffer : download_map) { async_buffers_death_ring.emplace_back(download_buffer); } committed_downloads.pop_front(); async_buffers.pop_front(); } else { const auto& download_ids = committed_downloads.front(); if (download_ids.empty()) { committed_downloads.pop_front(); return; } size_t total_size_bytes = 0; for (const PendingDownload& download_info : download_ids) { if (download_info.is_swizzle) { total_size_bytes += slot_images[download_info.object_id].unswizzled_size_bytes; } } auto download_map = runtime.DownloadStagingBuffer(total_size_bytes); const size_t original_offset = download_map.offset; for (const PendingDownload& download_info : download_ids) { if (!download_info.is_swizzle) { continue; } Image& image = slot_images[download_info.object_id]; const auto copies = FullDownloadCopies(image.info); image.DownloadMemory(download_map, copies); download_map.offset += image.unswizzled_size_bytes; } // Wait for downloads to finish runtime.Finish(); download_map.offset = original_offset; std::span download_span = download_map.mapped_span; for (const PendingDownload& download_info : download_ids) { if (!download_info.is_swizzle) { continue; } const ImageBase& image = slot_images[download_info.object_id]; const auto copies = FullDownloadCopies(image.info); SwizzleImage(*gpu_memory, image.gpu_addr, image.info, copies, download_span, swizzle_data_buffer); download_map.offset += image.unswizzled_size_bytes; download_span = download_span.subspan(image.unswizzled_size_bytes); } committed_downloads.pop_front(); } } template ImageId TextureCache

::DmaImageId(const Tegra::DMA::ImageOperand& operand, bool is_upload) { const ImageInfo dst_info(operand); const ImageId dst_id = FindDMAImage(dst_info, operand.address); if (!dst_id) { return NULL_IMAGE_ID; } auto& image = slot_images[dst_id]; if (False(image.flags & ImageFlagBits::GpuModified)) { // No need to waste time on an image that's synced with guest return NULL_IMAGE_ID; } if (!is_upload && !image.info.dma_downloaded) { // Force a full sync. image.info.dma_downloaded = true; return NULL_IMAGE_ID; } const auto base = image.TryFindBase(operand.address); if (!base) { return NULL_IMAGE_ID; } return dst_id; } template bool TextureCache

::IsRescaling() const noexcept { return is_rescaling; } template bool TextureCache

::IsRescaling(const ImageViewBase& image_view) const noexcept { if (image_view.type == ImageViewType::Buffer) { return false; } const ImageBase& image = slot_images[image_view.image_id]; return True(image.flags & ImageFlagBits::Rescaled); } template bool TextureCache

::IsRegionGpuModified(VAddr addr, size_t size) { bool is_modified = false; ForEachImageInRegion(addr, size, [&is_modified](ImageId, ImageBase& image) { if (False(image.flags & ImageFlagBits::GpuModified)) { return false; } is_modified = true; return true; }); return is_modified; } template std::pair::Image*, BufferImageCopy> TextureCache

::DmaBufferImageCopy( const Tegra::DMA::ImageCopy& copy_info, const Tegra::DMA::BufferOperand& buffer_operand, const Tegra::DMA::ImageOperand& image_operand, ImageId image_id, bool modifies_image) { const auto [level, base] = PrepareDmaImage(image_id, image_operand.address, modifies_image); auto* image = &slot_images[image_id]; const u32 buffer_size = static_cast(buffer_operand.pitch * buffer_operand.height); const u32 bpp = VideoCore::Surface::BytesPerBlock(image->info.format); const auto convert = [old_bpp = image_operand.bytes_per_pixel, bpp](u32 value) { return (old_bpp * value) / bpp; }; const u32 base_x = convert(image_operand.params.origin.x.Value()); const u32 base_y = image_operand.params.origin.y.Value(); const u32 length_x = convert(copy_info.length_x); const u32 length_y = copy_info.length_y; const BufferImageCopy copy{ .buffer_offset = 0, .buffer_size = buffer_size, .buffer_row_length = convert(buffer_operand.pitch), .buffer_image_height = buffer_operand.height, .image_subresource = { .base_level = static_cast(level), .base_layer = static_cast(base), .num_layers = 1, }, .image_offset = { .x = static_cast(base_x), .y = static_cast(base_y), .z = 0, }, .image_extent = { .width = length_x, .height = length_y, .depth = 1, }, }; return {image, copy}; } template void TextureCache

::DownloadImageIntoBuffer(typename TextureCache

::Image* image, typename TextureCache

::BufferType buffer, size_t buffer_offset, std::span copies, GPUVAddr address, size_t size) { if constexpr (IMPLEMENTS_ASYNC_DOWNLOADS) { const BufferDownload new_buffer_download{address, size}; auto slot = slot_buffer_downloads.insert(new_buffer_download); const PendingDownload new_download{false, uncommitted_async_buffers.size(), slot}; uncommitted_downloads.emplace_back(new_download); auto download_map = runtime.DownloadStagingBuffer(size, true); uncommitted_async_buffers.emplace_back(download_map); std::array buffers{ buffer, download_map.buffer, }; std::array buffer_offsets{ buffer_offset, download_map.offset, }; image->DownloadMemory(buffers, buffer_offsets, copies); } else { image->DownloadMemory(buffer, buffer_offset, copies); } } template void TextureCache

::RefreshContents(Image& image, ImageId image_id) { if (False(image.flags & ImageFlagBits::CpuModified)) { // Only upload modified images return; } image.flags &= ~ImageFlagBits::CpuModified; TrackImage(image, image_id); if (image.info.num_samples > 1 && !runtime.CanUploadMSAA()) { LOG_WARNING(HW_GPU, "MSAA image uploads are not implemented"); return; } if (True(image.flags & ImageFlagBits::AsynchronousDecode)) { QueueAsyncDecode(image, image_id); return; } auto staging = runtime.UploadStagingBuffer(MapSizeBytes(image)); UploadImageContents(image, staging); runtime.InsertUploadMemoryBarrier(); } template template void TextureCache

::UploadImageContents(Image& image, StagingBuffer& staging) { const std::span mapped_span = staging.mapped_span; const GPUVAddr gpu_addr = image.gpu_addr; if (True(image.flags & ImageFlagBits::AcceleratedUpload)) { gpu_memory->ReadBlock(gpu_addr, mapped_span.data(), mapped_span.size_bytes(), VideoCommon::CacheType::NoTextureCache); const auto uploads = FullUploadSwizzles(image.info); runtime.AccelerateImageUpload(image, staging, uploads); return; } const size_t guest_size_bytes = image.guest_size_bytes; swizzle_data_buffer.resize_destructive(guest_size_bytes); gpu_memory->ReadBlockUnsafe(gpu_addr, swizzle_data_buffer.data(), guest_size_bytes); if (True(image.flags & ImageFlagBits::Converted)) { unswizzle_data_buffer.resize_destructive(image.unswizzled_size_bytes); auto copies = UnswizzleImage(*gpu_memory, gpu_addr, image.info, swizzle_data_buffer, unswizzle_data_buffer); ConvertImage(unswizzle_data_buffer, image.info, mapped_span, copies); image.UploadMemory(staging, copies); } else { const auto copies = UnswizzleImage(*gpu_memory, gpu_addr, image.info, swizzle_data_buffer, mapped_span); image.UploadMemory(staging, copies); } } template ImageViewId TextureCache

::FindImageView(const TICEntry& config) { if (!IsValidEntry(*gpu_memory, config)) { return NULL_IMAGE_VIEW_ID; } const auto [pair, is_new] = channel_state->image_views.try_emplace(config); ImageViewId& image_view_id = pair->second; if (is_new) { image_view_id = CreateImageView(config); } return image_view_id; } template ImageViewId TextureCache

::CreateImageView(const TICEntry& config) { const ImageInfo info(config); if (info.type == ImageType::Buffer) { const ImageViewInfo view_info(config, 0); return slot_image_views.insert(runtime, info, view_info, config.Address()); } const u32 layer_offset = config.BaseLayer() * info.layer_stride; const GPUVAddr image_gpu_addr = config.Address() - layer_offset; const ImageId image_id = FindOrInsertImage(info, image_gpu_addr); if (!image_id) { return NULL_IMAGE_VIEW_ID; } ImageBase& image = slot_images[image_id]; const SubresourceBase base = image.TryFindBase(config.Address()).value(); ASSERT(base.level == 0); const ImageViewInfo view_info(config, base.layer); const ImageViewId image_view_id = FindOrEmplaceImageView(image_id, view_info); ImageViewBase& image_view = slot_image_views[image_view_id]; image_view.flags |= ImageViewFlagBits::Strong; image.flags |= ImageFlagBits::Strong; return image_view_id; } template ImageId TextureCache

::FindOrInsertImage(const ImageInfo& info, GPUVAddr gpu_addr, RelaxedOptions options) { if (const ImageId image_id = FindImage(info, gpu_addr, options); image_id) { return image_id; } return InsertImage(info, gpu_addr, options); } template ImageId TextureCache

::FindImage(const ImageInfo& info, GPUVAddr gpu_addr, RelaxedOptions options) { std::optional cpu_addr = gpu_memory->GpuToCpuAddress(gpu_addr); if (!cpu_addr) { cpu_addr = gpu_memory->GpuToCpuAddress(gpu_addr, CalculateGuestSizeInBytes(info)); if (!cpu_addr) { return ImageId{}; } } const bool broken_views = runtime.HasBrokenTextureViewFormats() || True(options & RelaxedOptions::ForceBrokenViews); const bool native_bgr = runtime.HasNativeBgr(); const bool flexible_formats = True(options & RelaxedOptions::Format); ImageId image_id{}; boost::container::small_vector image_ids; const auto lambda = [&](ImageId existing_image_id, ImageBase& existing_image) { if (True(existing_image.flags & ImageFlagBits::Remapped)) { return false; } if (info.type == ImageType::Linear || existing_image.info.type == ImageType::Linear) [[unlikely]] { const bool strict_size = False(options & RelaxedOptions::Size) && True(existing_image.flags & ImageFlagBits::Strong); const ImageInfo& existing = existing_image.info; if (existing_image.gpu_addr == gpu_addr && existing.type == info.type && existing.pitch == info.pitch && IsPitchLinearSameSize(existing, info, strict_size) && IsViewCompatible(existing.format, info.format, broken_views, native_bgr)) { image_id = existing_image_id; image_ids.push_back(existing_image_id); return !flexible_formats && existing.format == info.format; } } else if (IsSubresource(info, existing_image, gpu_addr, options, broken_views, native_bgr)) { image_id = existing_image_id; image_ids.push_back(existing_image_id); return !flexible_formats && existing_image.info.format == info.format; } return false; }; ForEachImageInRegion(*cpu_addr, CalculateGuestSizeInBytes(info), lambda); if (image_ids.size() <= 1) [[likely]] { return image_id; } auto image_ids_compare = [this](ImageId a, ImageId b) { auto& image_a = slot_images[a]; auto& image_b = slot_images[b]; return image_a.modification_tick < image_b.modification_tick; }; return *std::ranges::max_element(image_ids, image_ids_compare); } template bool TextureCache

::ImageCanRescale(ImageBase& image) { if (!image.info.rescaleable) { return false; } if (Settings::values.resolution_info.downscale && !image.info.downscaleable) { return false; } if (True(image.flags & (ImageFlagBits::Rescaled | ImageFlagBits::CheckingRescalable))) { return true; } if (True(image.flags & ImageFlagBits::IsRescalable)) { return true; } image.flags |= ImageFlagBits::CheckingRescalable; for (const auto& alias : image.aliased_images) { Image& other_image = slot_images[alias.id]; if (!ImageCanRescale(other_image)) { image.flags &= ~ImageFlagBits::CheckingRescalable; return false; } } image.flags &= ~ImageFlagBits::CheckingRescalable; image.flags |= ImageFlagBits::IsRescalable; return true; } template void TextureCache

::InvalidateScale(Image& image) { if (image.scale_tick <= frame_tick) { image.scale_tick = frame_tick + 1; } const std::span image_view_ids = image.image_view_ids; auto& dirty = maxwell3d->dirty.flags; dirty[Dirty::RenderTargets] = true; dirty[Dirty::ZetaBuffer] = true; for (size_t rt = 0; rt < NUM_RT; ++rt) { dirty[Dirty::ColorBuffer0 + rt] = true; } for (const ImageViewId image_view_id : image_view_ids) { std::ranges::replace(render_targets.color_buffer_ids, image_view_id, ImageViewId{}); if (render_targets.depth_buffer_id == image_view_id) { render_targets.depth_buffer_id = ImageViewId{}; } } RemoveImageViewReferences(image_view_ids); RemoveFramebuffers(image_view_ids); for (const ImageViewId image_view_id : image_view_ids) { sentenced_image_view.Push(std::move(slot_image_views[image_view_id])); slot_image_views.erase(image_view_id); } image.image_view_ids.clear(); image.image_view_infos.clear(); for (size_t c : active_channel_ids) { auto& channel_info = channel_storage[c]; if constexpr (ENABLE_VALIDATION) { std::ranges::fill(channel_info.graphics_image_view_ids, CORRUPT_ID); std::ranges::fill(channel_info.compute_image_view_ids, CORRUPT_ID); } channel_info.graphics_image_table.Invalidate(); channel_info.compute_image_table.Invalidate(); } has_deleted_images = true; } template u64 TextureCache

::GetScaledImageSizeBytes(const ImageBase& image) { const u64 scale_up = static_cast(Settings::values.resolution_info.up_scale * Settings::values.resolution_info.up_scale); const u64 down_shift = static_cast(Settings::values.resolution_info.down_shift + Settings::values.resolution_info.down_shift); const u64 image_size_bytes = static_cast(std::max(image.guest_size_bytes, image.unswizzled_size_bytes)); const u64 tentative_size = (image_size_bytes * scale_up) >> down_shift; const u64 fitted_size = Common::AlignUp(tentative_size, 1024); return fitted_size; } template void TextureCache

::QueueAsyncDecode(Image& image, ImageId image_id) { UNIMPLEMENTED_IF(False(image.flags & ImageFlagBits::Converted)); LOG_INFO(HW_GPU, "Queuing async texture decode"); image.flags |= ImageFlagBits::IsDecoding; auto decode = std::make_unique(); auto* decode_ptr = decode.get(); decode->image_id = image_id; async_decodes.push_back(std::move(decode)); Common::ScratchBuffer local_unswizzle_data_buffer(image.unswizzled_size_bytes); const size_t guest_size_bytes = image.guest_size_bytes; swizzle_data_buffer.resize_destructive(guest_size_bytes); gpu_memory->ReadBlockUnsafe(image.gpu_addr, swizzle_data_buffer.data(), guest_size_bytes); auto copies = UnswizzleImage(*gpu_memory, image.gpu_addr, image.info, swizzle_data_buffer, local_unswizzle_data_buffer); const size_t out_size = MapSizeBytes(image); auto func = [out_size, copies, info = image.info, input = std::move(local_unswizzle_data_buffer), async_decode = decode_ptr]() mutable { async_decode->decoded_data.resize_destructive(out_size); std::span copies_span{copies.data(), copies.size()}; ConvertImage(input, info, async_decode->decoded_data, copies_span); // TODO: Do we need this lock? std::unique_lock lock{async_decode->mutex}; async_decode->copies = std::move(copies); async_decode->complete = true; }; texture_decode_worker.QueueWork(std::move(func)); } template void TextureCache

::TickAsyncDecode() { bool has_uploads{}; auto i = async_decodes.begin(); while (i != async_decodes.end()) { auto* async_decode = i->get(); std::unique_lock lock{async_decode->mutex}; if (!async_decode->complete) { ++i; continue; } Image& image = slot_images[async_decode->image_id]; auto staging = runtime.UploadStagingBuffer(MapSizeBytes(image)); std::memcpy(staging.mapped_span.data(), async_decode->decoded_data.data(), async_decode->decoded_data.size()); image.UploadMemory(staging, async_decode->copies); image.flags &= ~ImageFlagBits::IsDecoding; has_uploads = true; i = async_decodes.erase(i); } if (has_uploads) { runtime.InsertUploadMemoryBarrier(); } } template bool TextureCache

::ScaleUp(Image& image) { const bool has_copy = image.HasScaled(); const bool rescaled = image.ScaleUp(); if (!rescaled) { return false; } if (!has_copy) { total_used_memory += GetScaledImageSizeBytes(image); } InvalidateScale(image); return true; } template bool TextureCache

::ScaleDown(Image& image) { const bool rescaled = image.ScaleDown(); if (!rescaled) { return false; } InvalidateScale(image); return true; } template ImageId TextureCache

::InsertImage(const ImageInfo& info, GPUVAddr gpu_addr, RelaxedOptions options) { std::optional cpu_addr = gpu_memory->GpuToCpuAddress(gpu_addr); if (!cpu_addr) { const auto size = CalculateGuestSizeInBytes(info); cpu_addr = gpu_memory->GpuToCpuAddress(gpu_addr, size); if (!cpu_addr) { const VAddr fake_addr = ~(1ULL << 40ULL) + virtual_invalid_space; virtual_invalid_space += Common::AlignUp(size, 32); cpu_addr = std::optional(fake_addr); } } ASSERT_MSG(cpu_addr, "Tried to insert an image to an invalid gpu_addr=0x{:x}", gpu_addr); const ImageId image_id = JoinImages(info, gpu_addr, *cpu_addr); const Image& image = slot_images[image_id]; // Using "image.gpu_addr" instead of "gpu_addr" is important because it might be different const auto [it, is_new] = image_allocs_table.try_emplace(image.gpu_addr); if (is_new) { it->second = slot_image_allocs.insert(); } slot_image_allocs[it->second].images.push_back(image_id); return image_id; } template ImageId TextureCache

::JoinImages(const ImageInfo& info, GPUVAddr gpu_addr, VAddr cpu_addr) { ImageInfo new_info = info; const size_t size_bytes = CalculateGuestSizeInBytes(new_info); const bool broken_views = runtime.HasBrokenTextureViewFormats(); const bool native_bgr = runtime.HasNativeBgr(); join_overlap_ids.clear(); join_overlaps_found.clear(); join_left_aliased_ids.clear(); join_right_aliased_ids.clear(); join_ignore_textures.clear(); join_bad_overlap_ids.clear(); join_copies_to_do.clear(); join_alias_indices.clear(); const bool this_is_linear = info.type == ImageType::Linear; const auto region_check = [&](ImageId overlap_id, ImageBase& overlap) { if (True(overlap.flags & ImageFlagBits::Remapped)) { join_ignore_textures.insert(overlap_id); return; } const bool overlap_is_linear = overlap.info.type == ImageType::Linear; if (this_is_linear != overlap_is_linear) { return; } if (this_is_linear && overlap_is_linear) { if (info.pitch == overlap.info.pitch && gpu_addr == overlap.gpu_addr) { // Alias linear images with the same pitch join_left_aliased_ids.push_back(overlap_id); } return; } join_overlaps_found.insert(overlap_id); static constexpr bool strict_size = true; const std::optional solution = ResolveOverlap( new_info, gpu_addr, cpu_addr, overlap, strict_size, broken_views, native_bgr); if (solution) { gpu_addr = solution->gpu_addr; cpu_addr = solution->cpu_addr; new_info.resources = solution->resources; join_overlap_ids.push_back(overlap_id); join_copies_to_do.emplace_back(JoinCopy{false, overlap_id}); return; } static constexpr auto options = RelaxedOptions::Size | RelaxedOptions::Format; const ImageBase new_image_base(new_info, gpu_addr, cpu_addr); if (IsSubresource(new_info, overlap, gpu_addr, options, broken_views, native_bgr)) { join_left_aliased_ids.push_back(overlap_id); overlap.flags |= ImageFlagBits::Alias; join_copies_to_do.emplace_back(JoinCopy{true, overlap_id}); } else if (IsSubresource(overlap.info, new_image_base, overlap.gpu_addr, options, broken_views, native_bgr)) { join_right_aliased_ids.push_back(overlap_id); overlap.flags |= ImageFlagBits::Alias; join_copies_to_do.emplace_back(JoinCopy{true, overlap_id}); } else { join_bad_overlap_ids.push_back(overlap_id); } }; ForEachImageInRegion(cpu_addr, size_bytes, region_check); const auto region_check_gpu = [&](ImageId overlap_id, ImageBase& overlap) { if (!join_overlaps_found.contains(overlap_id)) { if (True(overlap.flags & ImageFlagBits::Remapped)) { join_ignore_textures.insert(overlap_id); } if (overlap.gpu_addr == gpu_addr && overlap.guest_size_bytes == size_bytes) { join_ignore_textures.insert(overlap_id); } } }; ForEachSparseImageInRegion(gpu_addr, size_bytes, region_check_gpu); bool can_rescale = info.rescaleable; bool any_rescaled = false; for (const auto& copy : join_copies_to_do) { if (!can_rescale) { break; } Image& sibling = slot_images[copy.id]; can_rescale &= ImageCanRescale(sibling); any_rescaled |= True(sibling.flags & ImageFlagBits::Rescaled); } can_rescale &= any_rescaled; if (can_rescale) { for (const auto& copy : join_copies_to_do) { Image& sibling = slot_images[copy.id]; ScaleUp(sibling); } } else { for (const auto& copy : join_copies_to_do) { Image& sibling = slot_images[copy.id]; ScaleDown(sibling); } } const ImageId new_image_id = slot_images.insert(runtime, new_info, gpu_addr, cpu_addr); Image& new_image = slot_images[new_image_id]; if (!gpu_memory->IsContinuousRange(new_image.gpu_addr, new_image.guest_size_bytes)) { new_image.flags |= ImageFlagBits::Sparse; } for (const ImageId overlap_id : join_ignore_textures) { Image& overlap = slot_images[overlap_id]; if (True(overlap.flags & ImageFlagBits::GpuModified)) { UNIMPLEMENTED(); } if (True(overlap.flags & ImageFlagBits::Tracked)) { UntrackImage(overlap, overlap_id); } UnregisterImage(overlap_id); DeleteImage(overlap_id); } // TODO: Only upload what we need RefreshContents(new_image, new_image_id); if (can_rescale) { ScaleUp(new_image); } else { ScaleDown(new_image); } std::ranges::sort(join_copies_to_do, [this](const JoinCopy& lhs, const JoinCopy& rhs) { const ImageBase& lhs_image = slot_images[lhs.id]; const ImageBase& rhs_image = slot_images[rhs.id]; return lhs_image.modification_tick < rhs_image.modification_tick; }); ImageBase& new_image_base = new_image; for (const ImageId aliased_id : join_right_aliased_ids) { ImageBase& aliased = slot_images[aliased_id]; size_t alias_index = new_image_base.aliased_images.size(); if (!AddImageAlias(new_image_base, aliased, new_image_id, aliased_id)) { continue; } join_alias_indices.emplace(aliased_id, alias_index); new_image.flags |= ImageFlagBits::Alias; } for (const ImageId aliased_id : join_left_aliased_ids) { ImageBase& aliased = slot_images[aliased_id]; size_t alias_index = new_image_base.aliased_images.size(); if (!AddImageAlias(aliased, new_image_base, aliased_id, new_image_id)) { continue; } join_alias_indices.emplace(aliased_id, alias_index); new_image.flags |= ImageFlagBits::Alias; } for (const ImageId aliased_id : join_bad_overlap_ids) { ImageBase& aliased = slot_images[aliased_id]; aliased.overlapping_images.push_back(new_image_id); new_image.overlapping_images.push_back(aliased_id); if (aliased.info.resources.levels == 1 && aliased.info.block.depth == 0 && aliased.overlapping_images.size() > 1) { aliased.flags |= ImageFlagBits::BadOverlap; } if (new_image.info.resources.levels == 1 && new_image.info.block.depth == 0 && new_image.overlapping_images.size() > 1) { new_image.flags |= ImageFlagBits::BadOverlap; } } for (const auto& copy_object : join_copies_to_do) { Image& overlap = slot_images[copy_object.id]; if (copy_object.is_alias) { if (!overlap.IsSafeDownload()) { continue; } const auto alias_pointer = join_alias_indices.find(copy_object.id); if (alias_pointer == join_alias_indices.end()) { continue; } const AliasedImage& aliased = new_image.aliased_images[alias_pointer->second]; CopyImage(new_image_id, aliased.id, aliased.copies); new_image.modification_tick = overlap.modification_tick; continue; } if (True(overlap.flags & ImageFlagBits::GpuModified)) { new_image.flags |= ImageFlagBits::GpuModified; const auto& resolution = Settings::values.resolution_info; const SubresourceBase base = new_image.TryFindBase(overlap.gpu_addr).value(); const u32 up_scale = can_rescale ? resolution.up_scale : 1; const u32 down_shift = can_rescale ? resolution.down_shift : 0; auto copies = MakeShrinkImageCopies(new_info, overlap.info, base, up_scale, down_shift); if (overlap.info.num_samples != new_image.info.num_samples) { runtime.CopyImageMSAA(new_image, overlap, std::move(copies)); } else { runtime.CopyImage(new_image, overlap, std::move(copies)); } new_image.modification_tick = overlap.modification_tick; } if (True(overlap.flags & ImageFlagBits::Tracked)) { UntrackImage(overlap, copy_object.id); } UnregisterImage(copy_object.id); DeleteImage(copy_object.id); } RegisterImage(new_image_id); return new_image_id; } template std::optional::BlitImages> TextureCache

::GetBlitImages( const Tegra::Engines::Fermi2D::Surface& dst, const Tegra::Engines::Fermi2D::Surface& src, const Tegra::Engines::Fermi2D::Config& copy) { static constexpr auto FIND_OPTIONS = RelaxedOptions::Samples; const GPUVAddr dst_addr = dst.Address(); const GPUVAddr src_addr = src.Address(); ImageInfo dst_info(dst); ImageInfo src_info(src); const bool can_be_depth_blit = dst_info.format == src_info.format && copy.filter == Tegra::Engines::Fermi2D::Filter::Point; ImageId dst_id; ImageId src_id; RelaxedOptions try_options = FIND_OPTIONS; if (can_be_depth_blit) { try_options |= RelaxedOptions::Format; } do { has_deleted_images = false; src_id = FindImage(src_info, src_addr, try_options); dst_id = FindImage(dst_info, dst_addr, try_options); if (!copy.must_accelerate) { do { if (!src_id && !dst_id) { break; } if (src_id && True(slot_images[src_id].flags & ImageFlagBits::GpuModified)) { break; } if (dst_id && True(slot_images[dst_id].flags & ImageFlagBits::GpuModified)) { break; } return std::nullopt; } while (false); } const ImageBase* const src_image = src_id ? &slot_images[src_id] : nullptr; if (src_image && src_image->info.num_samples > 1) { RelaxedOptions find_options{FIND_OPTIONS | RelaxedOptions::ForceBrokenViews}; src_id = FindOrInsertImage(src_info, src_addr, find_options); dst_id = FindOrInsertImage(dst_info, dst_addr, find_options); if (has_deleted_images) { continue; } break; } if (can_be_depth_blit) { const ImageBase* const dst_image = dst_id ? &slot_images[dst_id] : nullptr; DeduceBlitImages(dst_info, src_info, dst_image, src_image); if (GetFormatType(dst_info.format) != GetFormatType(src_info.format)) { continue; } } if (!src_id) { src_id = InsertImage(src_info, src_addr, RelaxedOptions{}); } if (!dst_id) { dst_id = InsertImage(dst_info, dst_addr, RelaxedOptions{}); } } while (has_deleted_images); const ImageBase& src_image = slot_images[src_id]; const ImageBase& dst_image = slot_images[dst_id]; const bool native_bgr = runtime.HasNativeBgr(); if (GetFormatType(dst_info.format) != GetFormatType(dst_image.info.format) || GetFormatType(src_info.format) != GetFormatType(src_image.info.format) || !VideoCore::Surface::IsViewCompatible(dst_info.format, dst_image.info.format, false, native_bgr) || !VideoCore::Surface::IsViewCompatible(src_info.format, src_image.info.format, false, native_bgr)) { // Make sure the images match the expected format. do { has_deleted_images = false; src_id = FindOrInsertImage(src_info, src_addr, RelaxedOptions{}); dst_id = FindOrInsertImage(dst_info, dst_addr, RelaxedOptions{}); } while (has_deleted_images); } return {BlitImages{ .dst_id = dst_id, .src_id = src_id, .dst_format = dst_info.format, .src_format = src_info.format, }}; } template ImageId TextureCache

::FindDMAImage(const ImageInfo& info, GPUVAddr gpu_addr) { std::optional cpu_addr = gpu_memory->GpuToCpuAddress(gpu_addr); if (!cpu_addr) { cpu_addr = gpu_memory->GpuToCpuAddress(gpu_addr, CalculateGuestSizeInBytes(info)); if (!cpu_addr) { return ImageId{}; } } ImageId image_id{}; boost::container::small_vector image_ids; const auto lambda = [&](ImageId existing_image_id, ImageBase& existing_image) { if (True(existing_image.flags & ImageFlagBits::Remapped)) { return false; } if (info.type == ImageType::Linear || existing_image.info.type == ImageType::Linear) [[unlikely]] { const bool strict_size = True(existing_image.flags & ImageFlagBits::Strong); const ImageInfo& existing = existing_image.info; if (existing_image.gpu_addr == gpu_addr && existing.type == info.type && existing.pitch == info.pitch && IsPitchLinearSameSize(existing, info, strict_size) && IsViewCompatible(existing.format, info.format, false, true)) { image_id = existing_image_id; image_ids.push_back(existing_image_id); return true; } } else if (IsSubCopy(info, existing_image, gpu_addr)) { image_id = existing_image_id; image_ids.push_back(existing_image_id); return true; } return false; }; ForEachImageInRegion(*cpu_addr, CalculateGuestSizeInBytes(info), lambda); if (image_ids.size() <= 1) [[likely]] { return image_id; } auto image_ids_compare = [this](ImageId a, ImageId b) { auto& image_a = slot_images[a]; auto& image_b = slot_images[b]; return image_a.modification_tick < image_b.modification_tick; }; return *std::ranges::max_element(image_ids, image_ids_compare); } template std::pair TextureCache

::PrepareDmaImage(ImageId dst_id, GPUVAddr base_addr, bool mark_as_modified) { const auto& image = slot_images[dst_id]; const auto base = image.TryFindBase(base_addr); PrepareImage(dst_id, mark_as_modified, false); const auto& new_image = slot_images[dst_id]; lru_cache.Touch(new_image.lru_index, frame_tick); return std::make_pair(base->level, base->layer); } template SamplerId TextureCache

::FindSampler(const TSCEntry& config) { if (std::ranges::all_of(config.raw, [](u64 value) { return value == 0; })) { return NULL_SAMPLER_ID; } const auto [pair, is_new] = channel_state->samplers.try_emplace(config); if (is_new) { pair->second = slot_samplers.insert(runtime, config); } return pair->second; } template ImageViewId TextureCache

::FindColorBuffer(size_t index, bool is_clear) { const auto& regs = maxwell3d->regs; if (index >= regs.rt_control.count) { return ImageViewId{}; } const auto& rt = regs.rt[index]; const GPUVAddr gpu_addr = rt.Address(); if (gpu_addr == 0) { return ImageViewId{}; } if (rt.format == Tegra::RenderTargetFormat::NONE) { return ImageViewId{}; } const ImageInfo info(regs.rt[index], regs.anti_alias_samples_mode); return FindRenderTargetView(info, gpu_addr, is_clear); } template ImageViewId TextureCache

::FindDepthBuffer(bool is_clear) { const auto& regs = maxwell3d->regs; if (!regs.zeta_enable) { return ImageViewId{}; } const GPUVAddr gpu_addr = regs.zeta.Address(); if (gpu_addr == 0) { return ImageViewId{}; } const ImageInfo info(regs.zeta, regs.zeta_size, regs.anti_alias_samples_mode); return FindRenderTargetView(info, gpu_addr, is_clear); } template ImageViewId TextureCache

::FindRenderTargetView(const ImageInfo& info, GPUVAddr gpu_addr, bool is_clear) { const auto options = is_clear ? RelaxedOptions::Samples : RelaxedOptions{}; ImageId image_id{}; bool delete_state = has_deleted_images; do { has_deleted_images = false; image_id = FindOrInsertImage(info, gpu_addr, options); delete_state |= has_deleted_images; } while (has_deleted_images); has_deleted_images = delete_state; if (!image_id) { return NULL_IMAGE_VIEW_ID; } Image& image = slot_images[image_id]; const ImageViewType view_type = RenderTargetImageViewType(info); SubresourceBase base; if (image.info.type == ImageType::Linear) { base = SubresourceBase{.level = 0, .layer = 0}; } else { base = image.TryFindBase(gpu_addr).value(); } const s32 layers = image.info.type == ImageType::e3D ? info.size.depth : info.resources.layers; const SubresourceRange range{ .base = base, .extent = {.levels = 1, .layers = layers}, }; return FindOrEmplaceImageView(image_id, ImageViewInfo(view_type, info.format, range)); } template template void TextureCache

::ForEachImageInRegion(VAddr cpu_addr, size_t size, Func&& func) { using FuncReturn = typename std::invoke_result::type; static constexpr bool BOOL_BREAK = std::is_same_v; boost::container::small_vector images; boost::container::small_vector maps; ForEachCPUPage(cpu_addr, size, [this, &images, &maps, cpu_addr, size, func](u64 page) { const auto it = page_table.find(page); if (it == page_table.end()) { if constexpr (BOOL_BREAK) { return false; } else { return; } } for (const ImageMapId map_id : it->second) { ImageMapView& map = slot_map_views[map_id]; if (map.picked) { continue; } if (!map.Overlaps(cpu_addr, size)) { continue; } map.picked = true; maps.push_back(map_id); Image& image = slot_images[map.image_id]; if (True(image.flags & ImageFlagBits::Picked)) { continue; } image.flags |= ImageFlagBits::Picked; images.push_back(map.image_id); if constexpr (BOOL_BREAK) { if (func(map.image_id, image)) { return true; } } else { func(map.image_id, image); } } if constexpr (BOOL_BREAK) { return false; } }); for (const ImageId image_id : images) { slot_images[image_id].flags &= ~ImageFlagBits::Picked; } for (const ImageMapId map_id : maps) { slot_map_views[map_id].picked = false; } } template template void TextureCache

::ForEachImageInRegionGPU(size_t as_id, GPUVAddr gpu_addr, size_t size, Func&& func) { using FuncReturn = typename std::invoke_result::type; static constexpr bool BOOL_BREAK = std::is_same_v; boost::container::small_vector images; auto storage_id = getStorageID(as_id); if (!storage_id) { return; } auto& gpu_page_table = gpu_page_table_storage[*storage_id]; ForEachGPUPage(gpu_addr, size, [this, &gpu_page_table, &images, gpu_addr, size, func](u64 page) { const auto it = gpu_page_table.find(page); if (it == gpu_page_table.end()) { if constexpr (BOOL_BREAK) { return false; } else { return; } } for (const ImageId image_id : it->second) { Image& image = slot_images[image_id]; if (True(image.flags & ImageFlagBits::Picked)) { continue; } if (!image.OverlapsGPU(gpu_addr, size)) { continue; } image.flags |= ImageFlagBits::Picked; images.push_back(image_id); if constexpr (BOOL_BREAK) { if (func(image_id, image)) { return true; } } else { func(image_id, image); } } if constexpr (BOOL_BREAK) { return false; } }); for (const ImageId image_id : images) { slot_images[image_id].flags &= ~ImageFlagBits::Picked; } } template template void TextureCache

::ForEachSparseImageInRegion(GPUVAddr gpu_addr, size_t size, Func&& func) { using FuncReturn = typename std::invoke_result::type; static constexpr bool BOOL_BREAK = std::is_same_v; boost::container::small_vector images; ForEachGPUPage(gpu_addr, size, [this, &images, gpu_addr, size, func](u64 page) { const auto it = sparse_page_table.find(page); if (it == sparse_page_table.end()) { if constexpr (BOOL_BREAK) { return false; } else { return; } } for (const ImageId image_id : it->second) { Image& image = slot_images[image_id]; if (True(image.flags & ImageFlagBits::Picked)) { continue; } if (!image.OverlapsGPU(gpu_addr, size)) { continue; } image.flags |= ImageFlagBits::Picked; images.push_back(image_id); if constexpr (BOOL_BREAK) { if (func(image_id, image)) { return true; } } else { func(image_id, image); } } if constexpr (BOOL_BREAK) { return false; } }); for (const ImageId image_id : images) { slot_images[image_id].flags &= ~ImageFlagBits::Picked; } } template template void TextureCache

::ForEachSparseSegment(ImageBase& image, Func&& func) { using FuncReturn = typename std::invoke_result::type; static constexpr bool RETURNS_BOOL = std::is_same_v; const auto segments = gpu_memory->GetSubmappedRange(image.gpu_addr, image.guest_size_bytes); for (const auto& [gpu_addr, size] : segments) { std::optional cpu_addr = gpu_memory->GpuToCpuAddress(gpu_addr); ASSERT(cpu_addr); if constexpr (RETURNS_BOOL) { if (func(gpu_addr, *cpu_addr, size)) { break; } } else { func(gpu_addr, *cpu_addr, size); } } } template ImageViewId TextureCache

::FindOrEmplaceImageView(ImageId image_id, const ImageViewInfo& info) { Image& image = slot_images[image_id]; if (const ImageViewId image_view_id = image.FindView(info); image_view_id) { return image_view_id; } const ImageViewId image_view_id = slot_image_views.insert(runtime, info, image_id, image, slot_images); image.InsertView(info, image_view_id); return image_view_id; } template void TextureCache

::RegisterImage(ImageId image_id) { ImageBase& image = slot_images[image_id]; ASSERT_MSG(False(image.flags & ImageFlagBits::Registered), "Trying to register an already registered image"); image.flags |= ImageFlagBits::Registered; u64 tentative_size = std::max(image.guest_size_bytes, image.unswizzled_size_bytes); if ((IsPixelFormatASTC(image.info.format) && True(image.flags & ImageFlagBits::AcceleratedUpload)) || True(image.flags & ImageFlagBits::Converted)) { tentative_size = EstimatedDecompressedSize(tentative_size, image.info.format); } total_used_memory += Common::AlignUp(tentative_size, 1024); if (total_used_memory > critical_memory && critical_gc < GC_EMERGENCY_COUNTS) { RunGarbageCollector(); critical_gc++; } image.lru_index = lru_cache.Insert(image_id, frame_tick); ForEachGPUPage(image.gpu_addr, image.guest_size_bytes, [this, image_id](u64 page) { (*channel_state->gpu_page_table)[page].push_back(image_id); }); if (False(image.flags & ImageFlagBits::Sparse)) { auto map_id = slot_map_views.insert(image.gpu_addr, image.cpu_addr, image.guest_size_bytes, image_id); ForEachCPUPage(image.cpu_addr, image.guest_size_bytes, [this, map_id](u64 page) { page_table[page].push_back(map_id); }); image.map_view_id = map_id; return; } std::vector sparse_maps{}; ForEachSparseSegment( image, [this, image_id, &sparse_maps](GPUVAddr gpu_addr, VAddr cpu_addr, size_t size) { auto map_id = slot_map_views.insert(gpu_addr, cpu_addr, size, image_id); ForEachCPUPage(cpu_addr, size, [this, map_id](u64 page) { page_table[page].push_back(map_id); }); sparse_maps.push_back(map_id); }); sparse_views.emplace(image_id, std::move(sparse_maps)); ForEachGPUPage(image.gpu_addr, image.guest_size_bytes, [this, image_id](u64 page) { sparse_page_table[page].push_back(image_id); }); } template void TextureCache

::UnregisterImage(ImageId image_id) { Image& image = slot_images[image_id]; ASSERT_MSG(True(image.flags & ImageFlagBits::Registered), "Trying to unregister an already registered image"); image.flags &= ~ImageFlagBits::Registered; image.flags &= ~ImageFlagBits::BadOverlap; lru_cache.Free(image.lru_index); const auto& clear_page_table = [image_id](u64 page, std::unordered_map, Common::IdentityHash>& selected_page_table) { const auto page_it = selected_page_table.find(page); if (page_it == selected_page_table.end()) { ASSERT_MSG(false, "Unregistering unregistered page=0x{:x}", page << YUZU_PAGEBITS); return; } std::vector& image_ids = page_it->second; const auto vector_it = std::ranges::find(image_ids, image_id); if (vector_it == image_ids.end()) { ASSERT_MSG(false, "Unregistering unregistered image in page=0x{:x}", page << YUZU_PAGEBITS); return; } image_ids.erase(vector_it); }; ForEachGPUPage(image.gpu_addr, image.guest_size_bytes, [this, &clear_page_table](u64 page) { clear_page_table(page, (*channel_state->gpu_page_table)); }); if (False(image.flags & ImageFlagBits::Sparse)) { const auto map_id = image.map_view_id; ForEachCPUPage(image.cpu_addr, image.guest_size_bytes, [this, map_id](u64 page) { const auto page_it = page_table.find(page); if (page_it == page_table.end()) { ASSERT_MSG(false, "Unregistering unregistered page=0x{:x}", page << YUZU_PAGEBITS); return; } std::vector& image_map_ids = page_it->second; const auto vector_it = std::ranges::find(image_map_ids, map_id); if (vector_it == image_map_ids.end()) { ASSERT_MSG(false, "Unregistering unregistered image in page=0x{:x}", page << YUZU_PAGEBITS); return; } image_map_ids.erase(vector_it); }); slot_map_views.erase(map_id); return; } ForEachGPUPage(image.gpu_addr, image.guest_size_bytes, [this, &clear_page_table](u64 page) { clear_page_table(page, sparse_page_table); }); auto it = sparse_views.find(image_id); ASSERT(it != sparse_views.end()); auto& sparse_maps = it->second; for (auto& map_view_id : sparse_maps) { const auto& map_range = slot_map_views[map_view_id]; const VAddr cpu_addr = map_range.cpu_addr; const std::size_t size = map_range.size; ForEachCPUPage(cpu_addr, size, [this, image_id](u64 page) { const auto page_it = page_table.find(page); if (page_it == page_table.end()) { ASSERT_MSG(false, "Unregistering unregistered page=0x{:x}", page << YUZU_PAGEBITS); return; } std::vector& image_map_ids = page_it->second; auto vector_it = image_map_ids.begin(); while (vector_it != image_map_ids.end()) { ImageMapView& map = slot_map_views[*vector_it]; if (map.image_id != image_id) { vector_it++; continue; } if (!map.picked) { map.picked = true; } vector_it = image_map_ids.erase(vector_it); } }); slot_map_views.erase(map_view_id); } sparse_views.erase(it); } template void TextureCache

::TrackImage(ImageBase& image, ImageId image_id) { ASSERT(False(image.flags & ImageFlagBits::Tracked)); image.flags |= ImageFlagBits::Tracked; if (False(image.flags & ImageFlagBits::Sparse)) { rasterizer.UpdatePagesCachedCount(image.cpu_addr, image.guest_size_bytes, 1); return; } if (True(image.flags & ImageFlagBits::Registered)) { auto it = sparse_views.find(image_id); ASSERT(it != sparse_views.end()); auto& sparse_maps = it->second; for (auto& map_view_id : sparse_maps) { const auto& map = slot_map_views[map_view_id]; const VAddr cpu_addr = map.cpu_addr; const std::size_t size = map.size; rasterizer.UpdatePagesCachedCount(cpu_addr, size, 1); } return; } ForEachSparseSegment(image, [this]([[maybe_unused]] GPUVAddr gpu_addr, VAddr cpu_addr, size_t size) { rasterizer.UpdatePagesCachedCount(cpu_addr, size, 1); }); } template void TextureCache

::UntrackImage(ImageBase& image, ImageId image_id) { ASSERT(True(image.flags & ImageFlagBits::Tracked)); image.flags &= ~ImageFlagBits::Tracked; if (False(image.flags & ImageFlagBits::Sparse)) { rasterizer.UpdatePagesCachedCount(image.cpu_addr, image.guest_size_bytes, -1); return; } ASSERT(True(image.flags & ImageFlagBits::Registered)); auto it = sparse_views.find(image_id); ASSERT(it != sparse_views.end()); auto& sparse_maps = it->second; for (auto& map_view_id : sparse_maps) { const auto& map = slot_map_views[map_view_id]; const VAddr cpu_addr = map.cpu_addr; const std::size_t size = map.size; rasterizer.UpdatePagesCachedCount(cpu_addr, size, -1); } } template void TextureCache

::DeleteImage(ImageId image_id, bool immediate_delete) { ImageBase& image = slot_images[image_id]; if (image.HasScaled()) { total_used_memory -= GetScaledImageSizeBytes(image); } u64 tentative_size = std::max(image.guest_size_bytes, image.unswizzled_size_bytes); if ((IsPixelFormatASTC(image.info.format) && True(image.flags & ImageFlagBits::AcceleratedUpload)) || True(image.flags & ImageFlagBits::Converted)) { tentative_size = EstimatedDecompressedSize(tentative_size, image.info.format); } total_used_memory -= Common::AlignUp(tentative_size, 1024); const GPUVAddr gpu_addr = image.gpu_addr; const auto alloc_it = image_allocs_table.find(gpu_addr); if (alloc_it == image_allocs_table.end()) { ASSERT_MSG(false, "Trying to delete an image alloc that does not exist in address 0x{:x}", gpu_addr); return; } const ImageAllocId alloc_id = alloc_it->second; std::vector& alloc_images = slot_image_allocs[alloc_id].images; const auto alloc_image_it = std::ranges::find(alloc_images, image_id); if (alloc_image_it == alloc_images.end()) { ASSERT_MSG(false, "Trying to delete an image that does not exist"); return; } ASSERT_MSG(False(image.flags & ImageFlagBits::Tracked), "Image was not untracked"); ASSERT_MSG(False(image.flags & ImageFlagBits::Registered), "Image was not unregistered"); // Mark render targets as dirty auto& dirty = maxwell3d->dirty.flags; dirty[Dirty::RenderTargets] = true; dirty[Dirty::ZetaBuffer] = true; for (size_t rt = 0; rt < NUM_RT; ++rt) { dirty[Dirty::ColorBuffer0 + rt] = true; } const std::span image_view_ids = image.image_view_ids; for (const ImageViewId image_view_id : image_view_ids) { std::ranges::replace(render_targets.color_buffer_ids, image_view_id, ImageViewId{}); if (render_targets.depth_buffer_id == image_view_id) { render_targets.depth_buffer_id = ImageViewId{}; } } RemoveImageViewReferences(image_view_ids); RemoveFramebuffers(image_view_ids); for (const AliasedImage& alias : image.aliased_images) { ImageBase& other_image = slot_images[alias.id]; [[maybe_unused]] const size_t num_removed_aliases = std::erase_if(other_image.aliased_images, [image_id](const AliasedImage& other_alias) { return other_alias.id == image_id; }); other_image.CheckAliasState(); ASSERT_MSG(num_removed_aliases == 1, "Invalid number of removed aliases: {}", num_removed_aliases); } for (const ImageId overlap_id : image.overlapping_images) { ImageBase& other_image = slot_images[overlap_id]; [[maybe_unused]] const size_t num_removed_overlaps = std::erase_if( other_image.overlapping_images, [image_id](const ImageId other_overlap_id) { return other_overlap_id == image_id; }); other_image.CheckBadOverlapState(); ASSERT_MSG(num_removed_overlaps == 1, "Invalid number of removed overlapps: {}", num_removed_overlaps); } for (const ImageViewId image_view_id : image_view_ids) { if (!immediate_delete) { sentenced_image_view.Push(std::move(slot_image_views[image_view_id])); } slot_image_views.erase(image_view_id); } if (!immediate_delete) { sentenced_images.Push(std::move(slot_images[image_id])); } slot_images.erase(image_id); alloc_images.erase(alloc_image_it); if (alloc_images.empty()) { image_allocs_table.erase(alloc_it); } for (size_t c : active_channel_ids) { auto& channel_info = channel_storage[c]; if constexpr (ENABLE_VALIDATION) { std::ranges::fill(channel_info.graphics_image_view_ids, CORRUPT_ID); std::ranges::fill(channel_info.compute_image_view_ids, CORRUPT_ID); } channel_info.graphics_image_table.Invalidate(); channel_info.compute_image_table.Invalidate(); } has_deleted_images = true; } template void TextureCache

::RemoveImageViewReferences(std::span removed_views) { for (size_t c : active_channel_ids) { auto& channel_info = channel_storage[c]; auto it = channel_info.image_views.begin(); while (it != channel_info.image_views.end()) { const auto found = std::ranges::find(removed_views, it->second); if (found != removed_views.end()) { it = channel_info.image_views.erase(it); } else { ++it; } } } } template void TextureCache

::RemoveFramebuffers(std::span removed_views) { auto it = framebuffers.begin(); while (it != framebuffers.end()) { if (it->first.Contains(removed_views)) { auto framebuffer_id = it->second; ASSERT(framebuffer_id); sentenced_framebuffers.Push(std::move(slot_framebuffers[framebuffer_id])); it = framebuffers.erase(it); } else { ++it; } } } template void TextureCache

::MarkModification(ImageBase& image) noexcept { image.flags |= ImageFlagBits::GpuModified; image.modification_tick = ++modification_tick; } template void TextureCache

::SynchronizeAliases(ImageId image_id) { boost::container::small_vector aliased_images; Image& image = slot_images[image_id]; bool any_rescaled = True(image.flags & ImageFlagBits::Rescaled); bool any_modified = True(image.flags & ImageFlagBits::GpuModified); u64 most_recent_tick = image.modification_tick; for (const AliasedImage& aliased : image.aliased_images) { ImageBase& aliased_image = slot_images[aliased.id]; if (image.modification_tick < aliased_image.modification_tick) { most_recent_tick = std::max(most_recent_tick, aliased_image.modification_tick); aliased_images.push_back(&aliased); any_rescaled |= True(aliased_image.flags & ImageFlagBits::Rescaled); any_modified |= True(aliased_image.flags & ImageFlagBits::GpuModified); } } if (aliased_images.empty()) { return; } const bool can_rescale = ImageCanRescale(image); if (any_rescaled) { if (can_rescale) { ScaleUp(image); } else { ScaleDown(image); } } image.modification_tick = most_recent_tick; if (any_modified) { image.flags |= ImageFlagBits::GpuModified; } std::ranges::sort(aliased_images, [this](const AliasedImage* lhs, const AliasedImage* rhs) { const ImageBase& lhs_image = slot_images[lhs->id]; const ImageBase& rhs_image = slot_images[rhs->id]; return lhs_image.modification_tick < rhs_image.modification_tick; }); const auto& resolution = Settings::values.resolution_info; for (const AliasedImage* const aliased : aliased_images) { if (!resolution.active || !any_rescaled) { CopyImage(image_id, aliased->id, aliased->copies); continue; } Image& aliased_image = slot_images[aliased->id]; if (!can_rescale) { ScaleDown(aliased_image); CopyImage(image_id, aliased->id, aliased->copies); continue; } ScaleUp(aliased_image); CopyImage(image_id, aliased->id, aliased->copies); } } template void TextureCache

::PrepareImage(ImageId image_id, bool is_modification, bool invalidate) { Image& image = slot_images[image_id]; if (invalidate) { image.flags &= ~(ImageFlagBits::CpuModified | ImageFlagBits::GpuModified); if (False(image.flags & ImageFlagBits::Tracked)) { TrackImage(image, image_id); } } else { RefreshContents(image, image_id); SynchronizeAliases(image_id); } if (is_modification) { MarkModification(image); } lru_cache.Touch(image.lru_index, frame_tick); } template void TextureCache

::PrepareImageView(ImageViewId image_view_id, bool is_modification, bool invalidate) { if (!image_view_id) { return; } const ImageViewBase& image_view = slot_image_views[image_view_id]; if (image_view.IsBuffer()) { return; } PrepareImage(image_view.image_id, is_modification, invalidate); } template void TextureCache

::CopyImage(ImageId dst_id, ImageId src_id, std::vector copies) { Image& dst = slot_images[dst_id]; Image& src = slot_images[src_id]; const bool is_rescaled = True(src.flags & ImageFlagBits::Rescaled); if (is_rescaled) { ASSERT(True(dst.flags & ImageFlagBits::Rescaled)); const bool both_2d{src.info.type == ImageType::e2D && dst.info.type == ImageType::e2D}; const auto& resolution = Settings::values.resolution_info; for (auto& copy : copies) { copy.src_offset.x = resolution.ScaleUp(copy.src_offset.x); copy.dst_offset.x = resolution.ScaleUp(copy.dst_offset.x); copy.extent.width = resolution.ScaleUp(copy.extent.width); if (both_2d) { copy.src_offset.y = resolution.ScaleUp(copy.src_offset.y); copy.dst_offset.y = resolution.ScaleUp(copy.dst_offset.y); copy.extent.height = resolution.ScaleUp(copy.extent.height); } } } const auto dst_format_type = GetFormatType(dst.info.format); const auto src_format_type = GetFormatType(src.info.format); if (src_format_type == dst_format_type) { if constexpr (HAS_EMULATED_COPIES) { if (!runtime.CanImageBeCopied(dst, src)) { return runtime.EmulateCopyImage(dst, src, copies); } } return runtime.CopyImage(dst, src, copies); } UNIMPLEMENTED_IF(dst.info.type != ImageType::e2D); UNIMPLEMENTED_IF(src.info.type != ImageType::e2D); if (runtime.ShouldReinterpret(dst, src)) { return runtime.ReinterpretImage(dst, src, copies); } for (const ImageCopy& copy : copies) { UNIMPLEMENTED_IF(copy.dst_subresource.num_layers != 1); UNIMPLEMENTED_IF(copy.src_subresource.num_layers != 1); UNIMPLEMENTED_IF(copy.src_offset != Offset3D{}); UNIMPLEMENTED_IF(copy.dst_offset != Offset3D{}); const SubresourceBase dst_base{ .level = copy.dst_subresource.base_level, .layer = copy.dst_subresource.base_layer, }; const SubresourceBase src_base{ .level = copy.src_subresource.base_level, .layer = copy.src_subresource.base_layer, }; const SubresourceExtent dst_extent{.levels = 1, .layers = 1}; const SubresourceExtent src_extent{.levels = 1, .layers = 1}; const SubresourceRange dst_range{.base = dst_base, .extent = dst_extent}; const SubresourceRange src_range{.base = src_base, .extent = src_extent}; PixelFormat dst_format = dst.info.format; if (GetFormatType(src.info.format) == SurfaceType::DepthStencil && GetFormatType(dst_format) == SurfaceType::ColorTexture && BytesPerBlock(dst_format) == 4) { dst_format = PixelFormat::A8B8G8R8_UNORM; } const ImageViewInfo dst_view_info(ImageViewType::e2D, dst_format, dst_range); const ImageViewInfo src_view_info(ImageViewType::e2D, src.info.format, src_range); const auto [dst_framebuffer_id, dst_view_id] = RenderTargetFromImage(dst_id, dst_view_info); Framebuffer* const dst_framebuffer = &slot_framebuffers[dst_framebuffer_id]; const ImageViewId src_view_id = FindOrEmplaceImageView(src_id, src_view_info); ImageView& dst_view = slot_image_views[dst_view_id]; ImageView& src_view = slot_image_views[src_view_id]; [[maybe_unused]] const Extent3D expected_size{ .width = std::min(dst_view.size.width, src_view.size.width), .height = std::min(dst_view.size.height, src_view.size.height), .depth = std::min(dst_view.size.depth, src_view.size.depth), }; const Extent3D scaled_extent = [is_rescaled, expected_size]() { if (!is_rescaled) { return expected_size; } const auto& resolution = Settings::values.resolution_info; return Extent3D{ .width = resolution.ScaleUp(expected_size.width), .height = resolution.ScaleUp(expected_size.height), .depth = expected_size.depth, }; }(); UNIMPLEMENTED_IF(copy.extent != scaled_extent); runtime.ConvertImage(dst_framebuffer, dst_view, src_view); } } template void TextureCache

::BindRenderTarget(ImageViewId* old_id, ImageViewId new_id) { if (*old_id == new_id) { return; } if (new_id) { const ImageViewBase& old_view = slot_image_views[new_id]; if (True(old_view.flags & ImageViewFlagBits::PreemtiveDownload)) { const PendingDownload new_download{true, 0, old_view.image_id}; uncommitted_downloads.emplace_back(new_download); } } *old_id = new_id; } template std::pair TextureCache

::RenderTargetFromImage( ImageId image_id, const ImageViewInfo& view_info) { const ImageViewId view_id = FindOrEmplaceImageView(image_id, view_info); const ImageBase& image = slot_images[image_id]; const bool is_rescaled = True(image.flags & ImageFlagBits::Rescaled); const bool is_color = GetFormatType(image.info.format) == SurfaceType::ColorTexture; const ImageViewId color_view_id = is_color ? view_id : ImageViewId{}; const ImageViewId depth_view_id = is_color ? ImageViewId{} : view_id; Extent3D extent = MipSize(image.info.size, view_info.range.base.level); if (is_rescaled) { const auto& resolution = Settings::values.resolution_info; extent.width = resolution.ScaleUp(extent.width); if (image.info.type == ImageType::e2D) { extent.height = resolution.ScaleUp(extent.height); } } const u32 num_samples = image.info.num_samples; const auto [samples_x, samples_y] = SamplesLog2(num_samples); const FramebufferId framebuffer_id = GetFramebufferId(RenderTargets{ .color_buffer_ids = {color_view_id}, .depth_buffer_id = depth_view_id, .size = {extent.width >> samples_x, extent.height >> samples_y}, .is_rescaled = is_rescaled, }); return {framebuffer_id, view_id}; } template bool TextureCache

::IsFullClear(ImageViewId id) { if (!id) { return true; } const ImageViewBase& image_view = slot_image_views[id]; const ImageBase& image = slot_images[image_view.image_id]; const Extent3D size = image_view.size; const auto& regs = maxwell3d->regs; const auto& scissor = regs.scissor_test[0]; if (image.info.resources.levels > 1 || image.info.resources.layers > 1) { // Images with multiple resources can't be cleared in a single call return false; } if (regs.clear_control.use_scissor == 0) { // If scissor testing is disabled, the clear is always full return true; } // Make sure the clear covers all texels in the subresource return scissor.min_x == 0 && scissor.min_y == 0 && scissor.max_x >= size.width && scissor.max_y >= size.height; } template void TextureCache

::CreateChannel(struct Tegra::Control::ChannelState& channel) { VideoCommon::ChannelSetupCaches::CreateChannel(channel); const auto it = channel_map.find(channel.bind_id); auto* this_state = &channel_storage[it->second]; const auto& this_as_ref = address_spaces[channel.memory_manager->GetID()]; this_state->gpu_page_table = &gpu_page_table_storage[this_as_ref.storage_id]; } /// Bind a channel for execution. template void TextureCache

::OnGPUASRegister([[maybe_unused]] size_t map_id) { gpu_page_table_storage.emplace_back(); } } // namespace VideoCommon