diff options
Diffstat (limited to 'src/video_core')
52 files changed, 2165 insertions, 1864 deletions
diff --git a/src/video_core/CMakeLists.txt b/src/video_core/CMakeLists.txt index 9b931976a..47190c464 100644 --- a/src/video_core/CMakeLists.txt +++ b/src/video_core/CMakeLists.txt @@ -236,7 +236,6 @@ add_library(video_core STATIC texture_cache/types.h texture_cache/util.cpp texture_cache/util.h - textures/astc.cpp textures/astc.h textures/decoders.cpp textures/decoders.h diff --git a/src/video_core/buffer_cache/buffer_cache.h b/src/video_core/buffer_cache/buffer_cache.h index 4de1e37e5..32dcbd693 100644 --- a/src/video_core/buffer_cache/buffer_cache.h +++ b/src/video_core/buffer_cache/buffer_cache.h @@ -20,8 +20,8 @@ #include "common/div_ceil.h" #include "common/microprofile.h" #include "common/scope_exit.h" +#include "common/settings.h" #include "core/memory.h" -#include "core/settings.h" #include "video_core/buffer_cache/buffer_base.h" #include "video_core/delayed_destruction_ring.h" #include "video_core/dirty_flags.h" diff --git a/src/video_core/command_classes/codecs/vp9.cpp b/src/video_core/command_classes/codecs/vp9.cpp index 59e586695..29bb31418 100644 --- a/src/video_core/command_classes/codecs/vp9.cpp +++ b/src/video_core/command_classes/codecs/vp9.cpp @@ -2,8 +2,9 @@ // Licensed under GPLv2 or any later version // Refer to the license.txt file included. -#include <cstring> // for std::memcpy +#include <algorithm> // for std::copy #include <numeric> +#include "common/assert.h" #include "video_core/command_classes/codecs/vp9.h" #include "video_core/gpu.h" #include "video_core/memory_manager.h" @@ -362,7 +363,8 @@ Vp9PictureInfo VP9::GetVp9PictureInfo(const NvdecCommon::NvdecRegisters& state) // surface_luma_offset[0:3] contains the address of the reference frame offsets in the following // order: last, golden, altref, current. It may be worthwhile to track the updates done here // to avoid buffering frame data needed for reference frame updating in the header composition. - std::memcpy(vp9_info.frame_offsets.data(), state.surface_luma_offset.data(), 4 * sizeof(u64)); + std::copy(state.surface_luma_offset.begin(), state.surface_luma_offset.begin() + 4, + vp9_info.frame_offsets.begin()); return vp9_info; } @@ -821,11 +823,11 @@ const std::vector<u8>& VP9::ComposeFrameHeader(const NvdecCommon::NvdecRegisters // Write headers and frame to buffer frame.resize(uncompressed_header.size() + compressed_header.size() + bitstream.size()); - std::memcpy(frame.data(), uncompressed_header.data(), uncompressed_header.size()); - std::memcpy(frame.data() + uncompressed_header.size(), compressed_header.data(), - compressed_header.size()); - std::memcpy(frame.data() + uncompressed_header.size() + compressed_header.size(), - bitstream.data(), bitstream.size()); + std::copy(uncompressed_header.begin(), uncompressed_header.end(), frame.begin()); + std::copy(compressed_header.begin(), compressed_header.end(), + frame.begin() + uncompressed_header.size()); + std::copy(bitstream.begin(), bitstream.end(), + frame.begin() + uncompressed_header.size() + compressed_header.size()); // keep track of frame number current_frame_number++; diff --git a/src/video_core/engines/engine_interface.h b/src/video_core/engines/engine_interface.h index 18a9db7e6..c7ffd68c5 100644 --- a/src/video_core/engines/engine_interface.h +++ b/src/video_core/engines/engine_interface.h @@ -4,13 +4,14 @@ #pragma once -#include <type_traits> #include "common/common_types.h" namespace Tegra::Engines { class EngineInterface { public: + virtual ~EngineInterface() = default; + /// Write the value to the register identified by method. virtual void CallMethod(u32 method, u32 method_argument, bool is_last_call) = 0; diff --git a/src/video_core/engines/fermi_2d.h b/src/video_core/engines/fermi_2d.h index c808a577d..a4170ffff 100644 --- a/src/video_core/engines/fermi_2d.h +++ b/src/video_core/engines/fermi_2d.h @@ -35,7 +35,7 @@ namespace Tegra::Engines { class Fermi2D final : public EngineInterface { public: explicit Fermi2D(); - ~Fermi2D(); + ~Fermi2D() override; /// Binds a rasterizer to this engine. void BindRasterizer(VideoCore::RasterizerInterface* rasterizer); diff --git a/src/video_core/engines/kepler_memory.h b/src/video_core/engines/kepler_memory.h index 19808a5c6..0d8ea09a9 100644 --- a/src/video_core/engines/kepler_memory.h +++ b/src/video_core/engines/kepler_memory.h @@ -36,7 +36,7 @@ namespace Tegra::Engines { class KeplerMemory final : public EngineInterface { public: explicit KeplerMemory(Core::System& system_, MemoryManager& memory_manager); - ~KeplerMemory(); + ~KeplerMemory() override; /// Write the value to the register identified by method. void CallMethod(u32 method, u32 method_argument, bool is_last_call) override; diff --git a/src/video_core/engines/maxwell_dma.cpp b/src/video_core/engines/maxwell_dma.cpp index a2f19559f..2ee980bab 100644 --- a/src/video_core/engines/maxwell_dma.cpp +++ b/src/video_core/engines/maxwell_dma.cpp @@ -4,8 +4,8 @@ #include "common/assert.h" #include "common/logging/log.h" +#include "common/settings.h" #include "core/core.h" -#include "core/settings.h" #include "video_core/engines/maxwell_3d.h" #include "video_core/engines/maxwell_dma.h" #include "video_core/memory_manager.h" diff --git a/src/video_core/engines/maxwell_dma.h b/src/video_core/engines/maxwell_dma.h index 3c59eeb13..c77f02a22 100644 --- a/src/video_core/engines/maxwell_dma.h +++ b/src/video_core/engines/maxwell_dma.h @@ -188,7 +188,7 @@ public: static_assert(sizeof(RemapConst) == 12); explicit MaxwellDMA(Core::System& system_, MemoryManager& memory_manager_); - ~MaxwellDMA(); + ~MaxwellDMA() override; /// Write the value to the register identified by method. void CallMethod(u32 method, u32 method_argument, bool is_last_call) override; diff --git a/src/video_core/gpu.cpp b/src/video_core/gpu.cpp index c61f44619..9bdb282d2 100644 --- a/src/video_core/gpu.cpp +++ b/src/video_core/gpu.cpp @@ -6,13 +6,13 @@ #include "common/assert.h" #include "common/microprofile.h" +#include "common/settings.h" #include "core/core.h" #include "core/core_timing.h" #include "core/core_timing_util.h" #include "core/frontend/emu_window.h" #include "core/hardware_interrupt_manager.h" #include "core/memory.h" -#include "core/settings.h" #include "video_core/engines/fermi_2d.h" #include "video_core/engines/kepler_compute.h" #include "video_core/engines/kepler_memory.h" @@ -517,8 +517,8 @@ void GPU::TriggerCpuInterrupt(const u32 syncpoint_id, const u32 value) const { interrupt_manager.GPUInterruptSyncpt(syncpoint_id, value); } -void GPU::WaitIdle() const { - gpu_thread.WaitIdle(); +void GPU::ShutDown() { + gpu_thread.ShutDown(); } void GPU::OnCommandListEnd() { diff --git a/src/video_core/gpu.h b/src/video_core/gpu.h index b2ee45496..ecab35d3b 100644 --- a/src/video_core/gpu.h +++ b/src/video_core/gpu.h @@ -219,8 +219,8 @@ public: return *shader_notify; } - // Waits for the GPU to finish working - void WaitIdle() const; + // Stops the GPU execution and waits for the GPU to finish working + void ShutDown(); /// Allows the CPU/NvFlinger to wait on the GPU before presenting a frame. void WaitFence(u32 syncpoint_id, u32 value); diff --git a/src/video_core/gpu_thread.cpp b/src/video_core/gpu_thread.cpp index 99353f15f..cd1fbb9bf 100644 --- a/src/video_core/gpu_thread.cpp +++ b/src/video_core/gpu_thread.cpp @@ -5,10 +5,10 @@ #include "common/assert.h" #include "common/microprofile.h" #include "common/scope_exit.h" +#include "common/settings.h" #include "common/thread.h" #include "core/core.h" #include "core/frontend/emu_window.h" -#include "core/settings.h" #include "video_core/dma_pusher.h" #include "video_core/gpu.h" #include "video_core/gpu_thread.h" @@ -29,8 +29,7 @@ static void RunThread(Core::System& system, VideoCore::RendererBase& renderer, system.RegisterHostThread(); // Wait for first GPU command before acquiring the window context - while (state.queue.Empty()) - ; + state.queue.Wait(); // If emulation was stopped during disk shader loading, abort before trying to acquire context if (!state.is_running) { @@ -57,11 +56,17 @@ static void RunThread(Core::System& system, VideoCore::RendererBase& renderer, } else if (const auto* invalidate = std::get_if<InvalidateRegionCommand>(&next.data)) { rasterizer->OnCPUWrite(invalidate->addr, invalidate->size); } else if (std::holds_alternative<EndProcessingCommand>(next.data)) { - return; + ASSERT(state.is_running == false); } else { UNREACHABLE(); } state.signaled_fence.store(next.fence); + if (next.block) { + // We have to lock the write_lock to ensure that the condition_variable wait not get a + // race between the check and the lock itself. + std::lock_guard lk(state.write_lock); + state.cv.notify_all(); + } } } @@ -69,13 +74,7 @@ ThreadManager::ThreadManager(Core::System& system_, bool is_async_) : system{system_}, is_async{is_async_} {} ThreadManager::~ThreadManager() { - if (!thread.joinable()) { - return; - } - - // Notify GPU thread that a shutdown is pending - PushCommand(EndProcessingCommand()); - thread.join(); + ShutDown(); } void ThreadManager::StartThread(VideoCore::RendererBase& renderer, @@ -112,9 +111,8 @@ void ThreadManager::FlushRegion(VAddr addr, u64 size) { case Settings::GPUAccuracy::Extreme: { auto& gpu = system.GPU(); u64 fence = gpu.RequestFlush(addr, size); - PushCommand(GPUTickCommand()); - while (fence > gpu.CurrentFlushRequestFence()) { - } + PushCommand(GPUTickCommand(), true); + ASSERT(fence <= gpu.CurrentFlushRequestFence()); break; } default: @@ -131,23 +129,45 @@ void ThreadManager::FlushAndInvalidateRegion(VAddr addr, u64 size) { rasterizer->OnCPUWrite(addr, size); } -void ThreadManager::WaitIdle() const { - while (state.last_fence > state.signaled_fence.load(std::memory_order_relaxed) && - system.IsPoweredOn()) { +void ThreadManager::ShutDown() { + if (!state.is_running) { + return; } + + { + std::lock_guard lk(state.write_lock); + state.is_running = false; + state.cv.notify_all(); + } + + if (!thread.joinable()) { + return; + } + + // Notify GPU thread that a shutdown is pending + PushCommand(EndProcessingCommand()); + thread.join(); } void ThreadManager::OnCommandListEnd() { PushCommand(OnCommandListEndCommand()); } -u64 ThreadManager::PushCommand(CommandData&& command_data) { - const u64 fence{++state.last_fence}; - state.queue.Push(CommandDataContainer(std::move(command_data), fence)); - +u64 ThreadManager::PushCommand(CommandData&& command_data, bool block) { if (!is_async) { // In synchronous GPU mode, block the caller until the command has executed - WaitIdle(); + block = true; + } + + std::unique_lock lk(state.write_lock); + const u64 fence{++state.last_fence}; + state.queue.Push(CommandDataContainer(std::move(command_data), fence, block)); + + if (block) { + state.cv.wait(lk, [this, fence] { + return fence <= state.signaled_fence.load(std::memory_order_relaxed) || + !state.is_running; + }); } return fence; diff --git a/src/video_core/gpu_thread.h b/src/video_core/gpu_thread.h index 18269e51c..11a648f38 100644 --- a/src/video_core/gpu_thread.h +++ b/src/video_core/gpu_thread.h @@ -90,21 +90,24 @@ using CommandData = struct CommandDataContainer { CommandDataContainer() = default; - explicit CommandDataContainer(CommandData&& data_, u64 next_fence_) - : data{std::move(data_)}, fence{next_fence_} {} + explicit CommandDataContainer(CommandData&& data_, u64 next_fence_, bool block_) + : data{std::move(data_)}, fence{next_fence_}, block(block_) {} CommandData data; u64 fence{}; + bool block{}; }; /// Struct used to synchronize the GPU thread struct SynchState final { std::atomic_bool is_running{true}; - using CommandQueue = Common::MPSCQueue<CommandDataContainer>; + using CommandQueue = Common::SPSCQueue<CommandDataContainer>; + std::mutex write_lock; CommandQueue queue; u64 last_fence{}; std::atomic<u64> signaled_fence{}; + std::condition_variable cv; }; /// Class used to manage the GPU thread @@ -132,14 +135,14 @@ public: /// Notify rasterizer that any caches of the specified region should be flushed and invalidated void FlushAndInvalidateRegion(VAddr addr, u64 size); - // Wait until the gpu thread is idle. - void WaitIdle() const; + // Stops the GPU execution and waits for the GPU to finish working + void ShutDown(); void OnCommandListEnd(); private: /// Pushes a command to be executed by the GPU thread - u64 PushCommand(CommandData&& command_data); + u64 PushCommand(CommandData&& command_data, bool block = false); Core::System& system; const bool is_async; diff --git a/src/video_core/host_shaders/CMakeLists.txt b/src/video_core/host_shaders/CMakeLists.txt index 3494318ca..2208e1922 100644 --- a/src/video_core/host_shaders/CMakeLists.txt +++ b/src/video_core/host_shaders/CMakeLists.txt @@ -1,4 +1,5 @@ set(SHADER_FILES + astc_decoder.comp block_linear_unswizzle_2d.comp block_linear_unswizzle_3d.comp convert_depth_to_float.frag diff --git a/src/video_core/host_shaders/StringShaderHeader.cmake b/src/video_core/host_shaders/StringShaderHeader.cmake index c0fc49768..1b4bc6103 100644 --- a/src/video_core/host_shaders/StringShaderHeader.cmake +++ b/src/video_core/host_shaders/StringShaderHeader.cmake @@ -6,7 +6,27 @@ get_filename_component(CONTENTS_NAME ${SOURCE_FILE} NAME) string(REPLACE "." "_" CONTENTS_NAME ${CONTENTS_NAME}) string(TOUPPER ${CONTENTS_NAME} CONTENTS_NAME) -file(READ ${SOURCE_FILE} CONTENTS) +FILE(READ ${SOURCE_FILE} line_contents) + +# Replace double quotes with single quotes, +# as double quotes will be used to wrap the lines +STRING(REGEX REPLACE "\"" "'" line_contents "${line_contents}") + +# CMake separates list elements with semicolons, but semicolons +# are used extensively in the shader code. +# Replace with a temporary marker, to be reverted later. +STRING(REGEX REPLACE ";" "{{SEMICOLON}}" line_contents "${line_contents}") + +# Make every line an individual element in the CMake list. +STRING(REGEX REPLACE "\n" ";" line_contents "${line_contents}") + +# Build the shader string, wrapping each line in double quotes. +foreach(line IN LISTS line_contents) + string(CONCAT CONTENTS "${CONTENTS}" \"${line}\\n\"\n) +endforeach() + +# Revert the original semicolons in the source. +STRING(REGEX REPLACE "{{SEMICOLON}}" ";" CONTENTS "${CONTENTS}") get_filename_component(OUTPUT_DIR ${HEADER_FILE} DIRECTORY) make_directory(${OUTPUT_DIR}) diff --git a/src/video_core/host_shaders/astc_decoder.comp b/src/video_core/host_shaders/astc_decoder.comp new file mode 100644 index 000000000..703e34587 --- /dev/null +++ b/src/video_core/host_shaders/astc_decoder.comp @@ -0,0 +1,1339 @@ +// Copyright 2021 yuzu Emulator Project +// Licensed under GPLv2 or any later version +// Refer to the license.txt file included. + +#version 450 + +#ifdef VULKAN + +#define BEGIN_PUSH_CONSTANTS layout(push_constant) uniform PushConstants { +#define END_PUSH_CONSTANTS }; +#define UNIFORM(n) +#define BINDING_INPUT_BUFFER 0 +#define BINDING_ENC_BUFFER 1 +#define BINDING_6_TO_8_BUFFER 2 +#define BINDING_7_TO_8_BUFFER 3 +#define BINDING_8_TO_8_BUFFER 4 +#define BINDING_BYTE_TO_16_BUFFER 5 +#define BINDING_SWIZZLE_BUFFER 6 +#define BINDING_OUTPUT_IMAGE 7 + +#else // ^^^ Vulkan ^^^ // vvv OpenGL vvv + +#define BEGIN_PUSH_CONSTANTS +#define END_PUSH_CONSTANTS +#define UNIFORM(n) layout(location = n) uniform +#define BINDING_SWIZZLE_BUFFER 0 +#define BINDING_INPUT_BUFFER 1 +#define BINDING_ENC_BUFFER 2 +#define BINDING_6_TO_8_BUFFER 3 +#define BINDING_7_TO_8_BUFFER 4 +#define BINDING_8_TO_8_BUFFER 5 +#define BINDING_BYTE_TO_16_BUFFER 6 +#define BINDING_OUTPUT_IMAGE 0 + +#endif + +layout(local_size_x = 32, local_size_y = 32, local_size_z = 1) in; + +BEGIN_PUSH_CONSTANTS +UNIFORM(1) uvec2 block_dims; + +UNIFORM(2) uint bytes_per_block_log2; +UNIFORM(3) uint layer_stride; +UNIFORM(4) uint block_size; +UNIFORM(5) uint x_shift; +UNIFORM(6) uint block_height; +UNIFORM(7) uint block_height_mask; +END_PUSH_CONSTANTS + +struct EncodingData { + uint encoding; + uint num_bits; + uint bit_value; + uint quint_trit_value; +}; + +struct TexelWeightParams { + uvec2 size; + uint max_weight; + bool dual_plane; + bool error_state; + bool void_extent_ldr; + bool void_extent_hdr; +}; + +// Swizzle data +layout(binding = BINDING_SWIZZLE_BUFFER, std430) readonly buffer SwizzleTable { + uint swizzle_table[]; +}; + +layout(binding = BINDING_INPUT_BUFFER, std430) readonly buffer InputBufferU32 { + uint astc_data[]; +}; + +// ASTC Encodings data +layout(binding = BINDING_ENC_BUFFER, std430) readonly buffer EncodingsValues { + EncodingData encoding_values[]; +}; +// ASTC Precompiled tables +layout(binding = BINDING_6_TO_8_BUFFER, std430) readonly buffer REPLICATE_6_BIT_TO_8 { + uint REPLICATE_6_BIT_TO_8_TABLE[]; +}; +layout(binding = BINDING_7_TO_8_BUFFER, std430) readonly buffer REPLICATE_7_BIT_TO_8 { + uint REPLICATE_7_BIT_TO_8_TABLE[]; +}; +layout(binding = BINDING_8_TO_8_BUFFER, std430) readonly buffer REPLICATE_8_BIT_TO_8 { + uint REPLICATE_8_BIT_TO_8_TABLE[]; +}; +layout(binding = BINDING_BYTE_TO_16_BUFFER, std430) readonly buffer REPLICATE_BYTE_TO_16 { + uint REPLICATE_BYTE_TO_16_TABLE[]; +}; + +layout(binding = BINDING_OUTPUT_IMAGE, rgba8) uniform writeonly image2DArray dest_image; + +const uint GOB_SIZE_X = 64; +const uint GOB_SIZE_Y = 8; +const uint GOB_SIZE_Z = 1; +const uint GOB_SIZE = GOB_SIZE_X * GOB_SIZE_Y * GOB_SIZE_Z; + +const uint GOB_SIZE_X_SHIFT = 6; +const uint GOB_SIZE_Y_SHIFT = 3; +const uint GOB_SIZE_Z_SHIFT = 0; +const uint GOB_SIZE_SHIFT = GOB_SIZE_X_SHIFT + GOB_SIZE_Y_SHIFT + GOB_SIZE_Z_SHIFT; + +const uvec2 SWIZZLE_MASK = uvec2(GOB_SIZE_X - 1, GOB_SIZE_Y - 1); + +const int BLOCK_SIZE_IN_BYTES = 16; + +const int BLOCK_INFO_ERROR = 0; +const int BLOCK_INFO_VOID_EXTENT_HDR = 1; +const int BLOCK_INFO_VOID_EXTENT_LDR = 2; +const int BLOCK_INFO_NORMAL = 3; + +const int JUST_BITS = 0; +const int QUINT = 1; +const int TRIT = 2; + +// The following constants are expanded variants of the Replicate() +// function calls corresponding to the following arguments: +// value: index into the generated table +// num_bits: the after "REPLICATE" in the table name. i.e. 4 is num_bits in REPLICATE_4. +// to_bit: the integer after "TO_" +const uint REPLICATE_BIT_TO_7_TABLE[2] = uint[](0, 127); +const uint REPLICATE_1_BIT_TO_9_TABLE[2] = uint[](0, 511); + +const uint REPLICATE_1_BIT_TO_8_TABLE[2] = uint[](0, 255); +const uint REPLICATE_2_BIT_TO_8_TABLE[4] = uint[](0, 85, 170, 255); +const uint REPLICATE_3_BIT_TO_8_TABLE[8] = uint[](0, 36, 73, 109, 146, 182, 219, 255); +const uint REPLICATE_4_BIT_TO_8_TABLE[16] = + uint[](0, 17, 34, 51, 68, 85, 102, 119, 136, 153, 170, 187, 204, 221, 238, 255); +const uint REPLICATE_5_BIT_TO_8_TABLE[32] = + uint[](0, 8, 16, 24, 33, 41, 49, 57, 66, 74, 82, 90, 99, 107, 115, 123, 132, 140, 148, 156, 165, + 173, 181, 189, 198, 206, 214, 222, 231, 239, 247, 255); +const uint REPLICATE_1_BIT_TO_6_TABLE[2] = uint[](0, 63); +const uint REPLICATE_2_BIT_TO_6_TABLE[4] = uint[](0, 21, 42, 63); +const uint REPLICATE_3_BIT_TO_6_TABLE[8] = uint[](0, 9, 18, 27, 36, 45, 54, 63); +const uint REPLICATE_4_BIT_TO_6_TABLE[16] = + uint[](0, 4, 8, 12, 17, 21, 25, 29, 34, 38, 42, 46, 51, 55, 59, 63); +const uint REPLICATE_5_BIT_TO_6_TABLE[32] = + uint[](0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30, 33, 35, 37, 39, 41, 43, 45, + 47, 49, 51, 53, 55, 57, 59, 61, 63); + +// Input ASTC texture globals +uint current_index = 0; +int bitsread = 0; +uint total_bitsread = 0; +uint local_buff[16]; + +// Color data globals +uint color_endpoint_data[16]; +int color_bitsread = 0; +uint total_color_bitsread = 0; +int color_index = 0; + +// Four values, two endpoints, four maximum paritions +uint color_values[32]; +int colvals_index = 0; + +// Weight data globals +uint texel_weight_data[16]; +int texel_bitsread = 0; +uint total_texel_bitsread = 0; +int texel_index = 0; + +bool texel_flag = false; + +// Global "vectors" to be pushed into when decoding +EncodingData result_vector[100]; +int result_index = 0; + +EncodingData texel_vector[100]; +int texel_vector_index = 0; + +uint unquantized_texel_weights[2][144]; + +uint SwizzleOffset(uvec2 pos) { + pos = pos & SWIZZLE_MASK; + return swizzle_table[pos.y * 64 + pos.x]; +} + +uint ReadTexel(uint offset) { + // extract the 8-bit value from the 32-bit packed data. + return bitfieldExtract(astc_data[offset / 4], int((offset * 8) & 24), 8); +} + +// Replicates low num_bits such that [(to_bit - 1):(to_bit - 1 - from_bit)] +// is the same as [(num_bits - 1):0] and repeats all the way down. +uint Replicate(uint val, uint num_bits, uint to_bit) { + if (num_bits == 0 || to_bit == 0) { + return 0; + } + const uint v = val & uint((1 << num_bits) - 1); + uint res = v; + uint reslen = num_bits; + while (reslen < to_bit) { + uint comp = 0; + if (num_bits > to_bit - reslen) { + uint newshift = to_bit - reslen; + comp = num_bits - newshift; + num_bits = newshift; + } + res = uint(res << num_bits); + res = uint(res | (v >> comp)); + reslen += num_bits; + } + return res; +} + +uvec4 ReplicateByteTo16(uvec4 value) { + return uvec4(REPLICATE_BYTE_TO_16_TABLE[value.x], REPLICATE_BYTE_TO_16_TABLE[value.y], + REPLICATE_BYTE_TO_16_TABLE[value.z], REPLICATE_BYTE_TO_16_TABLE[value.w]); +} + +uint ReplicateBitTo7(uint value) { + return REPLICATE_BIT_TO_7_TABLE[value]; +} + +uint ReplicateBitTo9(uint value) { + return REPLICATE_1_BIT_TO_9_TABLE[value]; +} + +uint FastReplicateTo8(uint value, uint num_bits) { + switch (num_bits) { + case 1: + return REPLICATE_1_BIT_TO_8_TABLE[value]; + case 2: + return REPLICATE_2_BIT_TO_8_TABLE[value]; + case 3: + return REPLICATE_3_BIT_TO_8_TABLE[value]; + case 4: + return REPLICATE_4_BIT_TO_8_TABLE[value]; + case 5: + return REPLICATE_5_BIT_TO_8_TABLE[value]; + case 6: + return REPLICATE_6_BIT_TO_8_TABLE[value]; + case 7: + return REPLICATE_7_BIT_TO_8_TABLE[value]; + case 8: + return REPLICATE_8_BIT_TO_8_TABLE[value]; + } + return Replicate(value, num_bits, 8); +} + +uint FastReplicateTo6(uint value, uint num_bits) { + switch (num_bits) { + case 1: + return REPLICATE_1_BIT_TO_6_TABLE[value]; + case 2: + return REPLICATE_2_BIT_TO_6_TABLE[value]; + case 3: + return REPLICATE_3_BIT_TO_6_TABLE[value]; + case 4: + return REPLICATE_4_BIT_TO_6_TABLE[value]; + case 5: + return REPLICATE_5_BIT_TO_6_TABLE[value]; + } + return Replicate(value, num_bits, 6); +} + +uint Div3Floor(uint v) { + return (v * 0x5556) >> 16; +} + +uint Div3Ceil(uint v) { + return Div3Floor(v + 2); +} + +uint Div5Floor(uint v) { + return (v * 0x3334) >> 16; +} + +uint Div5Ceil(uint v) { + return Div5Floor(v + 4); +} + +uint Hash52(uint p) { + p ^= p >> 15; + p -= p << 17; + p += p << 7; + p += p << 4; + p ^= p >> 5; + p += p << 16; + p ^= p >> 7; + p ^= p >> 3; + p ^= p << 6; + p ^= p >> 17; + return p; +} + +uint SelectPartition(uint seed, uint x, uint y, uint z, uint partition_count, bool small_block) { + if (partition_count == 1) { + return 0; + } + if (small_block) { + x <<= 1; + y <<= 1; + z <<= 1; + } + + seed += (partition_count - 1) * 1024; + + uint rnum = Hash52(uint(seed)); + uint seed1 = uint(rnum & 0xF); + uint seed2 = uint((rnum >> 4) & 0xF); + uint seed3 = uint((rnum >> 8) & 0xF); + uint seed4 = uint((rnum >> 12) & 0xF); + uint seed5 = uint((rnum >> 16) & 0xF); + uint seed6 = uint((rnum >> 20) & 0xF); + uint seed7 = uint((rnum >> 24) & 0xF); + uint seed8 = uint((rnum >> 28) & 0xF); + uint seed9 = uint((rnum >> 18) & 0xF); + uint seed10 = uint((rnum >> 22) & 0xF); + uint seed11 = uint((rnum >> 26) & 0xF); + uint seed12 = uint(((rnum >> 30) | (rnum << 2)) & 0xF); + + seed1 = (seed1 * seed1); + seed2 = (seed2 * seed2); + seed3 = (seed3 * seed3); + seed4 = (seed4 * seed4); + seed5 = (seed5 * seed5); + seed6 = (seed6 * seed6); + seed7 = (seed7 * seed7); + seed8 = (seed8 * seed8); + seed9 = (seed9 * seed9); + seed10 = (seed10 * seed10); + seed11 = (seed11 * seed11); + seed12 = (seed12 * seed12); + + int sh1, sh2, sh3; + if ((seed & 1) > 0) { + sh1 = (seed & 2) > 0 ? 4 : 5; + sh2 = (partition_count == 3) ? 6 : 5; + } else { + sh1 = (partition_count == 3) ? 6 : 5; + sh2 = (seed & 2) > 0 ? 4 : 5; + } + sh3 = (seed & 0x10) > 0 ? sh1 : sh2; + + seed1 = (seed1 >> sh1); + seed2 = (seed2 >> sh2); + seed3 = (seed3 >> sh1); + seed4 = (seed4 >> sh2); + seed5 = (seed5 >> sh1); + seed6 = (seed6 >> sh2); + seed7 = (seed7 >> sh1); + seed8 = (seed8 >> sh2); + seed9 = (seed9 >> sh3); + seed10 = (seed10 >> sh3); + seed11 = (seed11 >> sh3); + seed12 = (seed12 >> sh3); + + uint a = seed1 * x + seed2 * y + seed11 * z + (rnum >> 14); + uint b = seed3 * x + seed4 * y + seed12 * z + (rnum >> 10); + uint c = seed5 * x + seed6 * y + seed9 * z + (rnum >> 6); + uint d = seed7 * x + seed8 * y + seed10 * z + (rnum >> 2); + + a &= 0x3F; + b &= 0x3F; + c &= 0x3F; + d &= 0x3F; + + if (partition_count < 4) { + d = 0; + } + if (partition_count < 3) { + c = 0; + } + + if (a >= b && a >= c && a >= d) { + return 0; + } else if (b >= c && b >= d) { + return 1; + } else if (c >= d) { + return 2; + } else { + return 3; + } +} + +uint Select2DPartition(uint seed, uint x, uint y, uint partition_count, bool small_block) { + return SelectPartition(seed, x, y, 0, partition_count, small_block); +} + +uint ReadBit() { + if (current_index >= local_buff.length()) { + return 0; + } + uint bit = bitfieldExtract(local_buff[current_index], bitsread, 1); + ++bitsread; + ++total_bitsread; + if (bitsread == 8) { + ++current_index; + bitsread = 0; + } + return bit; +} + +uint StreamBits(uint num_bits) { + uint ret = 0; + for (uint i = 0; i < num_bits; i++) { + ret |= ((ReadBit() & 1) << i); + } + return ret; +} + +uint ReadColorBit() { + uint bit = 0; + if (texel_flag) { + bit = bitfieldExtract(texel_weight_data[texel_index], texel_bitsread, 1); + ++texel_bitsread; + ++total_texel_bitsread; + if (texel_bitsread == 8) { + ++texel_index; + texel_bitsread = 0; + } + } else { + bit = bitfieldExtract(color_endpoint_data[color_index], color_bitsread, 1); + ++color_bitsread; + ++total_color_bitsread; + if (color_bitsread == 8) { + ++color_index; + color_bitsread = 0; + } + } + return bit; +} + +uint StreamColorBits(uint num_bits) { + uint ret = 0; + for (uint i = 0; i < num_bits; i++) { + ret |= ((ReadColorBit() & 1) << i); + } + return ret; +} + +void ResultEmplaceBack(EncodingData val) { + if (texel_flag) { + texel_vector[texel_vector_index] = val; + ++texel_vector_index; + } else { + result_vector[result_index] = val; + ++result_index; + } +} + +// Returns the number of bits required to encode n_vals values. +uint GetBitLength(uint n_vals, uint encoding_index) { + uint total_bits = encoding_values[encoding_index].num_bits * n_vals; + if (encoding_values[encoding_index].encoding == TRIT) { + total_bits += Div5Ceil(n_vals * 8); + } else if (encoding_values[encoding_index].encoding == QUINT) { + total_bits += Div3Ceil(n_vals * 7); + } + return total_bits; +} + +uint GetNumWeightValues(uvec2 size, bool dual_plane) { + uint n_vals = size.x * size.y; + if (dual_plane) { + n_vals *= 2; + } + return n_vals; +} + +uint GetPackedBitSize(uvec2 size, bool dual_plane, uint max_weight) { + uint n_vals = GetNumWeightValues(size, dual_plane); + return GetBitLength(n_vals, max_weight); +} + +uint BitsBracket(uint bits, uint pos) { + return ((bits >> pos) & 1); +} + +uint BitsOp(uint bits, uint start, uint end) { + if (start == end) { + return BitsBracket(bits, start); + } else if (start > end) { + uint t = start; + start = end; + end = t; + } + + uint mask = (1 << (end - start + 1)) - 1; + return ((bits >> start) & mask); +} + +void DecodeQuintBlock(uint num_bits) { + uint m[3]; + uint q[3]; + uint Q; + m[0] = StreamColorBits(num_bits); + Q = StreamColorBits(3); + m[1] = StreamColorBits(num_bits); + Q |= StreamColorBits(2) << 3; + m[2] = StreamColorBits(num_bits); + Q |= StreamColorBits(2) << 5; + if (BitsOp(Q, 1, 2) == 3 && BitsOp(Q, 5, 6) == 0) { + q[0] = 4; + q[1] = 4; + q[2] = (BitsBracket(Q, 0) << 2) | ((BitsBracket(Q, 4) & ~BitsBracket(Q, 0)) << 1) | + (BitsBracket(Q, 3) & ~BitsBracket(Q, 0)); + } else { + uint C = 0; + if (BitsOp(Q, 1, 2) == 3) { + q[2] = 4; + C = (BitsOp(Q, 3, 4) << 3) | ((~BitsOp(Q, 5, 6) & 3) << 1) | BitsBracket(Q, 0); + } else { + q[2] = BitsOp(Q, 5, 6); + C = BitsOp(Q, 0, 4); + } + if (BitsOp(C, 0, 2) == 5) { + q[1] = 4; + q[0] = BitsOp(C, 3, 4); + } else { + q[1] = BitsOp(C, 3, 4); + q[0] = BitsOp(C, 0, 2); + } + } + for (uint i = 0; i < 3; i++) { + EncodingData val; + val.encoding = QUINT; + val.num_bits = num_bits; + val.bit_value = m[i]; + val.quint_trit_value = q[i]; + ResultEmplaceBack(val); + } +} + +void DecodeTritBlock(uint num_bits) { + uint m[5]; + uint t[5]; + uint T; + m[0] = StreamColorBits(num_bits); + T = StreamColorBits(2); + m[1] = StreamColorBits(num_bits); + T |= StreamColorBits(2) << 2; + m[2] = StreamColorBits(num_bits); + T |= StreamColorBits(1) << 4; + m[3] = StreamColorBits(num_bits); + T |= StreamColorBits(2) << 5; + m[4] = StreamColorBits(num_bits); + T |= StreamColorBits(1) << 7; + uint C = 0; + if (BitsOp(T, 2, 4) == 7) { + C = (BitsOp(T, 5, 7) << 2) | BitsOp(T, 0, 1); + t[4] = 2; + t[3] = 2; + } else { + C = BitsOp(T, 0, 4); + if (BitsOp(T, 5, 6) == 3) { + t[4] = 2; + t[3] = BitsBracket(T, 7); + } else { + t[4] = BitsBracket(T, 7); + t[3] = BitsOp(T, 5, 6); + } + } + if (BitsOp(C, 0, 1) == 3) { + t[2] = 2; + t[1] = BitsBracket(C, 4); + t[0] = (BitsBracket(C, 3) << 1) | (BitsBracket(C, 2) & ~BitsBracket(C, 3)); + } else if (BitsOp(C, 2, 3) == 3) { + t[2] = 2; + t[1] = 2; + t[0] = BitsOp(C, 0, 1); + } else { + t[2] = BitsBracket(C, 4); + t[1] = BitsOp(C, 2, 3); + t[0] = (BitsBracket(C, 1) << 1) | (BitsBracket(C, 0) & ~BitsBracket(C, 1)); + } + for (uint i = 0; i < 5; i++) { + EncodingData val; + val.encoding = TRIT; + val.num_bits = num_bits; + val.bit_value = m[i]; + val.quint_trit_value = t[i]; + ResultEmplaceBack(val); + } +} + +void DecodeIntegerSequence(uint max_range, uint num_values) { + EncodingData val = encoding_values[max_range]; + uint vals_decoded = 0; + while (vals_decoded < num_values) { + switch (val.encoding) { + case QUINT: + DecodeQuintBlock(val.num_bits); + vals_decoded += 3; + break; + case TRIT: + DecodeTritBlock(val.num_bits); + vals_decoded += 5; + break; + case JUST_BITS: + val.bit_value = StreamColorBits(val.num_bits); + ResultEmplaceBack(val); + vals_decoded++; + break; + } + } +} + +void DecodeColorValues(uvec4 modes, uint num_partitions, uint color_data_bits) { + uint num_values = 0; + for (uint i = 0; i < num_partitions; i++) { + num_values += ((modes[i] >> 2) + 1) << 1; + } + int range = 256; + while (--range > 0) { + EncodingData val = encoding_values[range]; + uint bit_length = GetBitLength(num_values, range); + if (bit_length <= color_data_bits) { + while (--range > 0) { + EncodingData newval = encoding_values[range]; + if (newval.encoding != val.encoding && newval.num_bits != val.num_bits) { + break; + } + } + ++range; + break; + } + } + DecodeIntegerSequence(range, num_values); + uint out_index = 0; + for (int itr = 0; itr < result_index; ++itr) { + if (out_index >= num_values) { + break; + } + EncodingData val = result_vector[itr]; + uint bitlen = val.num_bits; + uint bitval = val.bit_value; + uint A = 0, B = 0, C = 0, D = 0; + A = ReplicateBitTo9((bitval & 1)); + switch (val.encoding) { + case JUST_BITS: + color_values[out_index++] = FastReplicateTo8(bitval, bitlen); + break; + case TRIT: { + D = val.quint_trit_value; + switch (bitlen) { + case 1: + C = 204; + break; + case 2: { + C = 93; + uint b = (bitval >> 1) & 1; + B = (b << 8) | (b << 4) | (b << 2) | (b << 1); + break; + } + case 3: { + C = 44; + uint cb = (bitval >> 1) & 3; + B = (cb << 7) | (cb << 2) | cb; + break; + } + case 4: { + C = 22; + uint dcb = (bitval >> 1) & 7; + B = (dcb << 6) | dcb; + break; + } + case 5: { + C = 11; + uint edcb = (bitval >> 1) & 0xF; + B = (edcb << 5) | (edcb >> 2); + break; + } + case 6: { + C = 5; + uint fedcb = (bitval >> 1) & 0x1F; + B = (fedcb << 4) | (fedcb >> 4); + break; + } + } + break; + } + case QUINT: { + D = val.quint_trit_value; + switch (bitlen) { + case 1: + C = 113; + break; + case 2: { + C = 54; + uint b = (bitval >> 1) & 1; + B = (b << 8) | (b << 3) | (b << 2); + break; + } + case 3: { + C = 26; + uint cb = (bitval >> 1) & 3; + B = (cb << 7) | (cb << 1) | (cb >> 1); + break; + } + case 4: { + C = 13; + uint dcb = (bitval >> 1) & 7; + B = (dcb << 6) | (dcb >> 1); + break; + } + case 5: { + C = 6; + uint edcb = (bitval >> 1) & 0xF; + B = (edcb << 5) | (edcb >> 3); + break; + } + } + break; + } + } + if (val.encoding != JUST_BITS) { + uint T = (D * C) + B; + T ^= A; + T = (A & 0x80) | (T >> 2); + color_values[out_index++] = T; + } + } +} + +ivec2 BitTransferSigned(int a, int b) { + ivec2 transferred; + transferred.y = b >> 1; + transferred.y |= a & 0x80; + transferred.x = a >> 1; + transferred.x &= 0x3F; + if ((transferred.x & 0x20) > 0) { + transferred.x -= 0x40; + } + return transferred; +} + +uvec4 ClampByte(ivec4 color) { + for (uint i = 0; i < 4; ++i) { + color[i] = (color[i] < 0) ? 0 : ((color[i] > 255) ? 255 : color[i]); + } + return uvec4(color); +} + +ivec4 BlueContract(int a, int r, int g, int b) { + return ivec4(a, (r + b) >> 1, (g + b) >> 1, b); +} + +void ComputeEndpoints(out uvec4 ep1, out uvec4 ep2, uint color_endpoint_mode) { +#define READ_UINT_VALUES(N) \ + uint v[N]; \ + for (uint i = 0; i < N; i++) { \ + v[i] = color_values[colvals_index++]; \ + } + +#define READ_INT_VALUES(N) \ + int v[N]; \ + for (uint i = 0; i < N; i++) { \ + v[i] = int(color_values[colvals_index++]); \ + } + + switch (color_endpoint_mode) { + case 0: { + READ_UINT_VALUES(2) + ep1 = uvec4(0xFF, v[0], v[0], v[0]); + ep2 = uvec4(0xFF, v[1], v[1], v[1]); + break; + } + case 1: { + READ_UINT_VALUES(2) + uint L0 = (v[0] >> 2) | (v[1] & 0xC0); + uint L1 = max(L0 + (v[1] & 0x3F), 0xFFU); + ep1 = uvec4(0xFF, L0, L0, L0); + ep2 = uvec4(0xFF, L1, L1, L1); + break; + } + case 4: { + READ_UINT_VALUES(4) + ep1 = uvec4(v[2], v[0], v[0], v[0]); + ep2 = uvec4(v[3], v[1], v[1], v[1]); + break; + } + case 5: { + READ_INT_VALUES(4) + ivec2 transferred = BitTransferSigned(v[1], v[0]); + v[1] = transferred.x; + v[0] = transferred.y; + transferred = BitTransferSigned(v[3], v[2]); + v[3] = transferred.x; + v[2] = transferred.y; + ep1 = ClampByte(ivec4(v[2], v[0], v[0], v[0])); + ep2 = ClampByte(ivec4(v[2] + v[3], v[0] + v[1], v[0] + v[1], v[0] + v[1])); + break; + } + case 6: { + READ_UINT_VALUES(4) + ep1 = uvec4(0xFF, (v[0] * v[3]) >> 8, (v[1] * v[3]) >> 8, (v[2] * v[3]) >> 8); + ep2 = uvec4(0xFF, v[0], v[1], v[2]); + break; + } + case 8: { + READ_UINT_VALUES(6) + if ((v[1] + v[3] + v[5]) >= (v[0] + v[2] + v[4])) { + ep1 = uvec4(0xFF, v[0], v[2], v[4]); + ep2 = uvec4(0xFF, v[1], v[3], v[5]); + } else { + ep1 = uvec4(BlueContract(0xFF, int(v[1]), int(v[3]), int(v[5]))); + ep2 = uvec4(BlueContract(0xFF, int(v[0]), int(v[2]), int(v[4]))); + } + break; + } + case 9: { + READ_INT_VALUES(6) + ivec2 transferred = BitTransferSigned(v[1], v[0]); + v[1] = transferred.x; + v[0] = transferred.y; + transferred = BitTransferSigned(v[3], v[2]); + v[3] = transferred.x; + v[2] = transferred.y; + transferred = BitTransferSigned(v[5], v[4]); + v[5] = transferred.x; + v[4] = transferred.y; + if ((v[1] + v[3] + v[5]) >= 0) { + ep1 = ClampByte(ivec4(0xFF, v[0], v[2], v[4])); + ep2 = ClampByte(ivec4(0xFF, v[0] + v[1], v[2] + v[3], v[4] + v[5])); + } else { + ep1 = ClampByte(BlueContract(0xFF, v[0] + v[1], v[2] + v[3], v[4] + v[5])); + ep2 = ClampByte(BlueContract(0xFF, v[0], v[2], v[4])); + } + break; + } + case 10: { + READ_UINT_VALUES(6) + ep1 = uvec4(v[4], (v[0] * v[3]) >> 8, (v[1] * v[3]) >> 8, (v[2] * v[3]) >> 8); + ep2 = uvec4(v[5], v[0], v[1], v[2]); + break; + } + case 12: { + READ_UINT_VALUES(8) + if ((v[1] + v[3] + v[5]) >= (v[0] + v[2] + v[4])) { + ep1 = uvec4(v[6], v[0], v[2], v[4]); + ep2 = uvec4(v[7], v[1], v[3], v[5]); + } else { + ep1 = uvec4(BlueContract(int(v[7]), int(v[1]), int(v[3]), int(v[5]))); + ep2 = uvec4(BlueContract(int(v[6]), int(v[0]), int(v[2]), int(v[4]))); + } + break; + } + case 13: { + READ_INT_VALUES(8) + ivec2 transferred = BitTransferSigned(v[1], v[0]); + v[1] = transferred.x; + v[0] = transferred.y; + transferred = BitTransferSigned(v[3], v[2]); + v[3] = transferred.x; + v[2] = transferred.y; + + transferred = BitTransferSigned(v[5], v[4]); + v[5] = transferred.x; + v[4] = transferred.y; + + transferred = BitTransferSigned(v[7], v[6]); + v[7] = transferred.x; + v[6] = transferred.y; + + if ((v[1] + v[3] + v[5]) >= 0) { + ep1 = ClampByte(ivec4(v[6], v[0], v[2], v[4])); + ep2 = ClampByte(ivec4(v[7] + v[6], v[0] + v[1], v[2] + v[3], v[4] + v[5])); + } else { + ep1 = ClampByte(BlueContract(v[6] + v[7], v[0] + v[1], v[2] + v[3], v[4] + v[5])); + ep2 = ClampByte(BlueContract(v[6], v[0], v[2], v[4])); + } + break; + } + default: { + // HDR mode, or more likely a bug computing the color_endpoint_mode + ep1 = uvec4(0xFF, 0xFF, 0, 0); + ep2 = uvec4(0xFF, 0xFF, 0, 0); + break; + } + } +#undef READ_UINT_VALUES +#undef READ_INT_VALUES +} + +uint UnquantizeTexelWeight(EncodingData val) { + uint bitval = val.bit_value; + uint bitlen = val.num_bits; + uint A = ReplicateBitTo7((bitval & 1)); + uint B = 0, C = 0, D = 0; + uint result = 0; + switch (val.encoding) { + case JUST_BITS: + result = FastReplicateTo6(bitval, bitlen); + break; + case TRIT: { + D = val.quint_trit_value; + switch (bitlen) { + case 0: { + uint results[3] = {0, 32, 63}; + result = results[D]; + break; + } + case 1: { + C = 50; + break; + } + case 2: { + C = 23; + uint b = (bitval >> 1) & 1; + B = (b << 6) | (b << 2) | b; + break; + } + case 3: { + C = 11; + uint cb = (bitval >> 1) & 3; + B = (cb << 5) | cb; + break; + } + default: + break; + } + break; + } + case QUINT: { + D = val.quint_trit_value; + switch (bitlen) { + case 0: { + uint results[5] = {0, 16, 32, 47, 63}; + result = results[D]; + break; + } + case 1: { + C = 28; + break; + } + case 2: { + C = 13; + uint b = (bitval >> 1) & 1; + B = (b << 6) | (b << 1); + break; + } + } + break; + } + } + if (val.encoding != JUST_BITS && bitlen > 0) { + result = D * C + B; + result ^= A; + result = (A & 0x20) | (result >> 2); + } + if (result > 32) { + result += 1; + } + return result; +} + +void UnquantizeTexelWeights(bool dual_plane, uvec2 size) { + uint weight_idx = 0; + uint unquantized[2][144]; + uint area = size.x * size.y; + for (uint itr = 0; itr < texel_vector_index; itr++) { + unquantized[0][weight_idx] = UnquantizeTexelWeight(texel_vector[itr]); + if (dual_plane) { + ++itr; + unquantized[1][weight_idx] = UnquantizeTexelWeight(texel_vector[itr]); + if (itr == texel_vector_index) { + break; + } + } + if (++weight_idx >= (area)) + break; + } + + const uint Ds = uint((block_dims.x * 0.5f + 1024) / (block_dims.x - 1)); + const uint Dt = uint((block_dims.y * 0.5f + 1024) / (block_dims.y - 1)); + const uint k_plane_scale = dual_plane ? 2 : 1; + for (uint plane = 0; plane < k_plane_scale; plane++) { + for (uint t = 0; t < block_dims.y; t++) { + for (uint s = 0; s < block_dims.x; s++) { + uint cs = Ds * s; + uint ct = Dt * t; + uint gs = (cs * (size.x - 1) + 32) >> 6; + uint gt = (ct * (size.y - 1) + 32) >> 6; + uint js = gs >> 4; + uint fs = gs & 0xF; + uint jt = gt >> 4; + uint ft = gt & 0x0F; + uint w11 = (fs * ft + 8) >> 4; + uint w10 = ft - w11; + uint w01 = fs - w11; + uint w00 = 16 - fs - ft + w11; + uvec4 w = uvec4(w00, w01, w10, w11); + uint v0 = jt * size.x + js; + + uvec4 p = uvec4(0); + if (v0 < area) { + p.x = unquantized[plane][v0]; + } + if ((v0 + 1) < (area)) { + p.y = unquantized[plane][v0 + 1]; + } + if ((v0 + size.x) < (area)) { + p.z = unquantized[plane][(v0 + size.x)]; + } + if ((v0 + size.x + 1) < (area)) { + p.w = unquantized[plane][(v0 + size.x + 1)]; + } + unquantized_texel_weights[plane][t * block_dims.x + s] = (uint(dot(p, w)) + 8) >> 4; + } + } + } +} + +int FindLayout(uint mode) { + if ((mode & 3) != 0) { + if ((mode & 8) != 0) { + if ((mode & 4) != 0) { + if ((mode & 0x100) != 0) { + return 4; + } + return 3; + } + return 2; + } + if ((mode & 4) != 0) { + return 1; + } + return 0; + } + if ((mode & 0x100) != 0) { + if ((mode & 0x80) != 0) { + if ((mode & 0x20) != 0) { + return 8; + } + return 7; + } + return 9; + } + if ((mode & 0x80) != 0) { + return 6; + } + return 5; +} + +TexelWeightParams DecodeBlockInfo(uint block_index) { + TexelWeightParams params = TexelWeightParams(uvec2(0), 0, false, false, false, false); + uint mode = StreamBits(11); + if ((mode & 0x1ff) == 0x1fc) { + if ((mode & 0x200) != 0) { + params.void_extent_hdr = true; + } else { + params.void_extent_ldr = true; + } + if ((mode & 0x400) == 0 || StreamBits(1) == 0) { + params.error_state = true; + } + return params; + } + if ((mode & 0xf) == 0) { + params.error_state = true; + return params; + } + if ((mode & 3) == 0 && (mode & 0x1c0) == 0x1c0) { + params.error_state = true; + return params; + } + uint A, B; + uint mode_layout = FindLayout(mode); + switch (mode_layout) { + case 0: + A = (mode >> 5) & 0x3; + B = (mode >> 7) & 0x3; + params.size = uvec2(B + 4, A + 2); + break; + case 1: + A = (mode >> 5) & 0x3; + B = (mode >> 7) & 0x3; + params.size = uvec2(B + 8, A + 2); + break; + case 2: + A = (mode >> 5) & 0x3; + B = (mode >> 7) & 0x3; + params.size = uvec2(A + 2, B + 8); + break; + case 3: + A = (mode >> 5) & 0x3; + B = (mode >> 7) & 0x1; + params.size = uvec2(A + 2, B + 6); + break; + case 4: + A = (mode >> 5) & 0x3; + B = (mode >> 7) & 0x1; + params.size = uvec2(B + 2, A + 2); + break; + case 5: + A = (mode >> 5) & 0x3; + params.size = uvec2(12, A + 2); + break; + case 6: + A = (mode >> 5) & 0x3; + params.size = uvec2(A + 2, 12); + break; + case 7: + params.size = uvec2(6, 10); + break; + case 8: + params.size = uvec2(10, 6); + break; + case 9: + A = (mode >> 5) & 0x3; + B = (mode >> 9) & 0x3; + params.size = uvec2(A + 6, B + 6); + break; + default: + params.error_state = true; + break; + } + params.dual_plane = (mode_layout != 9) && ((mode & 0x400) != 0); + uint weight_index = (mode & 0x10) != 0 ? 1 : 0; + if (mode_layout < 5) { + weight_index |= (mode & 0x3) << 1; + } else { + weight_index |= (mode & 0xc) >> 1; + } + weight_index -= 2; + if ((mode_layout != 9) && ((mode & 0x200) != 0)) { + const int max_weights[6] = int[6](9, 11, 15, 19, 23, 31); + params.max_weight = max_weights[weight_index]; + } else { + const int max_weights[6] = int[6](1, 2, 3, 4, 5, 7); + params.max_weight = max_weights[weight_index]; + } + return params; +} + +void FillError(ivec3 coord) { + for (uint j = 0; j < block_dims.y; j++) { + for (uint i = 0; i < block_dims.x; i++) { + imageStore(dest_image, coord + ivec3(i, j, 0), vec4(1.0, 1.0, 0.0, 1.0)); + } + } +} + +void FillVoidExtentLDR(ivec3 coord) { + StreamBits(52); + uint r_u = StreamBits(16); + uint g_u = StreamBits(16); + uint b_u = StreamBits(16); + uint a_u = StreamBits(16); + float a = float(a_u) / 65535.0f; + float r = float(r_u) / 65535.0f; + float g = float(g_u) / 65535.0f; + float b = float(b_u) / 65535.0f; + for (uint j = 0; j < block_dims.y; j++) { + for (uint i = 0; i < block_dims.x; i++) { + imageStore(dest_image, coord + ivec3(i, j, 0), vec4(r, g, b, a)); + } + } +} + +void DecompressBlock(ivec3 coord, uint block_index) { + TexelWeightParams params = DecodeBlockInfo(block_index); + if (params.error_state) { + FillError(coord); + return; + } + if (params.void_extent_hdr) { + FillError(coord); + return; + } + if (params.void_extent_ldr) { + FillVoidExtentLDR(coord); + return; + } + if ((params.size.x > block_dims.x) || (params.size.y > block_dims.y)) { + FillError(coord); + return; + } + uint num_partitions = StreamBits(2) + 1; + if (num_partitions > 4 || (num_partitions == 4 && params.dual_plane)) { + FillError(coord); + return; + } + int plane_index = -1; + uint partition_index = 1; + uvec4 color_endpoint_mode = uvec4(0); + uint ced_pointer = 0; + uint base_cem = 0; + if (num_partitions == 1) { + color_endpoint_mode.x = StreamBits(4); + partition_index = 0; + } else { + partition_index = StreamBits(10); + base_cem = StreamBits(6); + } + uint base_mode = base_cem & 3; + uint weight_bits = GetPackedBitSize(params.size, params.dual_plane, params.max_weight); + uint remaining_bits = 128 - weight_bits - total_bitsread; + uint extra_cem_bits = 0; + if (base_mode > 0) { + switch (num_partitions) { + case 2: + extra_cem_bits += 2; + break; + case 3: + extra_cem_bits += 5; + break; + case 4: + extra_cem_bits += 8; + break; + default: + return; + } + } + remaining_bits -= extra_cem_bits; + uint plane_selector_bits = 0; + if (params.dual_plane) { + plane_selector_bits = 2; + } + remaining_bits -= plane_selector_bits; + if (remaining_bits > 128) { + // Bad data, more remaining bits than 4 bytes + // return early + return; + } + // Read color data... + uint color_data_bits = remaining_bits; + while (remaining_bits > 0) { + int nb = int(min(remaining_bits, 8U)); + uint b = StreamBits(nb); + color_endpoint_data[ced_pointer] = uint(bitfieldExtract(b, 0, nb)); + ++ced_pointer; + remaining_bits -= nb; + } + plane_index = int(StreamBits(plane_selector_bits)); + if (base_mode > 0) { + uint extra_cem = StreamBits(extra_cem_bits); + uint cem = (extra_cem << 6) | base_cem; + cem >>= 2; + uvec4 C = uvec4(0); + for (uint i = 0; i < num_partitions; i++) { + C[i] = (cem & 1); + cem >>= 1; + } + uvec4 M = uvec4(0); + for (uint i = 0; i < num_partitions; i++) { + M[i] = cem & 3; + cem >>= 2; + } + for (uint i = 0; i < num_partitions; i++) { + color_endpoint_mode[i] = base_mode; + if (C[i] == 0) { + --color_endpoint_mode[i]; + } + color_endpoint_mode[i] <<= 2; + color_endpoint_mode[i] |= M[i]; + } + } else if (num_partitions > 1) { + uint cem = base_cem >> 2; + for (uint i = 0; i < num_partitions; i++) { + color_endpoint_mode[i] = cem; + } + } + DecodeColorValues(color_endpoint_mode, num_partitions, color_data_bits); + + uvec4 endpoints[4][2]; + for (uint i = 0; i < num_partitions; i++) { + ComputeEndpoints(endpoints[i][0], endpoints[i][1], color_endpoint_mode[i]); + } + + for (uint i = 0; i < 16; i++) { + texel_weight_data[i] = local_buff[i]; + } + for (uint i = 0; i < 8; i++) { +#define REVERSE_BYTE(b) ((b * 0x0802U & 0x22110U) | (b * 0x8020U & 0x88440U)) * 0x10101U >> 16 + uint a = REVERSE_BYTE(texel_weight_data[i]); + uint b = REVERSE_BYTE(texel_weight_data[15 - i]); +#undef REVERSE_BYTE + texel_weight_data[i] = uint(bitfieldExtract(b, 0, 8)); + texel_weight_data[15 - i] = uint(bitfieldExtract(a, 0, 8)); + } + uint clear_byte_start = + (GetPackedBitSize(params.size, params.dual_plane, params.max_weight) >> 3) + 1; + texel_weight_data[clear_byte_start - 1] = + texel_weight_data[clear_byte_start - 1] & + uint( + ((1 << (GetPackedBitSize(params.size, params.dual_plane, params.max_weight) % 8)) - 1)); + for (uint i = 0; i < 16 - clear_byte_start; i++) { + texel_weight_data[clear_byte_start + i] = 0U; + } + texel_flag = true; // use texel "vector" and bit stream in integer decoding + DecodeIntegerSequence(params.max_weight, GetNumWeightValues(params.size, params.dual_plane)); + + UnquantizeTexelWeights(params.dual_plane, params.size); + + for (uint j = 0; j < block_dims.y; j++) { + for (uint i = 0; i < block_dims.x; i++) { + uint local_partition = Select2DPartition(partition_index, i, j, num_partitions, + (block_dims.y * block_dims.x) < 32); + vec4 p; + uvec4 C0 = ReplicateByteTo16(endpoints[local_partition][0]); + uvec4 C1 = ReplicateByteTo16(endpoints[local_partition][1]); + uvec4 plane_vec = uvec4(0); + uvec4 weight_vec = uvec4(0); + for (uint c = 0; c < 4; c++) { + if (params.dual_plane && (((plane_index + 1) & 3) == c)) { + plane_vec[c] = 1; + } + weight_vec[c] = unquantized_texel_weights[plane_vec[c]][j * block_dims.x + i]; + } + vec4 Cf = vec4((C0 * (uvec4(64) - weight_vec) + C1 * weight_vec + uvec4(32)) / 64); + p = (Cf / 65535.0); + imageStore(dest_image, coord + ivec3(i, j, 0), p.gbar); + } + } +} + +void main() { + uvec3 pos = gl_GlobalInvocationID; + pos.x <<= bytes_per_block_log2; + + // Read as soon as possible due to its latency + const uint swizzle = SwizzleOffset(pos.xy); + + const uint block_y = pos.y >> GOB_SIZE_Y_SHIFT; + + uint offset = 0; + offset += pos.z * layer_stride; + offset += (block_y >> block_height) * block_size; + offset += (block_y & block_height_mask) << GOB_SIZE_SHIFT; + offset += (pos.x >> GOB_SIZE_X_SHIFT) << x_shift; + offset += swizzle; + + const ivec3 coord = ivec3(gl_GlobalInvocationID * uvec3(block_dims, 1)); + uint block_index = + pos.z * gl_WorkGroupSize.x * gl_WorkGroupSize.y + pos.y * gl_WorkGroupSize.x + pos.x; + + current_index = 0; + bitsread = 0; + for (int i = 0; i < 16; i++) { + local_buff[i] = ReadTexel(offset + i); + } + DecompressBlock(coord, block_index); +} diff --git a/src/video_core/host_shaders/source_shader.h.in b/src/video_core/host_shaders/source_shader.h.in index ccdb0d2a9..929dec39b 100644 --- a/src/video_core/host_shaders/source_shader.h.in +++ b/src/video_core/host_shaders/source_shader.h.in @@ -4,6 +4,8 @@ namespace HostShaders { -constexpr std::string_view @CONTENTS_NAME@ = R"(@CONTENTS@)"; +constexpr std::string_view @CONTENTS_NAME@ = { +@CONTENTS@ +}; } // namespace HostShaders diff --git a/src/video_core/macro/macro.cpp b/src/video_core/macro/macro.cpp index cd21a2112..d7fabe605 100644 --- a/src/video_core/macro/macro.cpp +++ b/src/video_core/macro/macro.cpp @@ -6,7 +6,7 @@ #include <boost/container_hash/hash.hpp> #include "common/assert.h" #include "common/logging/log.h" -#include "core/settings.h" +#include "common/settings.h" #include "video_core/engines/maxwell_3d.h" #include "video_core/macro/macro.h" #include "video_core/macro/macro_hle.h" diff --git a/src/video_core/query_cache.h b/src/video_core/query_cache.h index 203f2af05..aac851253 100644 --- a/src/video_core/query_cache.h +++ b/src/video_core/query_cache.h @@ -16,8 +16,8 @@ #include <vector> #include "common/assert.h" +#include "common/settings.h" #include "core/core.h" -#include "core/settings.h" #include "video_core/engines/maxwell_3d.h" #include "video_core/gpu.h" #include "video_core/memory_manager.h" @@ -208,9 +208,9 @@ public: private: /// Flushes a memory range to guest memory and removes it from the cache. void FlushAndRemoveRegion(VAddr addr, std::size_t size) { - const u64 addr_begin = static_cast<u64>(addr); - const u64 addr_end = addr_begin + static_cast<u64>(size); - const auto in_range = [addr_begin, addr_end](CachedQuery& query) { + const u64 addr_begin = addr; + const u64 addr_end = addr_begin + size; + const auto in_range = [addr_begin, addr_end](const CachedQuery& query) { const u64 cache_begin = query.GetCpuAddr(); const u64 cache_end = cache_begin + query.SizeInBytes(); return cache_begin < addr_end && addr_begin < cache_end; @@ -230,8 +230,7 @@ private: rasterizer.UpdatePagesCachedCount(query.GetCpuAddr(), query.SizeInBytes(), -1); query.Flush(); } - contents.erase(std::remove_if(std::begin(contents), std::end(contents), in_range), - std::end(contents)); + std::erase_if(contents, in_range); } } diff --git a/src/video_core/renderer_base.cpp b/src/video_core/renderer_base.cpp index a93a1732c..c9a360aaf 100644 --- a/src/video_core/renderer_base.cpp +++ b/src/video_core/renderer_base.cpp @@ -3,8 +3,8 @@ // Refer to the license.txt file included. #include "common/logging/log.h" +#include "common/settings.h" #include "core/frontend/emu_window.h" -#include "core/settings.h" #include "video_core/renderer_base.h" namespace VideoCore { diff --git a/src/video_core/renderer_opengl/gl_device.cpp b/src/video_core/renderer_opengl/gl_device.cpp index 5776fccdc..b113f54db 100644 --- a/src/video_core/renderer_opengl/gl_device.cpp +++ b/src/video_core/renderer_opengl/gl_device.cpp @@ -16,7 +16,7 @@ #include "common/logging/log.h" #include "common/scope_exit.h" -#include "core/settings.h" +#include "common/settings.h" #include "video_core/renderer_opengl/gl_device.h" #include "video_core/renderer_opengl/gl_resource_manager.h" diff --git a/src/video_core/renderer_opengl/gl_rasterizer.cpp b/src/video_core/renderer_opengl/gl_rasterizer.cpp index 4610fd160..0863904e9 100644 --- a/src/video_core/renderer_opengl/gl_rasterizer.cpp +++ b/src/video_core/renderer_opengl/gl_rasterizer.cpp @@ -17,10 +17,10 @@ #include "common/math_util.h" #include "common/microprofile.h" #include "common/scope_exit.h" +#include "common/settings.h" #include "core/core.h" #include "core/hle/kernel/process.h" #include "core/memory.h" -#include "core/settings.h" #include "video_core/engines/kepler_compute.h" #include "video_core/engines/maxwell_3d.h" #include "video_core/engines/shader_type.h" diff --git a/src/video_core/renderer_opengl/gl_shader_disk_cache.cpp b/src/video_core/renderer_opengl/gl_shader_disk_cache.cpp index 955b2abc4..97fb11ac6 100644 --- a/src/video_core/renderer_opengl/gl_shader_disk_cache.cpp +++ b/src/video_core/renderer_opengl/gl_shader_disk_cache.cpp @@ -12,10 +12,10 @@ #include "common/file_util.h" #include "common/logging/log.h" #include "common/scm_rev.h" +#include "common/settings.h" #include "common/zstd_compression.h" #include "core/core.h" #include "core/hle/kernel/process.h" -#include "core/settings.h" #include "video_core/engines/shader_type.h" #include "video_core/renderer_opengl/gl_shader_cache.h" #include "video_core/renderer_opengl/gl_shader_disk_cache.h" diff --git a/src/video_core/renderer_opengl/gl_texture_cache.cpp b/src/video_core/renderer_opengl/gl_texture_cache.cpp index e028677e9..623b43d8a 100644 --- a/src/video_core/renderer_opengl/gl_texture_cache.cpp +++ b/src/video_core/renderer_opengl/gl_texture_cache.cpp @@ -307,7 +307,8 @@ void ApplySwizzle(GLuint handle, PixelFormat format, std::array<SwizzleSource, 4 [[nodiscard]] bool CanBeAccelerated(const TextureCacheRuntime& runtime, const VideoCommon::ImageInfo& info) { - // Disable accelerated uploads for now as they don't implement swizzled uploads + return !runtime.HasNativeASTC() && IsPixelFormatASTC(info.format); + // Disable other accelerated uploads for now as they don't implement swizzled uploads return false; switch (info.type) { case ImageType::e2D: @@ -569,7 +570,11 @@ void TextureCacheRuntime::AccelerateImageUpload(Image& image, const ImageBufferM std::span<const SwizzleParameters> swizzles) { switch (image.info.type) { case ImageType::e2D: - return util_shaders.BlockLinearUpload2D(image, map, swizzles); + if (IsPixelFormatASTC(image.info.format)) { + return util_shaders.ASTCDecode(image, map, swizzles); + } else { + return util_shaders.BlockLinearUpload2D(image, map, swizzles); + } case ImageType::e3D: return util_shaders.BlockLinearUpload3D(image, map, swizzles); case ImageType::Linear: @@ -599,6 +604,10 @@ FormatProperties TextureCacheRuntime::FormatInfo(ImageType type, GLenum internal } } +bool TextureCacheRuntime::HasNativeASTC() const noexcept { + return device.HasASTC(); +} + TextureCacheRuntime::StagingBuffers::StagingBuffers(GLenum storage_flags_, GLenum map_flags_) : storage_flags{storage_flags_}, map_flags{map_flags_} {} diff --git a/src/video_core/renderer_opengl/gl_texture_cache.h b/src/video_core/renderer_opengl/gl_texture_cache.h index 3fbaa102f..3c871541b 100644 --- a/src/video_core/renderer_opengl/gl_texture_cache.h +++ b/src/video_core/renderer_opengl/gl_texture_cache.h @@ -95,6 +95,8 @@ public: return has_broken_texture_view_formats; } + bool HasNativeASTC() const noexcept; + private: struct StagingBuffers { explicit StagingBuffers(GLenum storage_flags_, GLenum map_flags_); diff --git a/src/video_core/renderer_opengl/renderer_opengl.cpp b/src/video_core/renderer_opengl/renderer_opengl.cpp index 9d2acd4d9..cc2e499f9 100644 --- a/src/video_core/renderer_opengl/renderer_opengl.cpp +++ b/src/video_core/renderer_opengl/renderer_opengl.cpp @@ -13,13 +13,13 @@ #include "common/assert.h" #include "common/logging/log.h" #include "common/microprofile.h" +#include "common/settings.h" #include "common/telemetry.h" #include "core/core.h" #include "core/core_timing.h" #include "core/frontend/emu_window.h" #include "core/memory.h" #include "core/perf_stats.h" -#include "core/settings.h" #include "core/telemetry_session.h" #include "video_core/host_shaders/opengl_present_frag.h" #include "video_core/host_shaders/opengl_present_vert.h" diff --git a/src/video_core/renderer_opengl/util_shaders.cpp b/src/video_core/renderer_opengl/util_shaders.cpp index 2fe4799bc..47fddcb6e 100644 --- a/src/video_core/renderer_opengl/util_shaders.cpp +++ b/src/video_core/renderer_opengl/util_shaders.cpp @@ -2,7 +2,6 @@ // Licensed under GPLv2 or any later version // Refer to the license.txt file included. -#include <bit> #include <span> #include <string_view> @@ -11,6 +10,7 @@ #include "common/assert.h" #include "common/common_types.h" #include "common/div_ceil.h" +#include "video_core/host_shaders/astc_decoder_comp.h" #include "video_core/host_shaders/block_linear_unswizzle_2d_comp.h" #include "video_core/host_shaders/block_linear_unswizzle_3d_comp.h" #include "video_core/host_shaders/opengl_copy_bc4_comp.h" @@ -20,16 +20,18 @@ #include "video_core/renderer_opengl/gl_shader_manager.h" #include "video_core/renderer_opengl/gl_texture_cache.h" #include "video_core/renderer_opengl/util_shaders.h" -#include "video_core/surface.h" #include "video_core/texture_cache/accelerated_swizzle.h" #include "video_core/texture_cache/types.h" #include "video_core/texture_cache/util.h" +#include "video_core/textures/astc.h" #include "video_core/textures/decoders.h" namespace OpenGL { using namespace HostShaders; +using namespace Tegra::Texture::ASTC; +using VideoCommon::Extent2D; using VideoCommon::Extent3D; using VideoCommon::ImageCopy; using VideoCommon::ImageType; @@ -57,7 +59,7 @@ size_t NumPixelsInCopy(const VideoCommon::ImageCopy& copy) { } // Anonymous namespace UtilShaders::UtilShaders(ProgramManager& program_manager_) - : program_manager{program_manager_}, + : program_manager{program_manager_}, astc_decoder_program(MakeProgram(ASTC_DECODER_COMP)), block_linear_unswizzle_2d_program(MakeProgram(BLOCK_LINEAR_UNSWIZZLE_2D_COMP)), block_linear_unswizzle_3d_program(MakeProgram(BLOCK_LINEAR_UNSWIZZLE_3D_COMP)), pitch_unswizzle_program(MakeProgram(PITCH_UNSWIZZLE_COMP)), @@ -65,11 +67,79 @@ UtilShaders::UtilShaders(ProgramManager& program_manager_) copy_bc4_program(MakeProgram(OPENGL_COPY_BC4_COMP)) { const auto swizzle_table = Tegra::Texture::MakeSwizzleTable(); swizzle_table_buffer.Create(); + astc_buffer.Create(); glNamedBufferStorage(swizzle_table_buffer.handle, sizeof(swizzle_table), &swizzle_table, 0); + glNamedBufferStorage(astc_buffer.handle, sizeof(ASTC_BUFFER_DATA), &ASTC_BUFFER_DATA, 0); } UtilShaders::~UtilShaders() = default; +void UtilShaders::ASTCDecode(Image& image, const ImageBufferMap& map, + std::span<const VideoCommon::SwizzleParameters> swizzles) { + static constexpr GLuint BINDING_SWIZZLE_BUFFER = 0; + static constexpr GLuint BINDING_INPUT_BUFFER = 1; + static constexpr GLuint BINDING_ENC_BUFFER = 2; + + static constexpr GLuint BINDING_6_TO_8_BUFFER = 3; + static constexpr GLuint BINDING_7_TO_8_BUFFER = 4; + static constexpr GLuint BINDING_8_TO_8_BUFFER = 5; + static constexpr GLuint BINDING_BYTE_TO_16_BUFFER = 6; + + static constexpr GLuint BINDING_OUTPUT_IMAGE = 0; + + const Extent2D tile_size{ + .width = VideoCore::Surface::DefaultBlockWidth(image.info.format), + .height = VideoCore::Surface::DefaultBlockHeight(image.info.format), + }; + program_manager.BindHostCompute(astc_decoder_program.handle); + glBindBufferBase(GL_SHADER_STORAGE_BUFFER, BINDING_SWIZZLE_BUFFER, swizzle_table_buffer.handle); + glBindBufferRange(GL_SHADER_STORAGE_BUFFER, BINDING_ENC_BUFFER, astc_buffer.handle, + offsetof(AstcBufferData, encoding_values), + sizeof(AstcBufferData::encoding_values)); + glBindBufferRange(GL_SHADER_STORAGE_BUFFER, BINDING_6_TO_8_BUFFER, astc_buffer.handle, + offsetof(AstcBufferData, replicate_6_to_8), + sizeof(AstcBufferData::replicate_6_to_8)); + glBindBufferRange(GL_SHADER_STORAGE_BUFFER, BINDING_7_TO_8_BUFFER, astc_buffer.handle, + offsetof(AstcBufferData, replicate_7_to_8), + sizeof(AstcBufferData::replicate_7_to_8)); + glBindBufferRange(GL_SHADER_STORAGE_BUFFER, BINDING_8_TO_8_BUFFER, astc_buffer.handle, + offsetof(AstcBufferData, replicate_8_to_8), + sizeof(AstcBufferData::replicate_8_to_8)); + glBindBufferRange(GL_SHADER_STORAGE_BUFFER, BINDING_BYTE_TO_16_BUFFER, astc_buffer.handle, + offsetof(AstcBufferData, replicate_byte_to_16), + sizeof(AstcBufferData::replicate_byte_to_16)); + + glFlushMappedNamedBufferRange(map.buffer, map.offset, image.guest_size_bytes); + glUniform2ui(1, tile_size.width, tile_size.height); + // Ensure buffer data is valid before dispatching + glFlush(); + for (const SwizzleParameters& swizzle : swizzles) { + const size_t input_offset = swizzle.buffer_offset + map.offset; + const u32 num_dispatches_x = Common::DivCeil(swizzle.num_tiles.width, 32U); + const u32 num_dispatches_y = Common::DivCeil(swizzle.num_tiles.height, 32U); + + const auto params = MakeBlockLinearSwizzle2DParams(swizzle, image.info); + ASSERT(params.origin == (std::array<u32, 3>{0, 0, 0})); + ASSERT(params.destination == (std::array<s32, 3>{0, 0, 0})); + + glUniform1ui(2, params.bytes_per_block_log2); + glUniform1ui(3, params.layer_stride); + glUniform1ui(4, params.block_size); + glUniform1ui(5, params.x_shift); + glUniform1ui(6, params.block_height); + glUniform1ui(7, params.block_height_mask); + + glBindImageTexture(BINDING_OUTPUT_IMAGE, image.StorageHandle(), swizzle.level, GL_TRUE, 0, + GL_WRITE_ONLY, GL_RGBA8); + // ASTC texture data + glBindBufferRange(GL_SHADER_STORAGE_BUFFER, BINDING_INPUT_BUFFER, map.buffer, input_offset, + image.guest_size_bytes - swizzle.buffer_offset); + + glDispatchCompute(num_dispatches_x, num_dispatches_y, image.info.resources.layers); + } + program_manager.RestoreGuestCompute(); +} + void UtilShaders::BlockLinearUpload2D(Image& image, const ImageBufferMap& map, std::span<const SwizzleParameters> swizzles) { static constexpr Extent3D WORKGROUP_SIZE{32, 32, 1}; diff --git a/src/video_core/renderer_opengl/util_shaders.h b/src/video_core/renderer_opengl/util_shaders.h index 93b009743..53d65f368 100644 --- a/src/video_core/renderer_opengl/util_shaders.h +++ b/src/video_core/renderer_opengl/util_shaders.h @@ -40,6 +40,9 @@ public: explicit UtilShaders(ProgramManager& program_manager); ~UtilShaders(); + void ASTCDecode(Image& image, const ImageBufferMap& map, + std::span<const VideoCommon::SwizzleParameters> swizzles); + void BlockLinearUpload2D(Image& image, const ImageBufferMap& map, std::span<const VideoCommon::SwizzleParameters> swizzles); @@ -59,7 +62,9 @@ private: ProgramManager& program_manager; OGLBuffer swizzle_table_buffer; + OGLBuffer astc_buffer; + OGLProgram astc_decoder_program; OGLProgram block_linear_unswizzle_2d_program; OGLProgram block_linear_unswizzle_3d_program; OGLProgram pitch_unswizzle_program; diff --git a/src/video_core/renderer_vulkan/maxwell_to_vk.cpp b/src/video_core/renderer_vulkan/maxwell_to_vk.cpp index 19aaf034f..f088447e9 100644 --- a/src/video_core/renderer_vulkan/maxwell_to_vk.cpp +++ b/src/video_core/renderer_vulkan/maxwell_to_vk.cpp @@ -166,7 +166,7 @@ struct FormatTuple { {VK_FORMAT_R16G16_SINT, Attachable | Storage}, // R16G16_SINT {VK_FORMAT_R16G16_SNORM, Attachable | Storage}, // R16G16_SNORM {VK_FORMAT_UNDEFINED}, // R32G32B32_FLOAT - {VK_FORMAT_R8G8B8A8_SRGB, Attachable}, // A8B8G8R8_SRGB + {VK_FORMAT_A8B8G8R8_SRGB_PACK32, Attachable}, // A8B8G8R8_SRGB {VK_FORMAT_R8G8_UNORM, Attachable | Storage}, // R8G8_UNORM {VK_FORMAT_R8G8_SNORM, Attachable | Storage}, // R8G8_SNORM {VK_FORMAT_R8G8_SINT, Attachable | Storage}, // R8G8_SINT diff --git a/src/video_core/renderer_vulkan/renderer_vulkan.cpp b/src/video_core/renderer_vulkan/renderer_vulkan.cpp index 1cc720ddd..2e0cf4232 100644 --- a/src/video_core/renderer_vulkan/renderer_vulkan.cpp +++ b/src/video_core/renderer_vulkan/renderer_vulkan.cpp @@ -13,11 +13,11 @@ #include <fmt/format.h> #include "common/logging/log.h" +#include "common/settings.h" #include "common/telemetry.h" #include "core/core.h" #include "core/core_timing.h" #include "core/frontend/emu_window.h" -#include "core/settings.h" #include "core/telemetry_session.h" #include "video_core/gpu.h" #include "video_core/renderer_vulkan/renderer_vulkan.h" @@ -143,7 +143,10 @@ void RendererVulkan::SwapBuffers(const Tegra::FramebufferConfig* framebuffer) { scheduler.WaitWorker(); - swapchain.AcquireNextImage(); + while (!swapchain.AcquireNextImage()) { + swapchain.Create(layout.width, layout.height, is_srgb); + blit_screen.Recreate(); + } const VkSemaphore render_semaphore = blit_screen.Draw(*framebuffer, use_accelerated); scheduler.Flush(render_semaphore); diff --git a/src/video_core/renderer_vulkan/vk_buffer_cache.cpp b/src/video_core/renderer_vulkan/vk_buffer_cache.cpp index 668633e7b..8cb65e588 100644 --- a/src/video_core/renderer_vulkan/vk_buffer_cache.cpp +++ b/src/video_core/renderer_vulkan/vk_buffer_cache.cpp @@ -176,7 +176,7 @@ void BufferCacheRuntime::BindVertexBuffer(u32 index, VkBuffer buffer, u32 offset u32 stride) { if (device.IsExtExtendedDynamicStateSupported()) { scheduler.Record([index, buffer, offset, size, stride](vk::CommandBuffer cmdbuf) { - const VkDeviceSize vk_offset = offset; + const VkDeviceSize vk_offset = buffer != VK_NULL_HANDLE ? offset : 0; const VkDeviceSize vk_size = buffer != VK_NULL_HANDLE ? size : VK_WHOLE_SIZE; const VkDeviceSize vk_stride = stride; cmdbuf.BindVertexBuffers2EXT(index, 1, &buffer, &vk_offset, &vk_size, &vk_stride); diff --git a/src/video_core/renderer_vulkan/vk_compute_pass.cpp b/src/video_core/renderer_vulkan/vk_compute_pass.cpp index 2f9a7b028..e11406e58 100644 --- a/src/video_core/renderer_vulkan/vk_compute_pass.cpp +++ b/src/video_core/renderer_vulkan/vk_compute_pass.cpp @@ -11,18 +11,39 @@ #include "common/assert.h" #include "common/common_types.h" #include "common/div_ceil.h" +#include "video_core/host_shaders/astc_decoder_comp_spv.h" #include "video_core/host_shaders/vulkan_quad_indexed_comp_spv.h" #include "video_core/host_shaders/vulkan_uint8_comp_spv.h" #include "video_core/renderer_vulkan/vk_compute_pass.h" #include "video_core/renderer_vulkan/vk_descriptor_pool.h" #include "video_core/renderer_vulkan/vk_scheduler.h" #include "video_core/renderer_vulkan/vk_staging_buffer_pool.h" +#include "video_core/renderer_vulkan/vk_texture_cache.h" #include "video_core/renderer_vulkan/vk_update_descriptor.h" +#include "video_core/texture_cache/accelerated_swizzle.h" +#include "video_core/texture_cache/types.h" +#include "video_core/textures/astc.h" +#include "video_core/textures/decoders.h" #include "video_core/vulkan_common/vulkan_device.h" #include "video_core/vulkan_common/vulkan_wrapper.h" namespace Vulkan { + +using Tegra::Texture::SWIZZLE_TABLE; +using Tegra::Texture::ASTC::EncodingsValues; +using namespace Tegra::Texture::ASTC; + namespace { + +constexpr u32 ASTC_BINDING_INPUT_BUFFER = 0; +constexpr u32 ASTC_BINDING_ENC_BUFFER = 1; +constexpr u32 ASTC_BINDING_6_TO_8_BUFFER = 2; +constexpr u32 ASTC_BINDING_7_TO_8_BUFFER = 3; +constexpr u32 ASTC_BINDING_8_TO_8_BUFFER = 4; +constexpr u32 ASTC_BINDING_BYTE_TO_16_BUFFER = 5; +constexpr u32 ASTC_BINDING_SWIZZLE_BUFFER = 6; +constexpr u32 ASTC_BINDING_OUTPUT_IMAGE = 7; + VkPushConstantRange BuildComputePushConstantRange(std::size_t size) { return { .stageFlags = VK_SHADER_STAGE_COMPUTE_BIT, @@ -50,6 +71,67 @@ std::array<VkDescriptorSetLayoutBinding, 2> BuildInputOutputDescriptorSetBinding }}; } +std::array<VkDescriptorSetLayoutBinding, 8> BuildASTCDescriptorSetBindings() { + return {{ + { + .binding = ASTC_BINDING_INPUT_BUFFER, + .descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER, + .descriptorCount = 1, + .stageFlags = VK_SHADER_STAGE_COMPUTE_BIT, + .pImmutableSamplers = nullptr, + }, + { + .binding = ASTC_BINDING_ENC_BUFFER, + .descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER, + .descriptorCount = 1, + .stageFlags = VK_SHADER_STAGE_COMPUTE_BIT, + .pImmutableSamplers = nullptr, + }, + { + .binding = ASTC_BINDING_6_TO_8_BUFFER, + .descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER, + .descriptorCount = 1, + .stageFlags = VK_SHADER_STAGE_COMPUTE_BIT, + .pImmutableSamplers = nullptr, + }, + { + .binding = ASTC_BINDING_7_TO_8_BUFFER, + .descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER, + .descriptorCount = 1, + .stageFlags = VK_SHADER_STAGE_COMPUTE_BIT, + .pImmutableSamplers = nullptr, + }, + { + .binding = ASTC_BINDING_8_TO_8_BUFFER, + .descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER, + .descriptorCount = 1, + .stageFlags = VK_SHADER_STAGE_COMPUTE_BIT, + .pImmutableSamplers = nullptr, + }, + { + .binding = ASTC_BINDING_BYTE_TO_16_BUFFER, + .descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER, + .descriptorCount = 1, + .stageFlags = VK_SHADER_STAGE_COMPUTE_BIT, + .pImmutableSamplers = nullptr, + }, + { + .binding = ASTC_BINDING_SWIZZLE_BUFFER, + .descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER, + .descriptorCount = 1, + .stageFlags = VK_SHADER_STAGE_COMPUTE_BIT, + .pImmutableSamplers = nullptr, + }, + { + .binding = ASTC_BINDING_OUTPUT_IMAGE, + .descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_IMAGE, + .descriptorCount = 1, + .stageFlags = VK_SHADER_STAGE_COMPUTE_BIT, + .pImmutableSamplers = nullptr, + }, + }}; +} + VkDescriptorUpdateTemplateEntryKHR BuildInputOutputDescriptorUpdateTemplate() { return { .dstBinding = 0, @@ -61,6 +143,94 @@ VkDescriptorUpdateTemplateEntryKHR BuildInputOutputDescriptorUpdateTemplate() { }; } +std::array<VkDescriptorUpdateTemplateEntryKHR, 8> BuildASTCPassDescriptorUpdateTemplateEntry() { + return {{ + { + .dstBinding = ASTC_BINDING_INPUT_BUFFER, + .dstArrayElement = 0, + .descriptorCount = 1, + .descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER, + .offset = ASTC_BINDING_INPUT_BUFFER * sizeof(DescriptorUpdateEntry), + .stride = sizeof(DescriptorUpdateEntry), + }, + { + .dstBinding = ASTC_BINDING_ENC_BUFFER, + .dstArrayElement = 0, + .descriptorCount = 1, + .descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER, + .offset = ASTC_BINDING_ENC_BUFFER * sizeof(DescriptorUpdateEntry), + .stride = sizeof(DescriptorUpdateEntry), + }, + { + .dstBinding = ASTC_BINDING_6_TO_8_BUFFER, + .dstArrayElement = 0, + .descriptorCount = 1, + .descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER, + .offset = ASTC_BINDING_6_TO_8_BUFFER * sizeof(DescriptorUpdateEntry), + .stride = sizeof(DescriptorUpdateEntry), + }, + { + .dstBinding = ASTC_BINDING_7_TO_8_BUFFER, + .dstArrayElement = 0, + .descriptorCount = 1, + .descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER, + .offset = ASTC_BINDING_7_TO_8_BUFFER * sizeof(DescriptorUpdateEntry), + .stride = sizeof(DescriptorUpdateEntry), + }, + { + .dstBinding = ASTC_BINDING_8_TO_8_BUFFER, + .dstArrayElement = 0, + .descriptorCount = 1, + .descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER, + .offset = ASTC_BINDING_8_TO_8_BUFFER * sizeof(DescriptorUpdateEntry), + .stride = sizeof(DescriptorUpdateEntry), + }, + { + .dstBinding = ASTC_BINDING_BYTE_TO_16_BUFFER, + .dstArrayElement = 0, + .descriptorCount = 1, + .descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER, + .offset = ASTC_BINDING_BYTE_TO_16_BUFFER * sizeof(DescriptorUpdateEntry), + .stride = sizeof(DescriptorUpdateEntry), + }, + { + .dstBinding = ASTC_BINDING_SWIZZLE_BUFFER, + .dstArrayElement = 0, + .descriptorCount = 1, + .descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER, + .offset = ASTC_BINDING_SWIZZLE_BUFFER * sizeof(DescriptorUpdateEntry), + .stride = sizeof(DescriptorUpdateEntry), + }, + { + .dstBinding = ASTC_BINDING_OUTPUT_IMAGE, + .dstArrayElement = 0, + .descriptorCount = 1, + .descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_IMAGE, + .offset = ASTC_BINDING_OUTPUT_IMAGE * sizeof(DescriptorUpdateEntry), + .stride = sizeof(DescriptorUpdateEntry), + }, + }}; +} + +struct AstcPushConstants { + std::array<u32, 2> blocks_dims; + u32 bytes_per_block_log2; + u32 layer_stride; + u32 block_size; + u32 x_shift; + u32 block_height; + u32 block_height_mask; +}; + +struct AstcBufferData { + decltype(SWIZZLE_TABLE) swizzle_table_buffer = SWIZZLE_TABLE; + decltype(EncodingsValues) encoding_values = EncodingsValues; + decltype(REPLICATE_6_BIT_TO_8_TABLE) replicate_6_to_8 = REPLICATE_6_BIT_TO_8_TABLE; + decltype(REPLICATE_7_BIT_TO_8_TABLE) replicate_7_to_8 = REPLICATE_7_BIT_TO_8_TABLE; + decltype(REPLICATE_8_BIT_TO_8_TABLE) replicate_8_to_8 = REPLICATE_8_BIT_TO_8_TABLE; + decltype(REPLICATE_BYTE_TO_16_TABLE) replicate_byte_to_16 = REPLICATE_BYTE_TO_16_TABLE; +} constexpr ASTC_BUFFER_DATA; + } // Anonymous namespace VKComputePass::VKComputePass(const Device& device, VKDescriptorPool& descriptor_pool, @@ -238,4 +408,167 @@ std::pair<VkBuffer, VkDeviceSize> QuadIndexedPass::Assemble( return {staging.buffer, staging.offset}; } +ASTCDecoderPass::ASTCDecoderPass(const Device& device_, VKScheduler& scheduler_, + VKDescriptorPool& descriptor_pool_, + StagingBufferPool& staging_buffer_pool_, + VKUpdateDescriptorQueue& update_descriptor_queue_, + MemoryAllocator& memory_allocator_) + : VKComputePass(device_, descriptor_pool_, BuildASTCDescriptorSetBindings(), + BuildASTCPassDescriptorUpdateTemplateEntry(), + BuildComputePushConstantRange(sizeof(AstcPushConstants)), + ASTC_DECODER_COMP_SPV), + device{device_}, scheduler{scheduler_}, staging_buffer_pool{staging_buffer_pool_}, + update_descriptor_queue{update_descriptor_queue_}, memory_allocator{memory_allocator_} {} + +ASTCDecoderPass::~ASTCDecoderPass() = default; + +void ASTCDecoderPass::MakeDataBuffer() { + constexpr size_t TOTAL_BUFFER_SIZE = sizeof(ASTC_BUFFER_DATA) + sizeof(SWIZZLE_TABLE); + data_buffer = device.GetLogical().CreateBuffer(VkBufferCreateInfo{ + .sType = VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO, + .pNext = nullptr, + .flags = 0, + .size = TOTAL_BUFFER_SIZE, + .usage = VK_BUFFER_USAGE_STORAGE_BUFFER_BIT | VK_BUFFER_USAGE_TRANSFER_DST_BIT, + .sharingMode = VK_SHARING_MODE_EXCLUSIVE, + .queueFamilyIndexCount = 0, + .pQueueFamilyIndices = nullptr, + }); + data_buffer_commit = memory_allocator.Commit(data_buffer, MemoryUsage::Upload); + + const auto staging_ref = staging_buffer_pool.Request(TOTAL_BUFFER_SIZE, MemoryUsage::Upload); + std::memcpy(staging_ref.mapped_span.data(), &ASTC_BUFFER_DATA, sizeof(ASTC_BUFFER_DATA)); + // Tack on the swizzle table at the end of the buffer + std::memcpy(staging_ref.mapped_span.data() + sizeof(ASTC_BUFFER_DATA), &SWIZZLE_TABLE, + sizeof(SWIZZLE_TABLE)); + + scheduler.Record([src = staging_ref.buffer, offset = staging_ref.offset, dst = *data_buffer, + TOTAL_BUFFER_SIZE](vk::CommandBuffer cmdbuf) { + cmdbuf.CopyBuffer(src, dst, + VkBufferCopy{ + .srcOffset = offset, + .dstOffset = 0, + .size = TOTAL_BUFFER_SIZE, + }); + cmdbuf.PipelineBarrier( + VK_PIPELINE_STAGE_ALL_COMMANDS_BIT, VK_PIPELINE_STAGE_ALL_COMMANDS_BIT, 0, + VkMemoryBarrier{ + .sType = VK_STRUCTURE_TYPE_MEMORY_BARRIER, + .pNext = nullptr, + .srcAccessMask = 0, + .dstAccessMask = VK_ACCESS_TRANSFER_READ_BIT | VK_ACCESS_TRANSFER_WRITE_BIT, + }); + }); +} + +void ASTCDecoderPass::Assemble(Image& image, const StagingBufferRef& map, + std::span<const VideoCommon::SwizzleParameters> swizzles) { + using namespace VideoCommon::Accelerated; + const std::array<u32, 2> block_dims{ + VideoCore::Surface::DefaultBlockWidth(image.info.format), + VideoCore::Surface::DefaultBlockHeight(image.info.format), + }; + scheduler.RequestOutsideRenderPassOperationContext(); + if (!data_buffer) { + MakeDataBuffer(); + } + const VkPipeline vk_pipeline = *pipeline; + const VkImageAspectFlags aspect_mask = image.AspectMask(); + const VkImage vk_image = image.Handle(); + const bool is_initialized = image.ExchangeInitialization(); + scheduler.Record( + [vk_pipeline, vk_image, aspect_mask, is_initialized](vk::CommandBuffer cmdbuf) { + const VkImageMemoryBarrier image_barrier{ + .sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER, + .pNext = nullptr, + .srcAccessMask = VK_ACCESS_SHADER_WRITE_BIT, + .dstAccessMask = VK_ACCESS_SHADER_READ_BIT | VK_ACCESS_SHADER_WRITE_BIT, + .oldLayout = is_initialized ? VK_IMAGE_LAYOUT_GENERAL : VK_IMAGE_LAYOUT_UNDEFINED, + .newLayout = VK_IMAGE_LAYOUT_GENERAL, + .srcQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED, + .dstQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED, + .image = vk_image, + .subresourceRange{ + .aspectMask = aspect_mask, + .baseMipLevel = 0, + .levelCount = VK_REMAINING_MIP_LEVELS, + .baseArrayLayer = 0, + .layerCount = VK_REMAINING_ARRAY_LAYERS, + }, + }; + cmdbuf.PipelineBarrier(is_initialized ? VK_PIPELINE_STAGE_ALL_COMMANDS_BIT : 0, + VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT, 0, image_barrier); + cmdbuf.BindPipeline(VK_PIPELINE_BIND_POINT_COMPUTE, vk_pipeline); + }); + for (const VideoCommon::SwizzleParameters& swizzle : swizzles) { + const size_t input_offset = swizzle.buffer_offset + map.offset; + const u32 num_dispatches_x = Common::DivCeil(swizzle.num_tiles.width, 32U); + const u32 num_dispatches_y = Common::DivCeil(swizzle.num_tiles.height, 32U); + const u32 num_dispatches_z = image.info.resources.layers; + + update_descriptor_queue.Acquire(); + update_descriptor_queue.AddBuffer(map.buffer, input_offset, + image.guest_size_bytes - swizzle.buffer_offset); + update_descriptor_queue.AddBuffer(*data_buffer, offsetof(AstcBufferData, encoding_values), + sizeof(AstcBufferData::encoding_values)); + update_descriptor_queue.AddBuffer(*data_buffer, offsetof(AstcBufferData, replicate_6_to_8), + sizeof(AstcBufferData::replicate_6_to_8)); + update_descriptor_queue.AddBuffer(*data_buffer, offsetof(AstcBufferData, replicate_7_to_8), + sizeof(AstcBufferData::replicate_7_to_8)); + update_descriptor_queue.AddBuffer(*data_buffer, offsetof(AstcBufferData, replicate_8_to_8), + sizeof(AstcBufferData::replicate_8_to_8)); + update_descriptor_queue.AddBuffer(*data_buffer, + offsetof(AstcBufferData, replicate_byte_to_16), + sizeof(AstcBufferData::replicate_byte_to_16)); + update_descriptor_queue.AddBuffer(*data_buffer, sizeof(AstcBufferData), + sizeof(SWIZZLE_TABLE)); + update_descriptor_queue.AddImage(image.StorageImageView(swizzle.level)); + + const VkDescriptorSet set = CommitDescriptorSet(update_descriptor_queue); + const VkPipelineLayout vk_layout = *layout; + + // To unswizzle the ASTC data + const auto params = MakeBlockLinearSwizzle2DParams(swizzle, image.info); + ASSERT(params.origin == (std::array<u32, 3>{0, 0, 0})); + ASSERT(params.destination == (std::array<s32, 3>{0, 0, 0})); + scheduler.Record([vk_layout, num_dispatches_x, num_dispatches_y, num_dispatches_z, + block_dims, params, set](vk::CommandBuffer cmdbuf) { + const AstcPushConstants uniforms{ + .blocks_dims = block_dims, + .bytes_per_block_log2 = params.bytes_per_block_log2, + .layer_stride = params.layer_stride, + .block_size = params.block_size, + .x_shift = params.x_shift, + .block_height = params.block_height, + .block_height_mask = params.block_height_mask, + }; + cmdbuf.BindDescriptorSets(VK_PIPELINE_BIND_POINT_COMPUTE, vk_layout, 0, set, {}); + cmdbuf.PushConstants(vk_layout, VK_SHADER_STAGE_COMPUTE_BIT, uniforms); + cmdbuf.Dispatch(num_dispatches_x, num_dispatches_y, num_dispatches_z); + }); + } + scheduler.Record([vk_image, aspect_mask](vk::CommandBuffer cmdbuf) { + const VkImageMemoryBarrier image_barrier{ + .sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER, + .pNext = nullptr, + .srcAccessMask = VK_ACCESS_SHADER_WRITE_BIT, + .dstAccessMask = VK_ACCESS_SHADER_READ_BIT | VK_ACCESS_SHADER_WRITE_BIT, + .oldLayout = VK_IMAGE_LAYOUT_GENERAL, + .newLayout = VK_IMAGE_LAYOUT_GENERAL, + .srcQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED, + .dstQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED, + .image = vk_image, + .subresourceRange{ + .aspectMask = aspect_mask, + .baseMipLevel = 0, + .levelCount = VK_REMAINING_MIP_LEVELS, + .baseArrayLayer = 0, + .layerCount = VK_REMAINING_ARRAY_LAYERS, + }, + }; + cmdbuf.PipelineBarrier(VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT, + VK_PIPELINE_STAGE_ALL_COMMANDS_BIT, 0, image_barrier); + }); +} + } // namespace Vulkan diff --git a/src/video_core/renderer_vulkan/vk_compute_pass.h b/src/video_core/renderer_vulkan/vk_compute_pass.h index 17d781d99..5ea187c30 100644 --- a/src/video_core/renderer_vulkan/vk_compute_pass.h +++ b/src/video_core/renderer_vulkan/vk_compute_pass.h @@ -11,14 +11,21 @@ #include "common/common_types.h" #include "video_core/engines/maxwell_3d.h" #include "video_core/renderer_vulkan/vk_descriptor_pool.h" +#include "video_core/vulkan_common/vulkan_memory_allocator.h" #include "video_core/vulkan_common/vulkan_wrapper.h" +namespace VideoCommon { +struct SwizzleParameters; +} + namespace Vulkan { class Device; class StagingBufferPool; class VKScheduler; class VKUpdateDescriptorQueue; +class Image; +struct StagingBufferRef; class VKComputePass { public: @@ -77,4 +84,29 @@ private: VKUpdateDescriptorQueue& update_descriptor_queue; }; +class ASTCDecoderPass final : public VKComputePass { +public: + explicit ASTCDecoderPass(const Device& device_, VKScheduler& scheduler_, + VKDescriptorPool& descriptor_pool_, + StagingBufferPool& staging_buffer_pool_, + VKUpdateDescriptorQueue& update_descriptor_queue_, + MemoryAllocator& memory_allocator_); + ~ASTCDecoderPass(); + + void Assemble(Image& image, const StagingBufferRef& map, + std::span<const VideoCommon::SwizzleParameters> swizzles); + +private: + void MakeDataBuffer(); + + const Device& device; + VKScheduler& scheduler; + StagingBufferPool& staging_buffer_pool; + VKUpdateDescriptorQueue& update_descriptor_queue; + MemoryAllocator& memory_allocator; + + vk::Buffer data_buffer; + MemoryCommit data_buffer_commit; +}; + } // namespace Vulkan diff --git a/src/video_core/renderer_vulkan/vk_master_semaphore.cpp b/src/video_core/renderer_vulkan/vk_master_semaphore.cpp index 56ec5e380..db78ce3d9 100644 --- a/src/video_core/renderer_vulkan/vk_master_semaphore.cpp +++ b/src/video_core/renderer_vulkan/vk_master_semaphore.cpp @@ -5,7 +5,7 @@ #include <atomic> #include <chrono> -#include "core/settings.h" +#include "common/settings.h" #include "video_core/renderer_vulkan/vk_master_semaphore.h" #include "video_core/vulkan_common/vulkan_device.h" #include "video_core/vulkan_common/vulkan_wrapper.h" diff --git a/src/video_core/renderer_vulkan/vk_master_semaphore.h b/src/video_core/renderer_vulkan/vk_master_semaphore.h index 2c7ed654d..4b6d64daa 100644 --- a/src/video_core/renderer_vulkan/vk_master_semaphore.h +++ b/src/video_core/renderer_vulkan/vk_master_semaphore.h @@ -35,8 +35,8 @@ public: } /// Returns true when a tick has been hit by the GPU. - [[nodiscard]] bool IsFree(u64 tick) { - return gpu_tick.load(std::memory_order_relaxed) >= tick; + [[nodiscard]] bool IsFree(u64 tick) const noexcept { + return KnownGpuTick() >= tick; } /// Advance to the logical tick. diff --git a/src/video_core/renderer_vulkan/vk_rasterizer.cpp b/src/video_core/renderer_vulkan/vk_rasterizer.cpp index dfd38f575..e9a0e7811 100644 --- a/src/video_core/renderer_vulkan/vk_rasterizer.cpp +++ b/src/video_core/renderer_vulkan/vk_rasterizer.cpp @@ -13,8 +13,8 @@ #include "common/logging/log.h" #include "common/microprofile.h" #include "common/scope_exit.h" +#include "common/settings.h" #include "core/core.h" -#include "core/settings.h" #include "video_core/engines/kepler_compute.h" #include "video_core/engines/maxwell_3d.h" #include "video_core/renderer_vulkan/blit_image.h" @@ -241,7 +241,10 @@ RasterizerVulkan::RasterizerVulkan(Core::Frontend::EmuWindow& emu_window_, Tegra staging_pool(device, memory_allocator, scheduler), descriptor_pool(device, scheduler), update_descriptor_queue(device, scheduler), blit_image(device, scheduler, state_tracker, descriptor_pool), - texture_cache_runtime{device, scheduler, memory_allocator, staging_pool, blit_image}, + astc_decoder_pass(device, scheduler, descriptor_pool, staging_pool, update_descriptor_queue, + memory_allocator), + texture_cache_runtime{device, scheduler, memory_allocator, + staging_pool, blit_image, astc_decoder_pass}, texture_cache(texture_cache_runtime, *this, maxwell3d, kepler_compute, gpu_memory), buffer_cache_runtime(device, memory_allocator, scheduler, staging_pool, update_descriptor_queue, descriptor_pool), diff --git a/src/video_core/renderer_vulkan/vk_rasterizer.h b/src/video_core/renderer_vulkan/vk_rasterizer.h index acea1ba2d..235afc6f3 100644 --- a/src/video_core/renderer_vulkan/vk_rasterizer.h +++ b/src/video_core/renderer_vulkan/vk_rasterizer.h @@ -173,6 +173,7 @@ private: VKDescriptorPool descriptor_pool; VKUpdateDescriptorQueue update_descriptor_queue; BlitImageHelper blit_image; + ASTCDecoderPass astc_decoder_pass; GraphicsPipelineCacheKey graphics_key; diff --git a/src/video_core/renderer_vulkan/vk_swapchain.cpp b/src/video_core/renderer_vulkan/vk_swapchain.cpp index 0b63bd6c8..dfd5c65ba 100644 --- a/src/video_core/renderer_vulkan/vk_swapchain.cpp +++ b/src/video_core/renderer_vulkan/vk_swapchain.cpp @@ -82,11 +82,13 @@ void VKSwapchain::Create(u32 width, u32 height, bool srgb) { resource_ticks.resize(image_count); } -void VKSwapchain::AcquireNextImage() { - device.GetLogical().AcquireNextImageKHR(*swapchain, std::numeric_limits<u64>::max(), - *present_semaphores[frame_index], {}, &image_index); +bool VKSwapchain::AcquireNextImage() { + const VkResult result = + device.GetLogical().AcquireNextImageKHR(*swapchain, std::numeric_limits<u64>::max(), + *present_semaphores[frame_index], {}, &image_index); scheduler.Wait(resource_ticks[image_index]); + return result == VK_SUCCESS || result == VK_SUBOPTIMAL_KHR; } bool VKSwapchain::Present(VkSemaphore render_semaphore) { diff --git a/src/video_core/renderer_vulkan/vk_swapchain.h b/src/video_core/renderer_vulkan/vk_swapchain.h index a728511e0..adc8d27cf 100644 --- a/src/video_core/renderer_vulkan/vk_swapchain.h +++ b/src/video_core/renderer_vulkan/vk_swapchain.h @@ -28,7 +28,7 @@ public: void Create(u32 width, u32 height, bool srgb); /// Acquires the next image in the swapchain, waits as needed. - void AcquireNextImage(); + bool AcquireNextImage(); /// Presents the rendered image to the swapchain. Returns true when the swapchains had to be /// recreated. Takes responsability for the ownership of fence. diff --git a/src/video_core/renderer_vulkan/vk_texture_cache.cpp b/src/video_core/renderer_vulkan/vk_texture_cache.cpp index 22a1014a9..bc2a53841 100644 --- a/src/video_core/renderer_vulkan/vk_texture_cache.cpp +++ b/src/video_core/renderer_vulkan/vk_texture_cache.cpp @@ -7,9 +7,12 @@ #include <span> #include <vector> +#include "common/bit_cast.h" + #include "video_core/engines/fermi_2d.h" #include "video_core/renderer_vulkan/blit_image.h" #include "video_core/renderer_vulkan/maxwell_to_vk.h" +#include "video_core/renderer_vulkan/vk_compute_pass.h" #include "video_core/renderer_vulkan/vk_rasterizer.h" #include "video_core/renderer_vulkan/vk_scheduler.h" #include "video_core/renderer_vulkan/vk_staging_buffer_pool.h" @@ -807,7 +810,7 @@ Image::Image(TextureCacheRuntime& runtime, const ImageInfo& info_, GPUVAddr gpu_ commit = runtime.memory_allocator.Commit(buffer, MemoryUsage::DeviceLocal); } if (IsPixelFormatASTC(info.format) && !runtime.device.IsOptimalAstcSupported()) { - flags |= VideoCommon::ImageFlagBits::Converted; + flags |= VideoCommon::ImageFlagBits::AcceleratedUpload; } if (runtime.device.HasDebuggingToolAttached()) { if (image) { @@ -816,6 +819,38 @@ Image::Image(TextureCacheRuntime& runtime, const ImageInfo& info_, GPUVAddr gpu_ buffer.SetObjectNameEXT(VideoCommon::Name(*this).c_str()); } } + static constexpr VkImageViewUsageCreateInfo storage_image_view_usage_create_info{ + .sType = VK_STRUCTURE_TYPE_IMAGE_VIEW_USAGE_CREATE_INFO, + .pNext = nullptr, + .usage = VK_IMAGE_USAGE_STORAGE_BIT, + }; + if (IsPixelFormatASTC(info.format) && !runtime.device.IsOptimalAstcSupported()) { + const auto& device = runtime.device.GetLogical(); + storage_image_views.reserve(info.resources.levels); + for (s32 level = 0; level < info.resources.levels; ++level) { + storage_image_views.push_back(device.CreateImageView(VkImageViewCreateInfo{ + .sType = VK_STRUCTURE_TYPE_IMAGE_VIEW_CREATE_INFO, + .pNext = &storage_image_view_usage_create_info, + .flags = 0, + .image = *image, + .viewType = VK_IMAGE_VIEW_TYPE_2D_ARRAY, + .format = VK_FORMAT_A8B8G8R8_UNORM_PACK32, + .components{ + .r = VK_COMPONENT_SWIZZLE_IDENTITY, + .g = VK_COMPONENT_SWIZZLE_IDENTITY, + .b = VK_COMPONENT_SWIZZLE_IDENTITY, + .a = VK_COMPONENT_SWIZZLE_IDENTITY, + }, + .subresourceRange{ + .aspectMask = aspect_mask, + .baseMipLevel = static_cast<u32>(level), + .levelCount = 1, + .baseArrayLayer = 0, + .layerCount = VK_REMAINING_ARRAY_LAYERS, + }, + })); + } + } } void Image::UploadMemory(const StagingBufferRef& map, std::span<const BufferImageCopy> copies) { @@ -918,7 +953,6 @@ ImageView::ImageView(TextureCacheRuntime& runtime, const VideoCommon::ImageViewI } } const auto format_info = MaxwellToVK::SurfaceFormat(*device, FormatType::Optimal, true, format); - const VkFormat vk_format = format_info.format; const VkImageViewUsageCreateInfo image_view_usage{ .sType = VK_STRUCTURE_TYPE_IMAGE_VIEW_USAGE_CREATE_INFO, .pNext = nullptr, @@ -930,7 +964,7 @@ ImageView::ImageView(TextureCacheRuntime& runtime, const VideoCommon::ImageViewI .flags = 0, .image = image.Handle(), .viewType = VkImageViewType{}, - .format = vk_format, + .format = format_info.format, .components{ .r = ComponentSwizzle(swizzle[0]), .g = ComponentSwizzle(swizzle[1]), @@ -982,7 +1016,7 @@ ImageView::ImageView(TextureCacheRuntime& runtime, const VideoCommon::ImageViewI .pNext = nullptr, .flags = 0, .buffer = image.Buffer(), - .format = vk_format, + .format = format_info.format, .offset = 0, // TODO: Redesign buffer cache to support this .range = image.guest_size_bytes, }); @@ -1030,14 +1064,13 @@ vk::ImageView ImageView::MakeDepthStencilView(VkImageAspectFlags aspect_mask) { Sampler::Sampler(TextureCacheRuntime& runtime, const Tegra::Texture::TSCEntry& tsc) { const auto& device = runtime.device; const bool arbitrary_borders = runtime.device.IsExtCustomBorderColorSupported(); - const std::array<float, 4> color = tsc.BorderColor(); - // C++20 bit_cast - VkClearColorValue border_color; - std::memcpy(&border_color, &color, sizeof(color)); + const auto color = tsc.BorderColor(); + const VkSamplerCustomBorderColorCreateInfoEXT border_ci{ .sType = VK_STRUCTURE_TYPE_SAMPLER_CUSTOM_BORDER_COLOR_CREATE_INFO_EXT, .pNext = nullptr, - .customBorderColor = border_color, + // TODO: Make use of std::bit_cast once libc++ supports it. + .customBorderColor = Common::BitCast<VkClearColorValue>(color), .format = VK_FORMAT_UNDEFINED, }; const void* pnext = nullptr; @@ -1167,4 +1200,13 @@ Framebuffer::Framebuffer(TextureCacheRuntime& runtime, std::span<ImageView*, NUM } } +void TextureCacheRuntime::AccelerateImageUpload( + Image& image, const StagingBufferRef& map, + std::span<const VideoCommon::SwizzleParameters> swizzles) { + if (IsPixelFormatASTC(image.info.format)) { + return astc_decoder_pass.Assemble(image, map, swizzles); + } + UNREACHABLE(); +} + } // namespace Vulkan diff --git a/src/video_core/renderer_vulkan/vk_texture_cache.h b/src/video_core/renderer_vulkan/vk_texture_cache.h index 3aee27ce0..628785d5e 100644 --- a/src/video_core/renderer_vulkan/vk_texture_cache.h +++ b/src/video_core/renderer_vulkan/vk_texture_cache.h @@ -20,6 +20,7 @@ using VideoCommon::Offset2D; using VideoCommon::RenderTargets; using VideoCore::Surface::PixelFormat; +class ASTCDecoderPass; class BlitImageHelper; class Device; class Image; @@ -60,6 +61,7 @@ struct TextureCacheRuntime { MemoryAllocator& memory_allocator; StagingBufferPool& staging_buffer_pool; BlitImageHelper& blit_image_helper; + ASTCDecoderPass& astc_decoder_pass; std::unordered_map<RenderPassKey, vk::RenderPass> renderpass_cache{}; void Finish(); @@ -83,9 +85,7 @@ struct TextureCacheRuntime { } void AccelerateImageUpload(Image&, const StagingBufferRef&, - std::span<const VideoCommon::SwizzleParameters>) { - UNREACHABLE(); - } + std::span<const VideoCommon::SwizzleParameters>); void InsertUploadMemoryBarrier() {} @@ -121,15 +121,26 @@ public: return *buffer; } - [[nodiscard]] VkImageCreateFlags AspectMask() const noexcept { + [[nodiscard]] VkImageAspectFlags AspectMask() const noexcept { return aspect_mask; } + [[nodiscard]] VkImageView StorageImageView(s32 level) const noexcept { + return *storage_image_views[level]; + } + + /// Returns true when the image is already initialized and mark it as initialized + [[nodiscard]] bool ExchangeInitialization() noexcept { + return std::exchange(initialized, true); + } + private: VKScheduler* scheduler; vk::Image image; vk::Buffer buffer; MemoryCommit commit; + vk::ImageView image_view; + std::vector<vk::ImageView> storage_image_views; VkImageAspectFlags aspect_mask = 0; bool initialized = false; }; diff --git a/src/video_core/texture_cache/accelerated_swizzle.h b/src/video_core/texture_cache/accelerated_swizzle.h index 6ec5c78c4..a11c924e1 100644 --- a/src/video_core/texture_cache/accelerated_swizzle.h +++ b/src/video_core/texture_cache/accelerated_swizzle.h @@ -13,8 +13,8 @@ namespace VideoCommon::Accelerated { struct BlockLinearSwizzle2DParams { - std::array<u32, 3> origin; - std::array<s32, 3> destination; + alignas(16) std::array<u32, 3> origin; + alignas(16) std::array<s32, 3> destination; u32 bytes_per_block_log2; u32 layer_stride; u32 block_size; diff --git a/src/video_core/texture_cache/image_view_base.cpp b/src/video_core/texture_cache/image_view_base.cpp index f89a40b4c..e8d632f9e 100644 --- a/src/video_core/texture_cache/image_view_base.cpp +++ b/src/video_core/texture_cache/image_view_base.cpp @@ -5,7 +5,7 @@ #include <algorithm> #include "common/assert.h" -#include "core/settings.h" +#include "common/settings.h" #include "video_core/compatible_formats.h" #include "video_core/surface.h" #include "video_core/texture_cache/formatter.h" diff --git a/src/video_core/texture_cache/util.cpp b/src/video_core/texture_cache/util.cpp index 2c42d1449..0ab297413 100644 --- a/src/video_core/texture_cache/util.cpp +++ b/src/video_core/texture_cache/util.cpp @@ -47,7 +47,6 @@ #include "video_core/texture_cache/formatter.h" #include "video_core/texture_cache/samples_helper.h" #include "video_core/texture_cache/util.h" -#include "video_core/textures/astc.h" #include "video_core/textures/decoders.h" namespace VideoCommon { @@ -269,16 +268,19 @@ template <u32 GOB_EXTENT> return num_tiles << shift; } -[[nodiscard]] constexpr std::array<u32, MAX_MIP_LEVELS> CalculateLevelSizes(const LevelInfo& info, - u32 num_levels) { +[[nodiscard]] constexpr LevelArray CalculateLevelSizes(const LevelInfo& info, u32 num_levels) { ASSERT(num_levels <= MAX_MIP_LEVELS); - std::array<u32, MAX_MIP_LEVELS> sizes{}; + LevelArray sizes{}; for (u32 level = 0; level < num_levels; ++level) { sizes[level] = CalculateLevelSize(info, level); } return sizes; } +[[nodiscard]] u32 CalculateLevelBytes(const LevelArray& sizes, u32 num_levels) { + return std::reduce(sizes.begin(), sizes.begin() + num_levels, 0U); +} + [[nodiscard]] constexpr LevelInfo MakeLevelInfo(PixelFormat format, Extent3D size, Extent3D block, u32 num_samples, u32 tile_width_spacing) { const auto [samples_x, samples_y] = Samples(num_samples); @@ -567,10 +569,10 @@ void SwizzleBlockLinearImage(Tegra::MemoryManager& gpu_memory, GPUVAddr gpu_addr const u32 num_levels = info.resources.levels; const std::array sizes = CalculateLevelSizes(level_info, num_levels); - size_t guest_offset = std::reduce(sizes.begin(), sizes.begin() + level, 0); + size_t guest_offset = CalculateLevelBytes(sizes, level); const size_t layer_stride = - AlignLayerSize(std::reduce(sizes.begin(), sizes.begin() + num_levels, 0), size, - level_info.block, tile_size.height, info.tile_width_spacing); + AlignLayerSize(CalculateLevelBytes(sizes, num_levels), size, level_info.block, + tile_size.height, info.tile_width_spacing); const size_t subresource_size = sizes[level]; const auto dst_data = std::make_unique<u8[]>(subresource_size); @@ -644,10 +646,10 @@ u32 CalculateLayerSize(const ImageInfo& info) noexcept { info.tile_width_spacing, info.resources.levels); } -std::array<u32, MAX_MIP_LEVELS> CalculateMipLevelOffsets(const ImageInfo& info) noexcept { +LevelArray CalculateMipLevelOffsets(const ImageInfo& info) noexcept { ASSERT(info.resources.levels <= static_cast<s32>(MAX_MIP_LEVELS)); const LevelInfo level_info = MakeLevelInfo(info); - std::array<u32, MAX_MIP_LEVELS> offsets{}; + LevelArray offsets{}; u32 offset = 0; for (s32 level = 0; level < info.resources.levels; ++level) { offsets[level] = offset; @@ -813,7 +815,7 @@ std::vector<BufferImageCopy> UnswizzleImage(Tegra::MemoryManager& gpu_memory, GP const Extent2D tile_size = DefaultBlockSize(info.format); const std::array level_sizes = CalculateLevelSizes(level_info, num_levels); const Extent2D gob = GobSize(bpp_log2, info.block.height, info.tile_width_spacing); - const u32 layer_size = std::reduce(level_sizes.begin(), level_sizes.begin() + num_levels, 0); + const u32 layer_size = CalculateLevelBytes(level_sizes, num_levels); const u32 layer_stride = AlignLayerSize(layer_size, size, level_info.block, tile_size.height, info.tile_width_spacing); size_t guest_offset = 0; @@ -879,17 +881,8 @@ void ConvertImage(std::span<const u8> input, const ImageInfo& info, std::span<u8 ASSERT(copy.image_extent == mip_size); ASSERT(copy.buffer_row_length == Common::AlignUp(mip_size.width, tile_size.width)); ASSERT(copy.buffer_image_height == Common::AlignUp(mip_size.height, tile_size.height)); - - if (IsPixelFormatASTC(info.format)) { - ASSERT(copy.image_extent.depth == 1); - Tegra::Texture::ASTC::Decompress(input.subspan(copy.buffer_offset), - copy.image_extent.width, copy.image_extent.height, - copy.image_subresource.num_layers, tile_size.width, - tile_size.height, output.subspan(output_offset)); - } else { - DecompressBC4(input.subspan(copy.buffer_offset), copy.image_extent, - output.subspan(output_offset)); - } + DecompressBC4(input.subspan(copy.buffer_offset), copy.image_extent, + output.subspan(output_offset)); copy.buffer_offset = output_offset; copy.buffer_row_length = mip_size.width; copy.buffer_image_height = mip_size.height; diff --git a/src/video_core/texture_cache/util.h b/src/video_core/texture_cache/util.h index 4d0072867..cdc5cbc75 100644 --- a/src/video_core/texture_cache/util.h +++ b/src/video_core/texture_cache/util.h @@ -20,6 +20,8 @@ namespace VideoCommon { using Tegra::Texture::TICEntry; +using LevelArray = std::array<u32, MAX_MIP_LEVELS>; + struct OverlapResult { GPUVAddr gpu_addr; VAddr cpu_addr; @@ -36,8 +38,7 @@ struct OverlapResult { [[nodiscard]] u32 CalculateLayerSize(const ImageInfo& info) noexcept; -[[nodiscard]] std::array<u32, MAX_MIP_LEVELS> CalculateMipLevelOffsets( - const ImageInfo& info) noexcept; +[[nodiscard]] LevelArray CalculateMipLevelOffsets(const ImageInfo& info) noexcept; [[nodiscard]] std::vector<u32> CalculateSliceOffsets(const ImageInfo& info); diff --git a/src/video_core/textures/astc.cpp b/src/video_core/textures/astc.cpp deleted file mode 100644 index 3625b666c..000000000 --- a/src/video_core/textures/astc.cpp +++ /dev/null @@ -1,1710 +0,0 @@ -// Copyright 2016 The University of North Carolina at Chapel Hill -// -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. -// -// Please send all BUG REPORTS to <pavel@cs.unc.edu>. -// <http://gamma.cs.unc.edu/FasTC/> - -#include <algorithm> -#include <cassert> -#include <cstring> -#include <span> -#include <vector> - -#include <boost/container/static_vector.hpp> - -#include "common/common_types.h" - -#include "video_core/textures/astc.h" - -namespace { - -/// Count the number of bits set in a number. -constexpr u32 Popcnt(u32 n) { - u32 c = 0; - for (; n; c++) { - n &= n - 1; - } - return c; -} - -} // Anonymous namespace - -class InputBitStream { -public: - constexpr explicit InputBitStream(std::span<const u8> data, size_t start_offset = 0) - : cur_byte{data.data()}, total_bits{data.size()}, next_bit{start_offset % 8} {} - - constexpr size_t GetBitsRead() const { - return bits_read; - } - - constexpr bool ReadBit() { - if (bits_read >= total_bits * 8) { - return 0; - } - const bool bit = ((*cur_byte >> next_bit) & 1) != 0; - ++next_bit; - while (next_bit >= 8) { - next_bit -= 8; - ++cur_byte; - } - ++bits_read; - return bit; - } - - constexpr u32 ReadBits(std::size_t nBits) { - u32 ret = 0; - for (std::size_t i = 0; i < nBits; ++i) { - ret |= (ReadBit() & 1) << i; - } - return ret; - } - - template <std::size_t nBits> - constexpr u32 ReadBits() { - u32 ret = 0; - for (std::size_t i = 0; i < nBits; ++i) { - ret |= (ReadBit() & 1) << i; - } - return ret; - } - -private: - const u8* cur_byte; - size_t total_bits = 0; - size_t next_bit = 0; - size_t bits_read = 0; -}; - -class OutputBitStream { -public: - constexpr explicit OutputBitStream(u8* ptr, std::size_t bits = 0, std::size_t start_offset = 0) - : cur_byte{ptr}, num_bits{bits}, next_bit{start_offset % 8} {} - - constexpr std::size_t GetBitsWritten() const { - return bits_written; - } - - constexpr void WriteBitsR(u32 val, u32 nBits) { - for (u32 i = 0; i < nBits; i++) { - WriteBit((val >> (nBits - i - 1)) & 1); - } - } - - constexpr void WriteBits(u32 val, u32 nBits) { - for (u32 i = 0; i < nBits; i++) { - WriteBit((val >> i) & 1); - } - } - -private: - constexpr void WriteBit(bool b) { - if (bits_written >= num_bits) { - return; - } - - const u32 mask = 1 << next_bit++; - - // clear the bit - *cur_byte &= static_cast<u8>(~mask); - - // Write the bit, if necessary - if (b) - *cur_byte |= static_cast<u8>(mask); - - // Next byte? - if (next_bit >= 8) { - cur_byte += 1; - next_bit = 0; - } - } - - u8* cur_byte; - std::size_t num_bits; - std::size_t bits_written = 0; - std::size_t next_bit = 0; -}; - -template <typename IntType> -class Bits { -public: - explicit Bits(const IntType& v) : m_Bits(v) {} - - Bits(const Bits&) = delete; - Bits& operator=(const Bits&) = delete; - - u8 operator[](u32 bitPos) const { - return static_cast<u8>((m_Bits >> bitPos) & 1); - } - - IntType operator()(u32 start, u32 end) const { - if (start == end) { - return (*this)[start]; - } else if (start > end) { - u32 t = start; - start = end; - end = t; - } - - u64 mask = (1 << (end - start + 1)) - 1; - return (m_Bits >> start) & static_cast<IntType>(mask); - } - -private: - const IntType& m_Bits; -}; - -enum class IntegerEncoding { JustBits, Qus32, Trit }; - -struct IntegerEncodedValue { - constexpr IntegerEncodedValue() = default; - - constexpr IntegerEncodedValue(IntegerEncoding encoding_, u32 num_bits_) - : encoding{encoding_}, num_bits{num_bits_} {} - - constexpr bool MatchesEncoding(const IntegerEncodedValue& other) const { - return encoding == other.encoding && num_bits == other.num_bits; - } - - // Returns the number of bits required to encode nVals values. - u32 GetBitLength(u32 nVals) const { - u32 totalBits = num_bits * nVals; - if (encoding == IntegerEncoding::Trit) { - totalBits += (nVals * 8 + 4) / 5; - } else if (encoding == IntegerEncoding::Qus32) { - totalBits += (nVals * 7 + 2) / 3; - } - return totalBits; - } - - IntegerEncoding encoding{}; - u32 num_bits = 0; - u32 bit_value = 0; - union { - u32 qus32_value = 0; - u32 trit_value; - }; -}; -using IntegerEncodedVector = boost::container::static_vector< - IntegerEncodedValue, 256, - boost::container::static_vector_options< - boost::container::inplace_alignment<alignof(IntegerEncodedValue)>, - boost::container::throw_on_overflow<false>>::type>; - -static void DecodeTritBlock(InputBitStream& bits, IntegerEncodedVector& result, u32 nBitsPerValue) { - // Implement the algorithm in section C.2.12 - std::array<u32, 5> m; - std::array<u32, 5> t; - u32 T; - - // Read the trit encoded block according to - // table C.2.14 - m[0] = bits.ReadBits(nBitsPerValue); - T = bits.ReadBits<2>(); - m[1] = bits.ReadBits(nBitsPerValue); - T |= bits.ReadBits<2>() << 2; - m[2] = bits.ReadBits(nBitsPerValue); - T |= bits.ReadBit() << 4; - m[3] = bits.ReadBits(nBitsPerValue); - T |= bits.ReadBits<2>() << 5; - m[4] = bits.ReadBits(nBitsPerValue); - T |= bits.ReadBit() << 7; - - u32 C = 0; - - Bits<u32> Tb(T); - if (Tb(2, 4) == 7) { - C = (Tb(5, 7) << 2) | Tb(0, 1); - t[4] = t[3] = 2; - } else { - C = Tb(0, 4); - if (Tb(5, 6) == 3) { - t[4] = 2; - t[3] = Tb[7]; - } else { - t[4] = Tb[7]; - t[3] = Tb(5, 6); - } - } - - Bits<u32> Cb(C); - if (Cb(0, 1) == 3) { - t[2] = 2; - t[1] = Cb[4]; - t[0] = (Cb[3] << 1) | (Cb[2] & ~Cb[3]); - } else if (Cb(2, 3) == 3) { - t[2] = 2; - t[1] = 2; - t[0] = Cb(0, 1); - } else { - t[2] = Cb[4]; - t[1] = Cb(2, 3); - t[0] = (Cb[1] << 1) | (Cb[0] & ~Cb[1]); - } - - for (std::size_t i = 0; i < 5; ++i) { - IntegerEncodedValue& val = result.emplace_back(IntegerEncoding::Trit, nBitsPerValue); - val.bit_value = m[i]; - val.trit_value = t[i]; - } -} - -static void DecodeQus32Block(InputBitStream& bits, IntegerEncodedVector& result, - u32 nBitsPerValue) { - // Implement the algorithm in section C.2.12 - u32 m[3]; - u32 q[3]; - u32 Q; - - // Read the trit encoded block according to - // table C.2.15 - m[0] = bits.ReadBits(nBitsPerValue); - Q = bits.ReadBits<3>(); - m[1] = bits.ReadBits(nBitsPerValue); - Q |= bits.ReadBits<2>() << 3; - m[2] = bits.ReadBits(nBitsPerValue); - Q |= bits.ReadBits<2>() << 5; - - Bits<u32> Qb(Q); - if (Qb(1, 2) == 3 && Qb(5, 6) == 0) { - q[0] = q[1] = 4; - q[2] = (Qb[0] << 2) | ((Qb[4] & ~Qb[0]) << 1) | (Qb[3] & ~Qb[0]); - } else { - u32 C = 0; - if (Qb(1, 2) == 3) { - q[2] = 4; - C = (Qb(3, 4) << 3) | ((~Qb(5, 6) & 3) << 1) | Qb[0]; - } else { - q[2] = Qb(5, 6); - C = Qb(0, 4); - } - - Bits<u32> Cb(C); - if (Cb(0, 2) == 5) { - q[1] = 4; - q[0] = Cb(3, 4); - } else { - q[1] = Cb(3, 4); - q[0] = Cb(0, 2); - } - } - - for (std::size_t i = 0; i < 3; ++i) { - IntegerEncodedValue& val = result.emplace_back(IntegerEncoding::Qus32, nBitsPerValue); - val.bit_value = m[i]; - val.qus32_value = q[i]; - } -} - -// Returns a new instance of this struct that corresponds to the -// can take no more than maxval values -static constexpr IntegerEncodedValue CreateEncoding(u32 maxVal) { - while (maxVal > 0) { - u32 check = maxVal + 1; - - // Is maxVal a power of two? - if (!(check & (check - 1))) { - return IntegerEncodedValue(IntegerEncoding::JustBits, Popcnt(maxVal)); - } - - // Is maxVal of the type 3*2^n - 1? - if ((check % 3 == 0) && !((check / 3) & ((check / 3) - 1))) { - return IntegerEncodedValue(IntegerEncoding::Trit, Popcnt(check / 3 - 1)); - } - - // Is maxVal of the type 5*2^n - 1? - if ((check % 5 == 0) && !((check / 5) & ((check / 5) - 1))) { - return IntegerEncodedValue(IntegerEncoding::Qus32, Popcnt(check / 5 - 1)); - } - - // Apparently it can't be represented with a bounded integer sequence... - // just iterate. - maxVal--; - } - return IntegerEncodedValue(IntegerEncoding::JustBits, 0); -} - -static constexpr std::array<IntegerEncodedValue, 256> MakeEncodedValues() { - std::array<IntegerEncodedValue, 256> encodings{}; - for (std::size_t i = 0; i < encodings.size(); ++i) { - encodings[i] = CreateEncoding(static_cast<u32>(i)); - } - return encodings; -} - -static constexpr std::array EncodingsValues = MakeEncodedValues(); - -// Fills result with the values that are encoded in the given -// bitstream. We must know beforehand what the maximum possible -// value is, and how many values we're decoding. -static void DecodeIntegerSequence(IntegerEncodedVector& result, InputBitStream& bits, u32 maxRange, - u32 nValues) { - // Determine encoding parameters - IntegerEncodedValue val = EncodingsValues[maxRange]; - - // Start decoding - u32 nValsDecoded = 0; - while (nValsDecoded < nValues) { - switch (val.encoding) { - case IntegerEncoding::Qus32: - DecodeQus32Block(bits, result, val.num_bits); - nValsDecoded += 3; - break; - - case IntegerEncoding::Trit: - DecodeTritBlock(bits, result, val.num_bits); - nValsDecoded += 5; - break; - - case IntegerEncoding::JustBits: - val.bit_value = bits.ReadBits(val.num_bits); - result.push_back(val); - nValsDecoded++; - break; - } - } -} - -namespace ASTCC { - -struct TexelWeightParams { - u32 m_Width = 0; - u32 m_Height = 0; - bool m_bDualPlane = false; - u32 m_MaxWeight = 0; - bool m_bError = false; - bool m_bVoidExtentLDR = false; - bool m_bVoidExtentHDR = false; - - u32 GetPackedBitSize() const { - // How many indices do we have? - u32 nIdxs = m_Height * m_Width; - if (m_bDualPlane) { - nIdxs *= 2; - } - - return EncodingsValues[m_MaxWeight].GetBitLength(nIdxs); - } - - u32 GetNumWeightValues() const { - u32 ret = m_Width * m_Height; - if (m_bDualPlane) { - ret *= 2; - } - return ret; - } -}; - -static TexelWeightParams DecodeBlockInfo(InputBitStream& strm) { - TexelWeightParams params; - - // Read the entire block mode all at once - u16 modeBits = static_cast<u16>(strm.ReadBits<11>()); - - // Does this match the void extent block mode? - if ((modeBits & 0x01FF) == 0x1FC) { - if (modeBits & 0x200) { - params.m_bVoidExtentHDR = true; - } else { - params.m_bVoidExtentLDR = true; - } - - // Next two bits must be one. - if (!(modeBits & 0x400) || !strm.ReadBit()) { - params.m_bError = true; - } - - return params; - } - - // First check if the last four bits are zero - if ((modeBits & 0xF) == 0) { - params.m_bError = true; - return params; - } - - // If the last two bits are zero, then if bits - // [6-8] are all ones, this is also reserved. - if ((modeBits & 0x3) == 0 && (modeBits & 0x1C0) == 0x1C0) { - params.m_bError = true; - return params; - } - - // Otherwise, there is no error... Figure out the layout - // of the block mode. Layout is determined by a number - // between 0 and 9 corresponding to table C.2.8 of the - // ASTC spec. - u32 layout = 0; - - if ((modeBits & 0x1) || (modeBits & 0x2)) { - // layout is in [0-4] - if (modeBits & 0x8) { - // layout is in [2-4] - if (modeBits & 0x4) { - // layout is in [3-4] - if (modeBits & 0x100) { - layout = 4; - } else { - layout = 3; - } - } else { - layout = 2; - } - } else { - // layout is in [0-1] - if (modeBits & 0x4) { - layout = 1; - } else { - layout = 0; - } - } - } else { - // layout is in [5-9] - if (modeBits & 0x100) { - // layout is in [7-9] - if (modeBits & 0x80) { - // layout is in [7-8] - assert((modeBits & 0x40) == 0U); - if (modeBits & 0x20) { - layout = 8; - } else { - layout = 7; - } - } else { - layout = 9; - } - } else { - // layout is in [5-6] - if (modeBits & 0x80) { - layout = 6; - } else { - layout = 5; - } - } - } - - assert(layout < 10); - - // Determine R - u32 R = !!(modeBits & 0x10); - if (layout < 5) { - R |= (modeBits & 0x3) << 1; - } else { - R |= (modeBits & 0xC) >> 1; - } - assert(2 <= R && R <= 7); - - // Determine width & height - switch (layout) { - case 0: { - u32 A = (modeBits >> 5) & 0x3; - u32 B = (modeBits >> 7) & 0x3; - params.m_Width = B + 4; - params.m_Height = A + 2; - break; - } - - case 1: { - u32 A = (modeBits >> 5) & 0x3; - u32 B = (modeBits >> 7) & 0x3; - params.m_Width = B + 8; - params.m_Height = A + 2; - break; - } - - case 2: { - u32 A = (modeBits >> 5) & 0x3; - u32 B = (modeBits >> 7) & 0x3; - params.m_Width = A + 2; - params.m_Height = B + 8; - break; - } - - case 3: { - u32 A = (modeBits >> 5) & 0x3; - u32 B = (modeBits >> 7) & 0x1; - params.m_Width = A + 2; - params.m_Height = B + 6; - break; - } - - case 4: { - u32 A = (modeBits >> 5) & 0x3; - u32 B = (modeBits >> 7) & 0x1; - params.m_Width = B + 2; - params.m_Height = A + 2; - break; - } - - case 5: { - u32 A = (modeBits >> 5) & 0x3; - params.m_Width = 12; - params.m_Height = A + 2; - break; - } - - case 6: { - u32 A = (modeBits >> 5) & 0x3; - params.m_Width = A + 2; - params.m_Height = 12; - break; - } - - case 7: { - params.m_Width = 6; - params.m_Height = 10; - break; - } - - case 8: { - params.m_Width = 10; - params.m_Height = 6; - break; - } - - case 9: { - u32 A = (modeBits >> 5) & 0x3; - u32 B = (modeBits >> 9) & 0x3; - params.m_Width = A + 6; - params.m_Height = B + 6; - break; - } - - default: - assert(false && "Don't know this layout..."); - params.m_bError = true; - break; - } - - // Determine whether or not we're using dual planes - // and/or high precision layouts. - bool D = (layout != 9) && (modeBits & 0x400); - bool H = (layout != 9) && (modeBits & 0x200); - - if (H) { - const u32 maxWeights[6] = {9, 11, 15, 19, 23, 31}; - params.m_MaxWeight = maxWeights[R - 2]; - } else { - const u32 maxWeights[6] = {1, 2, 3, 4, 5, 7}; - params.m_MaxWeight = maxWeights[R - 2]; - } - - params.m_bDualPlane = D; - - return params; -} - -static void FillVoidExtentLDR(InputBitStream& strm, std::span<u32> outBuf, u32 blockWidth, - u32 blockHeight) { - // Don't actually care about the void extent, just read the bits... - for (s32 i = 0; i < 4; ++i) { - strm.ReadBits<13>(); - } - - // Decode the RGBA components and renormalize them to the range [0, 255] - u16 r = static_cast<u16>(strm.ReadBits<16>()); - u16 g = static_cast<u16>(strm.ReadBits<16>()); - u16 b = static_cast<u16>(strm.ReadBits<16>()); - u16 a = static_cast<u16>(strm.ReadBits<16>()); - - u32 rgba = (r >> 8) | (g & 0xFF00) | (static_cast<u32>(b) & 0xFF00) << 8 | - (static_cast<u32>(a) & 0xFF00) << 16; - - for (u32 j = 0; j < blockHeight; j++) { - for (u32 i = 0; i < blockWidth; i++) { - outBuf[j * blockWidth + i] = rgba; - } - } -} - -static void FillError(std::span<u32> outBuf, u32 blockWidth, u32 blockHeight) { - for (u32 j = 0; j < blockHeight; j++) { - for (u32 i = 0; i < blockWidth; i++) { - outBuf[j * blockWidth + i] = 0xFFFF00FF; - } - } -} - -// Replicates low numBits such that [(toBit - 1):(toBit - 1 - fromBit)] -// is the same as [(numBits - 1):0] and repeats all the way down. -template <typename IntType> -static constexpr IntType Replicate(IntType val, u32 numBits, u32 toBit) { - if (numBits == 0) { - return 0; - } - if (toBit == 0) { - return 0; - } - const IntType v = val & static_cast<IntType>((1 << numBits) - 1); - IntType res = v; - u32 reslen = numBits; - while (reslen < toBit) { - u32 comp = 0; - if (numBits > toBit - reslen) { - u32 newshift = toBit - reslen; - comp = numBits - newshift; - numBits = newshift; - } - res = static_cast<IntType>(res << numBits); - res = static_cast<IntType>(res | (v >> comp)); - reslen += numBits; - } - return res; -} - -static constexpr std::size_t NumReplicateEntries(u32 num_bits) { - return std::size_t(1) << num_bits; -} - -template <typename IntType, u32 num_bits, u32 to_bit> -static constexpr auto MakeReplicateTable() { - std::array<IntType, NumReplicateEntries(num_bits)> table{}; - for (IntType value = 0; value < static_cast<IntType>(std::size(table)); ++value) { - table[value] = Replicate(value, num_bits, to_bit); - } - return table; -} - -static constexpr auto REPLICATE_BYTE_TO_16_TABLE = MakeReplicateTable<u32, 8, 16>(); -static constexpr u32 ReplicateByteTo16(std::size_t value) { - return REPLICATE_BYTE_TO_16_TABLE[value]; -} - -static constexpr auto REPLICATE_BIT_TO_7_TABLE = MakeReplicateTable<u32, 1, 7>(); -static constexpr u32 ReplicateBitTo7(std::size_t value) { - return REPLICATE_BIT_TO_7_TABLE[value]; -} - -static constexpr auto REPLICATE_BIT_TO_9_TABLE = MakeReplicateTable<u32, 1, 9>(); -static constexpr u32 ReplicateBitTo9(std::size_t value) { - return REPLICATE_BIT_TO_9_TABLE[value]; -} - -static constexpr auto REPLICATE_1_BIT_TO_8_TABLE = MakeReplicateTable<u32, 1, 8>(); -static constexpr auto REPLICATE_2_BIT_TO_8_TABLE = MakeReplicateTable<u32, 2, 8>(); -static constexpr auto REPLICATE_3_BIT_TO_8_TABLE = MakeReplicateTable<u32, 3, 8>(); -static constexpr auto REPLICATE_4_BIT_TO_8_TABLE = MakeReplicateTable<u32, 4, 8>(); -static constexpr auto REPLICATE_5_BIT_TO_8_TABLE = MakeReplicateTable<u32, 5, 8>(); -static constexpr auto REPLICATE_6_BIT_TO_8_TABLE = MakeReplicateTable<u32, 6, 8>(); -static constexpr auto REPLICATE_7_BIT_TO_8_TABLE = MakeReplicateTable<u32, 7, 8>(); -static constexpr auto REPLICATE_8_BIT_TO_8_TABLE = MakeReplicateTable<u32, 8, 8>(); -/// Use a precompiled table with the most common usages, if it's not in the expected range, fallback -/// to the runtime implementation -static constexpr u32 FastReplicateTo8(u32 value, u32 num_bits) { - switch (num_bits) { - case 1: - return REPLICATE_1_BIT_TO_8_TABLE[value]; - case 2: - return REPLICATE_2_BIT_TO_8_TABLE[value]; - case 3: - return REPLICATE_3_BIT_TO_8_TABLE[value]; - case 4: - return REPLICATE_4_BIT_TO_8_TABLE[value]; - case 5: - return REPLICATE_5_BIT_TO_8_TABLE[value]; - case 6: - return REPLICATE_6_BIT_TO_8_TABLE[value]; - case 7: - return REPLICATE_7_BIT_TO_8_TABLE[value]; - case 8: - return REPLICATE_8_BIT_TO_8_TABLE[value]; - default: - return Replicate(value, num_bits, 8); - } -} - -static constexpr auto REPLICATE_1_BIT_TO_6_TABLE = MakeReplicateTable<u32, 1, 6>(); -static constexpr auto REPLICATE_2_BIT_TO_6_TABLE = MakeReplicateTable<u32, 2, 6>(); -static constexpr auto REPLICATE_3_BIT_TO_6_TABLE = MakeReplicateTable<u32, 3, 6>(); -static constexpr auto REPLICATE_4_BIT_TO_6_TABLE = MakeReplicateTable<u32, 4, 6>(); -static constexpr auto REPLICATE_5_BIT_TO_6_TABLE = MakeReplicateTable<u32, 5, 6>(); -static constexpr u32 FastReplicateTo6(u32 value, u32 num_bits) { - switch (num_bits) { - case 1: - return REPLICATE_1_BIT_TO_6_TABLE[value]; - case 2: - return REPLICATE_2_BIT_TO_6_TABLE[value]; - case 3: - return REPLICATE_3_BIT_TO_6_TABLE[value]; - case 4: - return REPLICATE_4_BIT_TO_6_TABLE[value]; - case 5: - return REPLICATE_5_BIT_TO_6_TABLE[value]; - default: - return Replicate(value, num_bits, 6); - } -} - -class Pixel { -protected: - using ChannelType = s16; - u8 m_BitDepth[4] = {8, 8, 8, 8}; - s16 color[4] = {}; - -public: - Pixel() = default; - Pixel(u32 a, u32 r, u32 g, u32 b, u32 bitDepth = 8) - : m_BitDepth{u8(bitDepth), u8(bitDepth), u8(bitDepth), u8(bitDepth)}, - color{static_cast<ChannelType>(a), static_cast<ChannelType>(r), - static_cast<ChannelType>(g), static_cast<ChannelType>(b)} {} - - // Changes the depth of each pixel. This scales the values to - // the appropriate bit depth by either truncating the least - // significant bits when going from larger to smaller bit depth - // or by repeating the most significant bits when going from - // smaller to larger bit depths. - void ChangeBitDepth() { - for (u32 i = 0; i < 4; i++) { - Component(i) = ChangeBitDepth(Component(i), m_BitDepth[i]); - m_BitDepth[i] = 8; - } - } - - template <typename IntType> - static float ConvertChannelToFloat(IntType channel, u8 bitDepth) { - float denominator = static_cast<float>((1 << bitDepth) - 1); - return static_cast<float>(channel) / denominator; - } - - // Changes the bit depth of a single component. See the comment - // above for how we do this. - static ChannelType ChangeBitDepth(Pixel::ChannelType val, u8 oldDepth) { - assert(oldDepth <= 8); - - if (oldDepth == 8) { - // Do nothing - return val; - } else if (oldDepth == 0) { - return static_cast<ChannelType>((1 << 8) - 1); - } else if (8 > oldDepth) { - return static_cast<ChannelType>(FastReplicateTo8(static_cast<u32>(val), oldDepth)); - } else { - // oldDepth > newDepth - const u8 bitsWasted = static_cast<u8>(oldDepth - 8); - u16 v = static_cast<u16>(val); - v = static_cast<u16>((v + (1 << (bitsWasted - 1))) >> bitsWasted); - v = ::std::min<u16>(::std::max<u16>(0, v), static_cast<u16>((1 << 8) - 1)); - return static_cast<u8>(v); - } - - assert(false && "We shouldn't get here."); - return 0; - } - - const ChannelType& A() const { - return color[0]; - } - ChannelType& A() { - return color[0]; - } - const ChannelType& R() const { - return color[1]; - } - ChannelType& R() { - return color[1]; - } - const ChannelType& G() const { - return color[2]; - } - ChannelType& G() { - return color[2]; - } - const ChannelType& B() const { - return color[3]; - } - ChannelType& B() { - return color[3]; - } - const ChannelType& Component(u32 idx) const { - return color[idx]; - } - ChannelType& Component(u32 idx) { - return color[idx]; - } - - void GetBitDepth(u8 (&outDepth)[4]) const { - for (s32 i = 0; i < 4; i++) { - outDepth[i] = m_BitDepth[i]; - } - } - - // Take all of the components, transform them to their 8-bit variants, - // and then pack each channel into an R8G8B8A8 32-bit integer. We assume - // that the architecture is little-endian, so the alpha channel will end - // up in the most-significant byte. - u32 Pack() const { - Pixel eightBit(*this); - eightBit.ChangeBitDepth(); - - u32 r = 0; - r |= eightBit.A(); - r <<= 8; - r |= eightBit.B(); - r <<= 8; - r |= eightBit.G(); - r <<= 8; - r |= eightBit.R(); - return r; - } - - // Clamps the pixel to the range [0,255] - void ClampByte() { - for (u32 i = 0; i < 4; i++) { - color[i] = (color[i] < 0) ? 0 : ((color[i] > 255) ? 255 : color[i]); - } - } - - void MakeOpaque() { - A() = 255; - } -}; - -static void DecodeColorValues(u32* out, std::span<u8> data, const u32* modes, const u32 nPartitions, - const u32 nBitsForColorData) { - // First figure out how many color values we have - u32 nValues = 0; - for (u32 i = 0; i < nPartitions; i++) { - nValues += ((modes[i] >> 2) + 1) << 1; - } - - // Then based on the number of values and the remaining number of bits, - // figure out the max value for each of them... - u32 range = 256; - while (--range > 0) { - IntegerEncodedValue val = EncodingsValues[range]; - u32 bitLength = val.GetBitLength(nValues); - if (bitLength <= nBitsForColorData) { - // Find the smallest possible range that matches the given encoding - while (--range > 0) { - IntegerEncodedValue newval = EncodingsValues[range]; - if (!newval.MatchesEncoding(val)) { - break; - } - } - - // Return to last matching range. - range++; - break; - } - } - - // We now have enough to decode our integer sequence. - IntegerEncodedVector decodedColorValues; - - InputBitStream colorStream(data, 0); - DecodeIntegerSequence(decodedColorValues, colorStream, range, nValues); - - // Once we have the decoded values, we need to dequantize them to the 0-255 range - // This procedure is outlined in ASTC spec C.2.13 - u32 outIdx = 0; - for (auto itr = decodedColorValues.begin(); itr != decodedColorValues.end(); ++itr) { - // Have we already decoded all that we need? - if (outIdx >= nValues) { - break; - } - - const IntegerEncodedValue& val = *itr; - u32 bitlen = val.num_bits; - u32 bitval = val.bit_value; - - assert(bitlen >= 1); - - u32 A = 0, B = 0, C = 0, D = 0; - // A is just the lsb replicated 9 times. - A = ReplicateBitTo9(bitval & 1); - - switch (val.encoding) { - // Replicate bits - case IntegerEncoding::JustBits: - out[outIdx++] = FastReplicateTo8(bitval, bitlen); - break; - - // Use algorithm in C.2.13 - case IntegerEncoding::Trit: { - - D = val.trit_value; - - switch (bitlen) { - case 1: { - C = 204; - } break; - - case 2: { - C = 93; - // B = b000b0bb0 - u32 b = (bitval >> 1) & 1; - B = (b << 8) | (b << 4) | (b << 2) | (b << 1); - } break; - - case 3: { - C = 44; - // B = cb000cbcb - u32 cb = (bitval >> 1) & 3; - B = (cb << 7) | (cb << 2) | cb; - } break; - - case 4: { - C = 22; - // B = dcb000dcb - u32 dcb = (bitval >> 1) & 7; - B = (dcb << 6) | dcb; - } break; - - case 5: { - C = 11; - // B = edcb000ed - u32 edcb = (bitval >> 1) & 0xF; - B = (edcb << 5) | (edcb >> 2); - } break; - - case 6: { - C = 5; - // B = fedcb000f - u32 fedcb = (bitval >> 1) & 0x1F; - B = (fedcb << 4) | (fedcb >> 4); - } break; - - default: - assert(false && "Unsupported trit encoding for color values!"); - break; - } // switch(bitlen) - } // case IntegerEncoding::Trit - break; - - case IntegerEncoding::Qus32: { - - D = val.qus32_value; - - switch (bitlen) { - case 1: { - C = 113; - } break; - - case 2: { - C = 54; - // B = b0000bb00 - u32 b = (bitval >> 1) & 1; - B = (b << 8) | (b << 3) | (b << 2); - } break; - - case 3: { - C = 26; - // B = cb0000cbc - u32 cb = (bitval >> 1) & 3; - B = (cb << 7) | (cb << 1) | (cb >> 1); - } break; - - case 4: { - C = 13; - // B = dcb0000dc - u32 dcb = (bitval >> 1) & 7; - B = (dcb << 6) | (dcb >> 1); - } break; - - case 5: { - C = 6; - // B = edcb0000e - u32 edcb = (bitval >> 1) & 0xF; - B = (edcb << 5) | (edcb >> 3); - } break; - - default: - assert(false && "Unsupported quint encoding for color values!"); - break; - } // switch(bitlen) - } // case IntegerEncoding::Qus32 - break; - } // switch(val.encoding) - - if (val.encoding != IntegerEncoding::JustBits) { - u32 T = D * C + B; - T ^= A; - T = (A & 0x80) | (T >> 2); - out[outIdx++] = T; - } - } - - // Make sure that each of our values is in the proper range... - for (u32 i = 0; i < nValues; i++) { - assert(out[i] <= 255); - } -} - -static u32 UnquantizeTexelWeight(const IntegerEncodedValue& val) { - u32 bitval = val.bit_value; - u32 bitlen = val.num_bits; - - u32 A = ReplicateBitTo7(bitval & 1); - u32 B = 0, C = 0, D = 0; - - u32 result = 0; - switch (val.encoding) { - case IntegerEncoding::JustBits: - result = FastReplicateTo6(bitval, bitlen); - break; - - case IntegerEncoding::Trit: { - D = val.trit_value; - assert(D < 3); - - switch (bitlen) { - case 0: { - u32 results[3] = {0, 32, 63}; - result = results[D]; - } break; - - case 1: { - C = 50; - } break; - - case 2: { - C = 23; - u32 b = (bitval >> 1) & 1; - B = (b << 6) | (b << 2) | b; - } break; - - case 3: { - C = 11; - u32 cb = (bitval >> 1) & 3; - B = (cb << 5) | cb; - } break; - - default: - assert(false && "Invalid trit encoding for texel weight"); - break; - } - } break; - - case IntegerEncoding::Qus32: { - D = val.qus32_value; - assert(D < 5); - - switch (bitlen) { - case 0: { - u32 results[5] = {0, 16, 32, 47, 63}; - result = results[D]; - } break; - - case 1: { - C = 28; - } break; - - case 2: { - C = 13; - u32 b = (bitval >> 1) & 1; - B = (b << 6) | (b << 1); - } break; - - default: - assert(false && "Invalid quint encoding for texel weight"); - break; - } - } break; - } - - if (val.encoding != IntegerEncoding::JustBits && bitlen > 0) { - // Decode the value... - result = D * C + B; - result ^= A; - result = (A & 0x20) | (result >> 2); - } - - assert(result < 64); - - // Change from [0,63] to [0,64] - if (result > 32) { - result += 1; - } - - return result; -} - -static void UnquantizeTexelWeights(u32 out[2][144], const IntegerEncodedVector& weights, - const TexelWeightParams& params, const u32 blockWidth, - const u32 blockHeight) { - u32 weightIdx = 0; - u32 unquantized[2][144]; - - for (auto itr = weights.begin(); itr != weights.end(); ++itr) { - unquantized[0][weightIdx] = UnquantizeTexelWeight(*itr); - - if (params.m_bDualPlane) { - ++itr; - unquantized[1][weightIdx] = UnquantizeTexelWeight(*itr); - if (itr == weights.end()) { - break; - } - } - - if (++weightIdx >= (params.m_Width * params.m_Height)) - break; - } - - // Do infill if necessary (Section C.2.18) ... - u32 Ds = (1024 + (blockWidth / 2)) / (blockWidth - 1); - u32 Dt = (1024 + (blockHeight / 2)) / (blockHeight - 1); - - const u32 kPlaneScale = params.m_bDualPlane ? 2U : 1U; - for (u32 plane = 0; plane < kPlaneScale; plane++) - for (u32 t = 0; t < blockHeight; t++) - for (u32 s = 0; s < blockWidth; s++) { - u32 cs = Ds * s; - u32 ct = Dt * t; - - u32 gs = (cs * (params.m_Width - 1) + 32) >> 6; - u32 gt = (ct * (params.m_Height - 1) + 32) >> 6; - - u32 js = gs >> 4; - u32 fs = gs & 0xF; - - u32 jt = gt >> 4; - u32 ft = gt & 0x0F; - - u32 w11 = (fs * ft + 8) >> 4; - u32 w10 = ft - w11; - u32 w01 = fs - w11; - u32 w00 = 16 - fs - ft + w11; - - u32 v0 = js + jt * params.m_Width; - -#define FIND_TEXEL(tidx, bidx) \ - u32 p##bidx = 0; \ - do { \ - if ((tidx) < (params.m_Width * params.m_Height)) { \ - p##bidx = unquantized[plane][(tidx)]; \ - } \ - } while (0) - - FIND_TEXEL(v0, 00); - FIND_TEXEL(v0 + 1, 01); - FIND_TEXEL(v0 + params.m_Width, 10); - FIND_TEXEL(v0 + params.m_Width + 1, 11); - -#undef FIND_TEXEL - - out[plane][t * blockWidth + s] = - (p00 * w00 + p01 * w01 + p10 * w10 + p11 * w11 + 8) >> 4; - } -} - -// Transfers a bit as described in C.2.14 -static inline void BitTransferSigned(s32& a, s32& b) { - b >>= 1; - b |= a & 0x80; - a >>= 1; - a &= 0x3F; - if (a & 0x20) - a -= 0x40; -} - -// Adds more precision to the blue channel as described -// in C.2.14 -static inline Pixel BlueContract(s32 a, s32 r, s32 g, s32 b) { - return Pixel(static_cast<s16>(a), static_cast<s16>((r + b) >> 1), - static_cast<s16>((g + b) >> 1), static_cast<s16>(b)); -} - -// Partition selection functions as specified in -// C.2.21 -static inline u32 hash52(u32 p) { - p ^= p >> 15; - p -= p << 17; - p += p << 7; - p += p << 4; - p ^= p >> 5; - p += p << 16; - p ^= p >> 7; - p ^= p >> 3; - p ^= p << 6; - p ^= p >> 17; - return p; -} - -static u32 SelectPartition(s32 seed, s32 x, s32 y, s32 z, s32 partitionCount, s32 smallBlock) { - if (1 == partitionCount) - return 0; - - if (smallBlock) { - x <<= 1; - y <<= 1; - z <<= 1; - } - - seed += (partitionCount - 1) * 1024; - - u32 rnum = hash52(static_cast<u32>(seed)); - u8 seed1 = static_cast<u8>(rnum & 0xF); - u8 seed2 = static_cast<u8>((rnum >> 4) & 0xF); - u8 seed3 = static_cast<u8>((rnum >> 8) & 0xF); - u8 seed4 = static_cast<u8>((rnum >> 12) & 0xF); - u8 seed5 = static_cast<u8>((rnum >> 16) & 0xF); - u8 seed6 = static_cast<u8>((rnum >> 20) & 0xF); - u8 seed7 = static_cast<u8>((rnum >> 24) & 0xF); - u8 seed8 = static_cast<u8>((rnum >> 28) & 0xF); - u8 seed9 = static_cast<u8>((rnum >> 18) & 0xF); - u8 seed10 = static_cast<u8>((rnum >> 22) & 0xF); - u8 seed11 = static_cast<u8>((rnum >> 26) & 0xF); - u8 seed12 = static_cast<u8>(((rnum >> 30) | (rnum << 2)) & 0xF); - - seed1 = static_cast<u8>(seed1 * seed1); - seed2 = static_cast<u8>(seed2 * seed2); - seed3 = static_cast<u8>(seed3 * seed3); - seed4 = static_cast<u8>(seed4 * seed4); - seed5 = static_cast<u8>(seed5 * seed5); - seed6 = static_cast<u8>(seed6 * seed6); - seed7 = static_cast<u8>(seed7 * seed7); - seed8 = static_cast<u8>(seed8 * seed8); - seed9 = static_cast<u8>(seed9 * seed9); - seed10 = static_cast<u8>(seed10 * seed10); - seed11 = static_cast<u8>(seed11 * seed11); - seed12 = static_cast<u8>(seed12 * seed12); - - s32 sh1, sh2, sh3; - if (seed & 1) { - sh1 = (seed & 2) ? 4 : 5; - sh2 = (partitionCount == 3) ? 6 : 5; - } else { - sh1 = (partitionCount == 3) ? 6 : 5; - sh2 = (seed & 2) ? 4 : 5; - } - sh3 = (seed & 0x10) ? sh1 : sh2; - - seed1 = static_cast<u8>(seed1 >> sh1); - seed2 = static_cast<u8>(seed2 >> sh2); - seed3 = static_cast<u8>(seed3 >> sh1); - seed4 = static_cast<u8>(seed4 >> sh2); - seed5 = static_cast<u8>(seed5 >> sh1); - seed6 = static_cast<u8>(seed6 >> sh2); - seed7 = static_cast<u8>(seed7 >> sh1); - seed8 = static_cast<u8>(seed8 >> sh2); - seed9 = static_cast<u8>(seed9 >> sh3); - seed10 = static_cast<u8>(seed10 >> sh3); - seed11 = static_cast<u8>(seed11 >> sh3); - seed12 = static_cast<u8>(seed12 >> sh3); - - s32 a = seed1 * x + seed2 * y + seed11 * z + (rnum >> 14); - s32 b = seed3 * x + seed4 * y + seed12 * z + (rnum >> 10); - s32 c = seed5 * x + seed6 * y + seed9 * z + (rnum >> 6); - s32 d = seed7 * x + seed8 * y + seed10 * z + (rnum >> 2); - - a &= 0x3F; - b &= 0x3F; - c &= 0x3F; - d &= 0x3F; - - if (partitionCount < 4) - d = 0; - if (partitionCount < 3) - c = 0; - - if (a >= b && a >= c && a >= d) - return 0; - else if (b >= c && b >= d) - return 1; - else if (c >= d) - return 2; - return 3; -} - -static inline u32 Select2DPartition(s32 seed, s32 x, s32 y, s32 partitionCount, s32 smallBlock) { - return SelectPartition(seed, x, y, 0, partitionCount, smallBlock); -} - -// Section C.2.14 -static void ComputeEndpos32s(Pixel& ep1, Pixel& ep2, const u32*& colorValues, - u32 colorEndpos32Mode) { -#define READ_UINT_VALUES(N) \ - u32 v[N]; \ - for (u32 i = 0; i < N; i++) { \ - v[i] = *(colorValues++); \ - } - -#define READ_INT_VALUES(N) \ - s32 v[N]; \ - for (u32 i = 0; i < N; i++) { \ - v[i] = static_cast<s32>(*(colorValues++)); \ - } - - switch (colorEndpos32Mode) { - case 0: { - READ_UINT_VALUES(2) - ep1 = Pixel(0xFF, v[0], v[0], v[0]); - ep2 = Pixel(0xFF, v[1], v[1], v[1]); - } break; - - case 1: { - READ_UINT_VALUES(2) - u32 L0 = (v[0] >> 2) | (v[1] & 0xC0); - u32 L1 = std::max(L0 + (v[1] & 0x3F), 0xFFU); - ep1 = Pixel(0xFF, L0, L0, L0); - ep2 = Pixel(0xFF, L1, L1, L1); - } break; - - case 4: { - READ_UINT_VALUES(4) - ep1 = Pixel(v[2], v[0], v[0], v[0]); - ep2 = Pixel(v[3], v[1], v[1], v[1]); - } break; - - case 5: { - READ_INT_VALUES(4) - BitTransferSigned(v[1], v[0]); - BitTransferSigned(v[3], v[2]); - ep1 = Pixel(v[2], v[0], v[0], v[0]); - ep2 = Pixel(v[2] + v[3], v[0] + v[1], v[0] + v[1], v[0] + v[1]); - ep1.ClampByte(); - ep2.ClampByte(); - } break; - - case 6: { - READ_UINT_VALUES(4) - ep1 = Pixel(0xFF, v[0] * v[3] >> 8, v[1] * v[3] >> 8, v[2] * v[3] >> 8); - ep2 = Pixel(0xFF, v[0], v[1], v[2]); - } break; - - case 8: { - READ_UINT_VALUES(6) - if (v[1] + v[3] + v[5] >= v[0] + v[2] + v[4]) { - ep1 = Pixel(0xFF, v[0], v[2], v[4]); - ep2 = Pixel(0xFF, v[1], v[3], v[5]); - } else { - ep1 = BlueContract(0xFF, v[1], v[3], v[5]); - ep2 = BlueContract(0xFF, v[0], v[2], v[4]); - } - } break; - - case 9: { - READ_INT_VALUES(6) - BitTransferSigned(v[1], v[0]); - BitTransferSigned(v[3], v[2]); - BitTransferSigned(v[5], v[4]); - if (v[1] + v[3] + v[5] >= 0) { - ep1 = Pixel(0xFF, v[0], v[2], v[4]); - ep2 = Pixel(0xFF, v[0] + v[1], v[2] + v[3], v[4] + v[5]); - } else { - ep1 = BlueContract(0xFF, v[0] + v[1], v[2] + v[3], v[4] + v[5]); - ep2 = BlueContract(0xFF, v[0], v[2], v[4]); - } - ep1.ClampByte(); - ep2.ClampByte(); - } break; - - case 10: { - READ_UINT_VALUES(6) - ep1 = Pixel(v[4], v[0] * v[3] >> 8, v[1] * v[3] >> 8, v[2] * v[3] >> 8); - ep2 = Pixel(v[5], v[0], v[1], v[2]); - } break; - - case 12: { - READ_UINT_VALUES(8) - if (v[1] + v[3] + v[5] >= v[0] + v[2] + v[4]) { - ep1 = Pixel(v[6], v[0], v[2], v[4]); - ep2 = Pixel(v[7], v[1], v[3], v[5]); - } else { - ep1 = BlueContract(v[7], v[1], v[3], v[5]); - ep2 = BlueContract(v[6], v[0], v[2], v[4]); - } - } break; - - case 13: { - READ_INT_VALUES(8) - BitTransferSigned(v[1], v[0]); - BitTransferSigned(v[3], v[2]); - BitTransferSigned(v[5], v[4]); - BitTransferSigned(v[7], v[6]); - if (v[1] + v[3] + v[5] >= 0) { - ep1 = Pixel(v[6], v[0], v[2], v[4]); - ep2 = Pixel(v[7] + v[6], v[0] + v[1], v[2] + v[3], v[4] + v[5]); - } else { - ep1 = BlueContract(v[6] + v[7], v[0] + v[1], v[2] + v[3], v[4] + v[5]); - ep2 = BlueContract(v[6], v[0], v[2], v[4]); - } - ep1.ClampByte(); - ep2.ClampByte(); - } break; - - default: - assert(false && "Unsupported color endpoint mode (is it HDR?)"); - break; - } - -#undef READ_UINT_VALUES -#undef READ_INT_VALUES -} - -static void DecompressBlock(std::span<const u8, 16> inBuf, const u32 blockWidth, - const u32 blockHeight, std::span<u32, 12 * 12> outBuf) { - InputBitStream strm(inBuf); - TexelWeightParams weightParams = DecodeBlockInfo(strm); - - // Was there an error? - if (weightParams.m_bError) { - assert(false && "Invalid block mode"); - FillError(outBuf, blockWidth, blockHeight); - return; - } - - if (weightParams.m_bVoidExtentLDR) { - FillVoidExtentLDR(strm, outBuf, blockWidth, blockHeight); - return; - } - - if (weightParams.m_bVoidExtentHDR) { - assert(false && "HDR void extent blocks are unsupported!"); - FillError(outBuf, blockWidth, blockHeight); - return; - } - - if (weightParams.m_Width > blockWidth) { - assert(false && "Texel weight grid width should be smaller than block width"); - FillError(outBuf, blockWidth, blockHeight); - return; - } - - if (weightParams.m_Height > blockHeight) { - assert(false && "Texel weight grid height should be smaller than block height"); - FillError(outBuf, blockWidth, blockHeight); - return; - } - - // Read num partitions - u32 nPartitions = strm.ReadBits<2>() + 1; - assert(nPartitions <= 4); - - if (nPartitions == 4 && weightParams.m_bDualPlane) { - assert(false && "Dual plane mode is incompatible with four partition blocks"); - FillError(outBuf, blockWidth, blockHeight); - return; - } - - // Based on the number of partitions, read the color endpos32 mode for - // each partition. - - // Determine partitions, partition index, and color endpos32 modes - s32 planeIdx = -1; - u32 partitionIndex; - u32 colorEndpos32Mode[4] = {0, 0, 0, 0}; - - // Define color data. - u8 colorEndpos32Data[16]; - memset(colorEndpos32Data, 0, sizeof(colorEndpos32Data)); - OutputBitStream colorEndpos32Stream(colorEndpos32Data, 16 * 8, 0); - - // Read extra config data... - u32 baseCEM = 0; - if (nPartitions == 1) { - colorEndpos32Mode[0] = strm.ReadBits<4>(); - partitionIndex = 0; - } else { - partitionIndex = strm.ReadBits<10>(); - baseCEM = strm.ReadBits<6>(); - } - u32 baseMode = (baseCEM & 3); - - // Remaining bits are color endpos32 data... - u32 nWeightBits = weightParams.GetPackedBitSize(); - s32 remainingBits = 128 - nWeightBits - static_cast<s32>(strm.GetBitsRead()); - - // Consider extra bits prior to texel data... - u32 extraCEMbits = 0; - if (baseMode) { - switch (nPartitions) { - case 2: - extraCEMbits += 2; - break; - case 3: - extraCEMbits += 5; - break; - case 4: - extraCEMbits += 8; - break; - default: - assert(false); - break; - } - } - remainingBits -= extraCEMbits; - - // Do we have a dual plane situation? - u32 planeSelectorBits = 0; - if (weightParams.m_bDualPlane) { - planeSelectorBits = 2; - } - remainingBits -= planeSelectorBits; - - // Read color data... - u32 colorDataBits = remainingBits; - while (remainingBits > 0) { - u32 nb = std::min(remainingBits, 8); - u32 b = strm.ReadBits(nb); - colorEndpos32Stream.WriteBits(b, nb); - remainingBits -= 8; - } - - // Read the plane selection bits - planeIdx = strm.ReadBits(planeSelectorBits); - - // Read the rest of the CEM - if (baseMode) { - u32 extraCEM = strm.ReadBits(extraCEMbits); - u32 CEM = (extraCEM << 6) | baseCEM; - CEM >>= 2; - - bool C[4] = {0}; - for (u32 i = 0; i < nPartitions; i++) { - C[i] = CEM & 1; - CEM >>= 1; - } - - u8 M[4] = {0}; - for (u32 i = 0; i < nPartitions; i++) { - M[i] = CEM & 3; - CEM >>= 2; - assert(M[i] <= 3); - } - - for (u32 i = 0; i < nPartitions; i++) { - colorEndpos32Mode[i] = baseMode; - if (!(C[i])) - colorEndpos32Mode[i] -= 1; - colorEndpos32Mode[i] <<= 2; - colorEndpos32Mode[i] |= M[i]; - } - } else if (nPartitions > 1) { - u32 CEM = baseCEM >> 2; - for (u32 i = 0; i < nPartitions; i++) { - colorEndpos32Mode[i] = CEM; - } - } - - // Make sure everything up till here is sane. - for (u32 i = 0; i < nPartitions; i++) { - assert(colorEndpos32Mode[i] < 16); - } - assert(strm.GetBitsRead() + weightParams.GetPackedBitSize() == 128); - - // Decode both color data and texel weight data - u32 colorValues[32]; // Four values, two endpos32s, four maximum paritions - DecodeColorValues(colorValues, colorEndpos32Data, colorEndpos32Mode, nPartitions, - colorDataBits); - - Pixel endpos32s[4][2]; - const u32* colorValuesPtr = colorValues; - for (u32 i = 0; i < nPartitions; i++) { - ComputeEndpos32s(endpos32s[i][0], endpos32s[i][1], colorValuesPtr, colorEndpos32Mode[i]); - } - - // Read the texel weight data.. - std::array<u8, 16> texelWeightData; - std::ranges::copy(inBuf, texelWeightData.begin()); - - // Reverse everything - for (u32 i = 0; i < 8; i++) { -// Taken from http://graphics.stanford.edu/~seander/bithacks.html#ReverseByteWith64Bits -#define REVERSE_BYTE(b) (((b)*0x80200802ULL) & 0x0884422110ULL) * 0x0101010101ULL >> 32 - u8 a = static_cast<u8>(REVERSE_BYTE(texelWeightData[i])); - u8 b = static_cast<u8>(REVERSE_BYTE(texelWeightData[15 - i])); -#undef REVERSE_BYTE - - texelWeightData[i] = b; - texelWeightData[15 - i] = a; - } - - // Make sure that higher non-texel bits are set to zero - const u32 clearByteStart = (weightParams.GetPackedBitSize() >> 3) + 1; - if (clearByteStart > 0 && clearByteStart <= texelWeightData.size()) { - texelWeightData[clearByteStart - 1] &= - static_cast<u8>((1 << (weightParams.GetPackedBitSize() % 8)) - 1); - std::memset(texelWeightData.data() + clearByteStart, 0, - std::min(16U - clearByteStart, 16U)); - } - - IntegerEncodedVector texelWeightValues; - - InputBitStream weightStream(texelWeightData); - - DecodeIntegerSequence(texelWeightValues, weightStream, weightParams.m_MaxWeight, - weightParams.GetNumWeightValues()); - - // Blocks can be at most 12x12, so we can have as many as 144 weights - u32 weights[2][144]; - UnquantizeTexelWeights(weights, texelWeightValues, weightParams, blockWidth, blockHeight); - - // Now that we have endpos32s and weights, we can s32erpolate and generate - // the proper decoding... - for (u32 j = 0; j < blockHeight; j++) - for (u32 i = 0; i < blockWidth; i++) { - u32 partition = Select2DPartition(partitionIndex, i, j, nPartitions, - (blockHeight * blockWidth) < 32); - assert(partition < nPartitions); - - Pixel p; - for (u32 c = 0; c < 4; c++) { - u32 C0 = endpos32s[partition][0].Component(c); - C0 = ReplicateByteTo16(C0); - u32 C1 = endpos32s[partition][1].Component(c); - C1 = ReplicateByteTo16(C1); - - u32 plane = 0; - if (weightParams.m_bDualPlane && (((planeIdx + 1) & 3) == c)) { - plane = 1; - } - - u32 weight = weights[plane][j * blockWidth + i]; - u32 C = (C0 * (64 - weight) + C1 * weight + 32) / 64; - if (C == 65535) { - p.Component(c) = 255; - } else { - double Cf = static_cast<double>(C); - p.Component(c) = static_cast<u16>(255.0 * (Cf / 65536.0) + 0.5); - } - } - - outBuf[j * blockWidth + i] = p.Pack(); - } -} - -} // namespace ASTCC - -namespace Tegra::Texture::ASTC { - -void Decompress(std::span<const uint8_t> data, uint32_t width, uint32_t height, uint32_t depth, - uint32_t block_width, uint32_t block_height, std::span<uint8_t> output) { - u32 block_index = 0; - std::size_t depth_offset = 0; - for (u32 z = 0; z < depth; z++) { - for (u32 y = 0; y < height; y += block_height) { - for (u32 x = 0; x < width; x += block_width) { - const std::span<const u8, 16> blockPtr{data.subspan(block_index * 16, 16)}; - - // Blocks can be at most 12x12 - std::array<u32, 12 * 12> uncompData; - ASTCC::DecompressBlock(blockPtr, block_width, block_height, uncompData); - - u32 decompWidth = std::min(block_width, width - x); - u32 decompHeight = std::min(block_height, height - y); - - const std::span<u8> outRow = output.subspan(depth_offset + (y * width + x) * 4); - for (u32 jj = 0; jj < decompHeight; jj++) { - std::memcpy(outRow.data() + jj * width * 4, - uncompData.data() + jj * block_width, decompWidth * 4); - } - ++block_index; - } - } - depth_offset += height * width * 4; - } -} - -} // namespace Tegra::Texture::ASTC diff --git a/src/video_core/textures/astc.h b/src/video_core/textures/astc.h index 9105119bc..c1c73fda5 100644 --- a/src/video_core/textures/astc.h +++ b/src/video_core/textures/astc.h @@ -4,11 +4,129 @@ #pragma once -#include <cstdint> +#include <bit> +#include "common/common_types.h" namespace Tegra::Texture::ASTC { -void Decompress(std::span<const uint8_t> data, uint32_t width, uint32_t height, uint32_t depth, - uint32_t block_width, uint32_t block_height, std::span<uint8_t> output); +enum class IntegerEncoding { JustBits, Quint, Trit }; + +struct IntegerEncodedValue { + constexpr IntegerEncodedValue() = default; + + constexpr IntegerEncodedValue(IntegerEncoding encoding_, u32 num_bits_) + : encoding{encoding_}, num_bits{num_bits_} {} + + constexpr bool MatchesEncoding(const IntegerEncodedValue& other) const { + return encoding == other.encoding && num_bits == other.num_bits; + } + + // Returns the number of bits required to encode num_vals values. + u32 GetBitLength(u32 num_vals) const { + u32 total_bits = num_bits * num_vals; + if (encoding == IntegerEncoding::Trit) { + total_bits += (num_vals * 8 + 4) / 5; + } else if (encoding == IntegerEncoding::Quint) { + total_bits += (num_vals * 7 + 2) / 3; + } + return total_bits; + } + + IntegerEncoding encoding{}; + u32 num_bits = 0; + u32 bit_value = 0; + union { + u32 quint_value = 0; + u32 trit_value; + }; +}; + +// Returns a new instance of this struct that corresponds to the +// can take no more than mav_value values +constexpr IntegerEncodedValue CreateEncoding(u32 mav_value) { + while (mav_value > 0) { + u32 check = mav_value + 1; + + // Is mav_value a power of two? + if (!(check & (check - 1))) { + return IntegerEncodedValue(IntegerEncoding::JustBits, std::popcount(mav_value)); + } + + // Is mav_value of the type 3*2^n - 1? + if ((check % 3 == 0) && !((check / 3) & ((check / 3) - 1))) { + return IntegerEncodedValue(IntegerEncoding::Trit, std::popcount(check / 3 - 1)); + } + + // Is mav_value of the type 5*2^n - 1? + if ((check % 5 == 0) && !((check / 5) & ((check / 5) - 1))) { + return IntegerEncodedValue(IntegerEncoding::Quint, std::popcount(check / 5 - 1)); + } + + // Apparently it can't be represented with a bounded integer sequence... + // just iterate. + mav_value--; + } + return IntegerEncodedValue(IntegerEncoding::JustBits, 0); +} + +constexpr std::array<IntegerEncodedValue, 256> MakeEncodedValues() { + std::array<IntegerEncodedValue, 256> encodings{}; + for (std::size_t i = 0; i < encodings.size(); ++i) { + encodings[i] = CreateEncoding(static_cast<u32>(i)); + } + return encodings; +} + +constexpr std::array<IntegerEncodedValue, 256> EncodingsValues = MakeEncodedValues(); + +// Replicates low num_bits such that [(to_bit - 1):(to_bit - 1 - from_bit)] +// is the same as [(num_bits - 1):0] and repeats all the way down. +template <typename IntType> +constexpr IntType Replicate(IntType val, u32 num_bits, u32 to_bit) { + if (num_bits == 0 || to_bit == 0) { + return 0; + } + const IntType v = val & static_cast<IntType>((1 << num_bits) - 1); + IntType res = v; + u32 reslen = num_bits; + while (reslen < to_bit) { + u32 comp = 0; + if (num_bits > to_bit - reslen) { + u32 newshift = to_bit - reslen; + comp = num_bits - newshift; + num_bits = newshift; + } + res = static_cast<IntType>(res << num_bits); + res = static_cast<IntType>(res | (v >> comp)); + reslen += num_bits; + } + return res; +} + +constexpr std::size_t NumReplicateEntries(u32 num_bits) { + return std::size_t(1) << num_bits; +} + +template <typename IntType, u32 num_bits, u32 to_bit> +constexpr auto MakeReplicateTable() { + std::array<IntType, NumReplicateEntries(num_bits)> table{}; + for (IntType value = 0; value < static_cast<IntType>(std::size(table)); ++value) { + table[value] = Replicate(value, num_bits, to_bit); + } + return table; +} + +constexpr auto REPLICATE_BYTE_TO_16_TABLE = MakeReplicateTable<u32, 8, 16>(); +constexpr auto REPLICATE_6_BIT_TO_8_TABLE = MakeReplicateTable<u32, 6, 8>(); +constexpr auto REPLICATE_7_BIT_TO_8_TABLE = MakeReplicateTable<u32, 7, 8>(); +constexpr auto REPLICATE_8_BIT_TO_8_TABLE = MakeReplicateTable<u32, 8, 8>(); + +struct AstcBufferData { + decltype(EncodingsValues) encoding_values = EncodingsValues; + decltype(REPLICATE_6_BIT_TO_8_TABLE) replicate_6_to_8 = REPLICATE_6_BIT_TO_8_TABLE; + decltype(REPLICATE_7_BIT_TO_8_TABLE) replicate_7_to_8 = REPLICATE_7_BIT_TO_8_TABLE; + decltype(REPLICATE_8_BIT_TO_8_TABLE) replicate_8_to_8 = REPLICATE_8_BIT_TO_8_TABLE; + decltype(REPLICATE_BYTE_TO_16_TABLE) replicate_byte_to_16 = REPLICATE_BYTE_TO_16_TABLE; +} constexpr ASTC_BUFFER_DATA; } // namespace Tegra::Texture::ASTC diff --git a/src/video_core/textures/decoders.cpp b/src/video_core/textures/decoders.cpp index 62685a183..3a463d5db 100644 --- a/src/video_core/textures/decoders.cpp +++ b/src/video_core/textures/decoders.cpp @@ -17,26 +17,7 @@ #include "video_core/textures/texture.h" namespace Tegra::Texture { - namespace { -/** - * This table represents the internal swizzle of a gob, in format 16 bytes x 2 sector packing. - * Calculates the offset of an (x, y) position within a swizzled texture. - * Taken from the Tegra X1 Technical Reference Manual. pages 1187-1188 - */ -constexpr SwizzleTable MakeSwizzleTableConst() { - SwizzleTable table{}; - for (u32 y = 0; y < table.size(); ++y) { - for (u32 x = 0; x < table[0].size(); ++x) { - table[y][x] = ((x % 64) / 32) * 256 + ((y % 8) / 2) * 64 + ((x % 32) / 16) * 32 + - (y % 2) * 16 + (x % 16); - } - } - return table; -} - -constexpr SwizzleTable SWIZZLE_TABLE = MakeSwizzleTableConst(); - template <bool TO_LINEAR> void Swizzle(std::span<u8> output, std::span<const u8> input, u32 bytes_per_pixel, u32 width, u32 height, u32 depth, u32 block_height, u32 block_depth, u32 stride_alignment) { @@ -91,10 +72,6 @@ void Swizzle(std::span<u8> output, std::span<const u8> input, u32 bytes_per_pixe } } // Anonymous namespace -SwizzleTable MakeSwizzleTable() { - return SWIZZLE_TABLE; -} - void UnswizzleTexture(std::span<u8> output, std::span<const u8> input, u32 bytes_per_pixel, u32 width, u32 height, u32 depth, u32 block_height, u32 block_depth, u32 stride_alignment) { diff --git a/src/video_core/textures/decoders.h b/src/video_core/textures/decoders.h index d7cdc81e8..4c14cefbf 100644 --- a/src/video_core/textures/decoders.h +++ b/src/video_core/textures/decoders.h @@ -23,8 +23,22 @@ constexpr u32 GOB_SIZE_SHIFT = GOB_SIZE_X_SHIFT + GOB_SIZE_Y_SHIFT + GOB_SIZE_Z_ using SwizzleTable = std::array<std::array<u32, GOB_SIZE_X>, GOB_SIZE_Y>; -/// Returns a z-order swizzle table -SwizzleTable MakeSwizzleTable(); +/** + * This table represents the internal swizzle of a gob, in format 16 bytes x 2 sector packing. + * Calculates the offset of an (x, y) position within a swizzled texture. + * Taken from the Tegra X1 Technical Reference Manual. pages 1187-1188 + */ +constexpr SwizzleTable MakeSwizzleTable() { + SwizzleTable table{}; + for (u32 y = 0; y < table.size(); ++y) { + for (u32 x = 0; x < table[0].size(); ++x) { + table[y][x] = ((x % 64) / 32) * 256 + ((y % 8) / 2) * 64 + ((x % 32) / 16) * 32 + + (y % 2) * 16 + (x % 16); + } + } + return table; +} +constexpr SwizzleTable SWIZZLE_TABLE = MakeSwizzleTable(); /// Unswizzles a block linear texture into linear memory. void UnswizzleTexture(std::span<u8> output, std::span<const u8> input, u32 bytes_per_pixel, diff --git a/src/video_core/textures/texture.cpp b/src/video_core/textures/texture.cpp index ae5621a7d..a552543ed 100644 --- a/src/video_core/textures/texture.cpp +++ b/src/video_core/textures/texture.cpp @@ -6,7 +6,7 @@ #include <array> #include "common/cityhash.h" -#include "core/settings.h" +#include "common/settings.h" #include "video_core/textures/texture.h" using Tegra::Texture::TICEntry; diff --git a/src/video_core/video_core.cpp b/src/video_core/video_core.cpp index e1b38c6ac..3b575db4d 100644 --- a/src/video_core/video_core.cpp +++ b/src/video_core/video_core.cpp @@ -5,8 +5,8 @@ #include <memory> #include "common/logging/log.h" +#include "common/settings.h" #include "core/core.h" -#include "core/settings.h" #include "video_core/renderer_base.h" #include "video_core/renderer_opengl/renderer_opengl.h" #include "video_core/renderer_vulkan/renderer_vulkan.h" diff --git a/src/video_core/vulkan_common/vulkan_device.cpp b/src/video_core/vulkan_common/vulkan_device.cpp index 34d396434..64206b3d2 100644 --- a/src/video_core/vulkan_common/vulkan_device.cpp +++ b/src/video_core/vulkan_common/vulkan_device.cpp @@ -12,7 +12,7 @@ #include <vector> #include "common/assert.h" -#include "core/settings.h" +#include "common/settings.h" #include "video_core/vulkan_common/nsight_aftermath_tracker.h" #include "video_core/vulkan_common/vulkan_device.h" #include "video_core/vulkan_common/vulkan_wrapper.h" @@ -51,7 +51,7 @@ constexpr std::array REQUIRED_EXTENSIONS{ #ifdef _WIN32 VK_KHR_EXTERNAL_MEMORY_WIN32_EXTENSION_NAME, #endif -#ifdef __linux__ +#ifdef __unix__ VK_KHR_EXTERNAL_MEMORY_FD_EXTENSION_NAME, #endif }; @@ -294,6 +294,15 @@ Device::Device(VkInstance instance_, vk::PhysicalDevice physical_, VkSurfaceKHR }; SetNext(next, bit8_storage); + VkPhysicalDeviceRobustness2FeaturesEXT robustness2{ + .sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_ROBUSTNESS_2_FEATURES_EXT, + .pNext = nullptr, + .robustBufferAccess2 = true, + .robustImageAccess2 = true, + .nullDescriptor = true, + }; + SetNext(next, robustness2); + VkPhysicalDeviceHostQueryResetFeaturesEXT host_query_reset{ .sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_HOST_QUERY_RESET_FEATURES_EXT, .pNext = nullptr, diff --git a/src/video_core/vulkan_common/vulkan_memory_allocator.cpp b/src/video_core/vulkan_common/vulkan_memory_allocator.cpp index 2a8b7a907..fa37aa79a 100644 --- a/src/video_core/vulkan_common/vulkan_memory_allocator.cpp +++ b/src/video_core/vulkan_common/vulkan_memory_allocator.cpp @@ -62,7 +62,7 @@ public: : memory{std::move(memory_)}, allocation_size{allocation_size_}, property_flags{properties}, shifted_memory_type{1U << type} {} -#if defined(_WIN32) || defined(__linux__) +#if defined(_WIN32) || defined(__unix__) ~MemoryAllocation() { if (owning_opengl_handle != 0) { glDeleteMemoryObjectsEXT(1, &owning_opengl_handle); @@ -114,7 +114,7 @@ public: } return owning_opengl_handle; } -#elif __linux__ +#elif __unix__ [[nodiscard]] u32 ExportOpenGLHandle() { if (!owning_opengl_handle) { glCreateMemoryObjectsEXT(1, &owning_opengl_handle); @@ -165,7 +165,7 @@ private: const u32 shifted_memory_type; ///< Shifted Vulkan memory type. std::vector<Range> commits; ///< All commit ranges done from this allocation. std::span<u8> memory_mapped_span; ///< Memory mapped span. Empty if not queried before. -#if defined(_WIN32) || defined(__linux__) +#if defined(_WIN32) || defined(__unix__) u32 owning_opengl_handle{}; ///< Owning OpenGL memory object handle. #endif }; @@ -249,7 +249,7 @@ void MemoryAllocator::AllocMemory(VkMemoryPropertyFlags flags, u32 type_mask, u6 .pNext = nullptr, #ifdef _WIN32 .handleTypes = VK_EXTERNAL_MEMORY_HANDLE_TYPE_OPAQUE_WIN32_BIT, -#elif __linux__ +#elif __unix__ .handleTypes = VK_EXTERNAL_MEMORY_HANDLE_TYPE_OPAQUE_FD_BIT, #else .handleTypes = 0, |