yuzu/src/video_core/shader/async_shaders.cpp

222 lines
7.6 KiB
C++
Raw Normal View History

2020-07-10 03:36:38 +00:00
// Copyright 2020 yuzu Emulator Project
// Licensed under GPLv2 or any later version
// Refer to the license.txt file included.
#include <chrono>
2020-07-18 04:24:32 +00:00
#include <condition_variable>
#include <mutex>
#include <thread>
#include <vector>
2020-07-10 03:36:38 +00:00
#include "video_core/engines/maxwell_3d.h"
#include "video_core/renderer_base.h"
#include "video_core/renderer_opengl/gl_shader_cache.h"
#include "video_core/shader/async_shaders.h"
namespace VideoCommon::Shader {
2020-07-18 04:24:32 +00:00
2020-07-10 03:36:38 +00:00
AsyncShaders::AsyncShaders(Core::Frontend::EmuWindow& emu_window) : emu_window(emu_window) {}
2020-07-18 04:24:32 +00:00
2020-07-10 03:36:38 +00:00
AsyncShaders::~AsyncShaders() {
KillWorkers();
}
void AsyncShaders::AllocateWorkers(std::size_t num_workers) {
// If we're already have workers queued or don't want to queue workers, ignore
if (num_workers == worker_threads.size() || num_workers == 0) {
return;
}
// If workers already exist, clear them
if (!worker_threads.empty()) {
FreeWorkers();
}
// Create workers
for (std::size_t i = 0; i < num_workers; i++) {
context_list.push_back(emu_window.CreateSharedContext());
worker_threads.push_back(std::move(
std::thread(&AsyncShaders::ShaderCompilerThread, this, context_list[i].get())));
}
}
void AsyncShaders::FreeWorkers() {
// Mark all threads to quit
is_thread_exiting.store(true);
2020-07-16 08:51:32 +00:00
cv.notify_all();
2020-07-10 03:36:38 +00:00
for (auto& thread : worker_threads) {
thread.join();
}
// Clear our shared contexts
context_list.clear();
// Clear our worker threads
worker_threads.clear();
}
void AsyncShaders::KillWorkers() {
is_thread_exiting.store(true);
for (auto& thread : worker_threads) {
thread.detach();
}
// Clear our shared contexts
context_list.clear();
// Clear our worker threads
worker_threads.clear();
}
bool AsyncShaders::HasWorkQueued() {
return !pending_queue.empty();
}
bool AsyncShaders::HasCompletedWork() {
2020-07-18 04:24:32 +00:00
std::shared_lock lock{completed_mutex};
2020-07-10 03:36:38 +00:00
return !finished_work.empty();
}
bool AsyncShaders::IsShaderAsync(const Tegra::GPU& gpu) const {
const auto& regs = gpu.Maxwell3D().regs;
// If something is using depth, we can assume that games are not rendering anything which will
// be used one time.
if (regs.zeta_enable) {
return true;
}
// If games are using a small index count, we can assume these are full screen quads. Usually
// these shaders are only used once for building textures so we can assume they can't be built
// async
if (regs.index_array.count <= 6 || regs.vertex_buffer.count <= 6) {
return false;
}
return true;
}
std::vector<AsyncShaders::Result> AsyncShaders::GetCompletedWork() {
std::vector<AsyncShaders::Result> results;
{
2020-07-18 04:24:32 +00:00
std::unique_lock lock{completed_mutex};
2020-07-10 03:36:38 +00:00
results.assign(std::make_move_iterator(finished_work.begin()),
std::make_move_iterator(finished_work.end()));
finished_work.clear();
}
return results;
}
void AsyncShaders::QueueOpenGLShader(const OpenGL::Device& device,
Tegra::Engines::ShaderType shader_type, u64 uid,
std::vector<u64> code, std::vector<u64> code_b,
u32 main_offset,
VideoCommon::Shader::CompilerSettings compiler_settings,
const VideoCommon::Shader::Registry& registry,
VAddr cpu_addr) {
auto params = std::make_unique<WorkerParams>();
params->backend = device.UseAssemblyShaders() ? Backend::GLASM : Backend::OpenGL;
params->device = &device;
params->shader_type = shader_type;
params->uid = uid;
params->code = std::move(code);
params->code_b = std::move(code_b);
params->main_offset = main_offset;
params->compiler_settings = compiler_settings;
params->registry = &registry;
params->cpu_address = cpu_addr;
2020-07-28 04:08:02 +00:00
std::unique_lock lock(queue_mutex);
pending_queue.push(std::move(params));
2020-07-28 04:08:02 +00:00
cv.notify_one();
}
void AsyncShaders::QueueVulkanShader(
Vulkan::VKPipelineCache* pp_cache, std::vector<VkDescriptorSetLayoutBinding> bindings,
Vulkan::SPIRVProgram program, Vulkan::RenderPassParams renderpass_params, u32 padding,
std::array<GPUVAddr, Vulkan::Maxwell::MaxShaderProgram> shaders,
Vulkan::FixedPipelineState fixed_state) {
auto params = std::make_unique<WorkerParams>();
params->backend = Backend::Vulkan;
params->pp_cache = pp_cache;
params->bindings = bindings;
params->program = program;
params->renderpass_params = renderpass_params;
params->padding = padding;
params->shaders = shaders;
params->fixed_state = fixed_state;
2020-07-28 04:08:02 +00:00
2020-07-10 03:36:38 +00:00
std::unique_lock lock(queue_mutex);
pending_queue.push(std::move(params));
2020-07-16 08:38:35 +00:00
cv.notify_one();
2020-07-10 03:36:38 +00:00
}
void AsyncShaders::ShaderCompilerThread(Core::Frontend::GraphicsContext* context) {
using namespace std::chrono_literals;
while (!is_thread_exiting.load(std::memory_order_relaxed)) {
2020-07-18 04:24:32 +00:00
std::unique_lock lock{queue_mutex};
cv.wait(lock, [this] { return HasWorkQueued() || is_thread_exiting; });
2020-07-16 08:38:35 +00:00
if (is_thread_exiting) {
return;
}
2020-07-10 03:36:38 +00:00
// Partial lock to allow all threads to read at the same time
if (!HasWorkQueued()) {
continue;
}
// Another thread beat us, just unlock and wait for the next load
if (pending_queue.empty()) {
continue;
}
2020-07-16 08:38:35 +00:00
// Pull work from queue
auto work = std::move(pending_queue.front());
pending_queue.pop();
2020-07-16 08:38:35 +00:00
lock.unlock();
2020-07-10 03:36:38 +00:00
if (work->backend == Backend::OpenGL || work->backend == Backend::GLASM) {
VideoCommon::Shader::Registry registry = *work->registry;
const ShaderIR ir(work->code, work->main_offset, work->compiler_settings, registry);
2020-07-10 03:36:38 +00:00
const auto scope = context->Acquire();
auto program =
OpenGL::BuildShader(*work->device, work->shader_type, work->uid, ir, registry);
2020-07-10 03:36:38 +00:00
Result result{};
result.backend = work->backend;
result.cpu_address = work->cpu_address;
result.uid = work->uid;
result.code = std::move(work->code);
result.code_b = std::move(work->code_b);
result.shader_type = work->shader_type;
if (work->backend == Backend::OpenGL) {
2020-07-10 03:36:38 +00:00
result.program.opengl = std::move(program->source_program);
} else if (work->backend == Backend::GLASM) {
2020-07-10 03:36:38 +00:00
result.program.glasm = std::move(program->assembly_program);
}
work.reset();
2020-07-10 03:36:38 +00:00
{
std::unique_lock complete_lock(completed_mutex);
finished_work.push_back(std::move(result));
}
} else if (work->backend == Backend::Vulkan) {
2020-07-28 04:08:02 +00:00
Vulkan::GraphicsPipelineCacheKey params_key{
.renderpass_params = work->renderpass_params,
.padding = work->padding,
.shaders = work->shaders,
.fixed_state = work->fixed_state,
2020-07-28 04:08:02 +00:00
};
auto pipeline = std::make_unique<Vulkan::VKGraphicsPipeline>(
work->pp_cache->GetDevice(), work->pp_cache->GetScheduler(),
work->pp_cache->GetDescriptorPool(), work->pp_cache->GetUpdateDescriptorQueue(),
work->pp_cache->GetRenderpassCache(), params_key, work->bindings, work->program);
2020-07-28 04:08:02 +00:00
work->pp_cache->EmplacePipeline(std::move(pipeline));
work.reset();
2020-07-10 03:36:38 +00:00
}
// Give a chance for another thread to get work.
std::this_thread::yield();
2020-07-10 03:36:38 +00:00
}
}
} // namespace VideoCommon::Shader