mirror of
https://github.com/citra-emu/citra.git
synced 2024-11-25 13:40:14 +00:00
cleanup
This commit is contained in:
parent
9535169577
commit
3be6948ca4
@ -35,454 +35,448 @@
|
|||||||
|
|
||||||
namespace GPU {
|
namespace GPU {
|
||||||
|
|
||||||
Regs g_regs;
|
Regs g_regs;
|
||||||
|
|
||||||
/// True if the current frame was skipped
|
/// True if the current frame was skipped
|
||||||
bool g_skip_frame;
|
bool g_skip_frame;
|
||||||
/// 268MHz CPU clocks / 60Hz frames per second
|
/// 268MHz CPU clocks / 60Hz frames per second
|
||||||
const u64 frame_ticks = 268123480ull / 60;
|
const u64 frame_ticks = 268123480ull / 60;
|
||||||
/// Event id for CoreTiming
|
/// Event id for CoreTiming
|
||||||
static int vblank_event;
|
static int vblank_event;
|
||||||
/// Total number of frames drawn
|
/// Total number of frames drawn
|
||||||
static u64 frame_count;
|
static u64 frame_count;
|
||||||
/// True if the last frame was skipped
|
/// True if the last frame was skipped
|
||||||
static bool last_skip_frame;
|
static bool last_skip_frame;
|
||||||
|
|
||||||
template <typename T>
|
template <typename T>
|
||||||
inline void Read(T &var, const u32 raw_addr) {
|
inline void Read(T &var, const u32 raw_addr) {
|
||||||
u32 addr = raw_addr - HW::VADDR_GPU;
|
u32 addr = raw_addr - HW::VADDR_GPU;
|
||||||
u32 index = addr / 4;
|
u32 index = addr / 4;
|
||||||
|
|
||||||
// Reads other than u32 are untested, so I'd rather have them abort than silently fail
|
// Reads other than u32 are untested, so I'd rather have them abort than silently fail
|
||||||
if (index >= Regs::NumIds() || !std::is_same<T, u32>::value) {
|
if (index >= Regs::NumIds() || !std::is_same<T, u32>::value) {
|
||||||
LOG_ERROR(HW_GPU, "unknown Read%lu @ 0x%08X", sizeof(var) * 8, addr);
|
LOG_ERROR(HW_GPU, "unknown Read%lu @ 0x%08X", sizeof(var) * 8, addr);
|
||||||
return;
|
return;
|
||||||
}
|
|
||||||
|
|
||||||
var = g_regs[addr / 4];
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static Math::Vec4<u8> DecodePixel(Regs::PixelFormat input_format, const u8* src_pixel) {
|
var = g_regs[addr / 4];
|
||||||
switch (input_format) {
|
}
|
||||||
case Regs::PixelFormat::RGBA8:
|
|
||||||
return Color::DecodeRGBA8(src_pixel);
|
|
||||||
|
|
||||||
case Regs::PixelFormat::RGB8:
|
static Math::Vec4<u8> DecodePixel(Regs::PixelFormat input_format, const u8* src_pixel) {
|
||||||
return Color::DecodeRGB8(src_pixel);
|
switch (input_format) {
|
||||||
|
case Regs::PixelFormat::RGBA8:
|
||||||
|
return Color::DecodeRGBA8(src_pixel);
|
||||||
|
|
||||||
case Regs::PixelFormat::RGB565:
|
case Regs::PixelFormat::RGB8:
|
||||||
return Color::DecodeRGB565(src_pixel);
|
return Color::DecodeRGB8(src_pixel);
|
||||||
|
|
||||||
case Regs::PixelFormat::RGB5A1:
|
case Regs::PixelFormat::RGB565:
|
||||||
return Color::DecodeRGB5A1(src_pixel);
|
return Color::DecodeRGB565(src_pixel);
|
||||||
|
|
||||||
case Regs::PixelFormat::RGBA4:
|
case Regs::PixelFormat::RGB5A1:
|
||||||
return Color::DecodeRGBA4(src_pixel);
|
return Color::DecodeRGB5A1(src_pixel);
|
||||||
|
|
||||||
default:
|
case Regs::PixelFormat::RGBA4:
|
||||||
LOG_ERROR(HW_GPU, "Unknown source framebuffer format %x", input_format);
|
return Color::DecodeRGBA4(src_pixel);
|
||||||
return{ 0, 0, 0, 0 };
|
|
||||||
}
|
default:
|
||||||
|
LOG_ERROR(HW_GPU, "Unknown source framebuffer format %x", input_format);
|
||||||
|
return {0, 0, 0, 0};
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
MICROPROFILE_DEFINE(GPU_DisplayTransfer, "GPU", "DisplayTransfer", MP_RGB(100, 100, 255));
|
||||||
|
MICROPROFILE_DEFINE(GPU_CmdlistProcessing, "GPU", "Cmdlist Processing", MP_RGB(100, 255, 100));
|
||||||
|
|
||||||
|
template <typename T>
|
||||||
|
inline void Write(u32 addr, const T data) {
|
||||||
|
addr -= HW::VADDR_GPU;
|
||||||
|
u32 index = addr / 4;
|
||||||
|
|
||||||
|
// Writes other than u32 are untested, so I'd rather have them abort than silently fail
|
||||||
|
if (index >= Regs::NumIds() || !std::is_same<T, u32>::value) {
|
||||||
|
LOG_ERROR(HW_GPU, "unknown Write%lu 0x%08X @ 0x%08X", sizeof(data) * 8, (u32)data, addr);
|
||||||
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
MICROPROFILE_DEFINE(GPU_DisplayTransfer, "GPU", "DisplayTransfer", MP_RGB(100, 100, 255));
|
g_regs[index] = static_cast<u32>(data);
|
||||||
MICROPROFILE_DEFINE(GPU_CmdlistProcessing, "GPU", "Cmdlist Processing", MP_RGB(100, 255, 100));
|
|
||||||
|
|
||||||
template <typename T>
|
switch (index) {
|
||||||
inline void Write(u32 addr, const T data) {
|
|
||||||
addr -= HW::VADDR_GPU;
|
|
||||||
u32 index = addr / 4;
|
|
||||||
|
|
||||||
// Writes other than u32 are untested, so I'd rather have them abort than silently fail
|
// Memory fills are triggered once the fill value is written.
|
||||||
if (index >= Regs::NumIds() || !std::is_same<T, u32>::value) {
|
case GPU_REG_INDEX_WORKAROUND(memory_fill_config[0].trigger, 0x00004 + 0x3):
|
||||||
LOG_ERROR(HW_GPU, "unknown Write%lu 0x%08X @ 0x%08X", sizeof(data) * 8, (u32)data, addr);
|
case GPU_REG_INDEX_WORKAROUND(memory_fill_config[1].trigger, 0x00008 + 0x3):
|
||||||
return;
|
{
|
||||||
}
|
const bool is_second_filler = (index != GPU_REG_INDEX(memory_fill_config[0].trigger));
|
||||||
|
auto& config = g_regs.memory_fill_config[is_second_filler];
|
||||||
|
|
||||||
g_regs[index] = static_cast<u32>(data);
|
if (config.trigger) {
|
||||||
|
if (config.address_start) { // Some games pass invalid values here
|
||||||
|
u8* start = Memory::GetPhysicalPointer(config.GetStartAddress());
|
||||||
|
u8* end = Memory::GetPhysicalPointer(config.GetEndAddress());
|
||||||
|
|
||||||
switch (index) {
|
// TODO: Consider always accelerating and returning vector of
|
||||||
|
// regions that the accelerated fill did not cover to
|
||||||
|
// reduce/eliminate the fill that the cpu has to do.
|
||||||
|
// This would also mean that the flush below is not needed.
|
||||||
|
// Fill should first flush all surfaces that touch but are
|
||||||
|
// not completely within the fill range.
|
||||||
|
// Then fill all completely covered surfaces, and return the
|
||||||
|
// regions that were between surfaces or within the touching
|
||||||
|
// ones for cpu to manually fill here.
|
||||||
|
if (!VideoCore::g_renderer->Rasterizer()->AccelerateFill(config)) {
|
||||||
|
Memory::RasterizerFlushAndInvalidateRegion(config.GetStartAddress(), config.GetEndAddress() - config.GetStartAddress());
|
||||||
|
|
||||||
// Memory fills are triggered once the fill value is written.
|
if (config.fill_24bit) {
|
||||||
case GPU_REG_INDEX_WORKAROUND(memory_fill_config[0].trigger, 0x00004 + 0x3):
|
// fill with 24-bit values
|
||||||
case GPU_REG_INDEX_WORKAROUND(memory_fill_config[1].trigger, 0x00008 + 0x3):
|
for (u8* ptr = start; ptr < end; ptr += 3) {
|
||||||
{
|
ptr[0] = config.value_24bit_r;
|
||||||
const bool is_second_filler = (index != GPU_REG_INDEX(memory_fill_config[0].trigger));
|
ptr[1] = config.value_24bit_g;
|
||||||
auto& config = g_regs.memory_fill_config[is_second_filler];
|
ptr[2] = config.value_24bit_b;
|
||||||
|
|
||||||
if (config.trigger) {
|
|
||||||
if (config.address_start) { // Some games pass invalid values here
|
|
||||||
u8* start = Memory::GetPhysicalPointer(config.GetStartAddress());
|
|
||||||
u8* end = Memory::GetPhysicalPointer(config.GetEndAddress());
|
|
||||||
|
|
||||||
// TODO: Consider always accelerating and returning vector of
|
|
||||||
// regions that the accelerated fill did not cover to
|
|
||||||
// reduce/eliminate the fill that the cpu has to do.
|
|
||||||
// This would also mean that the flush below is not needed.
|
|
||||||
// Fill should first flush all surfaces that touch but are
|
|
||||||
// not completely within the fill range.
|
|
||||||
// Then fill all completely covered surfaces, and return the
|
|
||||||
// regions that were between surfaces or within the touching
|
|
||||||
// ones for cpu to manually fill here.
|
|
||||||
if (!VideoCore::g_renderer->Rasterizer()->AccelerateFill(config)) {
|
|
||||||
Memory::RasterizerFlushAndInvalidateRegion(config.GetStartAddress(), config.GetEndAddress() - config.GetStartAddress());
|
|
||||||
|
|
||||||
if (config.fill_24bit) {
|
|
||||||
// fill with 24-bit values
|
|
||||||
for (u8* ptr = start; ptr < end; ptr += 3) {
|
|
||||||
ptr[0] = config.value_24bit_r;
|
|
||||||
ptr[1] = config.value_24bit_g;
|
|
||||||
ptr[2] = config.value_24bit_b;
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
else if (config.fill_32bit) {
|
} else if (config.fill_32bit) {
|
||||||
// fill with 32-bit values
|
// fill with 32-bit values
|
||||||
if (end > start) {
|
if (end > start) {
|
||||||
u32 value = config.value_32bit;
|
u32 value = config.value_32bit;
|
||||||
size_t len = (end - start) / sizeof(u32);
|
size_t len = (end - start) / sizeof(u32);
|
||||||
for (size_t i = 0; i < len; ++i)
|
for (size_t i = 0; i < len; ++i)
|
||||||
memcpy(&start[i * sizeof(u32)], &value, sizeof(u32));
|
memcpy(&start[i * sizeof(u32)], &value, sizeof(u32));
|
||||||
}
|
|
||||||
}
|
}
|
||||||
else {
|
} else {
|
||||||
// fill with 16-bit values
|
// fill with 16-bit values
|
||||||
u16 value_16bit = config.value_16bit.Value();
|
u16 value_16bit = config.value_16bit.Value();
|
||||||
for (u8* ptr = start; ptr < end; ptr += sizeof(u16))
|
for (u8* ptr = start; ptr < end; ptr += sizeof(u16))
|
||||||
memcpy(ptr, &value_16bit, sizeof(u16));
|
memcpy(ptr, &value_16bit, sizeof(u16));
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
LOG_TRACE(HW_GPU, "MemoryFill from 0x%08x to 0x%08x", config.GetStartAddress(), config.GetEndAddress());
|
|
||||||
|
|
||||||
if (!is_second_filler) {
|
|
||||||
GSP_GPU::SignalInterrupt(GSP_GPU::InterruptId::PSC0);
|
|
||||||
}
|
|
||||||
else {
|
|
||||||
GSP_GPU::SignalInterrupt(GSP_GPU::InterruptId::PSC1);
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Reset "trigger" flag and set the "finish" flag
|
LOG_TRACE(HW_GPU, "MemoryFill from 0x%08x to 0x%08x", config.GetStartAddress(), config.GetEndAddress());
|
||||||
// NOTE: This was confirmed to happen on hardware even if "address_start" is zero.
|
|
||||||
config.trigger.Assign(0);
|
|
||||||
config.finished.Assign(1);
|
|
||||||
}
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
|
|
||||||
case GPU_REG_INDEX(display_transfer_config.trigger):
|
if (!is_second_filler) {
|
||||||
{
|
GSP_GPU::SignalInterrupt(GSP_GPU::InterruptId::PSC0);
|
||||||
MICROPROFILE_SCOPE(GPU_DisplayTransfer);
|
} else {
|
||||||
|
GSP_GPU::SignalInterrupt(GSP_GPU::InterruptId::PSC1);
|
||||||
const auto& config = g_regs.display_transfer_config;
|
|
||||||
if (config.trigger & 1) {
|
|
||||||
|
|
||||||
if (Pica::g_debug_context)
|
|
||||||
Pica::g_debug_context->OnEvent(Pica::DebugContext::Event::IncomingDisplayTransfer, nullptr);
|
|
||||||
|
|
||||||
if (!VideoCore::g_renderer->Rasterizer()->AccelerateDisplayTransfer(config)) {
|
|
||||||
u8* src_pointer = Memory::GetPhysicalPointer(config.GetPhysicalInputAddress());
|
|
||||||
u8* dst_pointer = Memory::GetPhysicalPointer(config.GetPhysicalOutputAddress());
|
|
||||||
|
|
||||||
if (config.is_texture_copy) {
|
|
||||||
u32 input_width = config.texture_copy.input_width * 16;
|
|
||||||
u32 input_gap = config.texture_copy.input_gap * 16;
|
|
||||||
u32 output_width = config.texture_copy.output_width * 16;
|
|
||||||
u32 output_gap = config.texture_copy.output_gap * 16;
|
|
||||||
|
|
||||||
size_t contiguous_input_size = config.texture_copy.size / input_width * (input_width + input_gap);
|
|
||||||
Memory::RasterizerFlushRegion(config.GetPhysicalInputAddress(), static_cast<u32>(contiguous_input_size));
|
|
||||||
|
|
||||||
size_t contiguous_output_size = config.texture_copy.size / output_width * (output_width + output_gap);
|
|
||||||
Memory::RasterizerFlushAndInvalidateRegion(config.GetPhysicalOutputAddress(), static_cast<u32>(contiguous_output_size));
|
|
||||||
|
|
||||||
u32 remaining_size = config.texture_copy.size;
|
|
||||||
u32 remaining_input = input_width;
|
|
||||||
u32 remaining_output = output_width;
|
|
||||||
while (remaining_size > 0) {
|
|
||||||
u32 copy_size = std::min({ remaining_input, remaining_output, remaining_size });
|
|
||||||
|
|
||||||
std::memcpy(dst_pointer, src_pointer, copy_size);
|
|
||||||
src_pointer += copy_size;
|
|
||||||
dst_pointer += copy_size;
|
|
||||||
|
|
||||||
remaining_input -= copy_size;
|
|
||||||
remaining_output -= copy_size;
|
|
||||||
remaining_size -= copy_size;
|
|
||||||
|
|
||||||
if (remaining_input == 0) {
|
|
||||||
remaining_input = input_width;
|
|
||||||
src_pointer += input_gap;
|
|
||||||
}
|
|
||||||
if (remaining_output == 0) {
|
|
||||||
remaining_output = output_width;
|
|
||||||
dst_pointer += output_gap;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
LOG_TRACE(HW_GPU, "TextureCopy: 0x%X bytes from 0x%08X(%u+%u)-> 0x%08X(%u+%u), flags 0x%08X",
|
|
||||||
config.texture_copy.size,
|
|
||||||
config.GetPhysicalInputAddress(), input_width, input_gap,
|
|
||||||
config.GetPhysicalOutputAddress(), output_width, output_gap,
|
|
||||||
config.flags);
|
|
||||||
|
|
||||||
GSP_GPU::SignalInterrupt(GSP_GPU::InterruptId::PPF);
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (config.scaling > config.ScaleXY) {
|
|
||||||
LOG_CRITICAL(HW_GPU, "Unimplemented display transfer scaling mode %u", config.scaling.Value());
|
|
||||||
UNIMPLEMENTED();
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (config.input_linear && config.scaling != config.NoScale) {
|
|
||||||
LOG_CRITICAL(HW_GPU, "Scaling is only implemented on tiled input");
|
|
||||||
UNIMPLEMENTED();
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
|
|
||||||
int horizontal_scale = config.scaling != config.NoScale ? 1 : 0;
|
|
||||||
int vertical_scale = config.scaling == config.ScaleXY ? 1 : 0;
|
|
||||||
|
|
||||||
u32 output_width = config.output_width >> horizontal_scale;
|
|
||||||
u32 output_height = config.output_height >> vertical_scale;
|
|
||||||
|
|
||||||
u32 input_size = config.input_width * config.input_height * GPU::Regs::BytesPerPixel(config.input_format);
|
|
||||||
u32 output_size = output_width * output_height * GPU::Regs::BytesPerPixel(config.output_format);
|
|
||||||
|
|
||||||
Memory::RasterizerFlushRegion(config.GetPhysicalInputAddress(), input_size);
|
|
||||||
Memory::RasterizerFlushAndInvalidateRegion(config.GetPhysicalOutputAddress(), output_size);
|
|
||||||
|
|
||||||
for (u32 y = 0; y < output_height; ++y) {
|
|
||||||
for (u32 x = 0; x < output_width; ++x) {
|
|
||||||
Math::Vec4<u8> src_color;
|
|
||||||
|
|
||||||
// Calculate the [x,y] position of the input image
|
|
||||||
// based on the current output position and the scale
|
|
||||||
u32 input_x = x << horizontal_scale;
|
|
||||||
u32 input_y = y << vertical_scale;
|
|
||||||
|
|
||||||
if (config.flip_vertically) {
|
|
||||||
// Flip the y value of the output data,
|
|
||||||
// we do this after calculating the [x,y] position of the input image
|
|
||||||
// to account for the scaling options.
|
|
||||||
y = output_height - y - 1;
|
|
||||||
}
|
|
||||||
|
|
||||||
u32 dst_bytes_per_pixel = GPU::Regs::BytesPerPixel(config.output_format);
|
|
||||||
u32 src_bytes_per_pixel = GPU::Regs::BytesPerPixel(config.input_format);
|
|
||||||
u32 src_offset;
|
|
||||||
u32 dst_offset;
|
|
||||||
|
|
||||||
if (config.input_linear) {
|
|
||||||
if (!config.dont_swizzle) {
|
|
||||||
// Interpret the input as linear and the output as tiled
|
|
||||||
u32 coarse_y = y & ~7;
|
|
||||||
u32 stride = output_width * dst_bytes_per_pixel;
|
|
||||||
|
|
||||||
src_offset = (input_x + input_y * config.input_width) * src_bytes_per_pixel;
|
|
||||||
dst_offset = VideoCore::GetMortonOffset(x, y, dst_bytes_per_pixel) + coarse_y * stride;
|
|
||||||
}
|
|
||||||
else {
|
|
||||||
// Both input and output are linear
|
|
||||||
src_offset = (input_x + input_y * config.input_width) * src_bytes_per_pixel;
|
|
||||||
dst_offset = (x + y * output_width) * dst_bytes_per_pixel;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
else {
|
|
||||||
if (!config.dont_swizzle) {
|
|
||||||
// Interpret the input as tiled and the output as linear
|
|
||||||
u32 coarse_y = input_y & ~7;
|
|
||||||
u32 stride = config.input_width * src_bytes_per_pixel;
|
|
||||||
|
|
||||||
src_offset = VideoCore::GetMortonOffset(input_x, input_y, src_bytes_per_pixel) + coarse_y * stride;
|
|
||||||
dst_offset = (x + y * output_width) * dst_bytes_per_pixel;
|
|
||||||
}
|
|
||||||
else {
|
|
||||||
// Both input and output are tiled
|
|
||||||
u32 out_coarse_y = y & ~7;
|
|
||||||
u32 out_stride = output_width * dst_bytes_per_pixel;
|
|
||||||
|
|
||||||
u32 in_coarse_y = input_y & ~7;
|
|
||||||
u32 in_stride = config.input_width * src_bytes_per_pixel;
|
|
||||||
|
|
||||||
src_offset = VideoCore::GetMortonOffset(input_x, input_y, src_bytes_per_pixel) + in_coarse_y * in_stride;
|
|
||||||
dst_offset = VideoCore::GetMortonOffset(x, y, dst_bytes_per_pixel) + out_coarse_y * out_stride;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
const u8* src_pixel = src_pointer + src_offset;
|
|
||||||
src_color = DecodePixel(config.input_format, src_pixel);
|
|
||||||
if (config.scaling == config.ScaleX) {
|
|
||||||
Math::Vec4<u8> pixel = DecodePixel(config.input_format, src_pixel + src_bytes_per_pixel);
|
|
||||||
src_color = ((src_color + pixel) / 2).Cast<u8>();
|
|
||||||
}
|
|
||||||
else if (config.scaling == config.ScaleXY) {
|
|
||||||
Math::Vec4<u8> pixel1 = DecodePixel(config.input_format, src_pixel + 1 * src_bytes_per_pixel);
|
|
||||||
Math::Vec4<u8> pixel2 = DecodePixel(config.input_format, src_pixel + 2 * src_bytes_per_pixel);
|
|
||||||
Math::Vec4<u8> pixel3 = DecodePixel(config.input_format, src_pixel + 3 * src_bytes_per_pixel);
|
|
||||||
src_color = (((src_color + pixel1) + (pixel2 + pixel3)) / 4).Cast<u8>();
|
|
||||||
}
|
|
||||||
|
|
||||||
u8* dst_pixel = dst_pointer + dst_offset;
|
|
||||||
switch (config.output_format) {
|
|
||||||
case Regs::PixelFormat::RGBA8:
|
|
||||||
Color::EncodeRGBA8(src_color, dst_pixel);
|
|
||||||
break;
|
|
||||||
|
|
||||||
case Regs::PixelFormat::RGB8:
|
|
||||||
Color::EncodeRGB8(src_color, dst_pixel);
|
|
||||||
break;
|
|
||||||
|
|
||||||
case Regs::PixelFormat::RGB565:
|
|
||||||
Color::EncodeRGB565(src_color, dst_pixel);
|
|
||||||
break;
|
|
||||||
|
|
||||||
case Regs::PixelFormat::RGB5A1:
|
|
||||||
Color::EncodeRGB5A1(src_color, dst_pixel);
|
|
||||||
break;
|
|
||||||
|
|
||||||
case Regs::PixelFormat::RGBA4:
|
|
||||||
Color::EncodeRGBA4(src_color, dst_pixel);
|
|
||||||
break;
|
|
||||||
|
|
||||||
default:
|
|
||||||
LOG_ERROR(HW_GPU, "Unknown destination framebuffer format %x", config.output_format.Value());
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
LOG_TRACE(HW_GPU, "DisplayTriggerTransfer: 0x%08x bytes from 0x%08x(%ux%u)-> 0x%08x(%ux%u), dst format %x, flags 0x%08X",
|
|
||||||
config.output_height * output_width * GPU::Regs::BytesPerPixel(config.output_format),
|
|
||||||
config.GetPhysicalInputAddress(), config.input_width.Value(), config.input_height.Value(),
|
|
||||||
config.GetPhysicalOutputAddress(), output_width, output_height,
|
|
||||||
config.output_format.Value(), config.flags);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
g_regs.display_transfer_config.trigger = 0;
|
|
||||||
GSP_GPU::SignalInterrupt(GSP_GPU::InterruptId::PPF);
|
|
||||||
}
|
}
|
||||||
break;
|
|
||||||
}
|
// Reset "trigger" flag and set the "finish" flag
|
||||||
|
// NOTE: This was confirmed to happen on hardware even if "address_start" is zero.
|
||||||
// Seems like writing to this register triggers processing
|
config.trigger.Assign(0);
|
||||||
case GPU_REG_INDEX(command_processor_config.trigger):
|
config.finished.Assign(1);
|
||||||
{
|
|
||||||
const auto& config = g_regs.command_processor_config;
|
|
||||||
if (config.trigger & 1) {
|
|
||||||
MICROPROFILE_SCOPE(GPU_CmdlistProcessing);
|
|
||||||
|
|
||||||
u32* buffer = (u32*)Memory::GetPhysicalPointer(config.GetPhysicalAddress());
|
|
||||||
|
|
||||||
if (Pica::g_debug_context && Pica::g_debug_context->recorder) {
|
|
||||||
Pica::g_debug_context->recorder->MemoryAccessed((u8*)buffer, config.size * sizeof(u32), config.GetPhysicalAddress());
|
|
||||||
}
|
|
||||||
|
|
||||||
Pica::CommandProcessor::ProcessCommandList(buffer, config.size);
|
|
||||||
|
|
||||||
g_regs.command_processor_config.trigger = 0;
|
|
||||||
}
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
|
|
||||||
default:
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Notify tracer about the register write
|
|
||||||
// This is happening *after* handling the write to make sure we properly catch all memory reads.
|
|
||||||
if (Pica::g_debug_context && Pica::g_debug_context->recorder) {
|
|
||||||
// addr + GPU VBase - IO VBase + IO PBase
|
|
||||||
Pica::g_debug_context->recorder->RegisterWritten<T>(addr + 0x1EF00000 - 0x1EC00000 + 0x10100000, data);
|
|
||||||
}
|
}
|
||||||
|
break;
|
||||||
}
|
}
|
||||||
|
|
||||||
// Explicitly instantiate template functions because we aren't defining this in the header:
|
case GPU_REG_INDEX(display_transfer_config.trigger):
|
||||||
|
{
|
||||||
|
MICROPROFILE_SCOPE(GPU_DisplayTransfer);
|
||||||
|
|
||||||
template void Read<u64>(u64 &var, const u32 addr);
|
const auto& config = g_regs.display_transfer_config;
|
||||||
template void Read<u32>(u32 &var, const u32 addr);
|
if (config.trigger & 1) {
|
||||||
template void Read<u16>(u16 &var, const u32 addr);
|
|
||||||
template void Read<u8>(u8 &var, const u32 addr);
|
|
||||||
|
|
||||||
template void Write<u64>(u32 addr, const u64 data);
|
if (Pica::g_debug_context)
|
||||||
template void Write<u32>(u32 addr, const u32 data);
|
Pica::g_debug_context->OnEvent(Pica::DebugContext::Event::IncomingDisplayTransfer, nullptr);
|
||||||
template void Write<u16>(u32 addr, const u16 data);
|
|
||||||
template void Write<u8>(u32 addr, const u8 data);
|
|
||||||
|
|
||||||
/// Update hardware
|
if (!VideoCore::g_renderer->Rasterizer()->AccelerateDisplayTransfer(config)) {
|
||||||
static void VBlankCallback(u64 userdata, int cycles_late) {
|
u8* src_pointer = Memory::GetPhysicalPointer(config.GetPhysicalInputAddress());
|
||||||
frame_count++;
|
u8* dst_pointer = Memory::GetPhysicalPointer(config.GetPhysicalOutputAddress());
|
||||||
last_skip_frame = g_skip_frame;
|
|
||||||
g_skip_frame = (frame_count & Settings::values.frame_skip) != 0;
|
|
||||||
|
|
||||||
// Swap buffers based on the frameskip mode, which is a little bit tricky. When
|
if (config.is_texture_copy) {
|
||||||
// a frame is being skipped, nothing is being rendered to the internal framebuffer(s).
|
u32 input_width = config.texture_copy.input_width * 16;
|
||||||
// So, we should only swap frames if the last frame was rendered. The rules are:
|
u32 input_gap = config.texture_copy.input_gap * 16;
|
||||||
// - If frameskip == 0 (disabled), always swap buffers
|
u32 output_width = config.texture_copy.output_width * 16;
|
||||||
// - If frameskip == 1, swap buffers every other frame (starting from the first frame)
|
u32 output_gap = config.texture_copy.output_gap * 16;
|
||||||
// - If frameskip > 1, swap buffers every frameskip^n frames (starting from the second frame)
|
|
||||||
if ((((Settings::values.frame_skip != 1) ^ last_skip_frame) && last_skip_frame != g_skip_frame) ||
|
size_t contiguous_input_size = config.texture_copy.size / input_width * (input_width + input_gap);
|
||||||
|
Memory::RasterizerFlushRegion(config.GetPhysicalInputAddress(), static_cast<u32>(contiguous_input_size));
|
||||||
|
|
||||||
|
size_t contiguous_output_size = config.texture_copy.size / output_width * (output_width + output_gap);
|
||||||
|
Memory::RasterizerFlushAndInvalidateRegion(config.GetPhysicalOutputAddress(), static_cast<u32>(contiguous_output_size));
|
||||||
|
|
||||||
|
u32 remaining_size = config.texture_copy.size;
|
||||||
|
u32 remaining_input = input_width;
|
||||||
|
u32 remaining_output = output_width;
|
||||||
|
while (remaining_size > 0) {
|
||||||
|
u32 copy_size = std::min({ remaining_input, remaining_output, remaining_size });
|
||||||
|
|
||||||
|
std::memcpy(dst_pointer, src_pointer, copy_size);
|
||||||
|
src_pointer += copy_size;
|
||||||
|
dst_pointer += copy_size;
|
||||||
|
|
||||||
|
remaining_input -= copy_size;
|
||||||
|
remaining_output -= copy_size;
|
||||||
|
remaining_size -= copy_size;
|
||||||
|
|
||||||
|
if (remaining_input == 0) {
|
||||||
|
remaining_input = input_width;
|
||||||
|
src_pointer += input_gap;
|
||||||
|
}
|
||||||
|
if (remaining_output == 0) {
|
||||||
|
remaining_output = output_width;
|
||||||
|
dst_pointer += output_gap;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
LOG_TRACE(HW_GPU, "TextureCopy: 0x%X bytes from 0x%08X(%u+%u)-> 0x%08X(%u+%u), flags 0x%08X",
|
||||||
|
config.texture_copy.size,
|
||||||
|
config.GetPhysicalInputAddress(), input_width, input_gap,
|
||||||
|
config.GetPhysicalOutputAddress(), output_width, output_gap,
|
||||||
|
config.flags);
|
||||||
|
|
||||||
|
GSP_GPU::SignalInterrupt(GSP_GPU::InterruptId::PPF);
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (config.scaling > config.ScaleXY) {
|
||||||
|
LOG_CRITICAL(HW_GPU, "Unimplemented display transfer scaling mode %u", config.scaling.Value());
|
||||||
|
UNIMPLEMENTED();
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (config.input_linear && config.scaling != config.NoScale) {
|
||||||
|
LOG_CRITICAL(HW_GPU, "Scaling is only implemented on tiled input");
|
||||||
|
UNIMPLEMENTED();
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
|
||||||
|
int horizontal_scale = config.scaling != config.NoScale ? 1 : 0;
|
||||||
|
int vertical_scale = config.scaling == config.ScaleXY ? 1 : 0;
|
||||||
|
|
||||||
|
u32 output_width = config.output_width >> horizontal_scale;
|
||||||
|
u32 output_height = config.output_height >> vertical_scale;
|
||||||
|
|
||||||
|
u32 input_size = config.input_width * config.input_height * GPU::Regs::BytesPerPixel(config.input_format);
|
||||||
|
u32 output_size = output_width * output_height * GPU::Regs::BytesPerPixel(config.output_format);
|
||||||
|
|
||||||
|
Memory::RasterizerFlushRegion(config.GetPhysicalInputAddress(), input_size);
|
||||||
|
Memory::RasterizerFlushAndInvalidateRegion(config.GetPhysicalOutputAddress(), output_size);
|
||||||
|
|
||||||
|
for (u32 y = 0; y < output_height; ++y) {
|
||||||
|
for (u32 x = 0; x < output_width; ++x) {
|
||||||
|
Math::Vec4<u8> src_color;
|
||||||
|
|
||||||
|
// Calculate the [x,y] position of the input image
|
||||||
|
// based on the current output position and the scale
|
||||||
|
u32 input_x = x << horizontal_scale;
|
||||||
|
u32 input_y = y << vertical_scale;
|
||||||
|
|
||||||
|
if (config.flip_vertically) {
|
||||||
|
// Flip the y value of the output data,
|
||||||
|
// we do this after calculating the [x,y] position of the input image
|
||||||
|
// to account for the scaling options.
|
||||||
|
y = output_height - y - 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
u32 dst_bytes_per_pixel = GPU::Regs::BytesPerPixel(config.output_format);
|
||||||
|
u32 src_bytes_per_pixel = GPU::Regs::BytesPerPixel(config.input_format);
|
||||||
|
u32 src_offset;
|
||||||
|
u32 dst_offset;
|
||||||
|
|
||||||
|
if (config.input_linear) {
|
||||||
|
if (!config.dont_swizzle) {
|
||||||
|
// Interpret the input as linear and the output as tiled
|
||||||
|
u32 coarse_y = y & ~7;
|
||||||
|
u32 stride = output_width * dst_bytes_per_pixel;
|
||||||
|
|
||||||
|
src_offset = (input_x + input_y * config.input_width) * src_bytes_per_pixel;
|
||||||
|
dst_offset = VideoCore::GetMortonOffset(x, y, dst_bytes_per_pixel) + coarse_y * stride;
|
||||||
|
} else {
|
||||||
|
// Both input and output are linear
|
||||||
|
src_offset = (input_x + input_y * config.input_width) * src_bytes_per_pixel;
|
||||||
|
dst_offset = (x + y * output_width) * dst_bytes_per_pixel;
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
if (!config.dont_swizzle) {
|
||||||
|
// Interpret the input as tiled and the output as linear
|
||||||
|
u32 coarse_y = input_y & ~7;
|
||||||
|
u32 stride = config.input_width * src_bytes_per_pixel;
|
||||||
|
|
||||||
|
src_offset = VideoCore::GetMortonOffset(input_x, input_y, src_bytes_per_pixel) + coarse_y * stride;
|
||||||
|
dst_offset = (x + y * output_width) * dst_bytes_per_pixel;
|
||||||
|
} else {
|
||||||
|
// Both input and output are tiled
|
||||||
|
u32 out_coarse_y = y & ~7;
|
||||||
|
u32 out_stride = output_width * dst_bytes_per_pixel;
|
||||||
|
|
||||||
|
u32 in_coarse_y = input_y & ~7;
|
||||||
|
u32 in_stride = config.input_width * src_bytes_per_pixel;
|
||||||
|
|
||||||
|
src_offset = VideoCore::GetMortonOffset(input_x, input_y, src_bytes_per_pixel) + in_coarse_y * in_stride;
|
||||||
|
dst_offset = VideoCore::GetMortonOffset(x, y, dst_bytes_per_pixel) + out_coarse_y * out_stride;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
const u8* src_pixel = src_pointer + src_offset;
|
||||||
|
src_color = DecodePixel(config.input_format, src_pixel);
|
||||||
|
if (config.scaling == config.ScaleX) {
|
||||||
|
Math::Vec4<u8> pixel = DecodePixel(config.input_format, src_pixel + src_bytes_per_pixel);
|
||||||
|
src_color = ((src_color + pixel) / 2).Cast<u8>();
|
||||||
|
} else if (config.scaling == config.ScaleXY) {
|
||||||
|
Math::Vec4<u8> pixel1 = DecodePixel(config.input_format, src_pixel + 1 * src_bytes_per_pixel);
|
||||||
|
Math::Vec4<u8> pixel2 = DecodePixel(config.input_format, src_pixel + 2 * src_bytes_per_pixel);
|
||||||
|
Math::Vec4<u8> pixel3 = DecodePixel(config.input_format, src_pixel + 3 * src_bytes_per_pixel);
|
||||||
|
src_color = (((src_color + pixel1) + (pixel2 + pixel3)) / 4).Cast<u8>();
|
||||||
|
}
|
||||||
|
|
||||||
|
u8* dst_pixel = dst_pointer + dst_offset;
|
||||||
|
switch (config.output_format) {
|
||||||
|
case Regs::PixelFormat::RGBA8:
|
||||||
|
Color::EncodeRGBA8(src_color, dst_pixel);
|
||||||
|
break;
|
||||||
|
|
||||||
|
case Regs::PixelFormat::RGB8:
|
||||||
|
Color::EncodeRGB8(src_color, dst_pixel);
|
||||||
|
break;
|
||||||
|
|
||||||
|
case Regs::PixelFormat::RGB565:
|
||||||
|
Color::EncodeRGB565(src_color, dst_pixel);
|
||||||
|
break;
|
||||||
|
|
||||||
|
case Regs::PixelFormat::RGB5A1:
|
||||||
|
Color::EncodeRGB5A1(src_color, dst_pixel);
|
||||||
|
break;
|
||||||
|
|
||||||
|
case Regs::PixelFormat::RGBA4:
|
||||||
|
Color::EncodeRGBA4(src_color, dst_pixel);
|
||||||
|
break;
|
||||||
|
|
||||||
|
default:
|
||||||
|
LOG_ERROR(HW_GPU, "Unknown destination framebuffer format %x", config.output_format.Value());
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
LOG_TRACE(HW_GPU, "DisplayTriggerTransfer: 0x%08x bytes from 0x%08x(%ux%u)-> 0x%08x(%ux%u), dst format %x, flags 0x%08X",
|
||||||
|
config.output_height * output_width * GPU::Regs::BytesPerPixel(config.output_format),
|
||||||
|
config.GetPhysicalInputAddress(), config.input_width.Value(), config.input_height.Value(),
|
||||||
|
config.GetPhysicalOutputAddress(), output_width, output_height,
|
||||||
|
config.output_format.Value(), config.flags);
|
||||||
|
}
|
||||||
|
|
||||||
|
g_regs.display_transfer_config.trigger = 0;
|
||||||
|
GSP_GPU::SignalInterrupt(GSP_GPU::InterruptId::PPF);
|
||||||
|
}
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Seems like writing to this register triggers processing
|
||||||
|
case GPU_REG_INDEX(command_processor_config.trigger):
|
||||||
|
{
|
||||||
|
const auto& config = g_regs.command_processor_config;
|
||||||
|
if (config.trigger & 1)
|
||||||
|
{
|
||||||
|
MICROPROFILE_SCOPE(GPU_CmdlistProcessing);
|
||||||
|
|
||||||
|
u32* buffer = (u32*)Memory::GetPhysicalPointer(config.GetPhysicalAddress());
|
||||||
|
|
||||||
|
if (Pica::g_debug_context && Pica::g_debug_context->recorder) {
|
||||||
|
Pica::g_debug_context->recorder->MemoryAccessed((u8*)buffer, config.size * sizeof(u32), config.GetPhysicalAddress());
|
||||||
|
}
|
||||||
|
|
||||||
|
Pica::CommandProcessor::ProcessCommandList(buffer, config.size);
|
||||||
|
|
||||||
|
g_regs.command_processor_config.trigger = 0;
|
||||||
|
}
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
|
||||||
|
default:
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Notify tracer about the register write
|
||||||
|
// This is happening *after* handling the write to make sure we properly catch all memory reads.
|
||||||
|
if (Pica::g_debug_context && Pica::g_debug_context->recorder) {
|
||||||
|
// addr + GPU VBase - IO VBase + IO PBase
|
||||||
|
Pica::g_debug_context->recorder->RegisterWritten<T>(addr + 0x1EF00000 - 0x1EC00000 + 0x10100000, data);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Explicitly instantiate template functions because we aren't defining this in the header:
|
||||||
|
|
||||||
|
template void Read<u64>(u64 &var, const u32 addr);
|
||||||
|
template void Read<u32>(u32 &var, const u32 addr);
|
||||||
|
template void Read<u16>(u16 &var, const u32 addr);
|
||||||
|
template void Read<u8>(u8 &var, const u32 addr);
|
||||||
|
|
||||||
|
template void Write<u64>(u32 addr, const u64 data);
|
||||||
|
template void Write<u32>(u32 addr, const u32 data);
|
||||||
|
template void Write<u16>(u32 addr, const u16 data);
|
||||||
|
template void Write<u8>(u32 addr, const u8 data);
|
||||||
|
|
||||||
|
/// Update hardware
|
||||||
|
static void VBlankCallback(u64 userdata, int cycles_late) {
|
||||||
|
frame_count++;
|
||||||
|
last_skip_frame = g_skip_frame;
|
||||||
|
g_skip_frame = (frame_count & Settings::values.frame_skip) != 0;
|
||||||
|
|
||||||
|
// Swap buffers based on the frameskip mode, which is a little bit tricky. When
|
||||||
|
// a frame is being skipped, nothing is being rendered to the internal framebuffer(s).
|
||||||
|
// So, we should only swap frames if the last frame was rendered. The rules are:
|
||||||
|
// - If frameskip == 0 (disabled), always swap buffers
|
||||||
|
// - If frameskip == 1, swap buffers every other frame (starting from the first frame)
|
||||||
|
// - If frameskip > 1, swap buffers every frameskip^n frames (starting from the second frame)
|
||||||
|
if ((((Settings::values.frame_skip != 1) ^ last_skip_frame) && last_skip_frame != g_skip_frame) ||
|
||||||
Settings::values.frame_skip == 0) {
|
Settings::values.frame_skip == 0) {
|
||||||
VideoCore::g_renderer->SwapBuffers();
|
VideoCore::g_renderer->SwapBuffers();
|
||||||
}
|
|
||||||
|
|
||||||
// Signal to GSP that GPU interrupt has occurred
|
|
||||||
// TODO(yuriks): hwtest to determine if PDC0 is for the Top screen and PDC1 for the Sub
|
|
||||||
// screen, or if both use the same interrupts and these two instead determine the
|
|
||||||
// beginning and end of the VBlank period. If needed, split the interrupt firing into
|
|
||||||
// two different intervals.
|
|
||||||
GSP_GPU::SignalInterrupt(GSP_GPU::InterruptId::PDC0);
|
|
||||||
GSP_GPU::SignalInterrupt(GSP_GPU::InterruptId::PDC1);
|
|
||||||
|
|
||||||
// Reschedule recurrent event
|
|
||||||
CoreTiming::ScheduleEvent(frame_ticks - cycles_late, vblank_event);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Initialize hardware
|
// Signal to GSP that GPU interrupt has occurred
|
||||||
void Init() {
|
// TODO(yuriks): hwtest to determine if PDC0 is for the Top screen and PDC1 for the Sub
|
||||||
memset(&g_regs, 0, sizeof(g_regs));
|
// screen, or if both use the same interrupts and these two instead determine the
|
||||||
|
// beginning and end of the VBlank period. If needed, split the interrupt firing into
|
||||||
|
// two different intervals.
|
||||||
|
GSP_GPU::SignalInterrupt(GSP_GPU::InterruptId::PDC0);
|
||||||
|
GSP_GPU::SignalInterrupt(GSP_GPU::InterruptId::PDC1);
|
||||||
|
|
||||||
auto& framebuffer_top = g_regs.framebuffer_config[0];
|
// Reschedule recurrent event
|
||||||
auto& framebuffer_sub = g_regs.framebuffer_config[1];
|
CoreTiming::ScheduleEvent(frame_ticks - cycles_late, vblank_event);
|
||||||
|
}
|
||||||
|
|
||||||
// Setup default framebuffer addresses (located in VRAM)
|
/// Initialize hardware
|
||||||
// .. or at least these are the ones used by system applets.
|
void Init() {
|
||||||
// There's probably a smarter way to come up with addresses
|
memset(&g_regs, 0, sizeof(g_regs));
|
||||||
// like this which does not require hardcoding.
|
|
||||||
framebuffer_top.address_left1 = 0x181E6000;
|
|
||||||
framebuffer_top.address_left2 = 0x1822C800;
|
|
||||||
framebuffer_top.address_right1 = 0x18273000;
|
|
||||||
framebuffer_top.address_right2 = 0x182B9800;
|
|
||||||
framebuffer_sub.address_left1 = 0x1848F000;
|
|
||||||
framebuffer_sub.address_left2 = 0x184C7800;
|
|
||||||
|
|
||||||
framebuffer_top.width.Assign(240);
|
auto& framebuffer_top = g_regs.framebuffer_config[0];
|
||||||
framebuffer_top.height.Assign(400);
|
auto& framebuffer_sub = g_regs.framebuffer_config[1];
|
||||||
framebuffer_top.stride = 3 * 240;
|
|
||||||
framebuffer_top.color_format.Assign(Regs::PixelFormat::RGB8);
|
|
||||||
framebuffer_top.active_fb = 0;
|
|
||||||
|
|
||||||
framebuffer_sub.width.Assign(240);
|
// Setup default framebuffer addresses (located in VRAM)
|
||||||
framebuffer_sub.height.Assign(320);
|
// .. or at least these are the ones used by system applets.
|
||||||
framebuffer_sub.stride = 3 * 240;
|
// There's probably a smarter way to come up with addresses
|
||||||
framebuffer_sub.color_format.Assign(Regs::PixelFormat::RGB8);
|
// like this which does not require hardcoding.
|
||||||
framebuffer_sub.active_fb = 0;
|
framebuffer_top.address_left1 = 0x181E6000;
|
||||||
|
framebuffer_top.address_left2 = 0x1822C800;
|
||||||
|
framebuffer_top.address_right1 = 0x18273000;
|
||||||
|
framebuffer_top.address_right2 = 0x182B9800;
|
||||||
|
framebuffer_sub.address_left1 = 0x1848F000;
|
||||||
|
framebuffer_sub.address_left2 = 0x184C7800;
|
||||||
|
|
||||||
last_skip_frame = false;
|
framebuffer_top.width.Assign(240);
|
||||||
g_skip_frame = false;
|
framebuffer_top.height.Assign(400);
|
||||||
frame_count = 0;
|
framebuffer_top.stride = 3 * 240;
|
||||||
|
framebuffer_top.color_format.Assign(Regs::PixelFormat::RGB8);
|
||||||
|
framebuffer_top.active_fb = 0;
|
||||||
|
|
||||||
vblank_event = CoreTiming::RegisterEvent("GPU::VBlankCallback", VBlankCallback);
|
framebuffer_sub.width.Assign(240);
|
||||||
CoreTiming::ScheduleEvent(frame_ticks, vblank_event);
|
framebuffer_sub.height.Assign(320);
|
||||||
|
framebuffer_sub.stride = 3 * 240;
|
||||||
|
framebuffer_sub.color_format.Assign(Regs::PixelFormat::RGB8);
|
||||||
|
framebuffer_sub.active_fb = 0;
|
||||||
|
|
||||||
LOG_DEBUG(HW_GPU, "initialized OK");
|
last_skip_frame = false;
|
||||||
}
|
g_skip_frame = false;
|
||||||
|
frame_count = 0;
|
||||||
|
|
||||||
/// Shutdown hardware
|
vblank_event = CoreTiming::RegisterEvent("GPU::VBlankCallback", VBlankCallback);
|
||||||
void Shutdown() {
|
CoreTiming::ScheduleEvent(frame_ticks, vblank_event);
|
||||||
LOG_DEBUG(HW_GPU, "shutdown OK");
|
|
||||||
}
|
LOG_DEBUG(HW_GPU, "initialized OK");
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Shutdown hardware
|
||||||
|
void Shutdown() {
|
||||||
|
LOG_DEBUG(HW_GPU, "shutdown OK");
|
||||||
|
}
|
||||||
|
|
||||||
} // namespace
|
} // namespace
|
||||||
|
Loading…
Reference in New Issue
Block a user