1
0
Fork 0
forked from suyu/suyu

Merge pull request #3535 from ReinUsesLisp/gcc-warnings

video_core: Silence misc warnings
This commit is contained in:
Mat M 2020-03-18 20:09:32 -04:00 committed by GitHub
commit 9418b983bd
No known key found for this signature in database
GPG key ID: 4AEE18F83AFDEB23
9 changed files with 19 additions and 36 deletions

View file

@ -828,7 +828,7 @@ inline MicroProfileLogEntry MicroProfileMakeLogIndex(uint64_t nBegin, MicroProfi
MicroProfileLogEntry Entry = (nBegin<<62) | ((0x3fff&nToken)<<48) | (MP_LOG_TICK_MASK&nTick); MicroProfileLogEntry Entry = (nBegin<<62) | ((0x3fff&nToken)<<48) | (MP_LOG_TICK_MASK&nTick);
int t = MicroProfileLogType(Entry); int t = MicroProfileLogType(Entry);
uint64_t nTimerIndex = MicroProfileLogTimerIndex(Entry); uint64_t nTimerIndex = MicroProfileLogTimerIndex(Entry);
MP_ASSERT(t == nBegin); MP_ASSERT((uint64_t)t == nBegin);
MP_ASSERT(nTimerIndex == (nToken&0x3fff)); MP_ASSERT(nTimerIndex == (nToken&0x3fff));
return Entry; return Entry;
@ -1556,10 +1556,10 @@ void MicroProfileFlip()
pFramePut->nFrameStartCpu = MP_TICK(); pFramePut->nFrameStartCpu = MP_TICK();
pFramePut->nFrameStartGpu = (uint32_t)MicroProfileGpuInsertTimeStamp(); pFramePut->nFrameStartGpu = (uint32_t)MicroProfileGpuInsertTimeStamp();
if(pFrameNext->nFrameStartGpu != (uint64_t)-1) if(pFrameNext->nFrameStartGpu != -1)
pFrameNext->nFrameStartGpu = MicroProfileGpuGetTimeStamp((uint32_t)pFrameNext->nFrameStartGpu); pFrameNext->nFrameStartGpu = MicroProfileGpuGetTimeStamp((uint32_t)pFrameNext->nFrameStartGpu);
if(pFrameCurrent->nFrameStartGpu == (uint64_t)-1) if(pFrameCurrent->nFrameStartGpu == -1)
pFrameCurrent->nFrameStartGpu = pFrameNext->nFrameStartGpu + 1; pFrameCurrent->nFrameStartGpu = pFrameNext->nFrameStartGpu + 1;
uint64_t nFrameStartCpu = pFrameCurrent->nFrameStartCpu; uint64_t nFrameStartCpu = pFrameCurrent->nFrameStartCpu;

View file

@ -119,14 +119,6 @@ Texture::TICEntry KeplerCompute::GetTICEntry(u32 tic_index) const {
Texture::TICEntry tic_entry; Texture::TICEntry tic_entry;
memory_manager.ReadBlockUnsafe(tic_address_gpu, &tic_entry, sizeof(Texture::TICEntry)); memory_manager.ReadBlockUnsafe(tic_address_gpu, &tic_entry, sizeof(Texture::TICEntry));
const auto r_type{tic_entry.r_type.Value()};
const auto g_type{tic_entry.g_type.Value()};
const auto b_type{tic_entry.b_type.Value()};
const auto a_type{tic_entry.a_type.Value()};
// TODO(Subv): Different data types for separate components are not supported
DEBUG_ASSERT(r_type == g_type && r_type == b_type && r_type == a_type);
return tic_entry; return tic_entry;
} }

View file

@ -93,10 +93,6 @@ void oglEnable(GLenum cap, bool state) {
(state ? glEnable : glDisable)(cap); (state ? glEnable : glDisable)(cap);
} }
void oglEnablei(GLenum cap, bool state, GLuint index) {
(state ? glEnablei : glDisablei)(cap, index);
}
} // Anonymous namespace } // Anonymous namespace
RasterizerOpenGL::RasterizerOpenGL(Core::System& system, Core::Frontend::EmuWindow& emu_window, RasterizerOpenGL::RasterizerOpenGL(Core::System& system, Core::Frontend::EmuWindow& emu_window,
@ -478,7 +474,6 @@ void RasterizerOpenGL::Clear() {
void RasterizerOpenGL::Draw(bool is_indexed, bool is_instanced) { void RasterizerOpenGL::Draw(bool is_indexed, bool is_instanced) {
MICROPROFILE_SCOPE(OpenGL_Drawing); MICROPROFILE_SCOPE(OpenGL_Drawing);
auto& gpu = system.GPU().Maxwell3D(); auto& gpu = system.GPU().Maxwell3D();
const auto& regs = gpu.regs;
query_cache.UpdateCounters(); query_cache.UpdateCounters();
@ -529,7 +524,7 @@ void RasterizerOpenGL::Draw(bool is_indexed, bool is_instanced) {
// Upload vertex and index data. // Upload vertex and index data.
SetupVertexBuffer(); SetupVertexBuffer();
SetupVertexInstances(); SetupVertexInstances();
GLintptr index_buffer_offset; GLintptr index_buffer_offset = 0;
if (is_indexed) { if (is_indexed) {
index_buffer_offset = SetupIndexBuffer(); index_buffer_offset = SetupIndexBuffer();
} }
@ -555,7 +550,7 @@ void RasterizerOpenGL::Draw(bool is_indexed, bool is_instanced) {
ConfigureFramebuffers(); ConfigureFramebuffers();
// Signal the buffer cache that we are not going to upload more things. // Signal the buffer cache that we are not going to upload more things.
const bool invalidate = buffer_cache.Unmap(); buffer_cache.Unmap();
// Now that we are no longer uploading data, we can safely bind the buffers to OpenGL. // Now that we are no longer uploading data, we can safely bind the buffers to OpenGL.
vertex_array_pushbuffer.Bind(); vertex_array_pushbuffer.Bind();

View file

@ -393,10 +393,6 @@ std::string FlowStackTopName(MetaStackClass stack) {
return fmt::format("{}_flow_stack_top", GetFlowStackPrefix(stack)); return fmt::format("{}_flow_stack_top", GetFlowStackPrefix(stack));
} }
[[deprecated]] constexpr bool IsVertexShader(ShaderType stage) {
return stage == ShaderType::Vertex;
}
struct GenericVaryingDescription { struct GenericVaryingDescription {
std::string name; std::string name;
u8 first_element = 0; u8 first_element = 0;
@ -529,8 +525,9 @@ private:
} }
void DeclareVertex() { void DeclareVertex() {
if (!IsVertexShader(stage)) if (stage != ShaderType::Vertex) {
return; return;
}
DeclareVertexRedeclarations(); DeclareVertexRedeclarations();
} }
@ -602,14 +599,14 @@ private:
break; break;
} }
} }
if (!IsVertexShader(stage) || device.HasVertexViewportLayer()) { if (stage != ShaderType::Vertex || device.HasVertexViewportLayer()) {
if (ir.UsesLayer()) { if (ir.UsesLayer()) {
code.AddLine("int gl_Layer;"); code.AddLine("int gl_Layer;");
} }
if (ir.UsesViewportIndex()) { if (ir.UsesViewportIndex()) {
code.AddLine("int gl_ViewportIndex;"); code.AddLine("int gl_ViewportIndex;");
} }
} else if ((ir.UsesLayer() || ir.UsesViewportIndex()) && IsVertexShader(stage) && } else if ((ir.UsesLayer() || ir.UsesViewportIndex()) && stage == ShaderType::Vertex &&
!device.HasVertexViewportLayer()) { !device.HasVertexViewportLayer()) {
LOG_ERROR( LOG_ERROR(
Render_OpenGL, Render_OpenGL,
@ -1147,7 +1144,7 @@ private:
// TODO(Subv): Find out what the values are for the first two elements when inside a // TODO(Subv): Find out what the values are for the first two elements when inside a
// vertex shader, and what's the value of the fourth element when inside a Tess Eval // vertex shader, and what's the value of the fourth element when inside a Tess Eval
// shader. // shader.
ASSERT(IsVertexShader(stage)); ASSERT(stage == ShaderType::Vertex);
switch (element) { switch (element) {
case 2: case 2:
// Config pack's first value is instance_id. // Config pack's first value is instance_id.
@ -1218,12 +1215,12 @@ private:
UNIMPLEMENTED(); UNIMPLEMENTED();
return {}; return {};
case 1: case 1:
if (IsVertexShader(stage) && !device.HasVertexViewportLayer()) { if (stage == ShaderType::Vertex && !device.HasVertexViewportLayer()) {
return {}; return {};
} }
return {{"gl_Layer", Type::Int}}; return {{"gl_Layer", Type::Int}};
case 2: case 2:
if (IsVertexShader(stage) && !device.HasVertexViewportLayer()) { if (stage == ShaderType::Vertex && !device.HasVertexViewportLayer()) {
return {}; return {};
} }
return {{"gl_ViewportIndex", Type::Int}}; return {{"gl_ViewportIndex", Type::Int}};
@ -2532,7 +2529,7 @@ private:
} }
u32 GetNumPhysicalInputAttributes() const { u32 GetNumPhysicalInputAttributes() const {
return IsVertexShader(stage) ? GetNumPhysicalAttributes() : GetNumPhysicalVaryings(); return stage == ShaderType::Vertex ? GetNumPhysicalAttributes() : GetNumPhysicalVaryings();
} }
u32 GetNumPhysicalAttributes() const { u32 GetNumPhysicalAttributes() const {

View file

@ -257,6 +257,8 @@ vk::ShaderStageFlagBits ShaderStage(Tegra::Engines::ShaderType stage) {
return vk::ShaderStageFlagBits::eGeometry; return vk::ShaderStageFlagBits::eGeometry;
case Tegra::Engines::ShaderType::Fragment: case Tegra::Engines::ShaderType::Fragment:
return vk::ShaderStageFlagBits::eFragment; return vk::ShaderStageFlagBits::eFragment;
case Tegra::Engines::ShaderType::Compute:
return vk::ShaderStageFlagBits::eCompute;
} }
UNIMPLEMENTED_MSG("Unimplemented shader stage={}", static_cast<u32>(stage)); UNIMPLEMENTED_MSG("Unimplemented shader stage={}", static_cast<u32>(stage));
return {}; return {};

View file

@ -192,7 +192,6 @@ std::array<Shader, Maxwell::MaxShaderProgram> VKPipelineCache::GetShaders() {
std::array<Shader, Maxwell::MaxShaderProgram> shaders; std::array<Shader, Maxwell::MaxShaderProgram> shaders;
for (std::size_t index = 0; index < Maxwell::MaxShaderProgram; ++index) { for (std::size_t index = 0; index < Maxwell::MaxShaderProgram; ++index) {
const auto& shader_config = gpu.regs.shader_config[index];
const auto program{static_cast<Maxwell::ShaderProgram>(index)}; const auto program{static_cast<Maxwell::ShaderProgram>(index)};
// Skip stages that are not enabled // Skip stages that are not enabled

View file

@ -548,8 +548,6 @@ bool RasterizerVulkan::AccelerateDisplay(const Tegra::FramebufferConfig& config,
// Verify that the cached surface is the same size and format as the requested framebuffer // Verify that the cached surface is the same size and format as the requested framebuffer
const auto& params{surface->GetSurfaceParams()}; const auto& params{surface->GetSurfaceParams()};
const auto& pixel_format{
VideoCore::Surface::PixelFormatFromGPUPixelFormat(config.pixel_format)};
ASSERT_MSG(params.width == config.width, "Framebuffer width is different"); ASSERT_MSG(params.width == config.width, "Framebuffer width is different");
ASSERT_MSG(params.height == config.height, "Framebuffer height is different"); ASSERT_MSG(params.height == config.height, "Framebuffer height is different");

View file

@ -100,7 +100,6 @@ void VKStagingBufferPool::ReleaseCache(bool host_visible) {
} }
u64 VKStagingBufferPool::ReleaseLevel(StagingBuffersCache& cache, std::size_t log2) { u64 VKStagingBufferPool::ReleaseLevel(StagingBuffersCache& cache, std::size_t log2) {
static constexpr u64 epochs_to_destroy = 180;
static constexpr std::size_t deletions_per_tick = 16; static constexpr std::size_t deletions_per_tick = 16;
auto& staging = cache[log2]; auto& staging = cache[log2];
@ -108,6 +107,7 @@ u64 VKStagingBufferPool::ReleaseLevel(StagingBuffersCache& cache, std::size_t lo
const std::size_t old_size = entries.size(); const std::size_t old_size = entries.size();
const auto is_deleteable = [this](const auto& entry) { const auto is_deleteable = [this](const auto& entry) {
static constexpr u64 epochs_to_destroy = 180;
return entry.last_epoch + epochs_to_destroy < epoch && !entry.watch.IsUsed(); return entry.last_epoch + epochs_to_destroy < epoch && !entry.watch.IsUsed();
}; };
const std::size_t begin_offset = staging.delete_index; const std::size_t begin_offset = staging.delete_index;

View file

@ -52,6 +52,9 @@ vk::ImageType SurfaceTargetToImage(SurfaceTarget target) {
return vk::ImageType::e2D; return vk::ImageType::e2D;
case SurfaceTarget::Texture3D: case SurfaceTarget::Texture3D:
return vk::ImageType::e3D; return vk::ImageType::e3D;
case SurfaceTarget::TextureBuffer:
UNREACHABLE();
return {};
} }
UNREACHABLE_MSG("Unknown texture target={}", static_cast<u32>(target)); UNREACHABLE_MSG("Unknown texture target={}", static_cast<u32>(target));
return {}; return {};
@ -273,7 +276,6 @@ void CachedSurface::UploadImage(const std::vector<u8>& staging_buffer) {
for (u32 level = 0; level < params.num_levels; ++level) { for (u32 level = 0; level < params.num_levels; ++level) {
vk::BufferImageCopy copy = GetBufferImageCopy(level); vk::BufferImageCopy copy = GetBufferImageCopy(level);
const auto& dld = device.GetDispatchLoader();
if (image->GetAspectMask() == if (image->GetAspectMask() ==
(vk::ImageAspectFlagBits::eDepth | vk::ImageAspectFlagBits::eStencil)) { (vk::ImageAspectFlagBits::eDepth | vk::ImageAspectFlagBits::eStencil)) {
vk::BufferImageCopy depth = copy; vk::BufferImageCopy depth = copy;
@ -422,7 +424,6 @@ void VKTextureCache::ImageCopy(Surface& src_surface, Surface& dst_surface,
dst_base_layer, num_layers, copy_params.dest_level, 1, vk::PipelineStageFlagBits::eTransfer, dst_base_layer, num_layers, copy_params.dest_level, 1, vk::PipelineStageFlagBits::eTransfer,
vk::AccessFlagBits::eTransferWrite, vk::ImageLayout::eTransferDstOptimal); vk::AccessFlagBits::eTransferWrite, vk::ImageLayout::eTransferDstOptimal);
const auto& dld{device.GetDispatchLoader()};
const vk::ImageSubresourceLayers src_subresource( const vk::ImageSubresourceLayers src_subresource(
src_surface->GetAspectMask(), copy_params.source_level, copy_params.source_z, num_layers); src_surface->GetAspectMask(), copy_params.source_level, copy_params.source_z, num_layers);
const vk::ImageSubresourceLayers dst_subresource( const vk::ImageSubresourceLayers dst_subresource(
@ -458,7 +459,6 @@ void VKTextureCache::ImageBlit(View& src_view, View& dst_view,
dst_view->GetImageSubresourceLayers(), {dst_top_left, dst_bot_right}); dst_view->GetImageSubresourceLayers(), {dst_top_left, dst_bot_right});
const bool is_linear = copy_config.filter == Tegra::Engines::Fermi2D::Filter::Linear; const bool is_linear = copy_config.filter == Tegra::Engines::Fermi2D::Filter::Linear;
const auto& dld{device.GetDispatchLoader()};
scheduler.Record([src_image = src_view->GetImage(), dst_image = dst_view->GetImage(), blit, scheduler.Record([src_image = src_view->GetImage(), dst_image = dst_view->GetImage(), blit,
is_linear](auto cmdbuf, auto& dld) { is_linear](auto cmdbuf, auto& dld) {
cmdbuf.blitImage(src_image, vk::ImageLayout::eTransferSrcOptimal, dst_image, cmdbuf.blitImage(src_image, vk::ImageLayout::eTransferSrcOptimal, dst_image,