DMAPusher: Propagate multimethod writes into the engines.
This commit is contained in:
parent
eb26e9e711
commit
3fedcc2f6e
14 changed files with 163 additions and 13 deletions
|
@ -71,16 +71,22 @@ bool DmaPusher::Step() {
|
||||||
gpu.MemoryManager().ReadBlockUnsafe(dma_get, command_headers.data(),
|
gpu.MemoryManager().ReadBlockUnsafe(dma_get, command_headers.data(),
|
||||||
command_list_header.size * sizeof(u32));
|
command_list_header.size * sizeof(u32));
|
||||||
|
|
||||||
for (const CommandHeader& command_header : command_headers) {
|
for (std::size_t index = 0; index < command_headers.size();) {
|
||||||
|
const CommandHeader& command_header = command_headers[index];
|
||||||
|
|
||||||
// now, see if we're in the middle of a command
|
if (dma_state.method_count) {
|
||||||
if (dma_state.length_pending) {
|
|
||||||
// Second word of long non-inc methods command - method count
|
|
||||||
dma_state.length_pending = 0;
|
|
||||||
dma_state.method_count = command_header.method_count_;
|
|
||||||
} else if (dma_state.method_count) {
|
|
||||||
// Data word of methods command
|
// Data word of methods command
|
||||||
|
if (dma_state.non_incrementing) {
|
||||||
|
const u32 max_write = static_cast<u32>(
|
||||||
|
std::min<std::size_t>(index + dma_state.method_count, command_headers.size()) -
|
||||||
|
index);
|
||||||
|
CallMultiMethod(&command_header.argument, max_write);
|
||||||
|
dma_state.method_count -= max_write;
|
||||||
|
index += max_write;
|
||||||
|
continue;
|
||||||
|
} else {
|
||||||
CallMethod(command_header.argument);
|
CallMethod(command_header.argument);
|
||||||
|
}
|
||||||
|
|
||||||
if (!dma_state.non_incrementing) {
|
if (!dma_state.non_incrementing) {
|
||||||
dma_state.method++;
|
dma_state.method++;
|
||||||
|
@ -120,6 +126,7 @@ bool DmaPusher::Step() {
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
index++;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!non_main) {
|
if (!non_main) {
|
||||||
|
@ -140,4 +147,9 @@ void DmaPusher::CallMethod(u32 argument) const {
|
||||||
gpu.CallMethod({dma_state.method, argument, dma_state.subchannel, dma_state.method_count});
|
gpu.CallMethod({dma_state.method, argument, dma_state.subchannel, dma_state.method_count});
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void DmaPusher::CallMultiMethod(const u32* base_start, u32 num_methods) const {
|
||||||
|
gpu.CallMultiMethod(dma_state.method, dma_state.subchannel, base_start, num_methods,
|
||||||
|
dma_state.method_count);
|
||||||
|
}
|
||||||
|
|
||||||
} // namespace Tegra
|
} // namespace Tegra
|
||||||
|
|
|
@ -75,6 +75,7 @@ private:
|
||||||
void SetState(const CommandHeader& command_header);
|
void SetState(const CommandHeader& command_header);
|
||||||
|
|
||||||
void CallMethod(u32 argument) const;
|
void CallMethod(u32 argument) const;
|
||||||
|
void CallMultiMethod(const u32* base_start, u32 num_methods) const;
|
||||||
|
|
||||||
std::vector<CommandHeader> command_headers; ///< Buffer for list of commands fetched at once
|
std::vector<CommandHeader> command_headers; ///< Buffer for list of commands fetched at once
|
||||||
|
|
||||||
|
|
|
@ -28,6 +28,12 @@ void Fermi2D::CallMethod(const GPU::MethodCall& method_call) {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void Fermi2D::CallMultiMethod(u32 method, const u32* base_start, u32 amount, u32 methods_pending) {
|
||||||
|
for (std::size_t i = 0; i < amount; i++) {
|
||||||
|
CallMethod({method, base_start[i], 0, methods_pending - static_cast<u32>(i)});
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
static std::pair<u32, u32> DelimitLine(u32 src_1, u32 src_2, u32 dst_1, u32 dst_2, u32 src_line) {
|
static std::pair<u32, u32> DelimitLine(u32 src_1, u32 src_2, u32 dst_1, u32 dst_2, u32 src_line) {
|
||||||
const u32 line_a = src_2 - src_1;
|
const u32 line_a = src_2 - src_1;
|
||||||
const u32 line_b = dst_2 - dst_1;
|
const u32 line_b = dst_2 - dst_1;
|
||||||
|
|
|
@ -39,6 +39,9 @@ public:
|
||||||
/// Write the value to the register identified by method.
|
/// Write the value to the register identified by method.
|
||||||
void CallMethod(const GPU::MethodCall& method_call);
|
void CallMethod(const GPU::MethodCall& method_call);
|
||||||
|
|
||||||
|
/// Write multiple values to the register identified by method.
|
||||||
|
void CallMultiMethod(u32 method, const u32* base_start, u32 amount, u32 methods_pending);
|
||||||
|
|
||||||
enum class Origin : u32 {
|
enum class Origin : u32 {
|
||||||
Center = 0,
|
Center = 0,
|
||||||
Corner = 1,
|
Corner = 1,
|
||||||
|
|
|
@ -51,6 +51,12 @@ void KeplerCompute::CallMethod(const GPU::MethodCall& method_call) {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void KeplerCompute::CallMultiMethod(u32 method, const u32* base_start, u32 amount, u32 methods_pending) {
|
||||||
|
for (std::size_t i = 0; i < amount; i++) {
|
||||||
|
CallMethod({method, base_start[i], 0, methods_pending - static_cast<u32>(i)});
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
Texture::FullTextureInfo KeplerCompute::GetTexture(std::size_t offset) const {
|
Texture::FullTextureInfo KeplerCompute::GetTexture(std::size_t offset) const {
|
||||||
const std::bitset<8> cbuf_mask = launch_description.const_buffer_enable_mask.Value();
|
const std::bitset<8> cbuf_mask = launch_description.const_buffer_enable_mask.Value();
|
||||||
ASSERT(cbuf_mask[regs.tex_cb_index]);
|
ASSERT(cbuf_mask[regs.tex_cb_index]);
|
||||||
|
|
|
@ -202,6 +202,9 @@ public:
|
||||||
/// Write the value to the register identified by method.
|
/// Write the value to the register identified by method.
|
||||||
void CallMethod(const GPU::MethodCall& method_call);
|
void CallMethod(const GPU::MethodCall& method_call);
|
||||||
|
|
||||||
|
/// Write multiple values to the register identified by method.
|
||||||
|
void CallMultiMethod(u32 method, const u32* base_start, u32 amount, u32 methods_pending);
|
||||||
|
|
||||||
Texture::FullTextureInfo GetTexture(std::size_t offset) const;
|
Texture::FullTextureInfo GetTexture(std::size_t offset) const;
|
||||||
|
|
||||||
/// Given a texture handle, returns the TSC and TIC entries.
|
/// Given a texture handle, returns the TSC and TIC entries.
|
||||||
|
|
|
@ -41,4 +41,10 @@ void KeplerMemory::CallMethod(const GPU::MethodCall& method_call) {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void KeplerMemory::CallMultiMethod(u32 method, const u32* base_start, u32 amount, u32 methods_pending) {
|
||||||
|
for (std::size_t i = 0; i < amount; i++) {
|
||||||
|
CallMethod({method, base_start[i], 0, methods_pending - static_cast<u32>(i)});
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
} // namespace Tegra::Engines
|
} // namespace Tegra::Engines
|
||||||
|
|
|
@ -40,6 +40,9 @@ public:
|
||||||
/// Write the value to the register identified by method.
|
/// Write the value to the register identified by method.
|
||||||
void CallMethod(const GPU::MethodCall& method_call);
|
void CallMethod(const GPU::MethodCall& method_call);
|
||||||
|
|
||||||
|
/// Write multiple values to the register identified by method.
|
||||||
|
void CallMultiMethod(u32 method, const u32* base_start, u32 amount, u32 methods_pending);
|
||||||
|
|
||||||
struct Regs {
|
struct Regs {
|
||||||
static constexpr size_t NUM_REGS = 0x7F;
|
static constexpr size_t NUM_REGS = 0x7F;
|
||||||
|
|
||||||
|
|
|
@ -280,6 +280,36 @@ void Maxwell3D::CallMethod(const GPU::MethodCall& method_call) {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void Maxwell3D::CallMultiMethod(u32 method, const u32* base_start, u32 amount, u32 methods_pending) {
|
||||||
|
switch (method) {
|
||||||
|
case MAXWELL3D_REG_INDEX(const_buffer.cb_data[0]):
|
||||||
|
case MAXWELL3D_REG_INDEX(const_buffer.cb_data[1]):
|
||||||
|
case MAXWELL3D_REG_INDEX(const_buffer.cb_data[2]):
|
||||||
|
case MAXWELL3D_REG_INDEX(const_buffer.cb_data[3]):
|
||||||
|
case MAXWELL3D_REG_INDEX(const_buffer.cb_data[4]):
|
||||||
|
case MAXWELL3D_REG_INDEX(const_buffer.cb_data[5]):
|
||||||
|
case MAXWELL3D_REG_INDEX(const_buffer.cb_data[6]):
|
||||||
|
case MAXWELL3D_REG_INDEX(const_buffer.cb_data[7]):
|
||||||
|
case MAXWELL3D_REG_INDEX(const_buffer.cb_data[8]):
|
||||||
|
case MAXWELL3D_REG_INDEX(const_buffer.cb_data[9]):
|
||||||
|
case MAXWELL3D_REG_INDEX(const_buffer.cb_data[10]):
|
||||||
|
case MAXWELL3D_REG_INDEX(const_buffer.cb_data[11]):
|
||||||
|
case MAXWELL3D_REG_INDEX(const_buffer.cb_data[12]):
|
||||||
|
case MAXWELL3D_REG_INDEX(const_buffer.cb_data[13]):
|
||||||
|
case MAXWELL3D_REG_INDEX(const_buffer.cb_data[14]):
|
||||||
|
case MAXWELL3D_REG_INDEX(const_buffer.cb_data[15]): {
|
||||||
|
ProcessCBMultiData(method, base_start, amount);
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
default: {
|
||||||
|
for (std::size_t i = 0; i < amount; i++) {
|
||||||
|
CallMethod({method, base_start[i], 0, methods_pending - static_cast<u32>(i)});
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
void Maxwell3D::StepInstance(const MMEDrawMode expected_mode, const u32 count) {
|
void Maxwell3D::StepInstance(const MMEDrawMode expected_mode, const u32 count) {
|
||||||
if (mme_draw.current_mode == MMEDrawMode::Undefined) {
|
if (mme_draw.current_mode == MMEDrawMode::Undefined) {
|
||||||
if (mme_draw.gl_begin_consume) {
|
if (mme_draw.gl_begin_consume) {
|
||||||
|
@ -570,6 +600,28 @@ void Maxwell3D::StartCBData(u32 method) {
|
||||||
ProcessCBData(regs.const_buffer.cb_data[cb_data_state.id]);
|
ProcessCBData(regs.const_buffer.cb_data[cb_data_state.id]);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void Maxwell3D::ProcessCBMultiData(u32 method, const u32* start_base, u32 amount) {
|
||||||
|
if (cb_data_state.current != method) {
|
||||||
|
if (cb_data_state.current != null_cb_data) {
|
||||||
|
FinishCBData();
|
||||||
|
}
|
||||||
|
constexpr u32 first_cb_data = MAXWELL3D_REG_INDEX(const_buffer.cb_data[0]);
|
||||||
|
cb_data_state.start_pos = regs.const_buffer.cb_pos;
|
||||||
|
cb_data_state.id = method - first_cb_data;
|
||||||
|
cb_data_state.current = method;
|
||||||
|
cb_data_state.counter = 0;
|
||||||
|
}
|
||||||
|
const std::size_t id = cb_data_state.id;
|
||||||
|
const std::size_t size = amount;
|
||||||
|
std::size_t i = 0;
|
||||||
|
for (; i < size; i++) {
|
||||||
|
cb_data_state.buffer[id][cb_data_state.counter] = start_base[i];
|
||||||
|
cb_data_state.counter++;
|
||||||
|
}
|
||||||
|
// Increment the current buffer position.
|
||||||
|
regs.const_buffer.cb_pos = regs.const_buffer.cb_pos + 4 * amount;
|
||||||
|
}
|
||||||
|
|
||||||
void Maxwell3D::FinishCBData() {
|
void Maxwell3D::FinishCBData() {
|
||||||
// Write the input value to the current const buffer at the current position.
|
// Write the input value to the current const buffer at the current position.
|
||||||
const GPUVAddr buffer_address = regs.const_buffer.BufferAddress();
|
const GPUVAddr buffer_address = regs.const_buffer.BufferAddress();
|
||||||
|
|
|
@ -1358,6 +1358,9 @@ public:
|
||||||
/// Write the value to the register identified by method.
|
/// Write the value to the register identified by method.
|
||||||
void CallMethod(const GPU::MethodCall& method_call);
|
void CallMethod(const GPU::MethodCall& method_call);
|
||||||
|
|
||||||
|
/// Write multiple values to the register identified by method.
|
||||||
|
void CallMultiMethod(u32 method, const u32* base_start, u32 amount, u32 methods_pending);
|
||||||
|
|
||||||
/// Write the value to the register identified by method.
|
/// Write the value to the register identified by method.
|
||||||
void CallMethodFromMME(const GPU::MethodCall& method_call);
|
void CallMethodFromMME(const GPU::MethodCall& method_call);
|
||||||
|
|
||||||
|
@ -1511,6 +1514,7 @@ private:
|
||||||
/// Handles a write to the CB_DATA[i] register.
|
/// Handles a write to the CB_DATA[i] register.
|
||||||
void StartCBData(u32 method);
|
void StartCBData(u32 method);
|
||||||
void ProcessCBData(u32 value);
|
void ProcessCBData(u32 value);
|
||||||
|
void ProcessCBMultiData(u32 method, const u32* start_base, u32 amount);
|
||||||
void FinishCBData();
|
void FinishCBData();
|
||||||
|
|
||||||
/// Handles a write to the CB_BIND register.
|
/// Handles a write to the CB_BIND register.
|
||||||
|
|
|
@ -36,6 +36,12 @@ void MaxwellDMA::CallMethod(const GPU::MethodCall& method_call) {
|
||||||
#undef MAXWELLDMA_REG_INDEX
|
#undef MAXWELLDMA_REG_INDEX
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void MaxwellDMA::CallMultiMethod(u32 method, const u32* base_start, u32 amount, u32 methods_pending) {
|
||||||
|
for (std::size_t i = 0; i < amount; i++) {
|
||||||
|
CallMethod({method, base_start[i], 0, methods_pending - static_cast<u32>(i)});
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
void MaxwellDMA::HandleCopy() {
|
void MaxwellDMA::HandleCopy() {
|
||||||
LOG_TRACE(HW_GPU, "Requested a DMA copy");
|
LOG_TRACE(HW_GPU, "Requested a DMA copy");
|
||||||
|
|
||||||
|
|
|
@ -35,6 +35,9 @@ public:
|
||||||
/// Write the value to the register identified by method.
|
/// Write the value to the register identified by method.
|
||||||
void CallMethod(const GPU::MethodCall& method_call);
|
void CallMethod(const GPU::MethodCall& method_call);
|
||||||
|
|
||||||
|
/// Write multiple values to the register identified by method.
|
||||||
|
void CallMultiMethod(u32 method, const u32* base_start, u32 amount, u32 methods_pending);
|
||||||
|
|
||||||
struct Regs {
|
struct Regs {
|
||||||
static constexpr std::size_t NUM_REGS = 0x1D6;
|
static constexpr std::size_t NUM_REGS = 0x1D6;
|
||||||
|
|
||||||
|
|
|
@ -209,16 +209,31 @@ void GPU::CallMethod(const MethodCall& method_call) {
|
||||||
|
|
||||||
ASSERT(method_call.subchannel < bound_engines.size());
|
ASSERT(method_call.subchannel < bound_engines.size());
|
||||||
|
|
||||||
if (ExecuteMethodOnEngine(method_call)) {
|
if (ExecuteMethodOnEngine(method_call.method)) {
|
||||||
CallEngineMethod(method_call);
|
CallEngineMethod(method_call);
|
||||||
} else {
|
} else {
|
||||||
CallPullerMethod(method_call);
|
CallPullerMethod(method_call);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
bool GPU::ExecuteMethodOnEngine(const MethodCall& method_call) {
|
void GPU::CallMultiMethod(u32 method, u32 subchannel, const u32* base_start, u32 amount, u32 methods_pending) {
|
||||||
const auto method = static_cast<BufferMethods>(method_call.method);
|
LOG_TRACE(HW_GPU, "Processing method {:08X} on subchannel {}", method,
|
||||||
return method >= BufferMethods::NonPullerMethods;
|
subchannel);
|
||||||
|
|
||||||
|
ASSERT(subchannel < bound_engines.size());
|
||||||
|
|
||||||
|
if (ExecuteMethodOnEngine(method)) {
|
||||||
|
CallEngineMultiMethod(method, subchannel, base_start, amount, methods_pending);
|
||||||
|
} else {
|
||||||
|
for (std::size_t i = 0; i < amount; i++) {
|
||||||
|
CallPullerMethod({method, base_start[i], subchannel, methods_pending - static_cast<u32>(i)});
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
bool GPU::ExecuteMethodOnEngine(u32 method) {
|
||||||
|
const auto buffer_method = static_cast<BufferMethods>(method);
|
||||||
|
return buffer_method >= BufferMethods::NonPullerMethods;
|
||||||
}
|
}
|
||||||
|
|
||||||
void GPU::CallPullerMethod(const MethodCall& method_call) {
|
void GPU::CallPullerMethod(const MethodCall& method_call) {
|
||||||
|
@ -298,6 +313,30 @@ void GPU::CallEngineMethod(const MethodCall& method_call) {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void GPU::CallEngineMultiMethod(u32 method, u32 subchannel, const u32* base_start, u32 amount, u32 methods_pending) {
|
||||||
|
const EngineID engine = bound_engines[subchannel];
|
||||||
|
|
||||||
|
switch (engine) {
|
||||||
|
case EngineID::FERMI_TWOD_A:
|
||||||
|
fermi_2d->CallMultiMethod(method, base_start, amount, methods_pending);
|
||||||
|
break;
|
||||||
|
case EngineID::MAXWELL_B:
|
||||||
|
maxwell_3d->CallMultiMethod(method, base_start, amount, methods_pending);
|
||||||
|
break;
|
||||||
|
case EngineID::KEPLER_COMPUTE_B:
|
||||||
|
kepler_compute->CallMultiMethod(method, base_start, amount, methods_pending);
|
||||||
|
break;
|
||||||
|
case EngineID::MAXWELL_DMA_COPY_A:
|
||||||
|
maxwell_dma->CallMultiMethod(method, base_start, amount, methods_pending);
|
||||||
|
break;
|
||||||
|
case EngineID::KEPLER_INLINE_TO_MEMORY_B:
|
||||||
|
kepler_memory->CallMultiMethod(method, base_start, amount, methods_pending);
|
||||||
|
break;
|
||||||
|
default:
|
||||||
|
UNIMPLEMENTED_MSG("Unimplemented engine");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
void GPU::ProcessBindMethod(const MethodCall& method_call) {
|
void GPU::ProcessBindMethod(const MethodCall& method_call) {
|
||||||
// Bind the current subchannel to the desired engine id.
|
// Bind the current subchannel to the desired engine id.
|
||||||
LOG_DEBUG(HW_GPU, "Binding subchannel {} to engine {}", method_call.subchannel,
|
LOG_DEBUG(HW_GPU, "Binding subchannel {} to engine {}", method_call.subchannel,
|
||||||
|
|
|
@ -155,6 +155,9 @@ public:
|
||||||
/// Calls a GPU method.
|
/// Calls a GPU method.
|
||||||
void CallMethod(const MethodCall& method_call);
|
void CallMethod(const MethodCall& method_call);
|
||||||
|
|
||||||
|
/// Calls a GPU multivalue method.
|
||||||
|
void CallMultiMethod(u32 method, u32 subchannel, const u32* base_start, u32 amount, u32 methods_pending);
|
||||||
|
|
||||||
/// Flush all current written commands into the host GPU for execution.
|
/// Flush all current written commands into the host GPU for execution.
|
||||||
void FlushCommands();
|
void FlushCommands();
|
||||||
/// Synchronizes CPU writes with Host GPU memory.
|
/// Synchronizes CPU writes with Host GPU memory.
|
||||||
|
@ -309,8 +312,11 @@ private:
|
||||||
/// Calls a GPU engine method.
|
/// Calls a GPU engine method.
|
||||||
void CallEngineMethod(const MethodCall& method_call);
|
void CallEngineMethod(const MethodCall& method_call);
|
||||||
|
|
||||||
|
/// Calls a GPU engine multivalue method.
|
||||||
|
void CallEngineMultiMethod(u32 method, u32 subchannel, const u32* base_start, u32 amount, u32 methods_pending);
|
||||||
|
|
||||||
/// Determines where the method should be executed.
|
/// Determines where the method should be executed.
|
||||||
bool ExecuteMethodOnEngine(const MethodCall& method_call);
|
bool ExecuteMethodOnEngine(u32 method);
|
||||||
|
|
||||||
protected:
|
protected:
|
||||||
std::unique_ptr<Tegra::DmaPusher> dma_pusher;
|
std::unique_ptr<Tegra::DmaPusher> dma_pusher;
|
||||||
|
|
Loading…
Reference in a new issue