1
0
Fork 0
mirror of https://github.com/Ryujinx/Ryujinx.git synced 2024-11-14 11:56:51 +00:00
Ryujinx/Ryujinx.Graphics.Gpu/NvGpuFifo.cs

172 lines
4.7 KiB
C#
Raw Normal View History

2019-11-22 02:46:14 +00:00
using Ryujinx.Graphics.Gpu.State;
2019-10-13 07:02:07 +01:00
namespace Ryujinx.Graphics.Gpu
{
class NvGpuFifo
{
private const int MacrosCount = 0x80;
private const int MacroIndexMask = MacrosCount - 1;
// Note: The size of the macro memory is unknown, we just make
// a guess here and use 256kb as the size. Increase if needed.
private const int MmeWords = 256 * 256;
private GpuContext _context;
private struct CachedMacro
{
public int Position { get; private set; }
private bool _executionPending;
private int _argument;
private MacroInterpreter _interpreter;
public CachedMacro(GpuContext context, NvGpuFifo fifo, int position)
{
Position = position;
_executionPending = false;
_argument = 0;
_interpreter = new MacroInterpreter(context, fifo);
}
public void StartExecution(int argument)
{
_argument = argument;
_executionPending = true;
}
2019-11-22 02:46:14 +00:00
public void Execute(int[] mme, GpuState state)
2019-10-13 07:02:07 +01:00
{
if (_executionPending)
{
_executionPending = false;
2019-11-22 02:46:14 +00:00
_interpreter?.Execute(mme, Position, _argument, state);
2019-10-13 07:02:07 +01:00
}
}
public void PushArgument(int argument)
{
_interpreter?.Fifo.Enqueue(argument);
}
}
private int _currMacroPosition;
private int _currMacroBindIndex;
private CachedMacro[] _macros;
private int[] _mme;
2019-11-22 02:46:14 +00:00
private class SubChannel
{
public GpuState State { get; }
public ClassId Class { get; set; }
public SubChannel()
{
State = new GpuState();
}
}
private SubChannel[] _subChannels;
2019-10-13 07:02:07 +01:00
public NvGpuFifo(GpuContext context)
{
_context = context;
_macros = new CachedMacro[MacrosCount];
_mme = new int[MmeWords];
2019-11-22 02:46:14 +00:00
_subChannels = new SubChannel[8];
for (int index = 0; index < _subChannels.Length; index++)
{
_subChannels[index] = new SubChannel();
context.Methods.RegisterCallbacks(_subChannels[index].State);
}
2019-10-13 07:02:07 +01:00
}
public void CallMethod(MethodParams meth)
{
if ((NvGpuFifoMeth)meth.Method == NvGpuFifoMeth.BindChannel)
{
2019-11-22 02:46:14 +00:00
_subChannels[meth.SubChannel].Class = (ClassId)meth.Argument;
2019-10-13 07:02:07 +01:00
}
else if (meth.Method < 0x60)
{
switch ((NvGpuFifoMeth)meth.Method)
{
case NvGpuFifoMeth.WaitForIdle:
{
_context.Methods.PerformDeferredDraws();
2019-10-13 07:02:07 +01:00
_context.Renderer.FlushPipelines();
break;
}
case NvGpuFifoMeth.SetMacroUploadAddress:
{
_currMacroPosition = meth.Argument;
break;
}
case NvGpuFifoMeth.SendMacroCodeData:
{
_mme[_currMacroPosition++] = meth.Argument;
break;
}
case NvGpuFifoMeth.SetMacroBindingIndex:
{
_currMacroBindIndex = meth.Argument;
break;
}
case NvGpuFifoMeth.BindMacro:
{
int position = meth.Argument;
_macros[_currMacroBindIndex++] = new CachedMacro(_context, this, position);
break;
}
}
}
else if (meth.Method < 0xe00)
{
2019-11-22 02:46:14 +00:00
_subChannels[meth.SubChannel].State.CallMethod(meth);
2019-10-13 07:02:07 +01:00
}
else
{
int macroIndex = (meth.Method >> 1) & MacroIndexMask;
if ((meth.Method & 1) != 0)
{
_macros[macroIndex].PushArgument(meth.Argument);
}
else
{
_macros[macroIndex].StartExecution(meth.Argument);
}
if (meth.IsLastCall)
{
2019-11-22 02:46:14 +00:00
_macros[macroIndex].Execute(_mme, _subChannels[meth.SubChannel].State);
2019-10-13 07:02:07 +01:00
_context.Methods.PerformDeferredDraws();
}
}
}
}
}