early-access version 2824
This commit is contained in:
@@ -128,8 +128,8 @@ void AudioRenderer::CreateSinkStreams() {
|
||||
u32 channels{sink.GetDeviceChannels()};
|
||||
for (u32 i = 0; i < MaxRendererSessions; i++) {
|
||||
std::string name{fmt::format("ADSP_RenderStream-{}", i)};
|
||||
streams[i] = sink.AcquireSinkStream(system, channels, name,
|
||||
::AudioCore::Sink::StreamType::Render, &render_event);
|
||||
streams[i] =
|
||||
sink.AcquireSinkStream(system, channels, name, ::AudioCore::Sink::StreamType::Render);
|
||||
streams[i]->SetSystemChannels(streams[i]->GetDeviceChannels());
|
||||
}
|
||||
}
|
||||
@@ -199,12 +199,9 @@ void AudioRenderer::ThreadFunc() {
|
||||
command_list_processor.Process(index) - start_time;
|
||||
}
|
||||
|
||||
// If the stream queue is building up too much, wait for a signal
|
||||
// from the backend that a buffer was consumed.
|
||||
// In practice this will wait longer than 1 buffer due to timing.
|
||||
auto stream{command_list_processor.GetOutputSinkStream()};
|
||||
if (stream->GetQueueSize() >= 4) {
|
||||
render_event.WaitFor(std::chrono::milliseconds(5));
|
||||
if (index == 0) {
|
||||
auto stream{command_list_processor.GetOutputSinkStream()};
|
||||
system.AudioCore().SetStreamQueue(stream->GetQueueSize());
|
||||
}
|
||||
|
||||
const auto end_time{system.CoreTiming().GetClockTicks()};
|
||||
|
@@ -14,8 +14,11 @@
|
||||
#include "common/thread.h"
|
||||
|
||||
namespace Core {
|
||||
class System;
|
||||
namespace Timing {
|
||||
struct EventType;
|
||||
}
|
||||
class System;
|
||||
} // namespace Core
|
||||
|
||||
namespace AudioCore {
|
||||
namespace Sink {
|
||||
@@ -194,8 +197,6 @@ private:
|
||||
Sink::Sink& sink;
|
||||
/// The streams which will receive the processed samples
|
||||
std::array<Sink::SinkStream*, MaxRendererSessions> streams;
|
||||
/// An event signalled from the backend when a buffer is consumed, used for timing.
|
||||
Common::Event render_event{};
|
||||
};
|
||||
|
||||
} // namespace AudioRenderer::ADSP
|
||||
|
@@ -63,7 +63,8 @@ static void ApplyLightLimiterEffect(const LightLimiterInfo::ParameterVersion2& p
|
||||
|
||||
for (u32 sample_index = 0; sample_index < sample_count; sample_index++) {
|
||||
for (u32 channel = 0; channel < params.channel_count; channel++) {
|
||||
auto sample{Common::FixedPoint<49, 15>(inputs[channel][sample_index]) *
|
||||
auto sample{(Common::FixedPoint<49, 15>(inputs[channel][sample_index]) /
|
||||
Common::FixedPoint<49, 15>::one) *
|
||||
params.input_gain};
|
||||
auto abs_sample{sample};
|
||||
if (sample < 0.0f) {
|
||||
@@ -85,15 +86,17 @@ static void ApplyLightLimiterEffect(const LightLimiterInfo::ParameterVersion2& p
|
||||
auto lookahead_sample{
|
||||
state.look_ahead_sample_buffers[channel]
|
||||
[state.look_ahead_sample_offsets[channel]]};
|
||||
|
||||
state.look_ahead_sample_buffers[channel][state.look_ahead_sample_offsets[channel]] =
|
||||
sample;
|
||||
state.look_ahead_sample_offsets[channel] =
|
||||
(state.look_ahead_sample_offsets[channel] + 1) % params.look_ahead_samples_min;
|
||||
|
||||
outputs[channel][sample_index] = static_cast<s32>(std::clamp(
|
||||
(lookahead_sample * state.compression_gain[channel] * params.output_gain)
|
||||
.to_long(),
|
||||
min, max));
|
||||
outputs[channel][sample_index] = static_cast<s32>(
|
||||
std::clamp((lookahead_sample * state.compression_gain[channel] *
|
||||
params.output_gain * Common::FixedPoint<49, 15>::one)
|
||||
.to_long(),
|
||||
min, max));
|
||||
|
||||
if (statistics) {
|
||||
statistics->channel_max_sample[channel] =
|
||||
|
@@ -15,12 +15,17 @@ MICROPROFILE_DEFINE(Audio_RenderSystemManager, "Audio", "Render System Manager",
|
||||
MP_RGB(60, 19, 97));
|
||||
|
||||
namespace AudioCore::AudioRenderer {
|
||||
constexpr std::chrono::nanoseconds BaseRenderTime{5'000'000UL};
|
||||
constexpr std::chrono::nanoseconds RenderTimeOffset{400'000UL};
|
||||
|
||||
SystemManager::SystemManager(Core::System& core_)
|
||||
: core{core_}, adsp{core.AudioCore().GetADSP()}, mailbox{adsp.GetRenderMailbox()},
|
||||
thread_event{Core::Timing::CreateEvent(
|
||||
"AudioRendererSystemManager",
|
||||
[this](std::uintptr_t userdata, std::chrono::nanoseconds ns_late) { ThreadFunc2(); })} {}
|
||||
"AudioRendererSystemManager", [this](std::uintptr_t, s64 time, std::chrono::nanoseconds) {
|
||||
return ThreadFunc2(time);
|
||||
})} {
|
||||
core.CoreTiming().RegisterPauseCallback([this](bool paused) { PauseCallback(paused); });
|
||||
}
|
||||
|
||||
SystemManager::~SystemManager() {
|
||||
Stop();
|
||||
@@ -30,9 +35,9 @@ bool SystemManager::InitializeUnsafe() {
|
||||
if (!active) {
|
||||
if (adsp.Start()) {
|
||||
active = true;
|
||||
core.CoreTiming().ScheduleLoopingEvent(std::chrono::nanoseconds(2'304'000ULL * 2),
|
||||
thread_event);
|
||||
thread = std::jthread([this](std::stop_token stop_token) { ThreadFunc(); });
|
||||
core.CoreTiming().ScheduleLoopingEvent(std::chrono::nanoseconds(0),
|
||||
BaseRenderTime - RenderTimeOffset, thread_event);
|
||||
}
|
||||
}
|
||||
|
||||
@@ -95,11 +100,13 @@ void SystemManager::ThreadFunc() {
|
||||
constexpr char name[]{"yuzu:AudioRenderSystemManager"};
|
||||
MicroProfileOnThreadCreate(name);
|
||||
Common::SetCurrentThreadName(name);
|
||||
Common::SetCurrentThreadPriority(Common::ThreadPriority::Critical);
|
||||
Common::SetCurrentThreadPriority(Common::ThreadPriority::High);
|
||||
while (active) {
|
||||
{
|
||||
std::scoped_lock l{mutex1};
|
||||
|
||||
MICROPROFILE_SCOPE(Audio_RenderSystemManager);
|
||||
|
||||
for (auto system : systems) {
|
||||
system->SendCommandToDsp();
|
||||
}
|
||||
@@ -113,9 +120,43 @@ void SystemManager::ThreadFunc() {
|
||||
}
|
||||
}
|
||||
|
||||
void SystemManager::ThreadFunc2() {
|
||||
std::optional<std::chrono::nanoseconds> SystemManager::ThreadFunc2(s64 time) {
|
||||
std::optional<std::chrono::nanoseconds> new_schedule_time{std::nullopt};
|
||||
const auto queue_size{core.AudioCore().GetStreamQueue()};
|
||||
switch (state) {
|
||||
case StreamState::Filling:
|
||||
if (queue_size >= 5) {
|
||||
new_schedule_time = BaseRenderTime;
|
||||
state = StreamState::Steady;
|
||||
}
|
||||
break;
|
||||
case StreamState::Steady:
|
||||
if (queue_size <= 2) {
|
||||
new_schedule_time = BaseRenderTime - RenderTimeOffset;
|
||||
state = StreamState::Filling;
|
||||
} else if (queue_size > 5) {
|
||||
new_schedule_time = BaseRenderTime + RenderTimeOffset;
|
||||
state = StreamState::Draining;
|
||||
}
|
||||
break;
|
||||
case StreamState::Draining:
|
||||
if (queue_size <= 5) {
|
||||
new_schedule_time = BaseRenderTime;
|
||||
state = StreamState::Steady;
|
||||
}
|
||||
break;
|
||||
}
|
||||
|
||||
update.store(true);
|
||||
update.notify_all();
|
||||
return new_schedule_time;
|
||||
}
|
||||
|
||||
void SystemManager::PauseCallback(bool paused) {
|
||||
if (paused && core.IsPoweredOn() && core.IsShuttingDown()) {
|
||||
update.store(true);
|
||||
update.notify_all();
|
||||
}
|
||||
}
|
||||
|
||||
} // namespace AudioCore::AudioRenderer
|
||||
|
@@ -6,6 +6,7 @@
|
||||
#include <list>
|
||||
#include <memory>
|
||||
#include <mutex>
|
||||
#include <optional>
|
||||
#include <thread>
|
||||
|
||||
#include "audio_core/renderer/system.h"
|
||||
@@ -70,7 +71,20 @@ private:
|
||||
/**
|
||||
* Signalling core timing thread to run ThreadFunc.
|
||||
*/
|
||||
void ThreadFunc2();
|
||||
std::optional<std::chrono::nanoseconds> ThreadFunc2(s64 time);
|
||||
|
||||
/**
|
||||
* Callback from core timing when pausing, used to detect shutdowns and stop ThreadFunc.
|
||||
*
|
||||
* @param paused - Are we pausing or resuming?
|
||||
*/
|
||||
void PauseCallback(bool paused);
|
||||
|
||||
enum class StreamState {
|
||||
Filling,
|
||||
Steady,
|
||||
Draining,
|
||||
};
|
||||
|
||||
/// Core system
|
||||
Core::System& core;
|
||||
@@ -92,6 +106,8 @@ private:
|
||||
std::shared_ptr<Core::Timing::EventType> thread_event;
|
||||
/// Atomic for main thread to wait on
|
||||
std::atomic<bool> update{};
|
||||
/// Current state of the streams
|
||||
StreamState state{StreamState::Filling};
|
||||
};
|
||||
|
||||
} // namespace AudioCore::AudioRenderer
|
||||
|
Reference in New Issue
Block a user