early-access version 3452
This commit is contained in:
parent
1d9cd8d5e7
commit
5e818ba7dd
@ -1,7 +1,7 @@
|
|||||||
yuzu emulator early access
|
yuzu emulator early access
|
||||||
=============
|
=============
|
||||||
|
|
||||||
This is the source code for early-access 3451.
|
This is the source code for early-access 3452.
|
||||||
|
|
||||||
## Legal Notice
|
## Legal Notice
|
||||||
|
|
||||||
|
@ -47,14 +47,15 @@ Scheduler::Scheduler(const Device& device_, StateTracker& state_tracker_)
|
|||||||
Scheduler::~Scheduler() = default;
|
Scheduler::~Scheduler() = default;
|
||||||
|
|
||||||
void Scheduler::Flush(VkSemaphore signal_semaphore, VkSemaphore wait_semaphore) {
|
void Scheduler::Flush(VkSemaphore signal_semaphore, VkSemaphore wait_semaphore) {
|
||||||
|
// When flushing, we only send data to the worker thread; no waiting is necessary.
|
||||||
SubmitExecution(signal_semaphore, wait_semaphore);
|
SubmitExecution(signal_semaphore, wait_semaphore);
|
||||||
AllocateNewContext();
|
AllocateNewContext();
|
||||||
}
|
}
|
||||||
|
|
||||||
void Scheduler::Finish(VkSemaphore signal_semaphore, VkSemaphore wait_semaphore) {
|
void Scheduler::Finish(VkSemaphore signal_semaphore, VkSemaphore wait_semaphore) {
|
||||||
|
// When finishing, we need to wait for the submission to have executed on the device.
|
||||||
const u64 presubmit_tick = CurrentTick();
|
const u64 presubmit_tick = CurrentTick();
|
||||||
SubmitExecution(signal_semaphore, wait_semaphore);
|
SubmitExecution(signal_semaphore, wait_semaphore);
|
||||||
WaitWorker();
|
|
||||||
Wait(presubmit_tick);
|
Wait(presubmit_tick);
|
||||||
AllocateNewContext();
|
AllocateNewContext();
|
||||||
}
|
}
|
||||||
@ -63,8 +64,13 @@ void Scheduler::WaitWorker() {
|
|||||||
MICROPROFILE_SCOPE(Vulkan_WaitForWorker);
|
MICROPROFILE_SCOPE(Vulkan_WaitForWorker);
|
||||||
DispatchWork();
|
DispatchWork();
|
||||||
|
|
||||||
std::unique_lock lock{work_mutex};
|
// Ensure the queue is drained.
|
||||||
wait_cv.wait(lock, [this] { return work_queue.empty(); });
|
std::unique_lock ql{queue_mutex};
|
||||||
|
event_cv.wait(ql, [this] { return work_queue.empty(); });
|
||||||
|
|
||||||
|
// Now wait for execution to finish.
|
||||||
|
// This needs to be done in the same order as WorkerThread.
|
||||||
|
std::unique_lock el{execution_mutex};
|
||||||
}
|
}
|
||||||
|
|
||||||
void Scheduler::DispatchWork() {
|
void Scheduler::DispatchWork() {
|
||||||
@ -72,10 +78,10 @@ void Scheduler::DispatchWork() {
|
|||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
{
|
{
|
||||||
std::scoped_lock lock{work_mutex};
|
std::scoped_lock ql{queue_mutex};
|
||||||
work_queue.push(std::move(chunk));
|
work_queue.push(std::move(chunk));
|
||||||
}
|
}
|
||||||
work_cv.notify_one();
|
event_cv.notify_all();
|
||||||
AcquireNewChunk();
|
AcquireNewChunk();
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -137,30 +143,55 @@ bool Scheduler::UpdateRescaling(bool is_rescaling) {
|
|||||||
|
|
||||||
void Scheduler::WorkerThread(std::stop_token stop_token) {
|
void Scheduler::WorkerThread(std::stop_token stop_token) {
|
||||||
Common::SetCurrentThreadName("VulkanWorker");
|
Common::SetCurrentThreadName("VulkanWorker");
|
||||||
do {
|
|
||||||
std::unique_ptr<CommandChunk> work;
|
const auto TryPopQueue{[this](auto& work) -> bool {
|
||||||
bool has_submit{false};
|
|
||||||
{
|
|
||||||
std::unique_lock lock{work_mutex};
|
|
||||||
if (work_queue.empty()) {
|
if (work_queue.empty()) {
|
||||||
wait_cv.notify_all();
|
return false;
|
||||||
}
|
|
||||||
Common::CondvarWait(work_cv, lock, stop_token, [&] { return !work_queue.empty(); });
|
|
||||||
if (stop_token.stop_requested()) {
|
|
||||||
continue;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
work = std::move(work_queue.front());
|
work = std::move(work_queue.front());
|
||||||
work_queue.pop();
|
work_queue.pop();
|
||||||
|
event_cv.notify_all();
|
||||||
|
return true;
|
||||||
|
}};
|
||||||
|
|
||||||
has_submit = work->HasSubmit();
|
while (!stop_token.stop_requested()) {
|
||||||
work->ExecuteAll(current_cmdbuf);
|
std::unique_ptr<CommandChunk> work;
|
||||||
|
|
||||||
|
{
|
||||||
|
std::unique_lock lk{queue_mutex};
|
||||||
|
|
||||||
|
// Wait for work.
|
||||||
|
Common::CondvarWait(event_cv, lk, stop_token, [&] { return TryPopQueue(work); });
|
||||||
|
|
||||||
|
// If we've been asked to stop, we're done.
|
||||||
|
if (stop_token.stop_requested()) {
|
||||||
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Exchange lock ownership so that we take the execution lock before
|
||||||
|
// the queue lock goes out of scope. This allows us to force execution
|
||||||
|
// to complete in the next step.
|
||||||
|
std::exchange(lk, std::unique_lock{execution_mutex});
|
||||||
|
|
||||||
|
// Perform the work, tracking whether the chunk was a submission
|
||||||
|
// before executing.
|
||||||
|
const bool has_submit = work->HasSubmit();
|
||||||
|
work->ExecuteAll(current_cmdbuf);
|
||||||
|
|
||||||
|
// If the chunk was a submission, reallocate the command buffer.
|
||||||
if (has_submit) {
|
if (has_submit) {
|
||||||
AllocateWorkerCommandBuffer();
|
AllocateWorkerCommandBuffer();
|
||||||
}
|
}
|
||||||
std::scoped_lock reserve_lock{reserve_mutex};
|
}
|
||||||
chunk_reserve.push_back(std::move(work));
|
|
||||||
} while (!stop_token.stop_requested());
|
{
|
||||||
|
std::scoped_lock rl{reserve_mutex};
|
||||||
|
|
||||||
|
// Recycle the chunk back to the reserve.
|
||||||
|
chunk_reserve.emplace_back(std::move(work));
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void Scheduler::AllocateWorkerCommandBuffer() {
|
void Scheduler::AllocateWorkerCommandBuffer() {
|
||||||
@ -289,13 +320,16 @@ void Scheduler::EndRenderPass() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
void Scheduler::AcquireNewChunk() {
|
void Scheduler::AcquireNewChunk() {
|
||||||
std::scoped_lock lock{reserve_mutex};
|
std::scoped_lock rl{reserve_mutex};
|
||||||
|
|
||||||
if (chunk_reserve.empty()) {
|
if (chunk_reserve.empty()) {
|
||||||
|
// If we don't have anything reserved, we need to make a new chunk.
|
||||||
|
chunk = std::make_unique<CommandChunk>();
|
||||||
|
} else {
|
||||||
|
// Otherwise, we can just take from the reserve.
|
||||||
chunk = std::make_unique<CommandChunk>();
|
chunk = std::make_unique<CommandChunk>();
|
||||||
return;
|
|
||||||
}
|
|
||||||
chunk = std::move(chunk_reserve.back());
|
|
||||||
chunk_reserve.pop_back();
|
chunk_reserve.pop_back();
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
|
||||||
} // namespace Vulkan
|
} // namespace Vulkan
|
||||||
|
@ -232,10 +232,10 @@ private:
|
|||||||
|
|
||||||
std::queue<std::unique_ptr<CommandChunk>> work_queue;
|
std::queue<std::unique_ptr<CommandChunk>> work_queue;
|
||||||
std::vector<std::unique_ptr<CommandChunk>> chunk_reserve;
|
std::vector<std::unique_ptr<CommandChunk>> chunk_reserve;
|
||||||
|
std::mutex execution_mutex;
|
||||||
std::mutex reserve_mutex;
|
std::mutex reserve_mutex;
|
||||||
std::mutex work_mutex;
|
std::mutex queue_mutex;
|
||||||
std::condition_variable_any work_cv;
|
std::condition_variable_any event_cv;
|
||||||
std::condition_variable wait_cv;
|
|
||||||
std::jthread worker_thread;
|
std::jthread worker_thread;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -1306,6 +1306,7 @@ ImageId TextureCache<P>::JoinImages(const ImageInfo& info, GPUVAddr gpu_addr, VA
|
|||||||
auto copies = MakeShrinkImageCopies(new_info, overlap.info, base, up_scale, down_shift);
|
auto copies = MakeShrinkImageCopies(new_info, overlap.info, base, up_scale, down_shift);
|
||||||
if (overlap.info.num_samples != new_image.info.num_samples) {
|
if (overlap.info.num_samples != new_image.info.num_samples) {
|
||||||
runtime.CopyImageMSAA(new_image, overlap, std::move(copies));
|
runtime.CopyImageMSAA(new_image, overlap, std::move(copies));
|
||||||
|
continue;
|
||||||
} else {
|
} else {
|
||||||
runtime.CopyImage(new_image, overlap, std::move(copies));
|
runtime.CopyImage(new_image, overlap, std::move(copies));
|
||||||
}
|
}
|
||||||
|
@ -1075,9 +1075,6 @@ std::optional<OverlapResult> ResolveOverlap(const ImageInfo& new_info, GPUVAddr
|
|||||||
return std::nullopt;
|
return std::nullopt;
|
||||||
}
|
}
|
||||||
if (gpu_addr == overlap.gpu_addr) {
|
if (gpu_addr == overlap.gpu_addr) {
|
||||||
if (new_info.num_samples != overlap.info.num_samples) {
|
|
||||||
return std::nullopt;
|
|
||||||
}
|
|
||||||
const std::optional solution = ResolveOverlapEqualAddress(new_info, overlap, strict_size);
|
const std::optional solution = ResolveOverlapEqualAddress(new_info, overlap, strict_size);
|
||||||
if (!solution) {
|
if (!solution) {
|
||||||
return std::nullopt;
|
return std::nullopt;
|
||||||
|
Loading…
Reference in New Issue
Block a user