early-access version 2527
This commit is contained in:
parent
68d478c863
commit
6aa91a516c
@ -1,7 +1,7 @@
|
|||||||
yuzu emulator early access
|
yuzu emulator early access
|
||||||
=============
|
=============
|
||||||
|
|
||||||
This is the source code for early-access 2526.
|
This is the source code for early-access 2527.
|
||||||
|
|
||||||
## Legal Notice
|
## Legal Notice
|
||||||
|
|
||||||
|
@ -327,8 +327,8 @@ private:
|
|||||||
bool IsNiechePlaceholder(size_t virtual_offset, size_t length) const {
|
bool IsNiechePlaceholder(size_t virtual_offset, size_t length) const {
|
||||||
const auto it = placeholders.upper_bound({virtual_offset, virtual_offset + length});
|
const auto it = placeholders.upper_bound({virtual_offset, virtual_offset + length});
|
||||||
if (it != placeholders.end() && it->lower() == virtual_offset + length) {
|
if (it != placeholders.end() && it->lower() == virtual_offset + length) {
|
||||||
const bool is_root = it == placeholders.begin() && virtual_offset == 0;
|
return it == placeholders.begin() ? virtual_offset == 0
|
||||||
return is_root || std::prev(it)->upper() == virtual_offset;
|
: std::prev(it)->upper() == virtual_offset;
|
||||||
}
|
}
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
@ -218,19 +218,17 @@ private:
|
|||||||
Impl(const std::filesystem::path& file_backend_filename, const Filter& filter_)
|
Impl(const std::filesystem::path& file_backend_filename, const Filter& filter_)
|
||||||
: filter{filter_}, file_backend{file_backend_filename} {}
|
: filter{filter_}, file_backend{file_backend_filename} {}
|
||||||
|
|
||||||
~Impl() {
|
~Impl() = default;
|
||||||
StopBackendThread();
|
|
||||||
}
|
|
||||||
|
|
||||||
void StartBackendThread() {
|
void StartBackendThread() {
|
||||||
backend_thread = std::thread([this] {
|
backend_thread = std::jthread([this](std::stop_token stop_token) {
|
||||||
Common::SetCurrentThreadName("yuzu:Log");
|
Common::SetCurrentThreadName("yuzu:Log");
|
||||||
Entry entry;
|
Entry entry;
|
||||||
const auto write_logs = [this, &entry]() {
|
const auto write_logs = [this, &entry]() {
|
||||||
ForEachBackend([&entry](Backend& backend) { backend.Write(entry); });
|
ForEachBackend([&entry](Backend& backend) { backend.Write(entry); });
|
||||||
};
|
};
|
||||||
while (!stop.stop_requested()) {
|
while (!stop_token.stop_requested()) {
|
||||||
entry = message_queue.PopWait(stop.get_token());
|
entry = message_queue.PopWait(stop_token);
|
||||||
if (entry.filename != nullptr) {
|
if (entry.filename != nullptr) {
|
||||||
write_logs();
|
write_logs();
|
||||||
}
|
}
|
||||||
@ -244,11 +242,6 @@ private:
|
|||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
void StopBackendThread() {
|
|
||||||
stop.request_stop();
|
|
||||||
backend_thread.join();
|
|
||||||
}
|
|
||||||
|
|
||||||
Entry CreateEntry(Class log_class, Level log_level, const char* filename, unsigned int line_nr,
|
Entry CreateEntry(Class log_class, Level log_level, const char* filename, unsigned int line_nr,
|
||||||
const char* function, std::string&& message) const {
|
const char* function, std::string&& message) const {
|
||||||
using std::chrono::duration_cast;
|
using std::chrono::duration_cast;
|
||||||
@ -283,8 +276,7 @@ private:
|
|||||||
ColorConsoleBackend color_console_backend{};
|
ColorConsoleBackend color_console_backend{};
|
||||||
FileBackend file_backend;
|
FileBackend file_backend;
|
||||||
|
|
||||||
std::stop_source stop;
|
std::jthread backend_thread;
|
||||||
std::thread backend_thread;
|
|
||||||
MPSCQueue<Entry, true> message_queue{};
|
MPSCQueue<Entry, true> message_queue{};
|
||||||
std::chrono::steady_clock::time_point time_origin{std::chrono::steady_clock::now()};
|
std::chrono::steady_clock::time_point time_origin{std::chrono::steady_clock::now()};
|
||||||
};
|
};
|
||||||
|
@ -285,39 +285,64 @@ ResultCode KPageTable::MapProcessCode(VAddr addr, std::size_t num_pages, KMemory
|
|||||||
return ResultSuccess;
|
return ResultSuccess;
|
||||||
}
|
}
|
||||||
|
|
||||||
ResultCode KPageTable::MapCodeMemory(VAddr dst_addr, VAddr src_addr, std::size_t size) {
|
ResultCode KPageTable::MapCodeMemory(VAddr dst_address, VAddr src_address, std::size_t size) {
|
||||||
|
// Validate the mapping request.
|
||||||
|
R_UNLESS(this->CanContain(dst_address, size, KMemoryState::AliasCode),
|
||||||
|
ResultInvalidMemoryRegion);
|
||||||
|
|
||||||
|
// Lock the table.
|
||||||
KScopedLightLock lk(general_lock);
|
KScopedLightLock lk(general_lock);
|
||||||
|
|
||||||
const std::size_t num_pages{size / PageSize};
|
// Verify that the source memory is normal heap.
|
||||||
|
KMemoryState src_state{};
|
||||||
KMemoryState state{};
|
KMemoryPermission src_perm{};
|
||||||
KMemoryPermission perm{};
|
std::size_t num_src_allocator_blocks{};
|
||||||
CASCADE_CODE(CheckMemoryState(&state, &perm, nullptr, nullptr, src_addr, size,
|
R_TRY(this->CheckMemoryState(std::addressof(src_state), std::addressof(src_perm), nullptr,
|
||||||
|
std::addressof(num_src_allocator_blocks), src_address, size,
|
||||||
KMemoryState::All, KMemoryState::Normal, KMemoryPermission::All,
|
KMemoryState::All, KMemoryState::Normal, KMemoryPermission::All,
|
||||||
KMemoryPermission::UserReadWrite, KMemoryAttribute::Mask,
|
KMemoryPermission::UserReadWrite, KMemoryAttribute::All,
|
||||||
KMemoryAttribute::None, KMemoryAttribute::IpcAndDeviceMapped));
|
KMemoryAttribute::None));
|
||||||
|
|
||||||
if (IsRegionMapped(dst_addr, size)) {
|
// Verify that the destination memory is unmapped.
|
||||||
return ResultInvalidCurrentMemory;
|
std::size_t num_dst_allocator_blocks{};
|
||||||
}
|
R_TRY(this->CheckMemoryState(std::addressof(num_dst_allocator_blocks), dst_address, size,
|
||||||
|
KMemoryState::All, KMemoryState::Free, KMemoryPermission::None,
|
||||||
KPageLinkedList page_linked_list;
|
KMemoryPermission::None, KMemoryAttribute::None,
|
||||||
AddRegionToPages(src_addr, num_pages, page_linked_list);
|
KMemoryAttribute::None));
|
||||||
|
|
||||||
|
// Map the code memory.
|
||||||
{
|
{
|
||||||
auto block_guard = detail::ScopeExit(
|
// Determine the number of pages being operated on.
|
||||||
[&] { Operate(src_addr, num_pages, perm, OperationType::ChangePermissions); });
|
const std::size_t num_pages = size / PageSize;
|
||||||
|
|
||||||
CASCADE_CODE(Operate(src_addr, num_pages, KMemoryPermission::None,
|
// Create page groups for the memory being mapped.
|
||||||
OperationType::ChangePermissions));
|
KPageLinkedList pg;
|
||||||
CASCADE_CODE(MapPages(dst_addr, page_linked_list, KMemoryPermission::None));
|
AddRegionToPages(src_address, num_pages, pg);
|
||||||
|
|
||||||
block_guard.Cancel();
|
// Reprotect the source as kernel-read/not mapped.
|
||||||
}
|
const KMemoryPermission new_perm = static_cast<KMemoryPermission>(
|
||||||
|
KMemoryPermission::KernelRead | KMemoryPermission::NotMapped);
|
||||||
|
R_TRY(Operate(src_address, num_pages, new_perm, OperationType::ChangePermissions));
|
||||||
|
|
||||||
block_manager->Update(src_addr, num_pages, state, KMemoryPermission::None,
|
// Ensure that we unprotect the source pages on failure.
|
||||||
|
auto unprot_guard = SCOPE_GUARD({
|
||||||
|
ASSERT(this->Operate(src_address, num_pages, src_perm, OperationType::ChangePermissions)
|
||||||
|
.IsSuccess());
|
||||||
|
});
|
||||||
|
|
||||||
|
// Map the alias pages.
|
||||||
|
R_TRY(MapPages(dst_address, pg, new_perm));
|
||||||
|
|
||||||
|
// We successfully mapped the alias pages, so we don't need to unprotect the src pages on
|
||||||
|
// failure.
|
||||||
|
unprot_guard.Cancel();
|
||||||
|
|
||||||
|
// Apply the memory block updates.
|
||||||
|
block_manager->Update(src_address, num_pages, src_state, new_perm,
|
||||||
KMemoryAttribute::Locked);
|
KMemoryAttribute::Locked);
|
||||||
block_manager->Update(dst_addr, num_pages, KMemoryState::AliasCode);
|
block_manager->Update(dst_address, num_pages, KMemoryState::AliasCode, new_perm,
|
||||||
|
KMemoryAttribute::None);
|
||||||
|
}
|
||||||
|
|
||||||
return ResultSuccess;
|
return ResultSuccess;
|
||||||
}
|
}
|
||||||
@ -330,12 +355,6 @@ ResultCode KPageTable::UnmapCodeMemory(VAddr dst_address, VAddr src_address, std
|
|||||||
// Lock the table.
|
// Lock the table.
|
||||||
KScopedLightLock lk(general_lock);
|
KScopedLightLock lk(general_lock);
|
||||||
|
|
||||||
if (!size) {
|
|
||||||
return ResultSuccess;
|
|
||||||
}
|
|
||||||
|
|
||||||
const std::size_t num_pages{size / PageSize};
|
|
||||||
|
|
||||||
// Verify that the source memory is locked normal heap.
|
// Verify that the source memory is locked normal heap.
|
||||||
std::size_t num_src_allocator_blocks{};
|
std::size_t num_src_allocator_blocks{};
|
||||||
R_TRY(this->CheckMemoryState(std::addressof(num_src_allocator_blocks), src_address, size,
|
R_TRY(this->CheckMemoryState(std::addressof(num_src_allocator_blocks), src_address, size,
|
||||||
@ -384,9 +403,15 @@ ResultCode KPageTable::UnmapCodeMemory(VAddr dst_address, VAddr src_address, std
|
|||||||
|
|
||||||
// Unmap.
|
// Unmap.
|
||||||
{
|
{
|
||||||
// TODO(bunnei): We free the virtual address space, but do not nullptr the pointers in the
|
// Determine the number of pages being operated on.
|
||||||
// backing page table. This is a workaround because of an issue where CPU emulation may have
|
const std::size_t num_pages = size / PageSize;
|
||||||
// not quite finished running code when NROs are unloaded.
|
|
||||||
|
// Unmap the aliased copy of the pages.
|
||||||
|
R_TRY(Operate(dst_address, num_pages, KMemoryPermission::None, OperationType::Unmap));
|
||||||
|
|
||||||
|
// Try to set the permissions for the source pages back to what they should be.
|
||||||
|
R_TRY(Operate(src_address, num_pages, KMemoryPermission::UserReadWrite,
|
||||||
|
OperationType::ChangePermissions));
|
||||||
|
|
||||||
// Apply the memory block updates.
|
// Apply the memory block updates.
|
||||||
block_manager->Update(dst_address, num_pages, KMemoryState::None);
|
block_manager->Update(dst_address, num_pages, KMemoryState::None);
|
||||||
|
@ -36,8 +36,8 @@ public:
|
|||||||
KMemoryManager::Pool pool);
|
KMemoryManager::Pool pool);
|
||||||
ResultCode MapProcessCode(VAddr addr, std::size_t pages_count, KMemoryState state,
|
ResultCode MapProcessCode(VAddr addr, std::size_t pages_count, KMemoryState state,
|
||||||
KMemoryPermission perm);
|
KMemoryPermission perm);
|
||||||
ResultCode MapCodeMemory(VAddr dst_addr, VAddr src_addr, std::size_t size);
|
ResultCode MapCodeMemory(VAddr dst_address, VAddr src_address, std::size_t size);
|
||||||
ResultCode UnmapCodeMemory(VAddr dst_addr, VAddr src_addr, std::size_t size);
|
ResultCode UnmapCodeMemory(VAddr dst_address, VAddr src_address, std::size_t size);
|
||||||
ResultCode UnmapProcessMemory(VAddr dst_addr, std::size_t size, KPageTable& src_page_table,
|
ResultCode UnmapProcessMemory(VAddr dst_addr, std::size_t size, KPageTable& src_page_table,
|
||||||
VAddr src_addr);
|
VAddr src_addr);
|
||||||
ResultCode MapPhysicalMemory(VAddr addr, std::size_t size);
|
ResultCode MapPhysicalMemory(VAddr addr, std::size_t size);
|
||||||
|
Loading…
Reference in New Issue
Block a user