2018-01-13 21:22:39 +00:00
|
|
|
// Copyright 2018 yuzu emulator team
|
2014-12-17 05:38:14 +00:00
|
|
|
// Licensed under GPLv2 or any later version
|
2014-11-19 08:49:13 +00:00
|
|
|
// Refer to the license.txt file included.
|
2014-04-10 23:58:28 +00:00
|
|
|
|
2018-01-12 03:36:56 +00:00
|
|
|
#include <algorithm>
|
2018-02-25 12:40:22 +00:00
|
|
|
#include <cinttypes>
|
2018-04-20 02:36:48 +00:00
|
|
|
#include <iterator>
|
2018-07-31 12:06:09 +00:00
|
|
|
#include <mutex>
|
|
|
|
#include <vector>
|
2018-01-12 03:36:56 +00:00
|
|
|
|
2018-10-18 16:55:27 +00:00
|
|
|
#include "common/alignment.h"
|
2018-07-31 12:06:09 +00:00
|
|
|
#include "common/assert.h"
|
2020-02-27 23:12:41 +00:00
|
|
|
#include "common/fiber.h"
|
2015-05-06 07:06:12 +00:00
|
|
|
#include "common/logging/log.h"
|
2015-08-17 21:25:21 +00:00
|
|
|
#include "common/microprofile.h"
|
2018-01-05 00:45:15 +00:00
|
|
|
#include "common/string_util.h"
|
2018-08-31 16:21:34 +00:00
|
|
|
#include "core/arm/exclusive_monitor.h"
|
2018-03-13 21:49:59 +00:00
|
|
|
#include "core/core.h"
|
2016-09-18 00:38:01 +00:00
|
|
|
#include "core/core_timing.h"
|
2019-11-23 21:01:06 +00:00
|
|
|
#include "core/core_timing_util.h"
|
2020-03-10 17:13:39 +00:00
|
|
|
#include "core/cpu_manager.h"
|
2018-06-21 06:49:43 +00:00
|
|
|
#include "core/hle/kernel/address_arbiter.h"
|
2016-05-22 17:30:13 +00:00
|
|
|
#include "core/hle/kernel/client_port.h"
|
2016-06-14 23:03:30 +00:00
|
|
|
#include "core/hle/kernel/client_session.h"
|
2019-03-05 14:20:11 +00:00
|
|
|
#include "core/hle/kernel/errors.h"
|
2017-05-29 23:45:42 +00:00
|
|
|
#include "core/hle/kernel/handle_table.h"
|
2020-12-03 02:08:35 +00:00
|
|
|
#include "core/hle/kernel/k_scheduler.h"
|
2020-12-04 05:56:02 +00:00
|
|
|
#include "core/hle/kernel/k_scoped_scheduler_lock_and_sleep.h"
|
2018-08-31 16:21:34 +00:00
|
|
|
#include "core/hle/kernel/kernel.h"
|
2020-04-09 03:14:18 +00:00
|
|
|
#include "core/hle/kernel/memory/memory_block.h"
|
|
|
|
#include "core/hle/kernel/memory/page_table.h"
|
2018-01-01 19:38:34 +00:00
|
|
|
#include "core/hle/kernel/mutex.h"
|
2020-02-25 17:22:11 +00:00
|
|
|
#include "core/hle/kernel/physical_core.h"
|
2015-05-11 14:15:10 +00:00
|
|
|
#include "core/hle/kernel/process.h"
|
2018-11-26 23:34:07 +00:00
|
|
|
#include "core/hle/kernel/readable_event.h"
|
2017-12-31 20:58:16 +00:00
|
|
|
#include "core/hle/kernel/resource_limit.h"
|
2018-01-14 22:15:31 +00:00
|
|
|
#include "core/hle/kernel/shared_memory.h"
|
2018-01-03 01:40:30 +00:00
|
|
|
#include "core/hle/kernel/svc.h"
|
2020-03-27 01:13:46 +00:00
|
|
|
#include "core/hle/kernel/svc_types.h"
|
2018-01-03 01:40:30 +00:00
|
|
|
#include "core/hle/kernel/svc_wrap.h"
|
2020-02-11 21:36:39 +00:00
|
|
|
#include "core/hle/kernel/synchronization.h"
|
2017-10-15 02:18:42 +00:00
|
|
|
#include "core/hle/kernel/thread.h"
|
2020-02-25 23:43:28 +00:00
|
|
|
#include "core/hle/kernel/time_manager.h"
|
2019-03-13 07:09:27 +00:00
|
|
|
#include "core/hle/kernel/transfer_memory.h"
|
2018-11-26 23:34:07 +00:00
|
|
|
#include "core/hle/kernel/writable_event.h"
|
2017-10-14 21:30:07 +00:00
|
|
|
#include "core/hle/lock.h"
|
2014-10-23 03:20:01 +00:00
|
|
|
#include "core/hle/result.h"
|
2014-04-13 01:55:36 +00:00
|
|
|
#include "core/hle/service/service.h"
|
svc: Handle memory writing explicitly within QueryProcessMemory
Moves the memory writes directly into QueryProcessMemory instead of
letting the wrapper function do it. It would be inaccurate to allow the
handler to do it because there's cases where memory shouldn't even be
written to. For example, if the given process handle is invalid.
HOWEVER, if the memory writing is within the wrapper, then we have no
control over if these memory writes occur, meaning in an error case, 68
bytes of memory randomly get trashed with zeroes, 64 of those being
written to wherever the memory info address points to, and the remaining
4 being written wherever the page info address points to.
One solution in this case would be to just conditionally check within
the handler itself, but this is kind of smelly, given the handler
shouldn't be performing conditional behavior itself, it's a behavior of
the managed function. In other words, if you remove the handler from the
equation entirely, does the function still retain its proper behavior?
In this case, no.
Now, we don't potentially trash memory from this function if an invalid
query is performed.
2018-12-12 16:48:06 +00:00
|
|
|
#include "core/memory.h"
|
2019-05-18 01:46:17 +00:00
|
|
|
#include "core/reporter.h"
|
2014-04-10 23:58:28 +00:00
|
|
|
|
2020-03-27 00:00:30 +00:00
|
|
|
namespace Kernel::Svc {
|
2018-09-13 23:14:50 +00:00
|
|
|
namespace {
|
2018-10-10 18:18:27 +00:00
|
|
|
|
|
|
|
// Checks if address + size is greater than the given address
|
|
|
|
// This can return false if the size causes an overflow of a 64-bit type
|
|
|
|
// or if the given size is zero.
|
|
|
|
constexpr bool IsValidAddressRange(VAddr address, u64 size) {
|
|
|
|
return address + size > address;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Helper function that performs the common sanity checks for svcMapMemory
|
|
|
|
// and svcUnmapMemory. This is doable, as both functions perform their sanitizing
|
|
|
|
// in the same order.
|
2020-04-09 03:14:18 +00:00
|
|
|
ResultCode MapUnmapMemorySanityChecks(const Memory::PageTable& manager, VAddr dst_addr,
|
|
|
|
VAddr src_addr, u64 size) {
|
2018-11-26 08:47:39 +00:00
|
|
|
if (!Common::Is4KBAligned(dst_addr)) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Destination address is not aligned to 4KB, 0x{:016X}", dst_addr);
|
2018-10-10 18:18:27 +00:00
|
|
|
return ERR_INVALID_ADDRESS;
|
|
|
|
}
|
|
|
|
|
2018-11-26 08:47:39 +00:00
|
|
|
if (!Common::Is4KBAligned(src_addr)) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Source address is not aligned to 4KB, 0x{:016X}", src_addr);
|
2018-11-27 01:29:06 +00:00
|
|
|
return ERR_INVALID_SIZE;
|
2018-11-26 08:47:39 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
if (size == 0) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Size is 0");
|
|
|
|
return ERR_INVALID_SIZE;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!Common::Is4KBAligned(size)) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Size is not aligned to 4KB, 0x{:016X}", size);
|
2018-10-10 18:18:27 +00:00
|
|
|
return ERR_INVALID_SIZE;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!IsValidAddressRange(dst_addr, size)) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC,
|
|
|
|
"Destination is not a valid address range, addr=0x{:016X}, size=0x{:016X}",
|
|
|
|
dst_addr, size);
|
2018-10-10 18:18:27 +00:00
|
|
|
return ERR_INVALID_ADDRESS_STATE;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!IsValidAddressRange(src_addr, size)) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Source is not a valid address range, addr=0x{:016X}, size=0x{:016X}",
|
|
|
|
src_addr, size);
|
2018-10-10 18:18:27 +00:00
|
|
|
return ERR_INVALID_ADDRESS_STATE;
|
|
|
|
}
|
|
|
|
|
2020-04-09 03:14:18 +00:00
|
|
|
if (!manager.IsInsideAddressSpace(src_addr, size)) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC,
|
|
|
|
"Source is not within the address space, addr=0x{:016X}, size=0x{:016X}",
|
|
|
|
src_addr, size);
|
2018-10-10 18:18:27 +00:00
|
|
|
return ERR_INVALID_ADDRESS_STATE;
|
|
|
|
}
|
|
|
|
|
2020-04-09 03:14:18 +00:00
|
|
|
if (manager.IsOutsideStackRegion(dst_addr, size)) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC,
|
2019-07-06 06:02:01 +00:00
|
|
|
"Destination is not within the stack region, addr=0x{:016X}, size=0x{:016X}",
|
2018-11-26 08:47:39 +00:00
|
|
|
dst_addr, size);
|
2018-10-10 18:18:27 +00:00
|
|
|
return ERR_INVALID_MEMORY_RANGE;
|
|
|
|
}
|
|
|
|
|
2020-04-09 03:14:18 +00:00
|
|
|
if (manager.IsInsideHeapRegion(dst_addr, size)) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC,
|
|
|
|
"Destination does not fit within the heap region, addr=0x{:016X}, "
|
2020-04-09 03:14:18 +00:00
|
|
|
"size=0x{:016X}",
|
|
|
|
dst_addr, size);
|
2018-10-10 18:18:27 +00:00
|
|
|
return ERR_INVALID_MEMORY_RANGE;
|
|
|
|
}
|
|
|
|
|
2020-04-09 03:14:18 +00:00
|
|
|
if (manager.IsInsideAliasRegion(dst_addr, size)) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC,
|
|
|
|
"Destination does not fit within the map region, addr=0x{:016X}, "
|
2020-04-09 03:14:18 +00:00
|
|
|
"size=0x{:016X}",
|
|
|
|
dst_addr, size);
|
2018-10-10 18:18:27 +00:00
|
|
|
return ERR_INVALID_MEMORY_RANGE;
|
|
|
|
}
|
|
|
|
|
|
|
|
return RESULT_SUCCESS;
|
|
|
|
}
|
2018-11-27 00:14:29 +00:00
|
|
|
|
|
|
|
enum class ResourceLimitValueType {
|
|
|
|
CurrentValue,
|
|
|
|
LimitValue,
|
|
|
|
};
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
ResultVal<s64> RetrieveResourceLimitValue(Core::System& system, Handle resource_limit,
|
|
|
|
u32 resource_type, ResourceLimitValueType value_type) {
|
2020-02-27 14:28:44 +00:00
|
|
|
std::lock_guard lock{HLE::g_hle_lock};
|
2018-11-27 00:14:29 +00:00
|
|
|
const auto type = static_cast<ResourceType>(resource_type);
|
|
|
|
if (!IsValidResourceType(type)) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Invalid resource limit type: '{}'", resource_type);
|
|
|
|
return ERR_INVALID_ENUM_VALUE;
|
|
|
|
}
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
const auto* const current_process = system.Kernel().CurrentProcess();
|
2018-11-27 00:14:29 +00:00
|
|
|
ASSERT(current_process != nullptr);
|
|
|
|
|
|
|
|
const auto resource_limit_object =
|
|
|
|
current_process->GetHandleTable().Get<ResourceLimit>(resource_limit);
|
|
|
|
if (!resource_limit_object) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Handle to non-existent resource limit instance used. Handle={:08X}",
|
|
|
|
resource_limit);
|
|
|
|
return ERR_INVALID_HANDLE;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (value_type == ResourceLimitValueType::CurrentValue) {
|
|
|
|
return MakeResult(resource_limit_object->GetCurrentResourceValue(type));
|
|
|
|
}
|
|
|
|
|
|
|
|
return MakeResult(resource_limit_object->GetMaxResourceValue(type));
|
|
|
|
}
|
2018-09-13 23:14:50 +00:00
|
|
|
} // Anonymous namespace
|
2014-04-11 03:26:12 +00:00
|
|
|
|
2017-12-28 20:29:52 +00:00
|
|
|
/// Set the process heap to a given Size. It can both extend and shrink the heap.
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode SetHeapSize(Core::System& system, VAddr* heap_addr, u64 heap_size) {
|
2020-02-27 14:28:44 +00:00
|
|
|
std::lock_guard lock{HLE::g_hle_lock};
|
2018-07-02 16:13:26 +00:00
|
|
|
LOG_TRACE(Kernel_SVC, "called, heap_size=0x{:X}", heap_size);
|
2018-09-13 23:09:04 +00:00
|
|
|
|
2018-11-27 01:29:06 +00:00
|
|
|
// Size must be a multiple of 0x200000 (2MB) and be equal to or less than 8GB.
|
2018-11-27 01:53:18 +00:00
|
|
|
if ((heap_size % 0x200000) != 0) {
|
2018-11-27 01:29:06 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "The heap size is not a multiple of 2MB, heap_size=0x{:016X}",
|
|
|
|
heap_size);
|
|
|
|
return ERR_INVALID_SIZE;
|
|
|
|
}
|
|
|
|
|
2018-11-27 01:53:18 +00:00
|
|
|
if (heap_size >= 0x200000000) {
|
2018-11-27 01:29:06 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "The heap size is not less than 8GB, heap_size=0x{:016X}", heap_size);
|
2018-09-13 23:09:04 +00:00
|
|
|
return ERR_INVALID_SIZE;
|
|
|
|
}
|
|
|
|
|
2020-04-09 03:14:18 +00:00
|
|
|
auto& page_table{system.Kernel().CurrentProcess()->PageTable()};
|
|
|
|
|
|
|
|
CASCADE_RESULT(*heap_addr, page_table.SetHeapSize(heap_size));
|
2018-12-27 23:31:31 +00:00
|
|
|
|
2017-12-28 20:29:52 +00:00
|
|
|
return RESULT_SUCCESS;
|
|
|
|
}
|
|
|
|
|
2020-03-02 05:06:41 +00:00
|
|
|
static ResultCode SetHeapSize32(Core::System& system, u32* heap_addr, u32 heap_size) {
|
|
|
|
VAddr temp_heap_addr{};
|
|
|
|
const ResultCode result{SetHeapSize(system, &temp_heap_addr, heap_size)};
|
|
|
|
*heap_addr = static_cast<u32>(temp_heap_addr);
|
|
|
|
return result;
|
|
|
|
}
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode SetMemoryAttribute(Core::System& system, VAddr address, u64 size, u32 mask,
|
|
|
|
u32 attribute) {
|
2020-02-27 14:28:44 +00:00
|
|
|
std::lock_guard lock{HLE::g_hle_lock};
|
2018-12-15 20:21:41 +00:00
|
|
|
LOG_DEBUG(Kernel_SVC,
|
|
|
|
"called, address=0x{:016X}, size=0x{:X}, mask=0x{:08X}, attribute=0x{:08X}", address,
|
|
|
|
size, mask, attribute);
|
|
|
|
|
|
|
|
if (!Common::Is4KBAligned(address)) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Address not page aligned (0x{:016X})", address);
|
|
|
|
return ERR_INVALID_ADDRESS;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (size == 0 || !Common::Is4KBAligned(size)) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Invalid size (0x{:X}). Size must be non-zero and page aligned.",
|
|
|
|
size);
|
|
|
|
return ERR_INVALID_ADDRESS;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!IsValidAddressRange(address, size)) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Address range overflowed (Address: 0x{:016X}, Size: 0x{:016X})",
|
|
|
|
address, size);
|
|
|
|
return ERR_INVALID_ADDRESS_STATE;
|
|
|
|
}
|
|
|
|
|
2020-04-09 03:14:18 +00:00
|
|
|
const auto attributes{static_cast<Memory::MemoryAttribute>(mask | attribute)};
|
|
|
|
if (attributes != static_cast<Memory::MemoryAttribute>(mask) ||
|
|
|
|
(attributes | Memory::MemoryAttribute::Uncached) != Memory::MemoryAttribute::Uncached) {
|
2018-12-15 20:21:41 +00:00
|
|
|
LOG_ERROR(Kernel_SVC,
|
|
|
|
"Memory attribute doesn't match the given mask (Attribute: 0x{:X}, Mask: {:X}",
|
|
|
|
attribute, mask);
|
|
|
|
return ERR_INVALID_COMBINATION;
|
|
|
|
}
|
|
|
|
|
2020-04-09 03:14:18 +00:00
|
|
|
auto& page_table{system.Kernel().CurrentProcess()->PageTable()};
|
2018-12-15 20:21:41 +00:00
|
|
|
|
2020-04-09 03:14:18 +00:00
|
|
|
return page_table.SetMemoryAttribute(address, size, static_cast<Memory::MemoryAttribute>(mask),
|
|
|
|
static_cast<Memory::MemoryAttribute>(attribute));
|
2018-01-08 02:23:42 +00:00
|
|
|
}
|
|
|
|
|
2020-06-19 00:33:04 +00:00
|
|
|
static ResultCode SetMemoryAttribute32(Core::System& system, u32 address, u32 size, u32 mask,
|
|
|
|
u32 attribute) {
|
|
|
|
return SetMemoryAttribute(system, static_cast<VAddr>(address), static_cast<std::size_t>(size),
|
|
|
|
mask, attribute);
|
|
|
|
}
|
|
|
|
|
2017-12-29 02:38:38 +00:00
|
|
|
/// Maps a memory range into a different range.
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode MapMemory(Core::System& system, VAddr dst_addr, VAddr src_addr, u64 size) {
|
2020-02-27 14:28:44 +00:00
|
|
|
std::lock_guard lock{HLE::g_hle_lock};
|
2018-07-02 16:13:26 +00:00
|
|
|
LOG_TRACE(Kernel_SVC, "called, dst_addr=0x{:X}, src_addr=0x{:X}, size=0x{:X}", dst_addr,
|
2018-07-02 16:20:50 +00:00
|
|
|
src_addr, size);
|
2018-09-13 23:14:50 +00:00
|
|
|
|
2020-04-09 03:14:18 +00:00
|
|
|
auto& page_table{system.Kernel().CurrentProcess()->PageTable()};
|
2018-12-27 23:31:31 +00:00
|
|
|
|
2020-04-09 03:14:18 +00:00
|
|
|
if (const ResultCode result{MapUnmapMemorySanityChecks(page_table, dst_addr, src_addr, size)};
|
|
|
|
result.IsError()) {
|
2018-10-10 18:18:27 +00:00
|
|
|
return result;
|
2018-09-13 23:14:50 +00:00
|
|
|
}
|
|
|
|
|
2020-04-09 03:14:18 +00:00
|
|
|
return page_table.Map(dst_addr, src_addr, size);
|
2017-12-29 02:38:38 +00:00
|
|
|
}
|
|
|
|
|
2020-06-19 23:40:07 +00:00
|
|
|
static ResultCode MapMemory32(Core::System& system, u32 dst_addr, u32 src_addr, u32 size) {
|
|
|
|
return MapMemory(system, static_cast<VAddr>(dst_addr), static_cast<VAddr>(src_addr),
|
|
|
|
static_cast<std::size_t>(size));
|
|
|
|
}
|
|
|
|
|
2017-12-31 20:22:49 +00:00
|
|
|
/// Unmaps a region that was previously mapped with svcMapMemory
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode UnmapMemory(Core::System& system, VAddr dst_addr, VAddr src_addr, u64 size) {
|
2020-02-27 14:28:44 +00:00
|
|
|
std::lock_guard lock{HLE::g_hle_lock};
|
2018-07-02 16:13:26 +00:00
|
|
|
LOG_TRACE(Kernel_SVC, "called, dst_addr=0x{:X}, src_addr=0x{:X}, size=0x{:X}", dst_addr,
|
2018-07-02 16:20:50 +00:00
|
|
|
src_addr, size);
|
2018-09-13 23:14:50 +00:00
|
|
|
|
2020-04-09 03:14:18 +00:00
|
|
|
auto& page_table{system.Kernel().CurrentProcess()->PageTable()};
|
2018-12-27 23:31:31 +00:00
|
|
|
|
2020-04-09 03:14:18 +00:00
|
|
|
if (const ResultCode result{MapUnmapMemorySanityChecks(page_table, dst_addr, src_addr, size)};
|
|
|
|
result.IsError()) {
|
2018-10-10 18:18:27 +00:00
|
|
|
return result;
|
2018-09-13 23:14:50 +00:00
|
|
|
}
|
|
|
|
|
2020-04-09 03:14:18 +00:00
|
|
|
return page_table.Unmap(dst_addr, src_addr, size);
|
2017-12-31 20:22:49 +00:00
|
|
|
}
|
|
|
|
|
2020-06-19 23:40:07 +00:00
|
|
|
static ResultCode UnmapMemory32(Core::System& system, u32 dst_addr, u32 src_addr, u32 size) {
|
|
|
|
return UnmapMemory(system, static_cast<VAddr>(dst_addr), static_cast<VAddr>(src_addr),
|
|
|
|
static_cast<std::size_t>(size));
|
|
|
|
}
|
|
|
|
|
2014-04-13 01:55:36 +00:00
|
|
|
/// Connect to an OS service given the port name, returns the handle to the port to out
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode ConnectToNamedPort(Core::System& system, Handle* out_handle,
|
|
|
|
VAddr port_name_address) {
|
2020-02-27 14:28:44 +00:00
|
|
|
std::lock_guard lock{HLE::g_hle_lock};
|
2019-11-26 20:48:19 +00:00
|
|
|
auto& memory = system.Memory();
|
|
|
|
|
|
|
|
if (!memory.IsValidVirtualAddress(port_name_address)) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC,
|
|
|
|
"Port Name Address is not a valid virtual address, port_name_address=0x{:016X}",
|
|
|
|
port_name_address);
|
2018-01-03 01:40:30 +00:00
|
|
|
return ERR_NOT_FOUND;
|
2018-09-02 15:58:58 +00:00
|
|
|
}
|
2017-10-14 21:30:07 +00:00
|
|
|
|
|
|
|
static constexpr std::size_t PortNameMaxLength = 11;
|
|
|
|
// Read 1 char beyond the max allowed port name to detect names that are too long.
|
2019-11-26 20:48:19 +00:00
|
|
|
const std::string port_name = memory.ReadCString(port_name_address, PortNameMaxLength + 1);
|
2018-09-02 15:58:58 +00:00
|
|
|
if (port_name.size() > PortNameMaxLength) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Port name is too long, expected {} but got {}", PortNameMaxLength,
|
|
|
|
port_name.size());
|
2018-11-16 19:24:27 +00:00
|
|
|
return ERR_OUT_OF_RANGE;
|
2018-09-02 15:58:58 +00:00
|
|
|
}
|
2014-06-02 00:48:29 +00:00
|
|
|
|
2018-07-02 16:13:26 +00:00
|
|
|
LOG_TRACE(Kernel_SVC, "called port_name={}", port_name);
|
2014-06-02 00:48:29 +00:00
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
auto& kernel = system.Kernel();
|
|
|
|
const auto it = kernel.FindNamedPort(port_name);
|
2018-09-02 15:58:58 +00:00
|
|
|
if (!kernel.IsValidNamedPort(it)) {
|
2018-07-02 16:13:26 +00:00
|
|
|
LOG_WARNING(Kernel_SVC, "tried to connect to unknown port: {}", port_name);
|
2018-01-03 01:40:30 +00:00
|
|
|
return ERR_NOT_FOUND;
|
2015-01-30 18:07:04 +00:00
|
|
|
}
|
2014-06-02 00:48:29 +00:00
|
|
|
|
2020-04-09 03:14:18 +00:00
|
|
|
ASSERT(kernel.CurrentProcess()->GetResourceLimit()->Reserve(ResourceType::Sessions, 1));
|
|
|
|
|
2016-12-05 16:02:08 +00:00
|
|
|
auto client_port = it->second;
|
2016-06-14 23:03:30 +00:00
|
|
|
|
2019-11-25 01:15:51 +00:00
|
|
|
std::shared_ptr<ClientSession> client_session;
|
2016-12-05 18:59:57 +00:00
|
|
|
CASCADE_RESULT(client_session, client_port->Connect());
|
2016-06-14 23:03:30 +00:00
|
|
|
|
|
|
|
// Return the client session
|
2019-04-06 22:46:18 +00:00
|
|
|
auto& handle_table = kernel.CurrentProcess()->GetHandleTable();
|
2018-10-20 18:34:41 +00:00
|
|
|
CASCADE_RESULT(*out_handle, handle_table.Create(client_session));
|
2015-01-23 05:36:58 +00:00
|
|
|
return RESULT_SUCCESS;
|
2014-04-13 01:55:36 +00:00
|
|
|
}
|
|
|
|
|
2020-03-02 05:06:41 +00:00
|
|
|
static ResultCode ConnectToNamedPort32(Core::System& system, Handle* out_handle,
|
|
|
|
u32 port_name_address) {
|
|
|
|
|
|
|
|
return ConnectToNamedPort(system, out_handle, port_name_address);
|
|
|
|
}
|
|
|
|
|
2016-12-08 16:06:19 +00:00
|
|
|
/// Makes a blocking IPC call to an OS service.
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode SendSyncRequest(Core::System& system, Handle handle) {
|
2020-12-03 02:08:35 +00:00
|
|
|
auto& kernel = system.Kernel();
|
|
|
|
const auto& handle_table = kernel.CurrentProcess()->GetHandleTable();
|
2019-11-25 01:15:51 +00:00
|
|
|
std::shared_ptr<ClientSession> session = handle_table.Get<ClientSession>(handle);
|
2017-12-30 18:40:28 +00:00
|
|
|
if (!session) {
|
2018-07-02 16:13:26 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "called with invalid handle=0x{:08X}", handle);
|
2015-01-23 05:44:52 +00:00
|
|
|
return ERR_INVALID_HANDLE;
|
2014-10-23 03:20:01 +00:00
|
|
|
}
|
2014-05-27 02:12:46 +00:00
|
|
|
|
2018-07-02 16:13:26 +00:00
|
|
|
LOG_TRACE(Kernel_SVC, "called handle=0x{:08X}({})", handle, session->GetName());
|
2014-05-27 02:12:46 +00:00
|
|
|
|
2020-12-03 02:08:35 +00:00
|
|
|
auto thread = kernel.CurrentScheduler()->GetCurrentThread();
|
2020-02-25 23:43:28 +00:00
|
|
|
{
|
2020-12-04 06:26:42 +00:00
|
|
|
KScopedSchedulerLock lock(kernel);
|
2020-02-25 23:43:28 +00:00
|
|
|
thread->InvalidateHLECallback();
|
|
|
|
thread->SetStatus(ThreadStatus::WaitIPC);
|
2020-09-14 18:03:10 +00:00
|
|
|
session->SendSyncRequest(SharedFrom(thread), system.Memory(), system.CoreTiming());
|
2020-02-25 23:43:28 +00:00
|
|
|
}
|
2020-03-31 01:50:05 +00:00
|
|
|
|
2020-02-25 23:43:28 +00:00
|
|
|
if (thread->HasHLECallback()) {
|
|
|
|
Handle event_handle = thread->GetHLETimeEvent();
|
|
|
|
if (event_handle != InvalidHandle) {
|
2020-12-03 02:08:35 +00:00
|
|
|
auto& time_manager = kernel.TimeManager();
|
2020-02-25 23:43:28 +00:00
|
|
|
time_manager.UnscheduleTimeEvent(event_handle);
|
|
|
|
}
|
2020-03-31 01:50:05 +00:00
|
|
|
|
|
|
|
{
|
2020-12-04 06:26:42 +00:00
|
|
|
KScopedSchedulerLock lock(kernel);
|
2020-03-31 01:50:05 +00:00
|
|
|
auto* sync_object = thread->GetHLESyncObject();
|
|
|
|
sync_object->RemoveWaitingThread(SharedFrom(thread));
|
|
|
|
}
|
2020-05-29 21:37:37 +00:00
|
|
|
|
|
|
|
thread->InvokeHLECallback(SharedFrom(thread));
|
2020-02-25 23:43:28 +00:00
|
|
|
}
|
2017-01-01 16:57:02 +00:00
|
|
|
|
2020-03-20 16:36:01 +00:00
|
|
|
return thread->GetSignalingResult();
|
2014-04-10 23:58:28 +00:00
|
|
|
}
|
|
|
|
|
2020-03-02 05:06:41 +00:00
|
|
|
static ResultCode SendSyncRequest32(Core::System& system, Handle handle) {
|
|
|
|
return SendSyncRequest(system, handle);
|
|
|
|
}
|
|
|
|
|
2017-10-23 04:15:45 +00:00
|
|
|
/// Get the ID for the specified thread.
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode GetThreadId(Core::System& system, u64* thread_id, Handle thread_handle) {
|
2018-07-02 16:13:26 +00:00
|
|
|
LOG_TRACE(Kernel_SVC, "called thread=0x{:08X}", thread_handle);
|
2017-10-23 04:15:45 +00:00
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
const auto& handle_table = system.Kernel().CurrentProcess()->GetHandleTable();
|
2019-11-25 01:15:51 +00:00
|
|
|
const std::shared_ptr<Thread> thread = handle_table.Get<Thread>(thread_handle);
|
2017-12-30 18:40:28 +00:00
|
|
|
if (!thread) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Thread handle does not exist, handle=0x{:08X}", thread_handle);
|
2017-10-23 04:15:45 +00:00
|
|
|
return ERR_INVALID_HANDLE;
|
|
|
|
}
|
|
|
|
|
2018-10-03 22:47:57 +00:00
|
|
|
*thread_id = thread->GetThreadID();
|
2017-10-23 04:15:45 +00:00
|
|
|
return RESULT_SUCCESS;
|
|
|
|
}
|
|
|
|
|
2020-03-02 05:06:41 +00:00
|
|
|
static ResultCode GetThreadId32(Core::System& system, u32* thread_id_low, u32* thread_id_high,
|
|
|
|
Handle thread_handle) {
|
|
|
|
u64 thread_id{};
|
|
|
|
const ResultCode result{GetThreadId(system, &thread_id, thread_handle)};
|
|
|
|
|
|
|
|
*thread_id_low = static_cast<u32>(thread_id >> 32);
|
|
|
|
*thread_id_high = static_cast<u32>(thread_id & std::numeric_limits<u32>::max());
|
|
|
|
|
|
|
|
return result;
|
|
|
|
}
|
|
|
|
|
2018-12-19 04:09:08 +00:00
|
|
|
/// Gets the ID of the specified process or a specified thread's owning process.
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode GetProcessId(Core::System& system, u64* process_id, Handle handle) {
|
2018-12-19 04:09:08 +00:00
|
|
|
LOG_DEBUG(Kernel_SVC, "called handle=0x{:08X}", handle);
|
2017-10-23 04:15:45 +00:00
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
const auto& handle_table = system.Kernel().CurrentProcess()->GetHandleTable();
|
2019-11-25 01:15:51 +00:00
|
|
|
const std::shared_ptr<Process> process = handle_table.Get<Process>(handle);
|
2018-12-19 04:09:08 +00:00
|
|
|
if (process) {
|
|
|
|
*process_id = process->GetProcessID();
|
|
|
|
return RESULT_SUCCESS;
|
2017-10-23 04:15:45 +00:00
|
|
|
}
|
|
|
|
|
2019-11-25 01:15:51 +00:00
|
|
|
const std::shared_ptr<Thread> thread = handle_table.Get<Thread>(handle);
|
2018-12-19 04:09:08 +00:00
|
|
|
if (thread) {
|
|
|
|
const Process* const owner_process = thread->GetOwnerProcess();
|
|
|
|
if (!owner_process) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Non-existent owning process encountered.");
|
|
|
|
return ERR_INVALID_HANDLE;
|
|
|
|
}
|
|
|
|
|
|
|
|
*process_id = owner_process->GetProcessID();
|
|
|
|
return RESULT_SUCCESS;
|
|
|
|
}
|
|
|
|
|
|
|
|
// NOTE: This should also handle debug objects before returning.
|
|
|
|
|
|
|
|
LOG_ERROR(Kernel_SVC, "Handle does not exist, handle=0x{:08X}", handle);
|
|
|
|
return ERR_INVALID_HANDLE;
|
2017-10-23 04:15:45 +00:00
|
|
|
}
|
|
|
|
|
2020-06-19 23:40:07 +00:00
|
|
|
static ResultCode GetProcessId32(Core::System& system, u32* process_id_low, u32* process_id_high,
|
|
|
|
Handle handle) {
|
|
|
|
u64 process_id{};
|
|
|
|
const auto result = GetProcessId(system, &process_id, handle);
|
|
|
|
*process_id_low = static_cast<u32>(process_id);
|
|
|
|
*process_id_high = static_cast<u32>(process_id >> 32);
|
|
|
|
return result;
|
|
|
|
}
|
|
|
|
|
2018-01-01 19:47:57 +00:00
|
|
|
/// Wait for the given handles to synchronize, timeout after the specified nanoseconds
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode WaitSynchronization(Core::System& system, Handle* index, VAddr handles_address,
|
|
|
|
u64 handle_count, s64 nano_seconds) {
|
2018-07-02 16:13:26 +00:00
|
|
|
LOG_TRACE(Kernel_SVC, "called handles_address=0x{:X}, handle_count={}, nano_seconds={}",
|
2018-07-02 16:20:50 +00:00
|
|
|
handles_address, handle_count, nano_seconds);
|
2018-01-06 19:34:32 +00:00
|
|
|
|
2019-11-26 21:29:34 +00:00
|
|
|
auto& memory = system.Memory();
|
|
|
|
if (!memory.IsValidVirtualAddress(handles_address)) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC,
|
|
|
|
"Handle address is not a valid virtual address, handle_address=0x{:016X}",
|
|
|
|
handles_address);
|
2018-01-06 19:34:32 +00:00
|
|
|
return ERR_INVALID_POINTER;
|
2018-11-26 06:06:13 +00:00
|
|
|
}
|
2018-01-06 19:34:32 +00:00
|
|
|
|
2018-01-09 20:02:43 +00:00
|
|
|
static constexpr u64 MaxHandles = 0x40;
|
|
|
|
|
2018-11-16 19:24:27 +00:00
|
|
|
if (handle_count > MaxHandles) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Handle count specified is too large, expected {} but got {}",
|
|
|
|
MaxHandles, handle_count);
|
2018-11-16 19:24:27 +00:00
|
|
|
return ERR_OUT_OF_RANGE;
|
|
|
|
}
|
2018-01-06 19:34:32 +00:00
|
|
|
|
2020-02-11 21:36:39 +00:00
|
|
|
auto& kernel = system.Kernel();
|
2020-02-11 14:46:25 +00:00
|
|
|
Thread::ThreadSynchronizationObjects objects(handle_count);
|
2020-02-11 21:36:39 +00:00
|
|
|
const auto& handle_table = kernel.CurrentProcess()->GetHandleTable();
|
2018-01-06 19:34:32 +00:00
|
|
|
|
2018-07-24 13:55:15 +00:00
|
|
|
for (u64 i = 0; i < handle_count; ++i) {
|
2019-11-26 21:29:34 +00:00
|
|
|
const Handle handle = memory.Read32(handles_address + i * sizeof(Handle));
|
2020-02-11 14:46:25 +00:00
|
|
|
const auto object = handle_table.Get<SynchronizationObject>(handle);
|
2018-07-24 13:55:15 +00:00
|
|
|
|
|
|
|
if (object == nullptr) {
|
2018-11-26 06:06:13 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Object is a nullptr");
|
2018-01-06 19:34:32 +00:00
|
|
|
return ERR_INVALID_HANDLE;
|
2018-07-24 13:55:15 +00:00
|
|
|
}
|
|
|
|
|
2018-01-06 19:34:32 +00:00
|
|
|
objects[i] = object;
|
|
|
|
}
|
2020-02-11 21:36:39 +00:00
|
|
|
auto& synchronization = kernel.Synchronization();
|
2020-02-13 21:01:44 +00:00
|
|
|
const auto [result, handle_result] = synchronization.WaitFor(objects, nano_seconds);
|
2020-02-11 21:36:39 +00:00
|
|
|
*index = handle_result;
|
|
|
|
return result;
|
2018-01-01 19:47:57 +00:00
|
|
|
}
|
|
|
|
|
2020-03-02 05:06:41 +00:00
|
|
|
static ResultCode WaitSynchronization32(Core::System& system, u32 timeout_low, u32 handles_address,
|
|
|
|
s32 handle_count, u32 timeout_high, Handle* index) {
|
|
|
|
const s64 nano_seconds{(static_cast<s64>(timeout_high) << 32) | static_cast<s64>(timeout_low)};
|
2020-10-21 02:07:39 +00:00
|
|
|
return WaitSynchronization(system, index, handles_address, handle_count, nano_seconds);
|
2020-03-02 05:06:41 +00:00
|
|
|
}
|
|
|
|
|
2018-01-09 20:02:04 +00:00
|
|
|
/// Resumes a thread waiting on WaitSynchronization
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode CancelSynchronization(Core::System& system, Handle thread_handle) {
|
2018-07-02 16:13:26 +00:00
|
|
|
LOG_TRACE(Kernel_SVC, "called thread=0x{:X}", thread_handle);
|
2018-01-09 20:02:04 +00:00
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
const auto& handle_table = system.Kernel().CurrentProcess()->GetHandleTable();
|
2019-11-25 01:15:51 +00:00
|
|
|
std::shared_ptr<Thread> thread = handle_table.Get<Thread>(thread_handle);
|
2018-01-09 20:02:04 +00:00
|
|
|
if (!thread) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Thread handle does not exist, thread_handle=0x{:08X}",
|
|
|
|
thread_handle);
|
2018-01-09 20:02:04 +00:00
|
|
|
return ERR_INVALID_HANDLE;
|
|
|
|
}
|
|
|
|
|
2019-04-17 11:08:12 +00:00
|
|
|
thread->CancelWait();
|
2018-01-09 20:02:04 +00:00
|
|
|
return RESULT_SUCCESS;
|
|
|
|
}
|
|
|
|
|
2020-06-19 23:40:07 +00:00
|
|
|
static ResultCode CancelSynchronization32(Core::System& system, Handle thread_handle) {
|
|
|
|
return CancelSynchronization(system, thread_handle);
|
|
|
|
}
|
|
|
|
|
2018-01-01 19:02:26 +00:00
|
|
|
/// Attempts to locks a mutex, creating it if it does not already exist
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode ArbitrateLock(Core::System& system, Handle holding_thread_handle,
|
|
|
|
VAddr mutex_addr, Handle requesting_thread_handle) {
|
2018-07-02 16:13:26 +00:00
|
|
|
LOG_TRACE(Kernel_SVC,
|
2018-07-02 16:20:50 +00:00
|
|
|
"called holding_thread_handle=0x{:08X}, mutex_addr=0x{:X}, "
|
|
|
|
"requesting_current_thread_handle=0x{:08X}",
|
|
|
|
holding_thread_handle, mutex_addr, requesting_thread_handle);
|
2018-01-01 19:02:26 +00:00
|
|
|
|
2020-03-31 19:10:44 +00:00
|
|
|
if (Core::Memory::IsKernelVirtualAddress(mutex_addr)) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Mutex Address is a kernel virtual address, mutex_addr={:016X}",
|
|
|
|
mutex_addr);
|
2018-09-17 22:49:51 +00:00
|
|
|
return ERR_INVALID_ADDRESS_STATE;
|
|
|
|
}
|
|
|
|
|
2018-10-18 17:01:26 +00:00
|
|
|
if (!Common::IsWordAligned(mutex_addr)) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Mutex Address is not word aligned, mutex_addr={:016X}", mutex_addr);
|
2018-10-18 17:01:26 +00:00
|
|
|
return ERR_INVALID_ADDRESS;
|
|
|
|
}
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
auto* const current_process = system.Kernel().CurrentProcess();
|
2019-03-14 04:29:54 +00:00
|
|
|
return current_process->GetMutex().TryAcquire(mutex_addr, holding_thread_handle,
|
|
|
|
requesting_thread_handle);
|
2018-01-01 19:02:26 +00:00
|
|
|
}
|
|
|
|
|
2020-06-19 23:40:07 +00:00
|
|
|
static ResultCode ArbitrateLock32(Core::System& system, Handle holding_thread_handle,
|
|
|
|
u32 mutex_addr, Handle requesting_thread_handle) {
|
|
|
|
return ArbitrateLock(system, holding_thread_handle, static_cast<VAddr>(mutex_addr),
|
|
|
|
requesting_thread_handle);
|
|
|
|
}
|
|
|
|
|
2018-01-01 19:04:36 +00:00
|
|
|
/// Unlock a mutex
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode ArbitrateUnlock(Core::System& system, VAddr mutex_addr) {
|
2018-07-02 16:13:26 +00:00
|
|
|
LOG_TRACE(Kernel_SVC, "called mutex_addr=0x{:X}", mutex_addr);
|
2018-01-01 19:04:36 +00:00
|
|
|
|
2020-03-31 19:10:44 +00:00
|
|
|
if (Core::Memory::IsKernelVirtualAddress(mutex_addr)) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Mutex Address is a kernel virtual address, mutex_addr={:016X}",
|
|
|
|
mutex_addr);
|
2018-09-17 22:49:51 +00:00
|
|
|
return ERR_INVALID_ADDRESS_STATE;
|
|
|
|
}
|
|
|
|
|
2018-10-18 17:01:26 +00:00
|
|
|
if (!Common::IsWordAligned(mutex_addr)) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Mutex Address is not word aligned, mutex_addr={:016X}", mutex_addr);
|
2018-10-18 17:01:26 +00:00
|
|
|
return ERR_INVALID_ADDRESS;
|
|
|
|
}
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
auto* const current_process = system.Kernel().CurrentProcess();
|
2019-03-14 04:29:54 +00:00
|
|
|
return current_process->GetMutex().Release(mutex_addr);
|
2018-01-01 19:04:36 +00:00
|
|
|
}
|
|
|
|
|
2020-06-19 23:40:07 +00:00
|
|
|
static ResultCode ArbitrateUnlock32(Core::System& system, u32 mutex_addr) {
|
|
|
|
return ArbitrateUnlock(system, static_cast<VAddr>(mutex_addr));
|
|
|
|
}
|
|
|
|
|
2018-10-23 04:17:13 +00:00
|
|
|
enum class BreakType : u32 {
|
2018-10-23 04:03:59 +00:00
|
|
|
Panic = 0,
|
2018-10-23 04:17:13 +00:00
|
|
|
AssertionFailed = 1,
|
2018-10-23 04:03:59 +00:00
|
|
|
PreNROLoad = 3,
|
|
|
|
PostNROLoad = 4,
|
|
|
|
PreNROUnload = 5,
|
|
|
|
PostNROUnload = 6,
|
2019-01-27 02:19:04 +00:00
|
|
|
CppException = 7,
|
2018-10-23 04:03:59 +00:00
|
|
|
};
|
|
|
|
|
2018-10-09 01:11:14 +00:00
|
|
|
struct BreakReason {
|
|
|
|
union {
|
2018-10-10 01:23:50 +00:00
|
|
|
u32 raw;
|
2018-10-23 04:03:59 +00:00
|
|
|
BitField<0, 30, BreakType> break_type;
|
2018-10-10 01:27:44 +00:00
|
|
|
BitField<31, 1, u32> signal_debugger;
|
2018-10-09 01:11:14 +00:00
|
|
|
};
|
|
|
|
};
|
|
|
|
|
2017-10-14 21:30:07 +00:00
|
|
|
/// Break program execution
|
2019-04-06 22:46:18 +00:00
|
|
|
static void Break(Core::System& system, u32 reason, u64 info1, u64 info2) {
|
2018-10-09 01:11:14 +00:00
|
|
|
BreakReason break_reason{reason};
|
2018-11-08 04:43:54 +00:00
|
|
|
bool has_dumped_buffer{};
|
2019-05-18 01:46:17 +00:00
|
|
|
std::vector<u8> debug_buffer;
|
2018-10-23 04:03:59 +00:00
|
|
|
|
2018-11-08 04:43:54 +00:00
|
|
|
const auto handle_debug_buffer = [&](VAddr addr, u64 sz) {
|
|
|
|
if (sz == 0 || addr == 0 || has_dumped_buffer) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2019-11-26 21:29:34 +00:00
|
|
|
auto& memory = system.Memory();
|
|
|
|
|
2018-11-08 04:43:54 +00:00
|
|
|
// This typically is an error code so we're going to assume this is the case
|
|
|
|
if (sz == sizeof(u32)) {
|
2019-11-26 21:29:34 +00:00
|
|
|
LOG_CRITICAL(Debug_Emulated, "debug_buffer_err_code={:X}", memory.Read32(addr));
|
2018-11-08 04:43:54 +00:00
|
|
|
} else {
|
|
|
|
// We don't know what's in here so we'll hexdump it
|
2019-05-18 01:46:17 +00:00
|
|
|
debug_buffer.resize(sz);
|
2019-11-26 21:29:34 +00:00
|
|
|
memory.ReadBlock(addr, debug_buffer.data(), sz);
|
2018-11-08 04:43:54 +00:00
|
|
|
std::string hexdump;
|
|
|
|
for (std::size_t i = 0; i < debug_buffer.size(); i++) {
|
|
|
|
hexdump += fmt::format("{:02X} ", debug_buffer[i]);
|
|
|
|
if (i != 0 && i % 16 == 0) {
|
|
|
|
hexdump += '\n';
|
|
|
|
}
|
|
|
|
}
|
|
|
|
LOG_CRITICAL(Debug_Emulated, "debug_buffer=\n{}", hexdump);
|
|
|
|
}
|
|
|
|
has_dumped_buffer = true;
|
|
|
|
};
|
2018-10-23 04:03:59 +00:00
|
|
|
switch (break_reason.break_type) {
|
|
|
|
case BreakType::Panic:
|
2018-10-23 04:17:13 +00:00
|
|
|
LOG_CRITICAL(Debug_Emulated, "Signalling debugger, PANIC! info1=0x{:016X}, info2=0x{:016X}",
|
|
|
|
info1, info2);
|
2018-11-08 04:43:54 +00:00
|
|
|
handle_debug_buffer(info1, info2);
|
2018-10-23 04:17:13 +00:00
|
|
|
break;
|
|
|
|
case BreakType::AssertionFailed:
|
|
|
|
LOG_CRITICAL(Debug_Emulated,
|
|
|
|
"Signalling debugger, Assertion failed! info1=0x{:016X}, info2=0x{:016X}",
|
|
|
|
info1, info2);
|
2018-11-08 04:43:54 +00:00
|
|
|
handle_debug_buffer(info1, info2);
|
2018-10-23 04:03:59 +00:00
|
|
|
break;
|
|
|
|
case BreakType::PreNROLoad:
|
2018-10-23 04:17:13 +00:00
|
|
|
LOG_WARNING(
|
2018-10-09 00:10:30 +00:00
|
|
|
Debug_Emulated,
|
2018-10-23 04:17:13 +00:00
|
|
|
"Signalling debugger, Attempting to load an NRO at 0x{:016X} with size 0x{:016X}",
|
|
|
|
info1, info2);
|
2018-10-23 04:03:59 +00:00
|
|
|
break;
|
|
|
|
case BreakType::PostNROLoad:
|
2018-10-23 04:17:13 +00:00
|
|
|
LOG_WARNING(Debug_Emulated,
|
|
|
|
"Signalling debugger, Loaded an NRO at 0x{:016X} with size 0x{:016X}", info1,
|
|
|
|
info2);
|
2018-10-23 04:03:59 +00:00
|
|
|
break;
|
|
|
|
case BreakType::PreNROUnload:
|
2018-10-23 04:17:13 +00:00
|
|
|
LOG_WARNING(
|
2018-10-09 00:10:30 +00:00
|
|
|
Debug_Emulated,
|
2018-10-23 04:03:59 +00:00
|
|
|
"Signalling debugger, Attempting to unload an NRO at 0x{:016X} with size 0x{:016X}",
|
|
|
|
info1, info2);
|
|
|
|
break;
|
|
|
|
case BreakType::PostNROUnload:
|
2018-10-23 04:17:13 +00:00
|
|
|
LOG_WARNING(Debug_Emulated,
|
|
|
|
"Signalling debugger, Unloaded an NRO at 0x{:016X} with size 0x{:016X}", info1,
|
|
|
|
info2);
|
2018-10-23 04:03:59 +00:00
|
|
|
break;
|
2019-01-27 02:19:04 +00:00
|
|
|
case BreakType::CppException:
|
|
|
|
LOG_CRITICAL(Debug_Emulated, "Signalling debugger. Uncaught C++ exception encountered.");
|
|
|
|
break;
|
2018-10-23 04:03:59 +00:00
|
|
|
default:
|
2018-10-23 04:17:13 +00:00
|
|
|
LOG_WARNING(
|
|
|
|
Debug_Emulated,
|
|
|
|
"Signalling debugger, Unknown break reason {}, info1=0x{:016X}, info2=0x{:016X}",
|
|
|
|
static_cast<u32>(break_reason.break_type.Value()), info1, info2);
|
2018-11-08 04:43:54 +00:00
|
|
|
handle_debug_buffer(info1, info2);
|
2018-10-23 04:03:59 +00:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2019-05-18 01:46:17 +00:00
|
|
|
system.GetReporter().SaveSvcBreakReport(
|
|
|
|
static_cast<u32>(break_reason.break_type.Value()), break_reason.signal_debugger, info1,
|
|
|
|
info2, has_dumped_buffer ? std::make_optional(debug_buffer) : std::nullopt);
|
|
|
|
|
2018-10-23 04:03:59 +00:00
|
|
|
if (!break_reason.signal_debugger) {
|
2018-10-09 01:11:14 +00:00
|
|
|
LOG_CRITICAL(
|
2018-10-09 00:10:30 +00:00
|
|
|
Debug_Emulated,
|
|
|
|
"Emulated program broke execution! reason=0x{:016X}, info1=0x{:016X}, info2=0x{:016X}",
|
|
|
|
reason, info1, info2);
|
2019-04-06 22:46:18 +00:00
|
|
|
|
2018-11-08 04:43:54 +00:00
|
|
|
handle_debug_buffer(info1, info2);
|
2019-04-06 22:46:18 +00:00
|
|
|
|
2020-12-03 02:08:35 +00:00
|
|
|
auto* const current_thread = system.Kernel().CurrentScheduler()->GetCurrentThread();
|
2019-04-06 22:46:18 +00:00
|
|
|
const auto thread_processor_id = current_thread->GetProcessorID();
|
|
|
|
system.ArmInterface(static_cast<std::size_t>(thread_processor_id)).LogBacktrace();
|
2018-10-09 00:10:30 +00:00
|
|
|
}
|
2014-04-17 00:41:33 +00:00
|
|
|
}
|
|
|
|
|
2020-06-19 23:40:07 +00:00
|
|
|
static void Break32(Core::System& system, u32 reason, u32 info1, u32 info2) {
|
|
|
|
Break(system, reason, static_cast<u64>(info1), static_cast<u64>(info2));
|
|
|
|
}
|
|
|
|
|
2017-10-14 21:30:07 +00:00
|
|
|
/// Used to output a message on a debug hardware unit - does nothing on a retail unit
|
2020-11-23 15:17:18 +00:00
|
|
|
static void OutputDebugString(Core::System& system, VAddr address, u64 len) {
|
2018-09-12 08:51:41 +00:00
|
|
|
if (len == 0) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2018-04-25 23:11:22 +00:00
|
|
|
std::string str(len, '\0');
|
2019-11-26 21:29:34 +00:00
|
|
|
system.Memory().ReadBlock(address, str.data(), str.size());
|
2018-07-02 16:13:26 +00:00
|
|
|
LOG_DEBUG(Debug_Emulated, "{}", str);
|
2014-05-18 03:37:25 +00:00
|
|
|
}
|
|
|
|
|
2018-01-01 21:01:06 +00:00
|
|
|
/// Gets system/memory information for the current process
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode GetInfo(Core::System& system, u64* result, u64 info_id, u64 handle,
|
|
|
|
u64 info_sub_id) {
|
2020-02-27 14:28:44 +00:00
|
|
|
std::lock_guard lock{HLE::g_hle_lock};
|
2018-07-02 16:13:26 +00:00
|
|
|
LOG_TRACE(Kernel_SVC, "called info_id=0x{:X}, info_sub_id=0x{:X}, handle=0x{:08X}", info_id,
|
2018-07-02 16:20:50 +00:00
|
|
|
info_sub_id, handle);
|
2018-01-01 21:01:06 +00:00
|
|
|
|
2018-10-26 04:37:14 +00:00
|
|
|
enum class GetInfoType : u64 {
|
|
|
|
// 1.0.0+
|
2018-12-31 02:09:00 +00:00
|
|
|
AllowedCPUCoreMask = 0,
|
|
|
|
AllowedThreadPriorityMask = 1,
|
2018-10-26 04:37:14 +00:00
|
|
|
MapRegionBaseAddr = 2,
|
|
|
|
MapRegionSize = 3,
|
|
|
|
HeapRegionBaseAddr = 4,
|
|
|
|
HeapRegionSize = 5,
|
2019-06-09 22:12:02 +00:00
|
|
|
TotalPhysicalMemoryAvailable = 6,
|
2019-03-29 02:59:17 +00:00
|
|
|
TotalPhysicalMemoryUsed = 7,
|
2018-10-26 04:37:14 +00:00
|
|
|
IsCurrentProcessBeingDebugged = 8,
|
2018-12-04 05:29:15 +00:00
|
|
|
RegisterResourceLimit = 9,
|
2018-10-26 04:37:14 +00:00
|
|
|
IdleTickCount = 10,
|
|
|
|
RandomEntropy = 11,
|
2019-06-09 22:08:37 +00:00
|
|
|
ThreadTickCount = 0xF0000002,
|
2018-10-26 04:37:14 +00:00
|
|
|
// 2.0.0+
|
|
|
|
ASLRRegionBaseAddr = 12,
|
|
|
|
ASLRRegionSize = 13,
|
2019-07-06 06:02:01 +00:00
|
|
|
StackRegionBaseAddr = 14,
|
|
|
|
StackRegionSize = 15,
|
2018-10-26 04:37:14 +00:00
|
|
|
// 3.0.0+
|
2019-07-07 16:42:54 +00:00
|
|
|
SystemResourceSize = 16,
|
|
|
|
SystemResourceUsage = 17,
|
2018-10-26 04:37:14 +00:00
|
|
|
TitleId = 18,
|
|
|
|
// 4.0.0+
|
|
|
|
PrivilegedProcessId = 19,
|
|
|
|
// 5.0.0+
|
|
|
|
UserExceptionContextAddr = 20,
|
2019-06-09 22:20:20 +00:00
|
|
|
// 6.0.0+
|
2019-07-07 18:48:11 +00:00
|
|
|
TotalPhysicalMemoryAvailableWithoutSystemResource = 21,
|
|
|
|
TotalPhysicalMemoryUsedWithoutSystemResource = 22,
|
2018-10-26 04:37:14 +00:00
|
|
|
};
|
|
|
|
|
2018-12-02 06:37:15 +00:00
|
|
|
const auto info_id_type = static_cast<GetInfoType>(info_id);
|
2018-01-10 05:58:25 +00:00
|
|
|
|
2018-12-02 06:37:15 +00:00
|
|
|
switch (info_id_type) {
|
2018-12-31 02:09:00 +00:00
|
|
|
case GetInfoType::AllowedCPUCoreMask:
|
|
|
|
case GetInfoType::AllowedThreadPriorityMask:
|
2018-01-16 22:06:45 +00:00
|
|
|
case GetInfoType::MapRegionBaseAddr:
|
|
|
|
case GetInfoType::MapRegionSize:
|
2018-01-15 20:42:57 +00:00
|
|
|
case GetInfoType::HeapRegionBaseAddr:
|
|
|
|
case GetInfoType::HeapRegionSize:
|
2018-12-02 06:37:15 +00:00
|
|
|
case GetInfoType::ASLRRegionBaseAddr:
|
|
|
|
case GetInfoType::ASLRRegionSize:
|
2019-07-06 06:02:01 +00:00
|
|
|
case GetInfoType::StackRegionBaseAddr:
|
|
|
|
case GetInfoType::StackRegionSize:
|
2019-06-09 22:12:02 +00:00
|
|
|
case GetInfoType::TotalPhysicalMemoryAvailable:
|
2019-03-29 02:59:17 +00:00
|
|
|
case GetInfoType::TotalPhysicalMemoryUsed:
|
2019-07-07 16:42:54 +00:00
|
|
|
case GetInfoType::SystemResourceSize:
|
|
|
|
case GetInfoType::SystemResourceUsage:
|
2018-12-02 06:37:15 +00:00
|
|
|
case GetInfoType::TitleId:
|
2019-06-09 22:20:20 +00:00
|
|
|
case GetInfoType::UserExceptionContextAddr:
|
2019-07-07 18:48:11 +00:00
|
|
|
case GetInfoType::TotalPhysicalMemoryAvailableWithoutSystemResource:
|
|
|
|
case GetInfoType::TotalPhysicalMemoryUsedWithoutSystemResource: {
|
2018-12-02 06:37:15 +00:00
|
|
|
if (info_sub_id != 0) {
|
2020-04-29 04:53:53 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Info sub id is non zero! info_id={}, info_sub_id={}", info_id,
|
|
|
|
info_sub_id);
|
2018-12-02 06:37:15 +00:00
|
|
|
return ERR_INVALID_ENUM_VALUE;
|
|
|
|
}
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
const auto& current_process_handle_table =
|
|
|
|
system.Kernel().CurrentProcess()->GetHandleTable();
|
svc: Use the current process' handle table for retrieving the process instance to act upon
The kernel uses the handle table of the current process to retrieve the
process that should be used to retrieve certain information. To someone
not familiar with the kernel, this might raise the question of "Ok,
sounds nice, but doesn't this make it impossible to retrieve information
about the current process?".
No, it doesn't, because HandleTable instances in the kernel have the
notion of a "pseudo-handle", where certain values allow the kernel to
lookup objects outside of a given handle table. Currently, there's only
a pseudo-handle for the current process (0xFFFF8001) and a pseudo-handle
for the current thread (0xFFFF8000), so to retrieve the current process,
one would just pass 0xFFFF8001 into svcGetInfo.
The lookup itself in the handle table would be something like:
template <typename T>
T* Lookup(Handle handle) {
if (handle == PSEUDO_HANDLE_CURRENT_PROCESS) {
return CurrentProcess();
}
if (handle == PSUEDO_HANDLE_CURRENT_THREAD) {
return CurrentThread();
}
return static_cast<T*>(&objects[handle]);
}
which, as is shown, allows accessing the current process or current
thread, even if those two objects aren't actually within the HandleTable
instance.
2018-12-02 07:00:11 +00:00
|
|
|
const auto process = current_process_handle_table.Get<Process>(static_cast<Handle>(handle));
|
2018-12-02 06:37:15 +00:00
|
|
|
if (!process) {
|
2020-04-29 04:53:53 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Process is not valid! info_id={}, info_sub_id={}, handle={:08X}",
|
|
|
|
info_id, info_sub_id, handle);
|
2018-12-02 06:37:15 +00:00
|
|
|
return ERR_INVALID_HANDLE;
|
|
|
|
}
|
|
|
|
|
|
|
|
switch (info_id_type) {
|
2018-12-31 02:09:00 +00:00
|
|
|
case GetInfoType::AllowedCPUCoreMask:
|
|
|
|
*result = process->GetCoreMask();
|
2018-12-02 06:37:15 +00:00
|
|
|
return RESULT_SUCCESS;
|
|
|
|
|
2018-12-31 02:09:00 +00:00
|
|
|
case GetInfoType::AllowedThreadPriorityMask:
|
|
|
|
*result = process->GetPriorityMask();
|
2018-12-02 06:37:15 +00:00
|
|
|
return RESULT_SUCCESS;
|
|
|
|
|
|
|
|
case GetInfoType::MapRegionBaseAddr:
|
2020-04-09 03:14:18 +00:00
|
|
|
*result = process->PageTable().GetAliasRegionStart();
|
2018-12-02 06:37:15 +00:00
|
|
|
return RESULT_SUCCESS;
|
|
|
|
|
|
|
|
case GetInfoType::MapRegionSize:
|
2020-04-09 03:14:18 +00:00
|
|
|
*result = process->PageTable().GetAliasRegionSize();
|
2018-12-02 06:37:15 +00:00
|
|
|
return RESULT_SUCCESS;
|
|
|
|
|
|
|
|
case GetInfoType::HeapRegionBaseAddr:
|
2020-04-09 03:14:18 +00:00
|
|
|
*result = process->PageTable().GetHeapRegionStart();
|
2018-12-02 06:37:15 +00:00
|
|
|
return RESULT_SUCCESS;
|
|
|
|
|
|
|
|
case GetInfoType::HeapRegionSize:
|
2020-04-09 03:14:18 +00:00
|
|
|
*result = process->PageTable().GetHeapRegionSize();
|
2018-12-02 06:37:15 +00:00
|
|
|
return RESULT_SUCCESS;
|
|
|
|
|
|
|
|
case GetInfoType::ASLRRegionBaseAddr:
|
2020-04-09 03:14:18 +00:00
|
|
|
*result = process->PageTable().GetAliasCodeRegionStart();
|
2018-12-02 06:37:15 +00:00
|
|
|
return RESULT_SUCCESS;
|
|
|
|
|
|
|
|
case GetInfoType::ASLRRegionSize:
|
2020-04-09 03:14:18 +00:00
|
|
|
*result = process->PageTable().GetAliasCodeRegionSize();
|
2018-12-02 06:37:15 +00:00
|
|
|
return RESULT_SUCCESS;
|
|
|
|
|
2019-07-06 06:02:01 +00:00
|
|
|
case GetInfoType::StackRegionBaseAddr:
|
2020-04-09 03:14:18 +00:00
|
|
|
*result = process->PageTable().GetStackRegionStart();
|
2018-12-02 06:37:15 +00:00
|
|
|
return RESULT_SUCCESS;
|
|
|
|
|
2019-07-06 06:02:01 +00:00
|
|
|
case GetInfoType::StackRegionSize:
|
2020-04-09 03:14:18 +00:00
|
|
|
*result = process->PageTable().GetStackRegionSize();
|
2018-12-02 06:37:15 +00:00
|
|
|
return RESULT_SUCCESS;
|
|
|
|
|
2019-06-09 22:12:02 +00:00
|
|
|
case GetInfoType::TotalPhysicalMemoryAvailable:
|
2019-06-09 22:20:20 +00:00
|
|
|
*result = process->GetTotalPhysicalMemoryAvailable();
|
2018-12-02 06:37:15 +00:00
|
|
|
return RESULT_SUCCESS;
|
|
|
|
|
2019-03-29 02:59:17 +00:00
|
|
|
case GetInfoType::TotalPhysicalMemoryUsed:
|
|
|
|
*result = process->GetTotalPhysicalMemoryUsed();
|
2018-12-02 06:37:15 +00:00
|
|
|
return RESULT_SUCCESS;
|
|
|
|
|
2019-07-07 16:42:54 +00:00
|
|
|
case GetInfoType::SystemResourceSize:
|
|
|
|
*result = process->GetSystemResourceSize();
|
|
|
|
return RESULT_SUCCESS;
|
|
|
|
|
|
|
|
case GetInfoType::SystemResourceUsage:
|
2019-07-07 19:08:29 +00:00
|
|
|
LOG_WARNING(Kernel_SVC, "(STUBBED) Attempted to query system resource usage");
|
2019-07-07 18:48:11 +00:00
|
|
|
*result = process->GetSystemResourceUsage();
|
2018-12-02 06:37:15 +00:00
|
|
|
return RESULT_SUCCESS;
|
|
|
|
|
|
|
|
case GetInfoType::TitleId:
|
|
|
|
*result = process->GetTitleID();
|
|
|
|
return RESULT_SUCCESS;
|
|
|
|
|
|
|
|
case GetInfoType::UserExceptionContextAddr:
|
2019-07-07 08:19:16 +00:00
|
|
|
*result = process->GetTLSRegionAddress();
|
2018-12-02 06:37:15 +00:00
|
|
|
return RESULT_SUCCESS;
|
|
|
|
|
2019-07-07 18:48:11 +00:00
|
|
|
case GetInfoType::TotalPhysicalMemoryAvailableWithoutSystemResource:
|
|
|
|
*result = process->GetTotalPhysicalMemoryAvailableWithoutSystemResource();
|
2019-06-09 22:20:20 +00:00
|
|
|
return RESULT_SUCCESS;
|
|
|
|
|
2019-07-07 18:48:11 +00:00
|
|
|
case GetInfoType::TotalPhysicalMemoryUsedWithoutSystemResource:
|
|
|
|
*result = process->GetTotalPhysicalMemoryUsedWithoutSystemResource();
|
2019-06-09 22:20:20 +00:00
|
|
|
return RESULT_SUCCESS;
|
|
|
|
|
2018-12-02 06:37:15 +00:00
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2020-04-28 15:17:59 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Unimplemented svcGetInfo id=0x{:016X}", info_id);
|
2018-12-02 06:37:15 +00:00
|
|
|
return ERR_INVALID_ENUM_VALUE;
|
|
|
|
}
|
|
|
|
|
2018-02-04 17:34:45 +00:00
|
|
|
case GetInfoType::IsCurrentProcessBeingDebugged:
|
|
|
|
*result = 0;
|
2018-12-02 06:37:15 +00:00
|
|
|
return RESULT_SUCCESS;
|
|
|
|
|
2018-12-04 05:29:15 +00:00
|
|
|
case GetInfoType::RegisterResourceLimit: {
|
|
|
|
if (handle != 0) {
|
2020-04-29 04:53:53 +00:00
|
|
|
LOG_ERROR(Kernel, "Handle is non zero! handle={:08X}", handle);
|
2018-12-04 05:29:15 +00:00
|
|
|
return ERR_INVALID_HANDLE;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (info_sub_id != 0) {
|
2020-04-29 04:53:53 +00:00
|
|
|
LOG_ERROR(Kernel, "Info sub id is non zero! info_id={}, info_sub_id={}", info_id,
|
|
|
|
info_sub_id);
|
2018-12-04 05:29:15 +00:00
|
|
|
return ERR_INVALID_COMBINATION;
|
|
|
|
}
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
Process* const current_process = system.Kernel().CurrentProcess();
|
2018-12-04 05:29:15 +00:00
|
|
|
HandleTable& handle_table = current_process->GetHandleTable();
|
|
|
|
const auto resource_limit = current_process->GetResourceLimit();
|
|
|
|
if (!resource_limit) {
|
|
|
|
*result = KernelHandle::InvalidHandle;
|
|
|
|
// Yes, the kernel considers this a successful operation.
|
|
|
|
return RESULT_SUCCESS;
|
|
|
|
}
|
|
|
|
|
|
|
|
const auto table_result = handle_table.Create(resource_limit);
|
|
|
|
if (table_result.Failed()) {
|
|
|
|
return table_result.Code();
|
|
|
|
}
|
|
|
|
|
|
|
|
*result = *table_result;
|
|
|
|
return RESULT_SUCCESS;
|
|
|
|
}
|
|
|
|
|
2018-01-01 21:01:06 +00:00
|
|
|
case GetInfoType::RandomEntropy:
|
2018-11-13 17:25:43 +00:00
|
|
|
if (handle != 0) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Process Handle is non zero, expected 0 result but got {:016X}",
|
|
|
|
handle);
|
2018-11-13 17:25:43 +00:00
|
|
|
return ERR_INVALID_HANDLE;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (info_sub_id >= Process::RANDOM_ENTROPY_SIZE) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Entropy size is out of range, expected {} but got {}",
|
|
|
|
Process::RANDOM_ENTROPY_SIZE, info_sub_id);
|
2018-11-16 19:24:27 +00:00
|
|
|
return ERR_INVALID_COMBINATION;
|
2018-11-13 17:25:43 +00:00
|
|
|
}
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
*result = system.Kernel().CurrentProcess()->GetRandomEntropy(info_sub_id);
|
2018-11-13 17:25:43 +00:00
|
|
|
return RESULT_SUCCESS;
|
2018-12-02 06:37:15 +00:00
|
|
|
|
2018-01-15 20:42:57 +00:00
|
|
|
case GetInfoType::PrivilegedProcessId:
|
2018-07-02 16:13:26 +00:00
|
|
|
LOG_WARNING(Kernel_SVC,
|
2018-07-02 16:20:50 +00:00
|
|
|
"(STUBBED) Attempted to query privileged process id bounds, returned 0");
|
2018-01-15 20:42:57 +00:00
|
|
|
*result = 0;
|
2018-12-02 06:37:15 +00:00
|
|
|
return RESULT_SUCCESS;
|
|
|
|
|
2018-10-25 22:42:50 +00:00
|
|
|
case GetInfoType::ThreadTickCount: {
|
|
|
|
constexpr u64 num_cpus = 4;
|
|
|
|
if (info_sub_id != 0xFFFFFFFFFFFFFFFF && info_sub_id >= num_cpus) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Core count is out of range, expected {} but got {}", num_cpus,
|
|
|
|
info_sub_id);
|
2018-11-16 19:24:27 +00:00
|
|
|
return ERR_INVALID_COMBINATION;
|
2018-10-25 22:42:50 +00:00
|
|
|
}
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
const auto thread = system.Kernel().CurrentProcess()->GetHandleTable().Get<Thread>(
|
|
|
|
static_cast<Handle>(handle));
|
2018-10-25 22:42:50 +00:00
|
|
|
if (!thread) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Thread handle does not exist, handle=0x{:08X}",
|
|
|
|
static_cast<Handle>(handle));
|
2018-10-25 22:42:50 +00:00
|
|
|
return ERR_INVALID_HANDLE;
|
|
|
|
}
|
|
|
|
|
2019-02-14 17:42:58 +00:00
|
|
|
const auto& core_timing = system.CoreTiming();
|
2020-12-03 02:08:35 +00:00
|
|
|
const auto& scheduler = *system.Kernel().CurrentScheduler();
|
2018-10-25 22:42:50 +00:00
|
|
|
const auto* const current_thread = scheduler.GetCurrentThread();
|
2019-11-25 01:15:51 +00:00
|
|
|
const bool same_thread = current_thread == thread.get();
|
2018-10-25 22:42:50 +00:00
|
|
|
|
|
|
|
const u64 prev_ctx_ticks = scheduler.GetLastContextSwitchTicks();
|
|
|
|
u64 out_ticks = 0;
|
|
|
|
if (same_thread && info_sub_id == 0xFFFFFFFFFFFFFFFF) {
|
|
|
|
const u64 thread_ticks = current_thread->GetTotalCPUTimeTicks();
|
|
|
|
|
2020-02-25 02:04:12 +00:00
|
|
|
out_ticks = thread_ticks + (core_timing.GetCPUTicks() - prev_ctx_ticks);
|
2018-10-25 22:42:50 +00:00
|
|
|
} else if (same_thread && info_sub_id == system.CurrentCoreIndex()) {
|
2020-02-25 02:04:12 +00:00
|
|
|
out_ticks = core_timing.GetCPUTicks() - prev_ctx_ticks;
|
2018-10-25 22:42:50 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
*result = out_ticks;
|
2018-12-02 06:37:15 +00:00
|
|
|
return RESULT_SUCCESS;
|
2018-10-25 22:42:50 +00:00
|
|
|
}
|
2018-12-02 06:37:15 +00:00
|
|
|
|
2018-01-01 21:01:06 +00:00
|
|
|
default:
|
2020-04-28 15:17:59 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Unimplemented svcGetInfo id=0x{:016X}", info_id);
|
2018-11-25 21:48:44 +00:00
|
|
|
return ERR_INVALID_ENUM_VALUE;
|
2015-05-17 05:06:59 +00:00
|
|
|
}
|
2014-05-01 22:50:36 +00:00
|
|
|
}
|
|
|
|
|
2020-03-02 05:06:41 +00:00
|
|
|
static ResultCode GetInfo32(Core::System& system, u32* result_low, u32* result_high, u32 sub_id_low,
|
|
|
|
u32 info_id, u32 handle, u32 sub_id_high) {
|
|
|
|
const u64 sub_id{static_cast<u64>(sub_id_low | (static_cast<u64>(sub_id_high) << 32))};
|
|
|
|
u64 res_value{};
|
|
|
|
|
|
|
|
const ResultCode result{GetInfo(system, &res_value, info_id, handle, sub_id)};
|
|
|
|
*result_high = static_cast<u32>(res_value >> 32);
|
|
|
|
*result_low = static_cast<u32>(res_value & std::numeric_limits<u32>::max());
|
|
|
|
|
|
|
|
return result;
|
|
|
|
}
|
|
|
|
|
2019-07-07 16:42:54 +00:00
|
|
|
/// Maps memory at a desired address
|
|
|
|
static ResultCode MapPhysicalMemory(Core::System& system, VAddr addr, u64 size) {
|
2020-02-27 14:28:44 +00:00
|
|
|
std::lock_guard lock{HLE::g_hle_lock};
|
2019-07-07 16:42:54 +00:00
|
|
|
LOG_DEBUG(Kernel_SVC, "called, addr=0x{:016X}, size=0x{:X}", addr, size);
|
|
|
|
|
|
|
|
if (!Common::Is4KBAligned(addr)) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Address is not aligned to 4KB, 0x{:016X}", addr);
|
|
|
|
return ERR_INVALID_ADDRESS;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!Common::Is4KBAligned(size)) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Size is not aligned to 4KB, 0x{:X}", size);
|
|
|
|
return ERR_INVALID_SIZE;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (size == 0) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Size is zero");
|
|
|
|
return ERR_INVALID_SIZE;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!(addr < addr + size)) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Size causes 64-bit overflow of address");
|
|
|
|
return ERR_INVALID_MEMORY_RANGE;
|
|
|
|
}
|
|
|
|
|
2020-04-09 03:14:18 +00:00
|
|
|
Process* const current_process{system.Kernel().CurrentProcess()};
|
|
|
|
auto& page_table{current_process->PageTable()};
|
2019-07-07 16:42:54 +00:00
|
|
|
|
|
|
|
if (current_process->GetSystemResourceSize() == 0) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "System Resource Size is zero");
|
|
|
|
return ERR_INVALID_STATE;
|
|
|
|
}
|
|
|
|
|
2020-04-09 03:14:18 +00:00
|
|
|
if (!page_table.IsInsideAddressSpace(addr, size)) {
|
|
|
|
LOG_ERROR(Kernel_SVC,
|
|
|
|
"Address is not within the address space, addr=0x{:016X}, size=0x{:016X}", addr,
|
|
|
|
size);
|
|
|
|
return ERR_INVALID_MEMORY_RANGE;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (page_table.IsOutsideAliasRegion(addr, size)) {
|
|
|
|
LOG_ERROR(Kernel_SVC,
|
|
|
|
"Address is not within the alias region, addr=0x{:016X}, size=0x{:016X}", addr,
|
|
|
|
size);
|
2019-07-07 16:42:54 +00:00
|
|
|
return ERR_INVALID_MEMORY_RANGE;
|
|
|
|
}
|
|
|
|
|
2020-04-09 03:14:18 +00:00
|
|
|
return page_table.MapPhysicalMemory(addr, size);
|
2019-07-07 16:42:54 +00:00
|
|
|
}
|
|
|
|
|
2020-06-19 23:40:07 +00:00
|
|
|
static ResultCode MapPhysicalMemory32(Core::System& system, u32 addr, u32 size) {
|
|
|
|
return MapPhysicalMemory(system, static_cast<VAddr>(addr), static_cast<std::size_t>(size));
|
|
|
|
}
|
|
|
|
|
2019-07-07 16:42:54 +00:00
|
|
|
/// Unmaps memory previously mapped via MapPhysicalMemory
|
|
|
|
static ResultCode UnmapPhysicalMemory(Core::System& system, VAddr addr, u64 size) {
|
2020-02-27 14:28:44 +00:00
|
|
|
std::lock_guard lock{HLE::g_hle_lock};
|
2019-07-07 16:42:54 +00:00
|
|
|
LOG_DEBUG(Kernel_SVC, "called, addr=0x{:016X}, size=0x{:X}", addr, size);
|
|
|
|
|
|
|
|
if (!Common::Is4KBAligned(addr)) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Address is not aligned to 4KB, 0x{:016X}", addr);
|
|
|
|
return ERR_INVALID_ADDRESS;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!Common::Is4KBAligned(size)) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Size is not aligned to 4KB, 0x{:X}", size);
|
|
|
|
return ERR_INVALID_SIZE;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (size == 0) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Size is zero");
|
|
|
|
return ERR_INVALID_SIZE;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!(addr < addr + size)) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Size causes 64-bit overflow of address");
|
|
|
|
return ERR_INVALID_MEMORY_RANGE;
|
|
|
|
}
|
|
|
|
|
2020-04-09 03:14:18 +00:00
|
|
|
Process* const current_process{system.Kernel().CurrentProcess()};
|
|
|
|
auto& page_table{current_process->PageTable()};
|
2019-07-07 16:42:54 +00:00
|
|
|
|
|
|
|
if (current_process->GetSystemResourceSize() == 0) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "System Resource Size is zero");
|
|
|
|
return ERR_INVALID_STATE;
|
|
|
|
}
|
|
|
|
|
2020-04-09 03:14:18 +00:00
|
|
|
if (!page_table.IsInsideAddressSpace(addr, size)) {
|
|
|
|
LOG_ERROR(Kernel_SVC,
|
|
|
|
"Address is not within the address space, addr=0x{:016X}, size=0x{:016X}", addr,
|
|
|
|
size);
|
|
|
|
return ERR_INVALID_MEMORY_RANGE;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (page_table.IsOutsideAliasRegion(addr, size)) {
|
|
|
|
LOG_ERROR(Kernel_SVC,
|
|
|
|
"Address is not within the alias region, addr=0x{:016X}, size=0x{:016X}", addr,
|
|
|
|
size);
|
2019-07-07 16:42:54 +00:00
|
|
|
return ERR_INVALID_MEMORY_RANGE;
|
|
|
|
}
|
|
|
|
|
2020-04-09 03:14:18 +00:00
|
|
|
return page_table.UnmapPhysicalMemory(addr, size);
|
2019-07-07 16:42:54 +00:00
|
|
|
}
|
|
|
|
|
2020-06-19 23:40:07 +00:00
|
|
|
static ResultCode UnmapPhysicalMemory32(Core::System& system, u32 addr, u32 size) {
|
|
|
|
return UnmapPhysicalMemory(system, static_cast<VAddr>(addr), static_cast<std::size_t>(size));
|
|
|
|
}
|
|
|
|
|
2018-04-03 03:50:17 +00:00
|
|
|
/// Sets the thread activity
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode SetThreadActivity(Core::System& system, Handle handle, u32 activity) {
|
2018-12-03 17:25:27 +00:00
|
|
|
LOG_DEBUG(Kernel_SVC, "called, handle=0x{:08X}, activity=0x{:08X}", handle, activity);
|
|
|
|
if (activity > static_cast<u32>(ThreadActivity::Paused)) {
|
|
|
|
return ERR_INVALID_ENUM_VALUE;
|
|
|
|
}
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
const auto* current_process = system.Kernel().CurrentProcess();
|
2019-11-25 01:15:51 +00:00
|
|
|
const std::shared_ptr<Thread> thread = current_process->GetHandleTable().Get<Thread>(handle);
|
2018-12-03 17:25:27 +00:00
|
|
|
if (!thread) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Thread handle does not exist, handle=0x{:08X}", handle);
|
|
|
|
return ERR_INVALID_HANDLE;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (thread->GetOwnerProcess() != current_process) {
|
|
|
|
LOG_ERROR(Kernel_SVC,
|
|
|
|
"The current process does not own the current thread, thread_handle={:08X} "
|
|
|
|
"thread_pid={}, "
|
|
|
|
"current_process_pid={}",
|
|
|
|
handle, thread->GetOwnerProcess()->GetProcessID(),
|
|
|
|
current_process->GetProcessID());
|
|
|
|
return ERR_INVALID_HANDLE;
|
|
|
|
}
|
|
|
|
|
2020-12-03 02:08:35 +00:00
|
|
|
if (thread.get() == system.Kernel().CurrentScheduler()->GetCurrentThread()) {
|
2018-12-03 17:25:27 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "The thread handle specified is the current running thread");
|
|
|
|
return ERR_BUSY;
|
|
|
|
}
|
|
|
|
|
2020-03-07 16:44:35 +00:00
|
|
|
return thread->SetActivity(static_cast<ThreadActivity>(activity));
|
2018-04-03 03:50:17 +00:00
|
|
|
}
|
|
|
|
|
2020-06-19 23:40:07 +00:00
|
|
|
static ResultCode SetThreadActivity32(Core::System& system, Handle handle, u32 activity) {
|
|
|
|
return SetThreadActivity(system, handle, activity);
|
|
|
|
}
|
|
|
|
|
2018-04-03 03:50:17 +00:00
|
|
|
/// Gets the thread context
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode GetThreadContext(Core::System& system, VAddr thread_context, Handle handle) {
|
2018-09-29 23:58:21 +00:00
|
|
|
LOG_DEBUG(Kernel_SVC, "called, context=0x{:08X}, thread=0x{:X}", thread_context, handle);
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
const auto* current_process = system.Kernel().CurrentProcess();
|
2019-11-25 01:15:51 +00:00
|
|
|
const std::shared_ptr<Thread> thread = current_process->GetHandleTable().Get<Thread>(handle);
|
2018-09-29 23:58:21 +00:00
|
|
|
if (!thread) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Thread handle does not exist, handle=0x{:08X}", handle);
|
2018-09-29 23:58:21 +00:00
|
|
|
return ERR_INVALID_HANDLE;
|
|
|
|
}
|
|
|
|
|
2018-10-03 22:47:57 +00:00
|
|
|
if (thread->GetOwnerProcess() != current_process) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC,
|
|
|
|
"The current process does not own the current thread, thread_handle={:08X} "
|
|
|
|
"thread_pid={}, "
|
|
|
|
"current_process_pid={}",
|
|
|
|
handle, thread->GetOwnerProcess()->GetProcessID(),
|
|
|
|
current_process->GetProcessID());
|
2018-09-29 23:58:21 +00:00
|
|
|
return ERR_INVALID_HANDLE;
|
|
|
|
}
|
|
|
|
|
2020-12-03 02:08:35 +00:00
|
|
|
if (thread.get() == system.Kernel().CurrentScheduler()->GetCurrentThread()) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "The thread handle specified is the current running thread");
|
2018-12-03 17:25:27 +00:00
|
|
|
return ERR_BUSY;
|
2018-09-29 23:58:21 +00:00
|
|
|
}
|
|
|
|
|
2020-03-02 05:06:41 +00:00
|
|
|
Core::ARM_Interface::ThreadContext64 ctx = thread->GetContext64();
|
2018-09-29 23:58:21 +00:00
|
|
|
// Mask away mode bits, interrupt bits, IL bit, and other reserved bits.
|
|
|
|
ctx.pstate &= 0xFF0FFE20;
|
|
|
|
|
|
|
|
// If 64-bit, we can just write the context registers directly and we're good.
|
|
|
|
// However, if 32-bit, we have to ensure some registers are zeroed out.
|
|
|
|
if (!current_process->Is64BitProcess()) {
|
|
|
|
std::fill(ctx.cpu_registers.begin() + 15, ctx.cpu_registers.end(), 0);
|
|
|
|
std::fill(ctx.vector_registers.begin() + 16, ctx.vector_registers.end(), u128{});
|
|
|
|
}
|
|
|
|
|
2019-11-26 22:39:57 +00:00
|
|
|
system.Memory().WriteBlock(thread_context, &ctx, sizeof(ctx));
|
2018-04-03 03:50:17 +00:00
|
|
|
return RESULT_SUCCESS;
|
|
|
|
}
|
|
|
|
|
2020-06-19 23:40:07 +00:00
|
|
|
static ResultCode GetThreadContext32(Core::System& system, u32 thread_context, Handle handle) {
|
|
|
|
return GetThreadContext(system, static_cast<VAddr>(thread_context), handle);
|
|
|
|
}
|
|
|
|
|
2014-06-02 02:12:54 +00:00
|
|
|
/// Gets the priority for the specified thread
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode GetThreadPriority(Core::System& system, u32* priority, Handle handle) {
|
2018-11-26 06:06:13 +00:00
|
|
|
LOG_TRACE(Kernel_SVC, "called");
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
const auto& handle_table = system.Kernel().CurrentProcess()->GetHandleTable();
|
2019-11-25 01:15:51 +00:00
|
|
|
const std::shared_ptr<Thread> thread = handle_table.Get<Thread>(handle);
|
2018-10-20 18:34:41 +00:00
|
|
|
if (!thread) {
|
2020-02-25 17:22:11 +00:00
|
|
|
*priority = 0;
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Thread handle does not exist, handle=0x{:08X}", handle);
|
2017-12-31 21:06:11 +00:00
|
|
|
return ERR_INVALID_HANDLE;
|
2018-10-20 18:34:41 +00:00
|
|
|
}
|
2017-12-31 21:06:11 +00:00
|
|
|
|
|
|
|
*priority = thread->GetPriority();
|
2015-01-23 05:36:58 +00:00
|
|
|
return RESULT_SUCCESS;
|
2014-12-03 23:49:51 +00:00
|
|
|
}
|
|
|
|
|
2020-03-02 05:06:41 +00:00
|
|
|
static ResultCode GetThreadPriority32(Core::System& system, u32* priority, Handle handle) {
|
|
|
|
return GetThreadPriority(system, priority, handle);
|
|
|
|
}
|
|
|
|
|
2017-12-31 20:58:16 +00:00
|
|
|
/// Sets the priority for the specified thread
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode SetThreadPriority(Core::System& system, Handle handle, u32 priority) {
|
2018-11-26 06:06:13 +00:00
|
|
|
LOG_TRACE(Kernel_SVC, "called");
|
|
|
|
|
2017-12-31 20:58:16 +00:00
|
|
|
if (priority > THREADPRIO_LOWEST) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(
|
|
|
|
Kernel_SVC,
|
|
|
|
"An invalid priority was specified, expected {} but got {} for thread_handle={:08X}",
|
|
|
|
THREADPRIO_LOWEST, priority, handle);
|
2018-09-12 08:25:53 +00:00
|
|
|
return ERR_INVALID_THREAD_PRIORITY;
|
2017-12-31 20:58:16 +00:00
|
|
|
}
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
const auto* const current_process = system.Kernel().CurrentProcess();
|
2017-12-31 20:58:16 +00:00
|
|
|
|
2019-11-25 01:15:51 +00:00
|
|
|
std::shared_ptr<Thread> thread = current_process->GetHandleTable().Get<Thread>(handle);
|
2018-10-24 18:07:53 +00:00
|
|
|
if (!thread) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Thread handle does not exist, handle=0x{:08X}", handle);
|
2018-10-24 18:07:53 +00:00
|
|
|
return ERR_INVALID_HANDLE;
|
2017-12-31 20:58:16 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
thread->SetPriority(priority);
|
|
|
|
|
|
|
|
return RESULT_SUCCESS;
|
|
|
|
}
|
|
|
|
|
2020-06-18 22:15:19 +00:00
|
|
|
static ResultCode SetThreadPriority32(Core::System& system, Handle handle, u32 priority) {
|
2020-06-19 00:33:04 +00:00
|
|
|
return SetThreadPriority(system, handle, priority);
|
2020-06-18 22:15:19 +00:00
|
|
|
}
|
|
|
|
|
2017-12-31 21:01:04 +00:00
|
|
|
/// Get which CPU core is executing the current thread
|
2019-04-06 22:46:18 +00:00
|
|
|
static u32 GetCurrentProcessorNumber(Core::System& system) {
|
2018-07-02 16:13:26 +00:00
|
|
|
LOG_TRACE(Kernel_SVC, "called");
|
2020-02-25 17:22:11 +00:00
|
|
|
return static_cast<u32>(system.CurrentPhysicalCore().CoreIndex());
|
2017-12-31 21:01:04 +00:00
|
|
|
}
|
|
|
|
|
2020-06-19 00:33:04 +00:00
|
|
|
static u32 GetCurrentProcessorNumber32(Core::System& system) {
|
|
|
|
return GetCurrentProcessorNumber(system);
|
|
|
|
}
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode MapSharedMemory(Core::System& system, Handle shared_memory_handle, VAddr addr,
|
|
|
|
u64 size, u32 permissions) {
|
2020-02-27 14:28:44 +00:00
|
|
|
std::lock_guard lock{HLE::g_hle_lock};
|
2018-07-02 16:20:50 +00:00
|
|
|
LOG_TRACE(Kernel_SVC,
|
|
|
|
"called, shared_memory_handle=0x{:X}, addr=0x{:X}, size=0x{:X}, permissions=0x{:08X}",
|
|
|
|
shared_memory_handle, addr, size, permissions);
|
2018-01-14 22:15:31 +00:00
|
|
|
|
2018-10-18 16:55:27 +00:00
|
|
|
if (!Common::Is4KBAligned(addr)) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Address is not aligned to 4KB, addr=0x{:016X}", addr);
|
2018-09-14 00:16:43 +00:00
|
|
|
return ERR_INVALID_ADDRESS;
|
|
|
|
}
|
|
|
|
|
2018-11-26 08:47:39 +00:00
|
|
|
if (size == 0) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Size is 0");
|
|
|
|
return ERR_INVALID_SIZE;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!Common::Is4KBAligned(size)) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Size is not aligned to 4KB, size=0x{:016X}", size);
|
2018-09-14 00:16:43 +00:00
|
|
|
return ERR_INVALID_SIZE;
|
|
|
|
}
|
|
|
|
|
2018-10-18 02:39:21 +00:00
|
|
|
if (!IsValidAddressRange(addr, size)) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Region is not a valid address range, addr=0x{:016X}, size=0x{:016X}",
|
|
|
|
addr, size);
|
2018-10-18 02:39:21 +00:00
|
|
|
return ERR_INVALID_ADDRESS_STATE;
|
|
|
|
}
|
|
|
|
|
2020-04-09 03:14:18 +00:00
|
|
|
const auto permission_type = static_cast<Memory::MemoryPermission>(permissions);
|
|
|
|
if ((permission_type | Memory::MemoryPermission::Write) !=
|
|
|
|
Memory::MemoryPermission::ReadAndWrite) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Expected Read or ReadWrite permission but got permissions=0x{:08X}",
|
|
|
|
permissions);
|
2018-09-14 00:16:43 +00:00
|
|
|
return ERR_INVALID_MEMORY_PERMISSIONS;
|
|
|
|
}
|
|
|
|
|
2020-04-09 03:14:18 +00:00
|
|
|
auto* const current_process{system.Kernel().CurrentProcess()};
|
|
|
|
auto& page_table{current_process->PageTable()};
|
2018-01-14 22:15:31 +00:00
|
|
|
|
2020-04-09 03:14:18 +00:00
|
|
|
if (page_table.IsInvalidRegion(addr, size)) {
|
|
|
|
LOG_ERROR(Kernel_SVC,
|
|
|
|
"Addr does not fit within the valid region, addr=0x{:016X}, "
|
|
|
|
"size=0x{:016X}",
|
2018-11-26 08:47:39 +00:00
|
|
|
addr, size);
|
2018-10-18 02:39:21 +00:00
|
|
|
return ERR_INVALID_MEMORY_RANGE;
|
|
|
|
}
|
|
|
|
|
2020-04-09 03:14:18 +00:00
|
|
|
if (page_table.IsInsideHeapRegion(addr, size)) {
|
|
|
|
LOG_ERROR(Kernel_SVC,
|
|
|
|
"Addr does not fit within the heap region, addr=0x{:016X}, "
|
|
|
|
"size=0x{:016X}",
|
|
|
|
addr, size);
|
|
|
|
return ERR_INVALID_MEMORY_RANGE;
|
2018-09-14 00:16:43 +00:00
|
|
|
}
|
|
|
|
|
2020-04-09 03:14:18 +00:00
|
|
|
if (page_table.IsInsideAliasRegion(addr, size)) {
|
|
|
|
LOG_ERROR(Kernel_SVC,
|
|
|
|
"Address does not fit within the map region, addr=0x{:016X}, "
|
|
|
|
"size=0x{:016X}",
|
2018-11-26 08:47:39 +00:00
|
|
|
addr, size);
|
2020-04-09 03:14:18 +00:00
|
|
|
return ERR_INVALID_MEMORY_RANGE;
|
2018-10-18 02:39:21 +00:00
|
|
|
}
|
|
|
|
|
2020-04-09 03:14:18 +00:00
|
|
|
auto shared_memory{current_process->GetHandleTable().Get<SharedMemory>(shared_memory_handle)};
|
2018-10-18 02:39:21 +00:00
|
|
|
if (!shared_memory) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Shared memory does not exist, shared_memory_handle=0x{:08X}",
|
|
|
|
shared_memory_handle);
|
2018-10-18 02:39:21 +00:00
|
|
|
return ERR_INVALID_HANDLE;
|
|
|
|
}
|
|
|
|
|
2020-04-09 03:14:18 +00:00
|
|
|
return shared_memory->Map(*current_process, addr, size, permission_type);
|
2018-02-22 19:16:43 +00:00
|
|
|
}
|
|
|
|
|
2020-06-19 23:40:07 +00:00
|
|
|
static ResultCode MapSharedMemory32(Core::System& system, Handle shared_memory_handle, u32 addr,
|
|
|
|
u32 size, u32 permissions) {
|
|
|
|
return MapSharedMemory(system, shared_memory_handle, static_cast<VAddr>(addr),
|
|
|
|
static_cast<std::size_t>(size), permissions);
|
|
|
|
}
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode QueryProcessMemory(Core::System& system, VAddr memory_info_address,
|
|
|
|
VAddr page_info_address, Handle process_handle,
|
|
|
|
VAddr address) {
|
2020-02-27 14:28:44 +00:00
|
|
|
std::lock_guard lock{HLE::g_hle_lock};
|
2018-12-12 16:34:01 +00:00
|
|
|
LOG_TRACE(Kernel_SVC, "called process=0x{:08X} address={:X}", process_handle, address);
|
2019-04-06 22:46:18 +00:00
|
|
|
const auto& handle_table = system.Kernel().CurrentProcess()->GetHandleTable();
|
2019-11-25 01:15:51 +00:00
|
|
|
std::shared_ptr<Process> process = handle_table.Get<Process>(process_handle);
|
2017-12-30 18:40:28 +00:00
|
|
|
if (!process) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Process handle does not exist, process_handle=0x{:08X}",
|
|
|
|
process_handle);
|
2015-07-17 19:45:12 +00:00
|
|
|
return ERR_INVALID_HANDLE;
|
2017-10-20 03:00:46 +00:00
|
|
|
}
|
2018-12-06 15:59:22 +00:00
|
|
|
|
2020-04-09 03:14:18 +00:00
|
|
|
auto& memory{system.Memory()};
|
2020-04-17 04:59:08 +00:00
|
|
|
const auto memory_info{process->PageTable().QueryInfo(address).GetSvcMemoryInfo()};
|
2020-04-09 03:14:18 +00:00
|
|
|
|
|
|
|
memory.Write64(memory_info_address + 0x00, memory_info.addr);
|
|
|
|
memory.Write64(memory_info_address + 0x08, memory_info.size);
|
|
|
|
memory.Write32(memory_info_address + 0x10, static_cast<u32>(memory_info.state) & 0xff);
|
|
|
|
memory.Write32(memory_info_address + 0x14, static_cast<u32>(memory_info.attr));
|
|
|
|
memory.Write32(memory_info_address + 0x18, static_cast<u32>(memory_info.perm));
|
|
|
|
memory.Write32(memory_info_address + 0x1c, memory_info.ipc_refcount);
|
|
|
|
memory.Write32(memory_info_address + 0x20, memory_info.device_refcount);
|
|
|
|
memory.Write32(memory_info_address + 0x24, 0);
|
svc: Handle memory writing explicitly within QueryProcessMemory
Moves the memory writes directly into QueryProcessMemory instead of
letting the wrapper function do it. It would be inaccurate to allow the
handler to do it because there's cases where memory shouldn't even be
written to. For example, if the given process handle is invalid.
HOWEVER, if the memory writing is within the wrapper, then we have no
control over if these memory writes occur, meaning in an error case, 68
bytes of memory randomly get trashed with zeroes, 64 of those being
written to wherever the memory info address points to, and the remaining
4 being written wherever the page info address points to.
One solution in this case would be to just conditionally check within
the handler itself, but this is kind of smelly, given the handler
shouldn't be performing conditional behavior itself, it's a behavior of
the managed function. In other words, if you remove the handler from the
equation entirely, does the function still retain its proper behavior?
In this case, no.
Now, we don't potentially trash memory from this function if an invalid
query is performed.
2018-12-12 16:48:06 +00:00
|
|
|
|
|
|
|
// Page info appears to be currently unused by the kernel and is always set to zero.
|
2019-11-26 22:39:57 +00:00
|
|
|
memory.Write32(page_info_address, 0);
|
2018-12-06 15:59:22 +00:00
|
|
|
|
2015-01-23 05:36:58 +00:00
|
|
|
return RESULT_SUCCESS;
|
2014-05-16 00:17:30 +00:00
|
|
|
}
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode QueryMemory(Core::System& system, VAddr memory_info_address,
|
|
|
|
VAddr page_info_address, VAddr query_address) {
|
svc: Handle memory writing explicitly within QueryProcessMemory
Moves the memory writes directly into QueryProcessMemory instead of
letting the wrapper function do it. It would be inaccurate to allow the
handler to do it because there's cases where memory shouldn't even be
written to. For example, if the given process handle is invalid.
HOWEVER, if the memory writing is within the wrapper, then we have no
control over if these memory writes occur, meaning in an error case, 68
bytes of memory randomly get trashed with zeroes, 64 of those being
written to wherever the memory info address points to, and the remaining
4 being written wherever the page info address points to.
One solution in this case would be to just conditionally check within
the handler itself, but this is kind of smelly, given the handler
shouldn't be performing conditional behavior itself, it's a behavior of
the managed function. In other words, if you remove the handler from the
equation entirely, does the function still retain its proper behavior?
In this case, no.
Now, we don't potentially trash memory from this function if an invalid
query is performed.
2018-12-12 16:48:06 +00:00
|
|
|
LOG_TRACE(Kernel_SVC,
|
|
|
|
"called, memory_info_address=0x{:016X}, page_info_address=0x{:016X}, "
|
|
|
|
"query_address=0x{:016X}",
|
|
|
|
memory_info_address, page_info_address, query_address);
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
return QueryProcessMemory(system, memory_info_address, page_info_address, CurrentProcess,
|
svc: Handle memory writing explicitly within QueryProcessMemory
Moves the memory writes directly into QueryProcessMemory instead of
letting the wrapper function do it. It would be inaccurate to allow the
handler to do it because there's cases where memory shouldn't even be
written to. For example, if the given process handle is invalid.
HOWEVER, if the memory writing is within the wrapper, then we have no
control over if these memory writes occur, meaning in an error case, 68
bytes of memory randomly get trashed with zeroes, 64 of those being
written to wherever the memory info address points to, and the remaining
4 being written wherever the page info address points to.
One solution in this case would be to just conditionally check within
the handler itself, but this is kind of smelly, given the handler
shouldn't be performing conditional behavior itself, it's a behavior of
the managed function. In other words, if you remove the handler from the
equation entirely, does the function still retain its proper behavior?
In this case, no.
Now, we don't potentially trash memory from this function if an invalid
query is performed.
2018-12-12 16:48:06 +00:00
|
|
|
query_address);
|
2015-07-17 19:45:12 +00:00
|
|
|
}
|
|
|
|
|
2020-03-02 05:06:41 +00:00
|
|
|
static ResultCode QueryMemory32(Core::System& system, u32 memory_info_address,
|
|
|
|
u32 page_info_address, u32 query_address) {
|
|
|
|
return QueryMemory(system, memory_info_address, page_info_address, query_address);
|
|
|
|
}
|
|
|
|
|
2020-04-23 22:05:09 +00:00
|
|
|
static ResultCode MapProcessCodeMemory(Core::System& system, Handle process_handle, u64 dst_address,
|
|
|
|
u64 src_address, u64 size) {
|
|
|
|
LOG_DEBUG(Kernel_SVC,
|
|
|
|
"called. process_handle=0x{:08X}, dst_address=0x{:016X}, "
|
|
|
|
"src_address=0x{:016X}, size=0x{:016X}",
|
|
|
|
process_handle, dst_address, src_address, size);
|
|
|
|
|
|
|
|
if (!Common::Is4KBAligned(src_address)) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "src_address is not page-aligned (src_address=0x{:016X}).",
|
|
|
|
src_address);
|
|
|
|
return ERR_INVALID_ADDRESS;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!Common::Is4KBAligned(dst_address)) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "dst_address is not page-aligned (dst_address=0x{:016X}).",
|
|
|
|
dst_address);
|
|
|
|
return ERR_INVALID_ADDRESS;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (size == 0 || !Common::Is4KBAligned(size)) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Size is zero or not page-aligned (size=0x{:016X})", size);
|
|
|
|
return ERR_INVALID_SIZE;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!IsValidAddressRange(dst_address, size)) {
|
|
|
|
LOG_ERROR(Kernel_SVC,
|
|
|
|
"Destination address range overflows the address space (dst_address=0x{:016X}, "
|
|
|
|
"size=0x{:016X}).",
|
|
|
|
dst_address, size);
|
|
|
|
return ERR_INVALID_ADDRESS_STATE;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!IsValidAddressRange(src_address, size)) {
|
|
|
|
LOG_ERROR(Kernel_SVC,
|
|
|
|
"Source address range overflows the address space (src_address=0x{:016X}, "
|
|
|
|
"size=0x{:016X}).",
|
|
|
|
src_address, size);
|
|
|
|
return ERR_INVALID_ADDRESS_STATE;
|
|
|
|
}
|
|
|
|
|
|
|
|
const auto& handle_table = system.Kernel().CurrentProcess()->GetHandleTable();
|
|
|
|
auto process = handle_table.Get<Process>(process_handle);
|
|
|
|
if (!process) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Invalid process handle specified (handle=0x{:08X}).",
|
|
|
|
process_handle);
|
|
|
|
return ERR_INVALID_HANDLE;
|
|
|
|
}
|
|
|
|
|
|
|
|
auto& page_table = process->PageTable();
|
|
|
|
if (!page_table.IsInsideAddressSpace(src_address, size)) {
|
|
|
|
LOG_ERROR(Kernel_SVC,
|
|
|
|
"Source address range is not within the address space (src_address=0x{:016X}, "
|
|
|
|
"size=0x{:016X}).",
|
|
|
|
src_address, size);
|
|
|
|
return ERR_INVALID_ADDRESS_STATE;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!page_table.IsInsideASLRRegion(dst_address, size)) {
|
|
|
|
LOG_ERROR(Kernel_SVC,
|
|
|
|
"Destination address range is not within the ASLR region (dst_address=0x{:016X}, "
|
|
|
|
"size=0x{:016X}).",
|
|
|
|
dst_address, size);
|
|
|
|
return ERR_INVALID_MEMORY_RANGE;
|
|
|
|
}
|
|
|
|
|
|
|
|
return page_table.MapProcessCodeMemory(dst_address, src_address, size);
|
|
|
|
}
|
|
|
|
|
|
|
|
static ResultCode UnmapProcessCodeMemory(Core::System& system, Handle process_handle,
|
|
|
|
u64 dst_address, u64 src_address, u64 size) {
|
|
|
|
LOG_DEBUG(Kernel_SVC,
|
|
|
|
"called. process_handle=0x{:08X}, dst_address=0x{:016X}, src_address=0x{:016X}, "
|
|
|
|
"size=0x{:016X}",
|
|
|
|
process_handle, dst_address, src_address, size);
|
|
|
|
|
|
|
|
if (!Common::Is4KBAligned(dst_address)) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "dst_address is not page-aligned (dst_address=0x{:016X}).",
|
|
|
|
dst_address);
|
|
|
|
return ERR_INVALID_ADDRESS;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!Common::Is4KBAligned(src_address)) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "src_address is not page-aligned (src_address=0x{:016X}).",
|
|
|
|
src_address);
|
|
|
|
return ERR_INVALID_ADDRESS;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (size == 0 || Common::Is4KBAligned(size)) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Size is zero or not page-aligned (size=0x{:016X}).", size);
|
|
|
|
return ERR_INVALID_SIZE;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!IsValidAddressRange(dst_address, size)) {
|
|
|
|
LOG_ERROR(Kernel_SVC,
|
|
|
|
"Destination address range overflows the address space (dst_address=0x{:016X}, "
|
|
|
|
"size=0x{:016X}).",
|
|
|
|
dst_address, size);
|
|
|
|
return ERR_INVALID_ADDRESS_STATE;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!IsValidAddressRange(src_address, size)) {
|
|
|
|
LOG_ERROR(Kernel_SVC,
|
|
|
|
"Source address range overflows the address space (src_address=0x{:016X}, "
|
|
|
|
"size=0x{:016X}).",
|
|
|
|
src_address, size);
|
|
|
|
return ERR_INVALID_ADDRESS_STATE;
|
|
|
|
}
|
|
|
|
|
|
|
|
const auto& handle_table = system.Kernel().CurrentProcess()->GetHandleTable();
|
|
|
|
auto process = handle_table.Get<Process>(process_handle);
|
|
|
|
if (!process) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Invalid process handle specified (handle=0x{:08X}).",
|
|
|
|
process_handle);
|
|
|
|
return ERR_INVALID_HANDLE;
|
|
|
|
}
|
|
|
|
|
|
|
|
auto& page_table = process->PageTable();
|
|
|
|
if (!page_table.IsInsideAddressSpace(src_address, size)) {
|
|
|
|
LOG_ERROR(Kernel_SVC,
|
|
|
|
"Source address range is not within the address space (src_address=0x{:016X}, "
|
|
|
|
"size=0x{:016X}).",
|
|
|
|
src_address, size);
|
|
|
|
return ERR_INVALID_ADDRESS_STATE;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!page_table.IsInsideASLRRegion(dst_address, size)) {
|
|
|
|
LOG_ERROR(Kernel_SVC,
|
|
|
|
"Destination address range is not within the ASLR region (dst_address=0x{:016X}, "
|
|
|
|
"size=0x{:016X}).",
|
|
|
|
dst_address, size);
|
|
|
|
return ERR_INVALID_MEMORY_RANGE;
|
|
|
|
}
|
|
|
|
|
|
|
|
return page_table.UnmapProcessCodeMemory(dst_address, src_address, size);
|
|
|
|
}
|
|
|
|
|
2018-01-01 19:38:34 +00:00
|
|
|
/// Exits the current process
|
2019-04-06 22:46:18 +00:00
|
|
|
static void ExitProcess(Core::System& system) {
|
|
|
|
auto* current_process = system.Kernel().CurrentProcess();
|
2020-02-26 22:55:11 +00:00
|
|
|
UNIMPLEMENTED();
|
2018-01-01 19:38:34 +00:00
|
|
|
|
2018-09-21 06:06:47 +00:00
|
|
|
LOG_INFO(Kernel_SVC, "Process {} exiting", current_process->GetProcessID());
|
|
|
|
ASSERT_MSG(current_process->GetStatus() == ProcessStatus::Running,
|
2018-03-13 21:49:59 +00:00
|
|
|
"Process has already exited");
|
2018-01-01 19:38:34 +00:00
|
|
|
|
2018-09-21 06:06:47 +00:00
|
|
|
current_process->PrepareForTermination();
|
2018-01-01 19:38:34 +00:00
|
|
|
|
|
|
|
// Kill the current thread
|
2020-12-03 02:08:35 +00:00
|
|
|
system.Kernel().CurrentScheduler()->GetCurrentThread()->Stop();
|
2018-01-01 19:38:34 +00:00
|
|
|
}
|
|
|
|
|
2020-06-19 23:40:07 +00:00
|
|
|
static void ExitProcess32(Core::System& system) {
|
|
|
|
ExitProcess(system);
|
|
|
|
}
|
|
|
|
|
2017-12-31 21:10:01 +00:00
|
|
|
/// Creates a new thread
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode CreateThread(Core::System& system, Handle* out_handle, VAddr entry_point, u64 arg,
|
|
|
|
VAddr stack_top, u32 priority, s32 processor_id) {
|
2019-04-16 01:33:07 +00:00
|
|
|
LOG_DEBUG(Kernel_SVC,
|
2018-12-04 03:13:50 +00:00
|
|
|
"called entrypoint=0x{:08X}, arg=0x{:08X}, stacktop=0x{:08X}, "
|
2018-11-26 06:06:13 +00:00
|
|
|
"threadpriority=0x{:08X}, processorid=0x{:08X} : created handle=0x{:08X}",
|
2018-12-01 07:05:19 +00:00
|
|
|
entry_point, arg, stack_top, priority, processor_id, *out_handle);
|
2018-11-26 06:06:13 +00:00
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
auto* const current_process = system.Kernel().CurrentProcess();
|
2017-12-31 21:10:01 +00:00
|
|
|
|
2018-12-28 02:14:59 +00:00
|
|
|
if (processor_id == THREADPROCESSORID_IDEAL) {
|
|
|
|
// Set the target CPU to the one specified by the process.
|
|
|
|
processor_id = current_process->GetIdealCore();
|
|
|
|
ASSERT(processor_id != THREADPROCESSORID_IDEAL);
|
2017-12-31 21:10:01 +00:00
|
|
|
}
|
|
|
|
|
2018-12-31 01:59:52 +00:00
|
|
|
if (processor_id < THREADPROCESSORID_0 || processor_id > THREADPROCESSORID_3) {
|
2018-09-12 08:27:35 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Invalid thread processor ID: {}", processor_id);
|
|
|
|
return ERR_INVALID_PROCESSOR_ID;
|
2017-12-31 21:10:01 +00:00
|
|
|
}
|
|
|
|
|
2018-12-31 02:20:07 +00:00
|
|
|
const u64 core_mask = current_process->GetCoreMask();
|
|
|
|
if ((core_mask | (1ULL << processor_id)) != core_mask) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Invalid thread core specified ({})", processor_id);
|
|
|
|
return ERR_INVALID_PROCESSOR_ID;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (priority > THREADPRIO_LOWEST) {
|
2018-12-31 02:27:30 +00:00
|
|
|
LOG_ERROR(Kernel_SVC,
|
|
|
|
"Invalid thread priority specified ({}). Must be within the range 0-64",
|
|
|
|
priority);
|
2018-12-31 02:20:07 +00:00
|
|
|
return ERR_INVALID_THREAD_PRIORITY;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (((1ULL << priority) & current_process->GetPriorityMask()) == 0) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Invalid thread priority specified ({})", priority);
|
|
|
|
return ERR_INVALID_THREAD_PRIORITY;
|
|
|
|
}
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
auto& kernel = system.Kernel();
|
2020-04-09 03:14:18 +00:00
|
|
|
|
|
|
|
ASSERT(kernel.CurrentProcess()->GetResourceLimit()->Reserve(ResourceType::Threads, 1));
|
|
|
|
|
2020-02-25 02:04:12 +00:00
|
|
|
ThreadType type = THREADTYPE_USER;
|
2019-11-25 01:15:51 +00:00
|
|
|
CASCADE_RESULT(std::shared_ptr<Thread> thread,
|
2020-02-25 17:22:11 +00:00
|
|
|
Thread::Create(system, type, "", entry_point, priority, arg, processor_id,
|
|
|
|
stack_top, current_process));
|
2018-10-20 18:34:41 +00:00
|
|
|
|
2019-04-14 10:06:04 +00:00
|
|
|
const auto new_thread_handle = current_process->GetHandleTable().Create(thread);
|
|
|
|
if (new_thread_handle.Failed()) {
|
2018-11-26 06:06:13 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Failed to create handle with error=0x{:X}",
|
2019-04-14 10:06:04 +00:00
|
|
|
new_thread_handle.Code().raw);
|
|
|
|
return new_thread_handle.Code();
|
2018-10-03 22:47:57 +00:00
|
|
|
}
|
2019-04-14 10:06:04 +00:00
|
|
|
*out_handle = *new_thread_handle;
|
2017-12-31 21:10:01 +00:00
|
|
|
|
2019-04-15 19:54:25 +00:00
|
|
|
// Set the thread name for debugging purposes.
|
|
|
|
thread->SetName(
|
|
|
|
fmt::format("thread[entry_point={:X}, handle={:X}]", entry_point, *new_thread_handle));
|
2017-12-31 21:10:01 +00:00
|
|
|
|
|
|
|
return RESULT_SUCCESS;
|
|
|
|
}
|
|
|
|
|
2020-06-19 23:40:07 +00:00
|
|
|
static ResultCode CreateThread32(Core::System& system, Handle* out_handle, u32 priority,
|
|
|
|
u32 entry_point, u32 arg, u32 stack_top, s32 processor_id) {
|
|
|
|
return CreateThread(system, out_handle, static_cast<VAddr>(entry_point), static_cast<u64>(arg),
|
|
|
|
static_cast<VAddr>(stack_top), priority, processor_id);
|
|
|
|
}
|
|
|
|
|
2017-12-30 18:40:28 +00:00
|
|
|
/// Starts the thread for the provided handle
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode StartThread(Core::System& system, Handle thread_handle) {
|
2019-04-16 01:33:07 +00:00
|
|
|
LOG_DEBUG(Kernel_SVC, "called thread=0x{:08X}", thread_handle);
|
2017-12-30 18:37:07 +00:00
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
const auto& handle_table = system.Kernel().CurrentProcess()->GetHandleTable();
|
2019-11-25 01:15:51 +00:00
|
|
|
const std::shared_ptr<Thread> thread = handle_table.Get<Thread>(thread_handle);
|
2017-12-30 18:37:07 +00:00
|
|
|
if (!thread) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Thread handle does not exist, thread_handle=0x{:08X}",
|
|
|
|
thread_handle);
|
2017-12-30 18:37:07 +00:00
|
|
|
return ERR_INVALID_HANDLE;
|
|
|
|
}
|
|
|
|
|
2018-10-03 22:47:57 +00:00
|
|
|
ASSERT(thread->GetStatus() == ThreadStatus::Dormant);
|
2018-05-19 21:57:44 +00:00
|
|
|
|
2020-02-25 16:40:33 +00:00
|
|
|
return thread->Start();
|
2017-12-30 18:37:07 +00:00
|
|
|
}
|
|
|
|
|
2020-06-19 23:40:07 +00:00
|
|
|
static ResultCode StartThread32(Core::System& system, Handle thread_handle) {
|
|
|
|
return StartThread(system, thread_handle);
|
|
|
|
}
|
|
|
|
|
2017-12-31 21:11:27 +00:00
|
|
|
/// Called when a thread exits
|
2019-04-06 22:46:18 +00:00
|
|
|
static void ExitThread(Core::System& system) {
|
2019-04-16 01:33:07 +00:00
|
|
|
LOG_DEBUG(Kernel_SVC, "called, pc=0x{:08X}", system.CurrentArmInterface().GetPC());
|
2019-03-16 03:38:51 +00:00
|
|
|
|
2020-12-03 02:08:35 +00:00
|
|
|
auto* const current_thread = system.Kernel().CurrentScheduler()->GetCurrentThread();
|
|
|
|
system.GlobalSchedulerContext().RemoveThread(SharedFrom(current_thread));
|
2020-02-25 16:40:33 +00:00
|
|
|
current_thread->Stop();
|
2017-12-31 21:11:27 +00:00
|
|
|
}
|
|
|
|
|
2020-06-19 23:40:07 +00:00
|
|
|
static void ExitThread32(Core::System& system) {
|
|
|
|
ExitThread(system);
|
|
|
|
}
|
|
|
|
|
2014-06-01 14:37:19 +00:00
|
|
|
/// Sleep the current thread
|
2019-04-06 22:46:18 +00:00
|
|
|
static void SleepThread(Core::System& system, s64 nanoseconds) {
|
2019-04-16 01:33:07 +00:00
|
|
|
LOG_DEBUG(Kernel_SVC, "called nanoseconds={}", nanoseconds);
|
2014-11-26 05:34:14 +00:00
|
|
|
|
2018-11-22 05:33:53 +00:00
|
|
|
enum class SleepType : s64 {
|
2020-12-03 02:08:35 +00:00
|
|
|
YieldWithoutCoreMigration = 0,
|
|
|
|
YieldWithCoreMigration = -1,
|
2018-12-02 05:44:40 +00:00
|
|
|
YieldAndWaitForLoadBalancing = -2,
|
2018-11-22 05:33:53 +00:00
|
|
|
};
|
2017-01-05 19:14:22 +00:00
|
|
|
|
2020-12-03 02:08:35 +00:00
|
|
|
auto& scheduler = *system.Kernel().CurrentScheduler();
|
2018-11-19 04:44:19 +00:00
|
|
|
if (nanoseconds <= 0) {
|
2018-11-22 05:33:53 +00:00
|
|
|
switch (static_cast<SleepType>(nanoseconds)) {
|
2020-12-03 02:08:35 +00:00
|
|
|
case SleepType::YieldWithoutCoreMigration: {
|
|
|
|
scheduler.YieldWithoutCoreMigration();
|
2018-11-19 04:44:19 +00:00
|
|
|
break;
|
2020-03-10 17:13:39 +00:00
|
|
|
}
|
2020-12-03 02:08:35 +00:00
|
|
|
case SleepType::YieldWithCoreMigration: {
|
|
|
|
scheduler.YieldWithCoreMigration();
|
2018-11-19 04:44:19 +00:00
|
|
|
break;
|
2020-03-10 17:13:39 +00:00
|
|
|
}
|
|
|
|
case SleepType::YieldAndWaitForLoadBalancing: {
|
2020-12-03 02:08:35 +00:00
|
|
|
scheduler.YieldToAnyThread();
|
2018-11-19 04:44:19 +00:00
|
|
|
break;
|
2020-03-10 17:13:39 +00:00
|
|
|
}
|
2018-11-19 04:44:19 +00:00
|
|
|
default:
|
2018-12-03 22:29:21 +00:00
|
|
|
UNREACHABLE_MSG("Unimplemented sleep yield type '{:016X}'!", nanoseconds);
|
2018-11-19 04:44:19 +00:00
|
|
|
}
|
2018-12-03 22:29:21 +00:00
|
|
|
} else {
|
2020-12-03 02:08:35 +00:00
|
|
|
scheduler.GetCurrentThread()->Sleep(nanoseconds);
|
2018-11-19 04:44:19 +00:00
|
|
|
}
|
2014-06-01 14:37:19 +00:00
|
|
|
}
|
|
|
|
|
2020-06-19 23:40:07 +00:00
|
|
|
static void SleepThread32(Core::System& system, u32 nanoseconds_low, u32 nanoseconds_high) {
|
|
|
|
const s64 nanoseconds = static_cast<s64>(static_cast<u64>(nanoseconds_low) |
|
|
|
|
(static_cast<u64>(nanoseconds_high) << 32));
|
|
|
|
SleepThread(system, nanoseconds);
|
|
|
|
}
|
|
|
|
|
2018-06-21 06:49:43 +00:00
|
|
|
/// Wait process wide key atomic
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode WaitProcessWideKeyAtomic(Core::System& system, VAddr mutex_addr,
|
|
|
|
VAddr condition_variable_addr, Handle thread_handle,
|
|
|
|
s64 nano_seconds) {
|
2018-07-02 16:13:26 +00:00
|
|
|
LOG_TRACE(
|
2018-01-09 02:41:37 +00:00
|
|
|
Kernel_SVC,
|
2018-05-02 13:14:28 +00:00
|
|
|
"called mutex_addr={:X}, condition_variable_addr={:X}, thread_handle=0x{:08X}, timeout={}",
|
2018-01-09 02:41:37 +00:00
|
|
|
mutex_addr, condition_variable_addr, thread_handle, nano_seconds);
|
2018-01-06 21:14:12 +00:00
|
|
|
|
2020-03-31 19:10:44 +00:00
|
|
|
if (Core::Memory::IsKernelVirtualAddress(mutex_addr)) {
|
2019-04-04 00:23:14 +00:00
|
|
|
LOG_ERROR(
|
|
|
|
Kernel_SVC,
|
|
|
|
"Given mutex address must not be within the kernel address space. address=0x{:016X}",
|
|
|
|
mutex_addr);
|
|
|
|
return ERR_INVALID_ADDRESS_STATE;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!Common::IsWordAligned(mutex_addr)) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Given mutex address must be word-aligned. address=0x{:016X}",
|
|
|
|
mutex_addr);
|
|
|
|
return ERR_INVALID_ADDRESS;
|
|
|
|
}
|
|
|
|
|
2019-10-12 11:57:32 +00:00
|
|
|
ASSERT(condition_variable_addr == Common::AlignDown(condition_variable_addr, 4));
|
2020-02-27 02:26:53 +00:00
|
|
|
auto& kernel = system.Kernel();
|
|
|
|
Handle event_handle;
|
2020-12-03 02:08:35 +00:00
|
|
|
Thread* current_thread = kernel.CurrentScheduler()->GetCurrentThread();
|
|
|
|
auto* const current_process = kernel.CurrentProcess();
|
2020-02-27 02:26:53 +00:00
|
|
|
{
|
2020-12-04 05:56:02 +00:00
|
|
|
KScopedSchedulerLockAndSleep lock(kernel, event_handle, current_thread, nano_seconds);
|
2020-02-27 02:26:53 +00:00
|
|
|
const auto& handle_table = current_process->GetHandleTable();
|
|
|
|
std::shared_ptr<Thread> thread = handle_table.Get<Thread>(thread_handle);
|
|
|
|
ASSERT(thread);
|
2018-01-06 21:14:12 +00:00
|
|
|
|
2020-02-27 02:26:53 +00:00
|
|
|
current_thread->SetSynchronizationResults(nullptr, RESULT_TIMEOUT);
|
|
|
|
|
2020-03-08 16:51:24 +00:00
|
|
|
if (thread->IsPendingTermination()) {
|
|
|
|
lock.CancelSleep();
|
|
|
|
return ERR_THREAD_TERMINATING;
|
|
|
|
}
|
|
|
|
|
2020-02-27 02:26:53 +00:00
|
|
|
const auto release_result = current_process->GetMutex().Release(mutex_addr);
|
|
|
|
if (release_result.IsError()) {
|
|
|
|
lock.CancelSleep();
|
|
|
|
return release_result;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (nano_seconds == 0) {
|
|
|
|
lock.CancelSleep();
|
|
|
|
return RESULT_TIMEOUT;
|
|
|
|
}
|
|
|
|
|
|
|
|
current_thread->SetCondVarWaitAddress(condition_variable_addr);
|
|
|
|
current_thread->SetMutexWaitAddress(mutex_addr);
|
|
|
|
current_thread->SetWaitHandle(thread_handle);
|
|
|
|
current_thread->SetStatus(ThreadStatus::WaitCondVar);
|
|
|
|
current_process->InsertConditionVariableThread(SharedFrom(current_thread));
|
2019-03-14 04:29:54 +00:00
|
|
|
}
|
2018-01-08 16:35:03 +00:00
|
|
|
|
2020-02-27 02:26:53 +00:00
|
|
|
if (event_handle != InvalidHandle) {
|
|
|
|
auto& time_manager = kernel.TimeManager();
|
|
|
|
time_manager.UnscheduleTimeEvent(event_handle);
|
|
|
|
}
|
2018-01-08 16:35:03 +00:00
|
|
|
|
2020-02-27 02:26:53 +00:00
|
|
|
{
|
2020-12-04 06:26:42 +00:00
|
|
|
KScopedSchedulerLock lock(kernel);
|
2018-01-06 21:14:12 +00:00
|
|
|
|
2020-03-08 16:51:24 +00:00
|
|
|
auto* owner = current_thread->GetLockOwner();
|
|
|
|
if (owner != nullptr) {
|
|
|
|
owner->RemoveMutexWaiter(SharedFrom(current_thread));
|
|
|
|
}
|
|
|
|
|
2020-02-27 02:26:53 +00:00
|
|
|
current_process->RemoveConditionVariableThread(SharedFrom(current_thread));
|
|
|
|
}
|
2018-04-21 01:15:16 +00:00
|
|
|
// Note: Deliberately don't attempt to inherit the lock owner's priority.
|
2018-01-06 21:14:12 +00:00
|
|
|
|
2020-02-27 02:26:53 +00:00
|
|
|
return current_thread->GetSignalingResult();
|
2018-01-06 21:14:12 +00:00
|
|
|
}
|
|
|
|
|
2020-06-19 23:40:07 +00:00
|
|
|
static ResultCode WaitProcessWideKeyAtomic32(Core::System& system, u32 mutex_addr,
|
|
|
|
u32 condition_variable_addr, Handle thread_handle,
|
|
|
|
u32 nanoseconds_low, u32 nanoseconds_high) {
|
|
|
|
const s64 nanoseconds =
|
|
|
|
static_cast<s64>(nanoseconds_low | (static_cast<u64>(nanoseconds_high) << 32));
|
|
|
|
return WaitProcessWideKeyAtomic(system, static_cast<VAddr>(mutex_addr),
|
|
|
|
static_cast<VAddr>(condition_variable_addr), thread_handle,
|
|
|
|
nanoseconds);
|
|
|
|
}
|
|
|
|
|
2017-10-14 21:30:07 +00:00
|
|
|
/// Signal process wide key
|
2019-12-11 12:11:47 +00:00
|
|
|
static void SignalProcessWideKey(Core::System& system, VAddr condition_variable_addr, s32 target) {
|
2018-07-02 16:13:26 +00:00
|
|
|
LOG_TRACE(Kernel_SVC, "called, condition_variable_addr=0x{:X}, target=0x{:08X}",
|
2018-07-02 16:20:50 +00:00
|
|
|
condition_variable_addr, target);
|
2018-01-07 21:55:17 +00:00
|
|
|
|
2019-10-12 11:57:32 +00:00
|
|
|
ASSERT(condition_variable_addr == Common::AlignDown(condition_variable_addr, 4));
|
|
|
|
|
2018-05-19 21:58:30 +00:00
|
|
|
// Retrieve a list of all threads that are waiting for this condition variable.
|
2020-02-27 02:26:53 +00:00
|
|
|
auto& kernel = system.Kernel();
|
2020-12-04 06:26:42 +00:00
|
|
|
KScopedSchedulerLock lock(kernel);
|
2020-02-27 02:26:53 +00:00
|
|
|
auto* const current_process = kernel.CurrentProcess();
|
2019-11-25 01:15:51 +00:00
|
|
|
std::vector<std::shared_ptr<Thread>> waiting_threads =
|
2019-11-15 00:13:18 +00:00
|
|
|
current_process->GetConditionVariableThreads(condition_variable_addr);
|
2018-05-19 21:58:30 +00:00
|
|
|
|
2019-11-16 17:55:21 +00:00
|
|
|
// Only process up to 'target' threads, unless 'target' is less equal 0, in which case process
|
2018-05-19 21:58:30 +00:00
|
|
|
// them all.
|
2018-09-15 13:21:06 +00:00
|
|
|
std::size_t last = waiting_threads.size();
|
2020-07-16 17:40:38 +00:00
|
|
|
if (target > 0) {
|
2019-03-13 23:04:40 +00:00
|
|
|
last = std::min(waiting_threads.size(), static_cast<std::size_t>(target));
|
2020-07-16 17:40:38 +00:00
|
|
|
}
|
2018-09-15 13:21:06 +00:00
|
|
|
for (std::size_t index = 0; index < last; ++index) {
|
2018-05-19 21:58:30 +00:00
|
|
|
auto& thread = waiting_threads[index];
|
2018-05-06 02:00:34 +00:00
|
|
|
|
2018-10-03 22:47:57 +00:00
|
|
|
ASSERT(thread->GetCondVarWaitAddress() == condition_variable_addr);
|
2018-05-06 02:00:34 +00:00
|
|
|
|
2019-03-20 00:57:31 +00:00
|
|
|
// liberate Cond Var Thread.
|
2019-11-15 00:13:18 +00:00
|
|
|
current_process->RemoveConditionVariableThread(thread);
|
2019-03-20 00:57:31 +00:00
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
const std::size_t current_core = system.CurrentCoreIndex();
|
|
|
|
auto& monitor = system.Monitor();
|
2018-07-22 17:27:24 +00:00
|
|
|
|
|
|
|
// Atomically read the value of the mutex.
|
|
|
|
u32 mutex_val = 0;
|
2019-10-07 22:57:13 +00:00
|
|
|
u32 update_val = 0;
|
|
|
|
const VAddr mutex_address = thread->GetMutexWaitAddress();
|
2018-07-22 17:27:24 +00:00
|
|
|
do {
|
|
|
|
// If the mutex is not yet acquired, acquire it.
|
2020-03-15 19:54:40 +00:00
|
|
|
mutex_val = monitor.ExclusiveRead32(current_core, mutex_address);
|
2018-07-22 17:27:24 +00:00
|
|
|
|
|
|
|
if (mutex_val != 0) {
|
2019-10-07 22:57:13 +00:00
|
|
|
update_val = mutex_val | Mutex::MutexHasWaitersFlag;
|
|
|
|
} else {
|
|
|
|
update_val = thread->GetWaitHandle();
|
2018-07-22 17:27:24 +00:00
|
|
|
}
|
2019-10-07 22:57:13 +00:00
|
|
|
} while (!monitor.ExclusiveWrite32(current_core, mutex_address, update_val));
|
2020-02-28 13:42:06 +00:00
|
|
|
monitor.ClearExclusive();
|
2018-05-19 21:58:30 +00:00
|
|
|
if (mutex_val == 0) {
|
|
|
|
// We were able to acquire the mutex, resume this thread.
|
2018-10-03 22:47:57 +00:00
|
|
|
auto* const lock_owner = thread->GetLockOwner();
|
|
|
|
if (lock_owner != nullptr) {
|
2018-05-19 21:58:30 +00:00
|
|
|
lock_owner->RemoveMutexWaiter(thread);
|
2018-10-03 22:47:57 +00:00
|
|
|
}
|
2018-05-19 21:58:30 +00:00
|
|
|
|
2018-10-03 22:47:57 +00:00
|
|
|
thread->SetLockOwner(nullptr);
|
2020-02-27 02:26:53 +00:00
|
|
|
thread->SetSynchronizationResults(nullptr, RESULT_SUCCESS);
|
2020-03-08 16:51:24 +00:00
|
|
|
thread->ResumeFromWait();
|
2018-05-19 21:58:30 +00:00
|
|
|
} else {
|
2018-07-22 17:27:24 +00:00
|
|
|
// The mutex is already owned by some other thread, make this thread wait on it.
|
2018-10-20 18:34:41 +00:00
|
|
|
const Handle owner_handle = static_cast<Handle>(mutex_val & Mutex::MutexOwnerMask);
|
2019-04-06 22:46:18 +00:00
|
|
|
const auto& handle_table = system.Kernel().CurrentProcess()->GetHandleTable();
|
2018-10-20 18:34:41 +00:00
|
|
|
auto owner = handle_table.Get<Thread>(owner_handle);
|
2018-05-19 21:58:30 +00:00
|
|
|
ASSERT(owner);
|
2020-03-08 16:51:24 +00:00
|
|
|
if (thread->GetStatus() == ThreadStatus::WaitCondVar) {
|
|
|
|
thread->SetStatus(ThreadStatus::WaitMutex);
|
|
|
|
}
|
2018-05-19 21:58:30 +00:00
|
|
|
|
|
|
|
owner->AddMutexWaiter(thread);
|
|
|
|
}
|
|
|
|
}
|
2014-12-09 04:52:27 +00:00
|
|
|
}
|
|
|
|
|
2020-03-02 05:06:41 +00:00
|
|
|
static void SignalProcessWideKey32(Core::System& system, u32 condition_variable_addr, s32 target) {
|
|
|
|
SignalProcessWideKey(system, condition_variable_addr, target);
|
|
|
|
}
|
|
|
|
|
2018-06-21 06:49:43 +00:00
|
|
|
// Wait for an address (via Address Arbiter)
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode WaitForAddress(Core::System& system, VAddr address, u32 type, s32 value,
|
|
|
|
s64 timeout) {
|
2019-07-18 02:05:08 +00:00
|
|
|
LOG_TRACE(Kernel_SVC, "called, address=0x{:X}, type=0x{:X}, value=0x{:X}, timeout={}", address,
|
|
|
|
type, value, timeout);
|
2019-04-06 22:46:18 +00:00
|
|
|
|
2018-06-21 06:49:43 +00:00
|
|
|
// If the passed address is a kernel virtual address, return invalid memory state.
|
2020-03-31 19:10:44 +00:00
|
|
|
if (Core::Memory::IsKernelVirtualAddress(address)) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Address is a kernel virtual address, address={:016X}", address);
|
2018-06-21 06:49:43 +00:00
|
|
|
return ERR_INVALID_ADDRESS_STATE;
|
|
|
|
}
|
2019-04-06 22:46:18 +00:00
|
|
|
|
2018-06-21 06:49:43 +00:00
|
|
|
// If the address is not properly aligned to 4 bytes, return invalid address.
|
2018-11-26 08:47:39 +00:00
|
|
|
if (!Common::IsWordAligned(address)) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Address is not word aligned, address={:016X}", address);
|
2018-06-21 06:49:43 +00:00
|
|
|
return ERR_INVALID_ADDRESS;
|
|
|
|
}
|
|
|
|
|
2019-03-07 23:34:22 +00:00
|
|
|
const auto arbitration_type = static_cast<AddressArbiter::ArbitrationType>(type);
|
2019-04-02 13:22:53 +00:00
|
|
|
auto& address_arbiter = system.Kernel().CurrentProcess()->GetAddressArbiter();
|
2019-06-19 13:11:18 +00:00
|
|
|
const ResultCode result =
|
|
|
|
address_arbiter.WaitForAddress(address, arbitration_type, value, timeout);
|
2019-03-29 21:11:25 +00:00
|
|
|
return result;
|
2018-06-21 06:49:43 +00:00
|
|
|
}
|
|
|
|
|
2020-06-19 23:40:07 +00:00
|
|
|
static ResultCode WaitForAddress32(Core::System& system, u32 address, u32 type, s32 value,
|
|
|
|
u32 timeout_low, u32 timeout_high) {
|
|
|
|
s64 timeout = static_cast<s64>(timeout_low | (static_cast<u64>(timeout_high) << 32));
|
|
|
|
return WaitForAddress(system, static_cast<VAddr>(address), type, value, timeout);
|
|
|
|
}
|
|
|
|
|
2018-06-21 06:49:43 +00:00
|
|
|
// Signals to an address (via Address Arbiter)
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode SignalToAddress(Core::System& system, VAddr address, u32 type, s32 value,
|
|
|
|
s32 num_to_wake) {
|
2019-07-18 02:05:08 +00:00
|
|
|
LOG_TRACE(Kernel_SVC, "called, address=0x{:X}, type=0x{:X}, value=0x{:X}, num_to_wake=0x{:X}",
|
|
|
|
address, type, value, num_to_wake);
|
2019-04-06 22:46:18 +00:00
|
|
|
|
2018-06-21 06:49:43 +00:00
|
|
|
// If the passed address is a kernel virtual address, return invalid memory state.
|
2020-03-31 19:10:44 +00:00
|
|
|
if (Core::Memory::IsKernelVirtualAddress(address)) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Address is a kernel virtual address, address={:016X}", address);
|
2018-06-21 06:49:43 +00:00
|
|
|
return ERR_INVALID_ADDRESS_STATE;
|
|
|
|
}
|
2019-04-06 22:46:18 +00:00
|
|
|
|
2018-06-21 06:49:43 +00:00
|
|
|
// If the address is not properly aligned to 4 bytes, return invalid address.
|
2018-11-26 08:47:39 +00:00
|
|
|
if (!Common::IsWordAligned(address)) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Address is not word aligned, address={:016X}", address);
|
2018-06-21 06:49:43 +00:00
|
|
|
return ERR_INVALID_ADDRESS;
|
|
|
|
}
|
|
|
|
|
2019-03-07 23:42:44 +00:00
|
|
|
const auto signal_type = static_cast<AddressArbiter::SignalType>(type);
|
2019-04-06 22:46:18 +00:00
|
|
|
auto& address_arbiter = system.Kernel().CurrentProcess()->GetAddressArbiter();
|
2019-03-07 23:42:44 +00:00
|
|
|
return address_arbiter.SignalToAddress(address, signal_type, value, num_to_wake);
|
2018-06-21 06:49:43 +00:00
|
|
|
}
|
|
|
|
|
2020-06-19 23:40:07 +00:00
|
|
|
static ResultCode SignalToAddress32(Core::System& system, u32 address, u32 type, s32 value,
|
|
|
|
s32 num_to_wake) {
|
|
|
|
return SignalToAddress(system, static_cast<VAddr>(address), type, value, num_to_wake);
|
|
|
|
}
|
|
|
|
|
2019-12-08 02:12:14 +00:00
|
|
|
static void KernelDebug([[maybe_unused]] Core::System& system,
|
|
|
|
[[maybe_unused]] u32 kernel_debug_type, [[maybe_unused]] u64 param1,
|
|
|
|
[[maybe_unused]] u64 param2, [[maybe_unused]] u64 param3) {
|
|
|
|
// Intentionally do nothing, as this does nothing in released kernel binaries.
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ChangeKernelTraceState([[maybe_unused]] Core::System& system,
|
|
|
|
[[maybe_unused]] u32 trace_state) {
|
|
|
|
// Intentionally do nothing, as this does nothing in released kernel binaries.
|
|
|
|
}
|
|
|
|
|
2018-01-12 02:59:31 +00:00
|
|
|
/// This returns the total CPU ticks elapsed since the CPU was powered-on
|
2019-04-06 22:46:18 +00:00
|
|
|
static u64 GetSystemTick(Core::System& system) {
|
2018-11-26 06:06:13 +00:00
|
|
|
LOG_TRACE(Kernel_SVC, "called");
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
auto& core_timing = system.CoreTiming();
|
2019-11-22 20:55:42 +00:00
|
|
|
|
|
|
|
// Returns the value of cntpct_el0 (https://switchbrew.org/wiki/SVC#svcGetSystemTick)
|
2020-02-25 02:04:12 +00:00
|
|
|
const u64 result{system.CoreTiming().GetClockTicks()};
|
2018-01-12 02:59:31 +00:00
|
|
|
|
2020-03-28 19:23:28 +00:00
|
|
|
if (!system.Kernel().IsMulticore()) {
|
|
|
|
core_timing.AddTicks(400U);
|
|
|
|
}
|
|
|
|
|
2018-01-12 02:59:31 +00:00
|
|
|
return result;
|
|
|
|
}
|
|
|
|
|
2020-06-19 23:40:07 +00:00
|
|
|
static void GetSystemTick32(Core::System& system, u32* time_low, u32* time_high) {
|
|
|
|
u64 time = GetSystemTick(system);
|
|
|
|
*time_low = static_cast<u32>(time);
|
|
|
|
*time_high = static_cast<u32>(time >> 32);
|
|
|
|
}
|
|
|
|
|
2017-10-14 21:30:07 +00:00
|
|
|
/// Close a handle
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode CloseHandle(Core::System& system, Handle handle) {
|
2018-07-02 16:13:26 +00:00
|
|
|
LOG_TRACE(Kernel_SVC, "Closing handle 0x{:08X}", handle);
|
2018-08-28 16:30:33 +00:00
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
auto& handle_table = system.Kernel().CurrentProcess()->GetHandleTable();
|
2018-10-20 18:34:41 +00:00
|
|
|
return handle_table.Close(handle);
|
2015-08-06 00:39:53 +00:00
|
|
|
}
|
|
|
|
|
2020-03-02 05:06:41 +00:00
|
|
|
static ResultCode CloseHandle32(Core::System& system, Handle handle) {
|
|
|
|
return CloseHandle(system, handle);
|
|
|
|
}
|
|
|
|
|
2018-12-05 00:59:29 +00:00
|
|
|
/// Clears the signaled state of an event or process.
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode ResetSignal(Core::System& system, Handle handle) {
|
2018-11-18 20:49:17 +00:00
|
|
|
LOG_DEBUG(Kernel_SVC, "called handle 0x{:08X}", handle);
|
2018-08-28 16:30:33 +00:00
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
const auto& handle_table = system.Kernel().CurrentProcess()->GetHandleTable();
|
2018-12-05 00:59:29 +00:00
|
|
|
|
2018-11-26 23:34:07 +00:00
|
|
|
auto event = handle_table.Get<ReadableEvent>(handle);
|
2018-12-05 00:59:29 +00:00
|
|
|
if (event) {
|
|
|
|
return event->Reset();
|
|
|
|
}
|
2018-08-28 16:30:33 +00:00
|
|
|
|
2018-12-05 00:59:29 +00:00
|
|
|
auto process = handle_table.Get<Process>(handle);
|
|
|
|
if (process) {
|
|
|
|
return process->ClearSignalState();
|
|
|
|
}
|
2018-08-28 16:30:33 +00:00
|
|
|
|
2018-12-05 00:59:29 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Invalid handle (0x{:08X})", handle);
|
|
|
|
return ERR_INVALID_HANDLE;
|
2018-01-08 02:24:19 +00:00
|
|
|
}
|
|
|
|
|
2020-06-19 23:40:07 +00:00
|
|
|
static ResultCode ResetSignal32(Core::System& system, Handle handle) {
|
|
|
|
return ResetSignal(system, handle);
|
|
|
|
}
|
|
|
|
|
2018-01-08 02:24:19 +00:00
|
|
|
/// Creates a TransferMemory object
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode CreateTransferMemory(Core::System& system, Handle* handle, VAddr addr, u64 size,
|
|
|
|
u32 permissions) {
|
2020-02-27 14:28:44 +00:00
|
|
|
std::lock_guard lock{HLE::g_hle_lock};
|
2018-11-09 22:02:50 +00:00
|
|
|
LOG_DEBUG(Kernel_SVC, "called addr=0x{:X}, size=0x{:X}, perms=0x{:08X}", addr, size,
|
|
|
|
permissions);
|
|
|
|
|
|
|
|
if (!Common::Is4KBAligned(addr)) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Address ({:016X}) is not page aligned!", addr);
|
|
|
|
return ERR_INVALID_ADDRESS;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!Common::Is4KBAligned(size) || size == 0) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Size ({:016X}) is not page aligned or equal to zero!", size);
|
|
|
|
return ERR_INVALID_ADDRESS;
|
|
|
|
}
|
|
|
|
|
2018-11-11 21:39:25 +00:00
|
|
|
if (!IsValidAddressRange(addr, size)) {
|
2018-11-09 22:02:50 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Address and size cause overflow! (address={:016X}, size={:016X})",
|
|
|
|
addr, size);
|
|
|
|
return ERR_INVALID_ADDRESS_STATE;
|
|
|
|
}
|
|
|
|
|
2020-04-09 03:14:18 +00:00
|
|
|
const auto perms{static_cast<Memory::MemoryPermission>(permissions)};
|
|
|
|
if (perms > Memory::MemoryPermission::ReadAndWrite ||
|
|
|
|
perms == Memory::MemoryPermission::Write) {
|
2018-11-09 22:02:50 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Invalid memory permissions for transfer memory! (perms={:08X})",
|
|
|
|
permissions);
|
|
|
|
return ERR_INVALID_MEMORY_PERMISSIONS;
|
|
|
|
}
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
auto& kernel = system.Kernel();
|
2020-01-31 03:39:07 +00:00
|
|
|
auto transfer_mem_handle = TransferMemory::Create(kernel, system.Memory(), addr, size, perms);
|
|
|
|
|
|
|
|
if (const auto reserve_result{transfer_mem_handle->Reserve()}; reserve_result.IsError()) {
|
|
|
|
return reserve_result;
|
|
|
|
}
|
2018-11-09 22:02:50 +00:00
|
|
|
|
2019-03-13 07:09:27 +00:00
|
|
|
auto& handle_table = kernel.CurrentProcess()->GetHandleTable();
|
2020-01-31 03:39:07 +00:00
|
|
|
const auto result{handle_table.Create(std::move(transfer_mem_handle))};
|
2019-03-13 07:09:27 +00:00
|
|
|
if (result.Failed()) {
|
|
|
|
return result.Code();
|
|
|
|
}
|
|
|
|
|
|
|
|
*handle = *result;
|
2018-01-08 02:24:19 +00:00
|
|
|
return RESULT_SUCCESS;
|
|
|
|
}
|
|
|
|
|
2020-06-19 00:33:04 +00:00
|
|
|
static ResultCode CreateTransferMemory32(Core::System& system, Handle* handle, u32 addr, u32 size,
|
|
|
|
u32 permissions) {
|
|
|
|
return CreateTransferMemory(system, handle, static_cast<VAddr>(addr),
|
|
|
|
static_cast<std::size_t>(size), permissions);
|
|
|
|
}
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode GetThreadCoreMask(Core::System& system, Handle thread_handle, u32* core,
|
|
|
|
u64* mask) {
|
2018-07-02 16:13:26 +00:00
|
|
|
LOG_TRACE(Kernel_SVC, "called, handle=0x{:08X}", thread_handle);
|
2018-05-06 03:13:15 +00:00
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
const auto& handle_table = system.Kernel().CurrentProcess()->GetHandleTable();
|
2019-11-25 01:15:51 +00:00
|
|
|
const std::shared_ptr<Thread> thread = handle_table.Get<Thread>(thread_handle);
|
2018-05-06 03:13:15 +00:00
|
|
|
if (!thread) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Thread handle does not exist, thread_handle=0x{:08X}",
|
|
|
|
thread_handle);
|
2020-02-25 17:22:11 +00:00
|
|
|
*core = 0;
|
|
|
|
*mask = 0;
|
2018-05-06 03:13:15 +00:00
|
|
|
return ERR_INVALID_HANDLE;
|
|
|
|
}
|
|
|
|
|
2020-10-21 02:07:39 +00:00
|
|
|
*core = thread->GetIdealCore();
|
2020-11-17 05:02:45 +00:00
|
|
|
*mask = thread->GetAffinityMask().GetAffinityMask();
|
2018-05-06 03:13:15 +00:00
|
|
|
|
2018-03-30 01:07:49 +00:00
|
|
|
return RESULT_SUCCESS;
|
|
|
|
}
|
|
|
|
|
2020-06-19 23:40:07 +00:00
|
|
|
static ResultCode GetThreadCoreMask32(Core::System& system, Handle thread_handle, u32* core,
|
|
|
|
u32* mask_low, u32* mask_high) {
|
|
|
|
u64 mask{};
|
|
|
|
const auto result = GetThreadCoreMask(system, thread_handle, core, &mask);
|
|
|
|
*mask_high = static_cast<u32>(mask >> 32);
|
|
|
|
*mask_low = static_cast<u32>(mask);
|
|
|
|
return result;
|
|
|
|
}
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode SetThreadCoreMask(Core::System& system, Handle thread_handle, u32 core,
|
2019-04-16 00:34:55 +00:00
|
|
|
u64 affinity_mask) {
|
|
|
|
LOG_DEBUG(Kernel_SVC, "called, handle=0x{:08X}, core=0x{:X}, affinity_mask=0x{:016X}",
|
|
|
|
thread_handle, core, affinity_mask);
|
2018-05-06 03:13:15 +00:00
|
|
|
|
2019-04-16 00:34:55 +00:00
|
|
|
const auto* const current_process = system.Kernel().CurrentProcess();
|
2018-05-06 03:13:15 +00:00
|
|
|
|
2018-12-28 02:14:59 +00:00
|
|
|
if (core == static_cast<u32>(THREADPROCESSORID_IDEAL)) {
|
2019-04-16 00:34:55 +00:00
|
|
|
const u8 ideal_cpu_core = current_process->GetIdealCore();
|
2018-10-03 22:47:57 +00:00
|
|
|
|
2018-12-28 02:14:59 +00:00
|
|
|
ASSERT(ideal_cpu_core != static_cast<u8>(THREADPROCESSORID_IDEAL));
|
2018-10-03 22:47:57 +00:00
|
|
|
|
2018-12-28 02:14:59 +00:00
|
|
|
// Set the target CPU to the ideal core specified by the process.
|
|
|
|
core = ideal_cpu_core;
|
2019-04-16 00:34:55 +00:00
|
|
|
affinity_mask = 1ULL << core;
|
|
|
|
} else {
|
|
|
|
const u64 core_mask = current_process->GetCoreMask();
|
|
|
|
|
|
|
|
if ((core_mask | affinity_mask) != core_mask) {
|
|
|
|
LOG_ERROR(
|
|
|
|
Kernel_SVC,
|
|
|
|
"Invalid processor ID specified (core_mask=0x{:08X}, affinity_mask=0x{:016X})",
|
|
|
|
core_mask, affinity_mask);
|
|
|
|
return ERR_INVALID_PROCESSOR_ID;
|
|
|
|
}
|
2018-05-30 17:03:19 +00:00
|
|
|
|
2019-04-16 00:34:55 +00:00
|
|
|
if (affinity_mask == 0) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Specfified affinity mask is zero.");
|
|
|
|
return ERR_INVALID_COMBINATION;
|
|
|
|
}
|
2018-05-30 17:03:19 +00:00
|
|
|
|
2020-04-01 21:28:49 +00:00
|
|
|
if (core < Core::Hardware::NUM_CPU_CORES) {
|
2019-04-16 00:34:55 +00:00
|
|
|
if ((affinity_mask & (1ULL << core)) == 0) {
|
|
|
|
LOG_ERROR(Kernel_SVC,
|
|
|
|
"Core is not enabled for the current mask, core={}, mask={:016X}", core,
|
|
|
|
affinity_mask);
|
|
|
|
return ERR_INVALID_COMBINATION;
|
|
|
|
}
|
|
|
|
} else if (core != static_cast<u32>(THREADPROCESSORID_DONT_CARE) &&
|
|
|
|
core != static_cast<u32>(THREADPROCESSORID_DONT_UPDATE)) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Invalid processor ID specified (core={}).", core);
|
|
|
|
return ERR_INVALID_PROCESSOR_ID;
|
|
|
|
}
|
2018-05-30 17:03:19 +00:00
|
|
|
}
|
|
|
|
|
2019-04-16 00:34:55 +00:00
|
|
|
const auto& handle_table = current_process->GetHandleTable();
|
2019-11-25 01:15:51 +00:00
|
|
|
const std::shared_ptr<Thread> thread = handle_table.Get<Thread>(thread_handle);
|
2019-04-16 00:34:55 +00:00
|
|
|
if (!thread) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Thread handle does not exist, thread_handle=0x{:08X}",
|
|
|
|
thread_handle);
|
|
|
|
return ERR_INVALID_HANDLE;
|
2018-05-30 17:03:19 +00:00
|
|
|
}
|
|
|
|
|
2020-10-21 02:07:39 +00:00
|
|
|
return thread->SetCoreAndAffinityMask(core, affinity_mask);
|
2018-01-16 22:23:53 +00:00
|
|
|
}
|
|
|
|
|
2020-06-18 22:15:19 +00:00
|
|
|
static ResultCode SetThreadCoreMask32(Core::System& system, Handle thread_handle, u32 core,
|
2020-06-19 00:33:04 +00:00
|
|
|
u32 affinity_mask_low, u32 affinity_mask_high) {
|
|
|
|
const u64 affinity_mask =
|
|
|
|
static_cast<u64>(affinity_mask_low) | (static_cast<u64>(affinity_mask_high) << 32);
|
2020-06-18 22:15:19 +00:00
|
|
|
return SetThreadCoreMask(system, thread_handle, core, affinity_mask);
|
|
|
|
}
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode CreateEvent(Core::System& system, Handle* write_handle, Handle* read_handle) {
|
2018-12-04 20:11:18 +00:00
|
|
|
LOG_DEBUG(Kernel_SVC, "called");
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
auto& kernel = system.Kernel();
|
2018-12-04 20:11:18 +00:00
|
|
|
const auto [readable_event, writable_event] =
|
2019-11-03 09:10:12 +00:00
|
|
|
WritableEvent::CreateEventPair(kernel, "CreateEvent");
|
2018-12-04 20:11:18 +00:00
|
|
|
|
|
|
|
HandleTable& handle_table = kernel.CurrentProcess()->GetHandleTable();
|
|
|
|
|
|
|
|
const auto write_create_result = handle_table.Create(writable_event);
|
|
|
|
if (write_create_result.Failed()) {
|
|
|
|
return write_create_result.Code();
|
|
|
|
}
|
|
|
|
*write_handle = *write_create_result;
|
|
|
|
|
|
|
|
const auto read_create_result = handle_table.Create(readable_event);
|
|
|
|
if (read_create_result.Failed()) {
|
|
|
|
handle_table.Close(*write_create_result);
|
|
|
|
return read_create_result.Code();
|
|
|
|
}
|
|
|
|
*read_handle = *read_create_result;
|
|
|
|
|
|
|
|
LOG_DEBUG(Kernel_SVC,
|
|
|
|
"successful. Writable event handle=0x{:08X}, Readable event handle=0x{:08X}",
|
|
|
|
*write_create_result, *read_create_result);
|
|
|
|
return RESULT_SUCCESS;
|
|
|
|
}
|
|
|
|
|
2020-06-19 23:40:07 +00:00
|
|
|
static ResultCode CreateEvent32(Core::System& system, Handle* write_handle, Handle* read_handle) {
|
|
|
|
return CreateEvent(system, write_handle, read_handle);
|
|
|
|
}
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode ClearEvent(Core::System& system, Handle handle) {
|
2018-07-02 16:13:26 +00:00
|
|
|
LOG_TRACE(Kernel_SVC, "called, event=0x{:08X}", handle);
|
2018-02-22 14:28:15 +00:00
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
const auto& handle_table = system.Kernel().CurrentProcess()->GetHandleTable();
|
2018-12-04 03:50:16 +00:00
|
|
|
|
|
|
|
auto writable_event = handle_table.Get<WritableEvent>(handle);
|
|
|
|
if (writable_event) {
|
|
|
|
writable_event->Clear();
|
|
|
|
return RESULT_SUCCESS;
|
2018-10-20 18:34:41 +00:00
|
|
|
}
|
|
|
|
|
2018-12-04 03:50:16 +00:00
|
|
|
auto readable_event = handle_table.Get<ReadableEvent>(handle);
|
|
|
|
if (readable_event) {
|
|
|
|
readable_event->Clear();
|
|
|
|
return RESULT_SUCCESS;
|
|
|
|
}
|
|
|
|
|
|
|
|
LOG_ERROR(Kernel_SVC, "Event handle does not exist, handle=0x{:08X}", handle);
|
|
|
|
return ERR_INVALID_HANDLE;
|
2018-02-22 14:28:15 +00:00
|
|
|
}
|
|
|
|
|
2020-06-19 23:40:07 +00:00
|
|
|
static ResultCode ClearEvent32(Core::System& system, Handle handle) {
|
|
|
|
return ClearEvent(system, handle);
|
|
|
|
}
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode SignalEvent(Core::System& system, Handle handle) {
|
2018-12-04 20:39:49 +00:00
|
|
|
LOG_DEBUG(Kernel_SVC, "called. Handle=0x{:08X}", handle);
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
HandleTable& handle_table = system.Kernel().CurrentProcess()->GetHandleTable();
|
2018-12-04 20:39:49 +00:00
|
|
|
auto writable_event = handle_table.Get<WritableEvent>(handle);
|
|
|
|
|
|
|
|
if (!writable_event) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Non-existent writable event handle used (0x{:08X})", handle);
|
|
|
|
return ERR_INVALID_HANDLE;
|
|
|
|
}
|
|
|
|
|
|
|
|
writable_event->Signal();
|
|
|
|
return RESULT_SUCCESS;
|
|
|
|
}
|
|
|
|
|
2020-06-19 23:40:07 +00:00
|
|
|
static ResultCode SignalEvent32(Core::System& system, Handle handle) {
|
|
|
|
return SignalEvent(system, handle);
|
|
|
|
}
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode GetProcessInfo(Core::System& system, u64* out, Handle process_handle, u32 type) {
|
2018-10-13 18:31:46 +00:00
|
|
|
LOG_DEBUG(Kernel_SVC, "called, handle=0x{:08X}, type=0x{:X}", process_handle, type);
|
|
|
|
|
|
|
|
// This function currently only allows retrieving a process' status.
|
|
|
|
enum class InfoType {
|
|
|
|
Status,
|
|
|
|
};
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
const auto& handle_table = system.Kernel().CurrentProcess()->GetHandleTable();
|
2018-10-20 18:34:41 +00:00
|
|
|
const auto process = handle_table.Get<Process>(process_handle);
|
2018-10-13 18:31:46 +00:00
|
|
|
if (!process) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Process handle does not exist, process_handle=0x{:08X}",
|
|
|
|
process_handle);
|
2018-10-13 18:31:46 +00:00
|
|
|
return ERR_INVALID_HANDLE;
|
|
|
|
}
|
|
|
|
|
|
|
|
const auto info_type = static_cast<InfoType>(type);
|
|
|
|
if (info_type != InfoType::Status) {
|
2018-11-26 08:47:39 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Expected info_type to be Status but got {} instead", type);
|
2018-10-13 18:31:46 +00:00
|
|
|
return ERR_INVALID_ENUM_VALUE;
|
|
|
|
}
|
|
|
|
|
|
|
|
*out = static_cast<u64>(process->GetStatus());
|
|
|
|
return RESULT_SUCCESS;
|
|
|
|
}
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode CreateResourceLimit(Core::System& system, Handle* out_handle) {
|
2020-02-27 14:28:44 +00:00
|
|
|
std::lock_guard lock{HLE::g_hle_lock};
|
2018-11-26 23:23:12 +00:00
|
|
|
LOG_DEBUG(Kernel_SVC, "called");
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
auto& kernel = system.Kernel();
|
2018-11-26 23:23:12 +00:00
|
|
|
auto resource_limit = ResourceLimit::Create(kernel);
|
|
|
|
|
|
|
|
auto* const current_process = kernel.CurrentProcess();
|
|
|
|
ASSERT(current_process != nullptr);
|
|
|
|
|
|
|
|
const auto handle = current_process->GetHandleTable().Create(std::move(resource_limit));
|
|
|
|
if (handle.Failed()) {
|
|
|
|
return handle.Code();
|
|
|
|
}
|
|
|
|
|
|
|
|
*out_handle = *handle;
|
|
|
|
return RESULT_SUCCESS;
|
|
|
|
}
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode GetResourceLimitLimitValue(Core::System& system, u64* out_value,
|
|
|
|
Handle resource_limit, u32 resource_type) {
|
2018-11-26 23:48:07 +00:00
|
|
|
LOG_DEBUG(Kernel_SVC, "called. Handle={:08X}, Resource type={}", resource_limit, resource_type);
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
const auto limit_value = RetrieveResourceLimitValue(system, resource_limit, resource_type,
|
2018-11-27 00:14:29 +00:00
|
|
|
ResourceLimitValueType::LimitValue);
|
|
|
|
if (limit_value.Failed()) {
|
|
|
|
return limit_value.Code();
|
2018-11-26 23:48:07 +00:00
|
|
|
}
|
|
|
|
|
2018-11-27 00:14:29 +00:00
|
|
|
*out_value = static_cast<u64>(*limit_value);
|
|
|
|
return RESULT_SUCCESS;
|
|
|
|
}
|
2018-11-26 23:48:07 +00:00
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode GetResourceLimitCurrentValue(Core::System& system, u64* out_value,
|
|
|
|
Handle resource_limit, u32 resource_type) {
|
2018-11-27 00:14:29 +00:00
|
|
|
LOG_DEBUG(Kernel_SVC, "called. Handle={:08X}, Resource type={}", resource_limit, resource_type);
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
const auto current_value = RetrieveResourceLimitValue(system, resource_limit, resource_type,
|
2018-11-27 00:14:29 +00:00
|
|
|
ResourceLimitValueType::CurrentValue);
|
|
|
|
if (current_value.Failed()) {
|
|
|
|
return current_value.Code();
|
2018-11-26 23:48:07 +00:00
|
|
|
}
|
|
|
|
|
2018-11-27 00:14:29 +00:00
|
|
|
*out_value = static_cast<u64>(*current_value);
|
2018-11-26 23:48:07 +00:00
|
|
|
return RESULT_SUCCESS;
|
|
|
|
}
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode SetResourceLimitLimitValue(Core::System& system, Handle resource_limit,
|
|
|
|
u32 resource_type, u64 value) {
|
2018-11-27 00:51:09 +00:00
|
|
|
LOG_DEBUG(Kernel_SVC, "called. Handle={:08X}, Resource type={}, Value={}", resource_limit,
|
|
|
|
resource_type, value);
|
|
|
|
|
|
|
|
const auto type = static_cast<ResourceType>(resource_type);
|
|
|
|
if (!IsValidResourceType(type)) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Invalid resource limit type: '{}'", resource_type);
|
|
|
|
return ERR_INVALID_ENUM_VALUE;
|
|
|
|
}
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
auto* const current_process = system.Kernel().CurrentProcess();
|
2018-11-27 00:51:09 +00:00
|
|
|
ASSERT(current_process != nullptr);
|
|
|
|
|
|
|
|
auto resource_limit_object =
|
|
|
|
current_process->GetHandleTable().Get<ResourceLimit>(resource_limit);
|
|
|
|
if (!resource_limit_object) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Handle to non-existent resource limit instance used. Handle={:08X}",
|
|
|
|
resource_limit);
|
|
|
|
return ERR_INVALID_HANDLE;
|
|
|
|
}
|
|
|
|
|
|
|
|
const auto set_result = resource_limit_object->SetLimitValue(type, static_cast<s64>(value));
|
|
|
|
if (set_result.IsError()) {
|
|
|
|
LOG_ERROR(
|
|
|
|
Kernel_SVC,
|
|
|
|
"Attempted to lower resource limit ({}) for category '{}' below its current value ({})",
|
|
|
|
resource_limit_object->GetMaxResourceValue(type), resource_type,
|
|
|
|
resource_limit_object->GetCurrentResourceValue(type));
|
|
|
|
return set_result;
|
|
|
|
}
|
|
|
|
|
|
|
|
return RESULT_SUCCESS;
|
|
|
|
}
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
static ResultCode GetProcessList(Core::System& system, u32* out_num_processes,
|
|
|
|
VAddr out_process_ids, u32 out_process_ids_size) {
|
2019-03-20 19:03:52 +00:00
|
|
|
LOG_DEBUG(Kernel_SVC, "called. out_process_ids=0x{:016X}, out_process_ids_size={}",
|
|
|
|
out_process_ids, out_process_ids_size);
|
|
|
|
|
|
|
|
// If the supplied size is negative or greater than INT32_MAX / sizeof(u64), bail.
|
|
|
|
if ((out_process_ids_size & 0xF0000000) != 0) {
|
|
|
|
LOG_ERROR(Kernel_SVC,
|
|
|
|
"Supplied size outside [0, 0x0FFFFFFF] range. out_process_ids_size={}",
|
|
|
|
out_process_ids_size);
|
|
|
|
return ERR_OUT_OF_RANGE;
|
|
|
|
}
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
const auto& kernel = system.Kernel();
|
2019-03-20 19:03:52 +00:00
|
|
|
const auto total_copy_size = out_process_ids_size * sizeof(u64);
|
|
|
|
|
2020-04-09 03:14:18 +00:00
|
|
|
if (out_process_ids_size > 0 && !kernel.CurrentProcess()->PageTable().IsInsideAddressSpace(
|
|
|
|
out_process_ids, total_copy_size)) {
|
2019-03-20 19:03:52 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Address range outside address space. begin=0x{:016X}, end=0x{:016X}",
|
|
|
|
out_process_ids, out_process_ids + total_copy_size);
|
|
|
|
return ERR_INVALID_ADDRESS_STATE;
|
|
|
|
}
|
|
|
|
|
2019-11-26 22:39:57 +00:00
|
|
|
auto& memory = system.Memory();
|
2019-03-20 19:03:52 +00:00
|
|
|
const auto& process_list = kernel.GetProcessList();
|
|
|
|
const auto num_processes = process_list.size();
|
|
|
|
const auto copy_amount = std::min(std::size_t{out_process_ids_size}, num_processes);
|
|
|
|
|
|
|
|
for (std::size_t i = 0; i < copy_amount; ++i) {
|
2019-11-26 22:39:57 +00:00
|
|
|
memory.Write64(out_process_ids, process_list[i]->GetProcessID());
|
2019-03-20 19:03:52 +00:00
|
|
|
out_process_ids += sizeof(u64);
|
|
|
|
}
|
|
|
|
|
|
|
|
*out_num_processes = static_cast<u32>(num_processes);
|
|
|
|
return RESULT_SUCCESS;
|
|
|
|
}
|
|
|
|
|
2019-05-18 23:01:27 +00:00
|
|
|
static ResultCode GetThreadList(Core::System& system, u32* out_num_threads, VAddr out_thread_ids,
|
|
|
|
u32 out_thread_ids_size, Handle debug_handle) {
|
2019-03-20 22:53:48 +00:00
|
|
|
// TODO: Handle this case when debug events are supported.
|
|
|
|
UNIMPLEMENTED_IF(debug_handle != InvalidHandle);
|
|
|
|
|
|
|
|
LOG_DEBUG(Kernel_SVC, "called. out_thread_ids=0x{:016X}, out_thread_ids_size={}",
|
|
|
|
out_thread_ids, out_thread_ids_size);
|
|
|
|
|
|
|
|
// If the size is negative or larger than INT32_MAX / sizeof(u64)
|
|
|
|
if ((out_thread_ids_size & 0xF0000000) != 0) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Supplied size outside [0, 0x0FFFFFFF] range. size={}",
|
|
|
|
out_thread_ids_size);
|
|
|
|
return ERR_OUT_OF_RANGE;
|
|
|
|
}
|
|
|
|
|
2019-04-06 22:46:18 +00:00
|
|
|
const auto* const current_process = system.Kernel().CurrentProcess();
|
2019-03-20 22:53:48 +00:00
|
|
|
const auto total_copy_size = out_thread_ids_size * sizeof(u64);
|
|
|
|
|
|
|
|
if (out_thread_ids_size > 0 &&
|
2020-04-09 03:14:18 +00:00
|
|
|
!current_process->PageTable().IsInsideAddressSpace(out_thread_ids, total_copy_size)) {
|
2019-03-20 22:53:48 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Address range outside address space. begin=0x{:016X}, end=0x{:016X}",
|
|
|
|
out_thread_ids, out_thread_ids + total_copy_size);
|
|
|
|
return ERR_INVALID_ADDRESS_STATE;
|
|
|
|
}
|
|
|
|
|
2019-11-26 22:39:57 +00:00
|
|
|
auto& memory = system.Memory();
|
2019-03-20 22:53:48 +00:00
|
|
|
const auto& thread_list = current_process->GetThreadList();
|
|
|
|
const auto num_threads = thread_list.size();
|
|
|
|
const auto copy_amount = std::min(std::size_t{out_thread_ids_size}, num_threads);
|
|
|
|
|
|
|
|
auto list_iter = thread_list.cbegin();
|
|
|
|
for (std::size_t i = 0; i < copy_amount; ++i, ++list_iter) {
|
2019-11-26 22:39:57 +00:00
|
|
|
memory.Write64(out_thread_ids, (*list_iter)->GetThreadID());
|
2019-03-20 22:53:48 +00:00
|
|
|
out_thread_ids += sizeof(u64);
|
|
|
|
}
|
|
|
|
|
|
|
|
*out_num_threads = static_cast<u32>(num_threads);
|
|
|
|
return RESULT_SUCCESS;
|
|
|
|
}
|
|
|
|
|
2020-06-19 23:40:07 +00:00
|
|
|
static ResultCode FlushProcessDataCache32(Core::System& system, Handle handle, u32 address,
|
|
|
|
u32 size) {
|
|
|
|
// Note(Blinkhawk): For emulation purposes of the data cache this is mostly a nope
|
|
|
|
// as all emulation is done in the same cache level in host architecture, thus data cache
|
|
|
|
// does not need flushing.
|
|
|
|
LOG_DEBUG(Kernel_SVC, "called");
|
|
|
|
return RESULT_SUCCESS;
|
|
|
|
}
|
|
|
|
|
2015-05-06 03:04:25 +00:00
|
|
|
namespace {
|
2016-09-18 00:38:01 +00:00
|
|
|
struct FunctionDef {
|
2019-04-06 22:46:18 +00:00
|
|
|
using Func = void(Core::System&);
|
2015-05-06 03:04:25 +00:00
|
|
|
|
2016-09-18 00:38:01 +00:00
|
|
|
u32 id;
|
|
|
|
Func* func;
|
|
|
|
const char* name;
|
|
|
|
};
|
2017-10-14 21:30:07 +00:00
|
|
|
} // namespace
|
2015-05-06 03:04:25 +00:00
|
|
|
|
2020-03-02 05:06:41 +00:00
|
|
|
static const FunctionDef SVC_Table_32[] = {
|
2016-09-18 00:38:01 +00:00
|
|
|
{0x00, nullptr, "Unknown"},
|
2020-03-02 05:06:41 +00:00
|
|
|
{0x01, SvcWrap32<SetHeapSize32>, "SetHeapSize32"},
|
|
|
|
{0x02, nullptr, "Unknown"},
|
2020-06-19 00:33:04 +00:00
|
|
|
{0x03, SvcWrap32<SetMemoryAttribute32>, "SetMemoryAttribute32"},
|
2020-06-19 23:40:07 +00:00
|
|
|
{0x04, SvcWrap32<MapMemory32>, "MapMemory32"},
|
|
|
|
{0x05, SvcWrap32<UnmapMemory32>, "UnmapMemory32"},
|
2020-03-02 05:06:41 +00:00
|
|
|
{0x06, SvcWrap32<QueryMemory32>, "QueryMemory32"},
|
2020-06-19 23:40:07 +00:00
|
|
|
{0x07, SvcWrap32<ExitProcess32>, "ExitProcess32"},
|
|
|
|
{0x08, SvcWrap32<CreateThread32>, "CreateThread32"},
|
|
|
|
{0x09, SvcWrap32<StartThread32>, "StartThread32"},
|
|
|
|
{0x0a, SvcWrap32<ExitThread32>, "ExitThread32"},
|
|
|
|
{0x0b, SvcWrap32<SleepThread32>, "SleepThread32"},
|
2020-03-02 05:06:41 +00:00
|
|
|
{0x0c, SvcWrap32<GetThreadPriority32>, "GetThreadPriority32"},
|
2020-06-18 22:15:19 +00:00
|
|
|
{0x0d, SvcWrap32<SetThreadPriority32>, "SetThreadPriority32"},
|
2020-06-19 23:40:07 +00:00
|
|
|
{0x0e, SvcWrap32<GetThreadCoreMask32>, "GetThreadCoreMask32"},
|
2020-06-18 22:15:19 +00:00
|
|
|
{0x0f, SvcWrap32<SetThreadCoreMask32>, "SetThreadCoreMask32"},
|
2020-06-19 00:33:04 +00:00
|
|
|
{0x10, SvcWrap32<GetCurrentProcessorNumber32>, "GetCurrentProcessorNumber32"},
|
2020-06-19 23:40:07 +00:00
|
|
|
{0x11, SvcWrap32<SignalEvent32>, "SignalEvent32"},
|
|
|
|
{0x12, SvcWrap32<ClearEvent32>, "ClearEvent32"},
|
|
|
|
{0x13, SvcWrap32<MapSharedMemory32>, "MapSharedMemory32"},
|
2020-03-02 05:06:41 +00:00
|
|
|
{0x14, nullptr, "UnmapSharedMemory32"},
|
2020-06-19 00:33:04 +00:00
|
|
|
{0x15, SvcWrap32<CreateTransferMemory32>, "CreateTransferMemory32"},
|
2020-03-02 05:06:41 +00:00
|
|
|
{0x16, SvcWrap32<CloseHandle32>, "CloseHandle32"},
|
2020-06-19 23:40:07 +00:00
|
|
|
{0x17, SvcWrap32<ResetSignal32>, "ResetSignal32"},
|
2020-03-02 05:06:41 +00:00
|
|
|
{0x18, SvcWrap32<WaitSynchronization32>, "WaitSynchronization32"},
|
2020-06-19 23:40:07 +00:00
|
|
|
{0x19, SvcWrap32<CancelSynchronization32>, "CancelSynchronization32"},
|
|
|
|
{0x1a, SvcWrap32<ArbitrateLock32>, "ArbitrateLock32"},
|
|
|
|
{0x1b, SvcWrap32<ArbitrateUnlock32>, "ArbitrateUnlock32"},
|
|
|
|
{0x1c, SvcWrap32<WaitProcessWideKeyAtomic32>, "WaitProcessWideKeyAtomic32"},
|
2020-03-02 05:06:41 +00:00
|
|
|
{0x1d, SvcWrap32<SignalProcessWideKey32>, "SignalProcessWideKey32"},
|
2020-06-19 23:40:07 +00:00
|
|
|
{0x1e, SvcWrap32<GetSystemTick32>, "GetSystemTick32"},
|
2020-03-02 05:06:41 +00:00
|
|
|
{0x1f, SvcWrap32<ConnectToNamedPort32>, "ConnectToNamedPort32"},
|
|
|
|
{0x20, nullptr, "Unknown"},
|
|
|
|
{0x21, SvcWrap32<SendSyncRequest32>, "SendSyncRequest32"},
|
|
|
|
{0x22, nullptr, "SendSyncRequestWithUserBuffer32"},
|
|
|
|
{0x23, nullptr, "Unknown"},
|
2020-06-19 23:40:07 +00:00
|
|
|
{0x24, SvcWrap32<GetProcessId32>, "GetProcessId32"},
|
2020-03-02 05:06:41 +00:00
|
|
|
{0x25, SvcWrap32<GetThreadId32>, "GetThreadId32"},
|
2020-06-19 23:40:07 +00:00
|
|
|
{0x26, SvcWrap32<Break32>, "Break32"},
|
2020-03-02 05:06:41 +00:00
|
|
|
{0x27, nullptr, "OutputDebugString32"},
|
|
|
|
{0x28, nullptr, "Unknown"},
|
|
|
|
{0x29, SvcWrap32<GetInfo32>, "GetInfo32"},
|
|
|
|
{0x2a, nullptr, "Unknown"},
|
|
|
|
{0x2b, nullptr, "Unknown"},
|
2020-06-19 23:40:07 +00:00
|
|
|
{0x2c, SvcWrap32<MapPhysicalMemory32>, "MapPhysicalMemory32"},
|
|
|
|
{0x2d, SvcWrap32<UnmapPhysicalMemory32>, "UnmapPhysicalMemory32"},
|
2020-03-02 05:06:41 +00:00
|
|
|
{0x2e, nullptr, "Unknown"},
|
|
|
|
{0x2f, nullptr, "Unknown"},
|
|
|
|
{0x30, nullptr, "Unknown"},
|
|
|
|
{0x31, nullptr, "Unknown"},
|
2020-06-19 23:40:07 +00:00
|
|
|
{0x32, SvcWrap32<SetThreadActivity32>, "SetThreadActivity32"},
|
|
|
|
{0x33, SvcWrap32<GetThreadContext32>, "GetThreadContext32"},
|
|
|
|
{0x34, SvcWrap32<WaitForAddress32>, "WaitForAddress32"},
|
|
|
|
{0x35, SvcWrap32<SignalToAddress32>, "SignalToAddress32"},
|
2020-03-02 05:06:41 +00:00
|
|
|
{0x36, nullptr, "Unknown"},
|
|
|
|
{0x37, nullptr, "Unknown"},
|
|
|
|
{0x38, nullptr, "Unknown"},
|
|
|
|
{0x39, nullptr, "Unknown"},
|
|
|
|
{0x3a, nullptr, "Unknown"},
|
|
|
|
{0x3b, nullptr, "Unknown"},
|
|
|
|
{0x3c, nullptr, "Unknown"},
|
|
|
|
{0x3d, nullptr, "Unknown"},
|
|
|
|
{0x3e, nullptr, "Unknown"},
|
|
|
|
{0x3f, nullptr, "Unknown"},
|
|
|
|
{0x40, nullptr, "CreateSession32"},
|
|
|
|
{0x41, nullptr, "AcceptSession32"},
|
|
|
|
{0x42, nullptr, "Unknown"},
|
|
|
|
{0x43, nullptr, "ReplyAndReceive32"},
|
|
|
|
{0x44, nullptr, "Unknown"},
|
2020-06-19 23:40:07 +00:00
|
|
|
{0x45, SvcWrap32<CreateEvent32>, "CreateEvent32"},
|
2020-03-02 05:06:41 +00:00
|
|
|
{0x46, nullptr, "Unknown"},
|
|
|
|
{0x47, nullptr, "Unknown"},
|
|
|
|
{0x48, nullptr, "Unknown"},
|
|
|
|
{0x49, nullptr, "Unknown"},
|
|
|
|
{0x4a, nullptr, "Unknown"},
|
|
|
|
{0x4b, nullptr, "Unknown"},
|
|
|
|
{0x4c, nullptr, "Unknown"},
|
|
|
|
{0x4d, nullptr, "Unknown"},
|
|
|
|
{0x4e, nullptr, "Unknown"},
|
|
|
|
{0x4f, nullptr, "Unknown"},
|
|
|
|
{0x50, nullptr, "Unknown"},
|
|
|
|
{0x51, nullptr, "Unknown"},
|
|
|
|
{0x52, nullptr, "Unknown"},
|
|
|
|
{0x53, nullptr, "Unknown"},
|
|
|
|
{0x54, nullptr, "Unknown"},
|
|
|
|
{0x55, nullptr, "Unknown"},
|
|
|
|
{0x56, nullptr, "Unknown"},
|
|
|
|
{0x57, nullptr, "Unknown"},
|
|
|
|
{0x58, nullptr, "Unknown"},
|
|
|
|
{0x59, nullptr, "Unknown"},
|
|
|
|
{0x5a, nullptr, "Unknown"},
|
|
|
|
{0x5b, nullptr, "Unknown"},
|
|
|
|
{0x5c, nullptr, "Unknown"},
|
|
|
|
{0x5d, nullptr, "Unknown"},
|
|
|
|
{0x5e, nullptr, "Unknown"},
|
2020-06-19 23:40:07 +00:00
|
|
|
{0x5F, SvcWrap32<FlushProcessDataCache32>, "FlushProcessDataCache32"},
|
2020-03-02 05:06:41 +00:00
|
|
|
{0x60, nullptr, "Unknown"},
|
|
|
|
{0x61, nullptr, "Unknown"},
|
|
|
|
{0x62, nullptr, "Unknown"},
|
|
|
|
{0x63, nullptr, "Unknown"},
|
|
|
|
{0x64, nullptr, "Unknown"},
|
|
|
|
{0x65, nullptr, "GetProcessList32"},
|
|
|
|
{0x66, nullptr, "Unknown"},
|
|
|
|
{0x67, nullptr, "Unknown"},
|
|
|
|
{0x68, nullptr, "Unknown"},
|
|
|
|
{0x69, nullptr, "Unknown"},
|
|
|
|
{0x6A, nullptr, "Unknown"},
|
|
|
|
{0x6B, nullptr, "Unknown"},
|
|
|
|
{0x6C, nullptr, "Unknown"},
|
|
|
|
{0x6D, nullptr, "Unknown"},
|
|
|
|
{0x6E, nullptr, "Unknown"},
|
|
|
|
{0x6f, nullptr, "GetSystemInfo32"},
|
|
|
|
{0x70, nullptr, "CreatePort32"},
|
|
|
|
{0x71, nullptr, "ManageNamedPort32"},
|
|
|
|
{0x72, nullptr, "ConnectToPort32"},
|
|
|
|
{0x73, nullptr, "SetProcessMemoryPermission32"},
|
|
|
|
{0x74, nullptr, "Unknown"},
|
|
|
|
{0x75, nullptr, "Unknown"},
|
|
|
|
{0x76, nullptr, "Unknown"},
|
|
|
|
{0x77, nullptr, "MapProcessCodeMemory32"},
|
|
|
|
{0x78, nullptr, "UnmapProcessCodeMemory32"},
|
|
|
|
{0x79, nullptr, "Unknown"},
|
|
|
|
{0x7A, nullptr, "Unknown"},
|
|
|
|
{0x7B, nullptr, "TerminateProcess32"},
|
|
|
|
};
|
|
|
|
|
|
|
|
static const FunctionDef SVC_Table_64[] = {
|
|
|
|
{0x00, nullptr, "Unknown"},
|
|
|
|
{0x01, SvcWrap64<SetHeapSize>, "SetHeapSize"},
|
2020-04-09 03:14:18 +00:00
|
|
|
{0x02, nullptr, "SetMemoryPermission"},
|
2020-03-02 05:06:41 +00:00
|
|
|
{0x03, SvcWrap64<SetMemoryAttribute>, "SetMemoryAttribute"},
|
|
|
|
{0x04, SvcWrap64<MapMemory>, "MapMemory"},
|
|
|
|
{0x05, SvcWrap64<UnmapMemory>, "UnmapMemory"},
|
|
|
|
{0x06, SvcWrap64<QueryMemory>, "QueryMemory"},
|
|
|
|
{0x07, SvcWrap64<ExitProcess>, "ExitProcess"},
|
|
|
|
{0x08, SvcWrap64<CreateThread>, "CreateThread"},
|
|
|
|
{0x09, SvcWrap64<StartThread>, "StartThread"},
|
|
|
|
{0x0A, SvcWrap64<ExitThread>, "ExitThread"},
|
|
|
|
{0x0B, SvcWrap64<SleepThread>, "SleepThread"},
|
|
|
|
{0x0C, SvcWrap64<GetThreadPriority>, "GetThreadPriority"},
|
|
|
|
{0x0D, SvcWrap64<SetThreadPriority>, "SetThreadPriority"},
|
|
|
|
{0x0E, SvcWrap64<GetThreadCoreMask>, "GetThreadCoreMask"},
|
|
|
|
{0x0F, SvcWrap64<SetThreadCoreMask>, "SetThreadCoreMask"},
|
|
|
|
{0x10, SvcWrap64<GetCurrentProcessorNumber>, "GetCurrentProcessorNumber"},
|
|
|
|
{0x11, SvcWrap64<SignalEvent>, "SignalEvent"},
|
|
|
|
{0x12, SvcWrap64<ClearEvent>, "ClearEvent"},
|
|
|
|
{0x13, SvcWrap64<MapSharedMemory>, "MapSharedMemory"},
|
2020-04-09 03:14:18 +00:00
|
|
|
{0x14, nullptr, "UnmapSharedMemory"},
|
2020-03-02 05:06:41 +00:00
|
|
|
{0x15, SvcWrap64<CreateTransferMemory>, "CreateTransferMemory"},
|
|
|
|
{0x16, SvcWrap64<CloseHandle>, "CloseHandle"},
|
|
|
|
{0x17, SvcWrap64<ResetSignal>, "ResetSignal"},
|
|
|
|
{0x18, SvcWrap64<WaitSynchronization>, "WaitSynchronization"},
|
|
|
|
{0x19, SvcWrap64<CancelSynchronization>, "CancelSynchronization"},
|
|
|
|
{0x1A, SvcWrap64<ArbitrateLock>, "ArbitrateLock"},
|
|
|
|
{0x1B, SvcWrap64<ArbitrateUnlock>, "ArbitrateUnlock"},
|
|
|
|
{0x1C, SvcWrap64<WaitProcessWideKeyAtomic>, "WaitProcessWideKeyAtomic"},
|
|
|
|
{0x1D, SvcWrap64<SignalProcessWideKey>, "SignalProcessWideKey"},
|
|
|
|
{0x1E, SvcWrap64<GetSystemTick>, "GetSystemTick"},
|
|
|
|
{0x1F, SvcWrap64<ConnectToNamedPort>, "ConnectToNamedPort"},
|
2018-01-03 01:47:26 +00:00
|
|
|
{0x20, nullptr, "SendSyncRequestLight"},
|
2020-03-02 05:06:41 +00:00
|
|
|
{0x21, SvcWrap64<SendSyncRequest>, "SendSyncRequest"},
|
2018-01-03 01:47:26 +00:00
|
|
|
{0x22, nullptr, "SendSyncRequestWithUserBuffer"},
|
|
|
|
{0x23, nullptr, "SendAsyncRequestWithUserBuffer"},
|
2020-03-02 05:06:41 +00:00
|
|
|
{0x24, SvcWrap64<GetProcessId>, "GetProcessId"},
|
|
|
|
{0x25, SvcWrap64<GetThreadId>, "GetThreadId"},
|
|
|
|
{0x26, SvcWrap64<Break>, "Break"},
|
|
|
|
{0x27, SvcWrap64<OutputDebugString>, "OutputDebugString"},
|
2018-01-03 01:47:26 +00:00
|
|
|
{0x28, nullptr, "ReturnFromException"},
|
2020-03-02 05:06:41 +00:00
|
|
|
{0x29, SvcWrap64<GetInfo>, "GetInfo"},
|
2018-01-03 01:47:26 +00:00
|
|
|
{0x2A, nullptr, "FlushEntireDataCache"},
|
|
|
|
{0x2B, nullptr, "FlushDataCache"},
|
2020-03-02 05:06:41 +00:00
|
|
|
{0x2C, SvcWrap64<MapPhysicalMemory>, "MapPhysicalMemory"},
|
|
|
|
{0x2D, SvcWrap64<UnmapPhysicalMemory>, "UnmapPhysicalMemory"},
|
2018-09-24 00:03:38 +00:00
|
|
|
{0x2E, nullptr, "GetFutureThreadInfo"},
|
2018-01-03 01:47:26 +00:00
|
|
|
{0x2F, nullptr, "GetLastThreadInfo"},
|
2020-03-02 05:06:41 +00:00
|
|
|
{0x30, SvcWrap64<GetResourceLimitLimitValue>, "GetResourceLimitLimitValue"},
|
|
|
|
{0x31, SvcWrap64<GetResourceLimitCurrentValue>, "GetResourceLimitCurrentValue"},
|
|
|
|
{0x32, SvcWrap64<SetThreadActivity>, "SetThreadActivity"},
|
|
|
|
{0x33, SvcWrap64<GetThreadContext>, "GetThreadContext"},
|
|
|
|
{0x34, SvcWrap64<WaitForAddress>, "WaitForAddress"},
|
|
|
|
{0x35, SvcWrap64<SignalToAddress>, "SignalToAddress"},
|
2019-04-19 18:33:01 +00:00
|
|
|
{0x36, nullptr, "SynchronizePreemptionState"},
|
2017-10-14 21:30:07 +00:00
|
|
|
{0x37, nullptr, "Unknown"},
|
|
|
|
{0x38, nullptr, "Unknown"},
|
|
|
|
{0x39, nullptr, "Unknown"},
|
|
|
|
{0x3A, nullptr, "Unknown"},
|
|
|
|
{0x3B, nullptr, "Unknown"},
|
2020-03-02 05:06:41 +00:00
|
|
|
{0x3C, SvcWrap64<KernelDebug>, "KernelDebug"},
|
|
|
|
{0x3D, SvcWrap64<ChangeKernelTraceState>, "ChangeKernelTraceState"},
|
2017-10-14 21:30:07 +00:00
|
|
|
{0x3E, nullptr, "Unknown"},
|
2016-09-18 00:38:01 +00:00
|
|
|
{0x3F, nullptr, "Unknown"},
|
2018-01-03 01:47:26 +00:00
|
|
|
{0x40, nullptr, "CreateSession"},
|
|
|
|
{0x41, nullptr, "AcceptSession"},
|
|
|
|
{0x42, nullptr, "ReplyAndReceiveLight"},
|
|
|
|
{0x43, nullptr, "ReplyAndReceive"},
|
|
|
|
{0x44, nullptr, "ReplyAndReceiveWithUserBuffer"},
|
2020-03-02 05:06:41 +00:00
|
|
|
{0x45, SvcWrap64<CreateEvent>, "CreateEvent"},
|
2016-09-18 00:38:01 +00:00
|
|
|
{0x46, nullptr, "Unknown"},
|
2017-10-14 21:30:07 +00:00
|
|
|
{0x47, nullptr, "Unknown"},
|
2018-09-24 00:03:38 +00:00
|
|
|
{0x48, nullptr, "MapPhysicalMemoryUnsafe"},
|
|
|
|
{0x49, nullptr, "UnmapPhysicalMemoryUnsafe"},
|
|
|
|
{0x4A, nullptr, "SetUnsafeLimit"},
|
|
|
|
{0x4B, nullptr, "CreateCodeMemory"},
|
|
|
|
{0x4C, nullptr, "ControlCodeMemory"},
|
2018-01-03 01:47:26 +00:00
|
|
|
{0x4D, nullptr, "SleepSystem"},
|
|
|
|
{0x4E, nullptr, "ReadWriteRegister"},
|
|
|
|
{0x4F, nullptr, "SetProcessActivity"},
|
2020-04-09 03:14:18 +00:00
|
|
|
{0x50, nullptr, "CreateSharedMemory"},
|
|
|
|
{0x51, nullptr, "MapTransferMemory"},
|
|
|
|
{0x52, nullptr, "UnmapTransferMemory"},
|
2018-01-03 01:47:26 +00:00
|
|
|
{0x53, nullptr, "CreateInterruptEvent"},
|
|
|
|
{0x54, nullptr, "QueryPhysicalAddress"},
|
|
|
|
{0x55, nullptr, "QueryIoMapping"},
|
|
|
|
{0x56, nullptr, "CreateDeviceAddressSpace"},
|
|
|
|
{0x57, nullptr, "AttachDeviceAddressSpace"},
|
|
|
|
{0x58, nullptr, "DetachDeviceAddressSpace"},
|
|
|
|
{0x59, nullptr, "MapDeviceAddressSpaceByForce"},
|
|
|
|
{0x5A, nullptr, "MapDeviceAddressSpaceAligned"},
|
|
|
|
{0x5B, nullptr, "MapDeviceAddressSpace"},
|
|
|
|
{0x5C, nullptr, "UnmapDeviceAddressSpace"},
|
|
|
|
{0x5D, nullptr, "InvalidateProcessDataCache"},
|
|
|
|
{0x5E, nullptr, "StoreProcessDataCache"},
|
|
|
|
{0x5F, nullptr, "FlushProcessDataCache"},
|
|
|
|
{0x60, nullptr, "DebugActiveProcess"},
|
|
|
|
{0x61, nullptr, "BreakDebugProcess"},
|
|
|
|
{0x62, nullptr, "TerminateDebugProcess"},
|
|
|
|
{0x63, nullptr, "GetDebugEvent"},
|
|
|
|
{0x64, nullptr, "ContinueDebugEvent"},
|
2020-03-02 05:06:41 +00:00
|
|
|
{0x65, SvcWrap64<GetProcessList>, "GetProcessList"},
|
|
|
|
{0x66, SvcWrap64<GetThreadList>, "GetThreadList"},
|
2018-01-03 01:47:26 +00:00
|
|
|
{0x67, nullptr, "GetDebugThreadContext"},
|
|
|
|
{0x68, nullptr, "SetDebugThreadContext"},
|
|
|
|
{0x69, nullptr, "QueryDebugProcessMemory"},
|
|
|
|
{0x6A, nullptr, "ReadDebugProcessMemory"},
|
|
|
|
{0x6B, nullptr, "WriteDebugProcessMemory"},
|
|
|
|
{0x6C, nullptr, "SetHardwareBreakPoint"},
|
|
|
|
{0x6D, nullptr, "GetDebugThreadParam"},
|
2016-09-18 00:38:01 +00:00
|
|
|
{0x6E, nullptr, "Unknown"},
|
2018-09-24 00:03:38 +00:00
|
|
|
{0x6F, nullptr, "GetSystemInfo"},
|
2018-01-03 01:47:26 +00:00
|
|
|
{0x70, nullptr, "CreatePort"},
|
|
|
|
{0x71, nullptr, "ManageNamedPort"},
|
|
|
|
{0x72, nullptr, "ConnectToPort"},
|
|
|
|
{0x73, nullptr, "SetProcessMemoryPermission"},
|
|
|
|
{0x74, nullptr, "MapProcessMemory"},
|
|
|
|
{0x75, nullptr, "UnmapProcessMemory"},
|
2020-03-02 05:06:41 +00:00
|
|
|
{0x76, SvcWrap64<QueryProcessMemory>, "QueryProcessMemory"},
|
2020-04-23 22:05:09 +00:00
|
|
|
{0x77, SvcWrap64<MapProcessCodeMemory>, "MapProcessCodeMemory"},
|
|
|
|
{0x78, SvcWrap64<UnmapProcessCodeMemory>, "UnmapProcessCodeMemory"},
|
2018-01-03 01:47:26 +00:00
|
|
|
{0x79, nullptr, "CreateProcess"},
|
|
|
|
{0x7A, nullptr, "StartProcess"},
|
|
|
|
{0x7B, nullptr, "TerminateProcess"},
|
2020-03-02 05:06:41 +00:00
|
|
|
{0x7C, SvcWrap64<GetProcessInfo>, "GetProcessInfo"},
|
|
|
|
{0x7D, SvcWrap64<CreateResourceLimit>, "CreateResourceLimit"},
|
|
|
|
{0x7E, SvcWrap64<SetResourceLimitLimitValue>, "SetResourceLimitLimitValue"},
|
2018-01-03 01:47:26 +00:00
|
|
|
{0x7F, nullptr, "CallSecureMonitor"},
|
2014-04-10 23:58:28 +00:00
|
|
|
};
|
|
|
|
|
2020-03-02 05:06:41 +00:00
|
|
|
static const FunctionDef* GetSVCInfo32(u32 func_num) {
|
|
|
|
if (func_num >= std::size(SVC_Table_32)) {
|
|
|
|
LOG_ERROR(Kernel_SVC, "Unknown svc=0x{:02X}", func_num);
|
|
|
|
return nullptr;
|
|
|
|
}
|
|
|
|
return &SVC_Table_32[func_num];
|
|
|
|
}
|
|
|
|
|
|
|
|
static const FunctionDef* GetSVCInfo64(u32 func_num) {
|
|
|
|
if (func_num >= std::size(SVC_Table_64)) {
|
2018-07-02 16:13:26 +00:00
|
|
|
LOG_ERROR(Kernel_SVC, "Unknown svc=0x{:02X}", func_num);
|
2015-05-06 03:04:25 +00:00
|
|
|
return nullptr;
|
|
|
|
}
|
2020-03-02 05:06:41 +00:00
|
|
|
return &SVC_Table_64[func_num];
|
2015-05-06 03:04:25 +00:00
|
|
|
}
|
|
|
|
|
2020-03-27 00:00:30 +00:00
|
|
|
void Call(Core::System& system, u32 immediate) {
|
2020-03-12 20:48:43 +00:00
|
|
|
system.ExitDynarmicProfile();
|
|
|
|
auto& kernel = system.Kernel();
|
|
|
|
kernel.EnterSVCProfile();
|
2015-05-06 03:04:25 +00:00
|
|
|
|
2020-12-03 02:08:35 +00:00
|
|
|
auto* thread = kernel.CurrentScheduler()->GetCurrentThread();
|
2020-11-13 19:11:12 +00:00
|
|
|
thread->SetContinuousOnSVC(true);
|
|
|
|
|
2020-03-02 05:06:41 +00:00
|
|
|
const FunctionDef* info = system.CurrentProcess()->Is64BitProcess() ? GetSVCInfo64(immediate)
|
|
|
|
: GetSVCInfo32(immediate);
|
2015-05-06 03:04:25 +00:00
|
|
|
if (info) {
|
|
|
|
if (info->func) {
|
2019-04-06 22:46:18 +00:00
|
|
|
info->func(system);
|
2015-05-06 03:04:25 +00:00
|
|
|
} else {
|
2018-07-02 16:13:26 +00:00
|
|
|
LOG_CRITICAL(Kernel_SVC, "Unimplemented SVC function {}(..)", info->name);
|
2015-05-06 03:04:25 +00:00
|
|
|
}
|
2017-10-14 21:30:07 +00:00
|
|
|
} else {
|
2018-07-02 16:13:26 +00:00
|
|
|
LOG_CRITICAL(Kernel_SVC, "Unknown SVC function 0x{:X}", immediate);
|
2015-05-06 03:04:25 +00:00
|
|
|
}
|
2020-03-10 22:41:11 +00:00
|
|
|
|
2020-03-12 20:48:43 +00:00
|
|
|
kernel.ExitSVCProfile();
|
2020-11-13 19:11:12 +00:00
|
|
|
|
|
|
|
if (!thread->IsContinuousOnSVC()) {
|
|
|
|
auto* host_context = thread->GetHostContext().get();
|
|
|
|
host_context->Rewind();
|
|
|
|
}
|
|
|
|
|
2020-03-12 20:48:43 +00:00
|
|
|
system.EnterDynarmicProfile();
|
2014-04-10 23:58:28 +00:00
|
|
|
}
|
2014-04-11 22:44:21 +00:00
|
|
|
|
2020-03-27 00:00:30 +00:00
|
|
|
} // namespace Kernel::Svc
|