diff options
Diffstat (limited to 'src/core/hle/kernel/process.cpp')
| -rw-r--r-- | src/core/hle/kernel/process.cpp | 120 |
1 files changed, 115 insertions, 5 deletions
diff --git a/src/core/hle/kernel/process.cpp b/src/core/hle/kernel/process.cpp index ad953cdbf..1db763999 100644 --- a/src/core/hle/kernel/process.cpp +++ b/src/core/hle/kernel/process.cpp | |||
| @@ -36,8 +36,7 @@ SharedPtr<Process> Process::Create(SharedPtr<CodeSet> code_set) { | |||
| 36 | process->codeset = std::move(code_set); | 36 | process->codeset = std::move(code_set); |
| 37 | process->flags.raw = 0; | 37 | process->flags.raw = 0; |
| 38 | process->flags.memory_region = MemoryRegion::APPLICATION; | 38 | process->flags.memory_region = MemoryRegion::APPLICATION; |
| 39 | process->address_space = Common::make_unique<VMManager>(); | 39 | Memory::InitLegacyAddressSpace(process->vm_manager); |
| 40 | Memory::InitLegacyAddressSpace(*process->address_space); | ||
| 41 | 40 | ||
| 42 | return process; | 41 | return process; |
| 43 | } | 42 | } |
| @@ -104,19 +103,130 @@ void Process::ParseKernelCaps(const u32* kernel_caps, size_t len) { | |||
| 104 | 103 | ||
| 105 | void Process::Run(s32 main_thread_priority, u32 stack_size) { | 104 | void Process::Run(s32 main_thread_priority, u32 stack_size) { |
| 106 | auto MapSegment = [&](CodeSet::Segment& segment, VMAPermission permissions, MemoryState memory_state) { | 105 | auto MapSegment = [&](CodeSet::Segment& segment, VMAPermission permissions, MemoryState memory_state) { |
| 107 | auto vma = address_space->MapMemoryBlock(segment.addr, codeset->memory, | 106 | auto vma = vm_manager.MapMemoryBlock(segment.addr, codeset->memory, |
| 108 | segment.offset, segment.size, memory_state).Unwrap(); | 107 | segment.offset, segment.size, memory_state).Unwrap(); |
| 109 | address_space->Reprotect(vma, permissions); | 108 | vm_manager.Reprotect(vma, permissions); |
| 110 | }; | 109 | }; |
| 111 | 110 | ||
| 111 | // Map CodeSet segments | ||
| 112 | MapSegment(codeset->code, VMAPermission::ReadExecute, MemoryState::Code); | 112 | MapSegment(codeset->code, VMAPermission::ReadExecute, MemoryState::Code); |
| 113 | MapSegment(codeset->rodata, VMAPermission::Read, MemoryState::Code); | 113 | MapSegment(codeset->rodata, VMAPermission::Read, MemoryState::Code); |
| 114 | MapSegment(codeset->data, VMAPermission::ReadWrite, MemoryState::Private); | 114 | MapSegment(codeset->data, VMAPermission::ReadWrite, MemoryState::Private); |
| 115 | 115 | ||
| 116 | address_space->LogLayout(Log::Level::Debug); | 116 | // Allocate and map stack |
| 117 | vm_manager.MapMemoryBlock(Memory::HEAP_VADDR_END - stack_size, | ||
| 118 | std::make_shared<std::vector<u8>>(stack_size, 0), 0, stack_size, MemoryState::Locked | ||
| 119 | ).Unwrap(); | ||
| 120 | |||
| 121 | vm_manager.LogLayout(Log::Level::Debug); | ||
| 117 | Kernel::SetupMainThread(codeset->entrypoint, main_thread_priority); | 122 | Kernel::SetupMainThread(codeset->entrypoint, main_thread_priority); |
| 118 | } | 123 | } |
| 119 | 124 | ||
| 125 | ResultVal<VAddr> Process::HeapAllocate(VAddr target, u32 size, VMAPermission perms) { | ||
| 126 | if (target < Memory::HEAP_VADDR || target + size > Memory::HEAP_VADDR_END || target + size < target) { | ||
| 127 | return ERR_INVALID_ADDRESS; | ||
| 128 | } | ||
| 129 | |||
| 130 | if (heap_memory == nullptr) { | ||
| 131 | // Initialize heap | ||
| 132 | heap_memory = std::make_shared<std::vector<u8>>(); | ||
| 133 | heap_start = heap_end = target; | ||
| 134 | } | ||
| 135 | |||
| 136 | // If necessary, expand backing vector to cover new heap extents. | ||
| 137 | if (target < heap_start) { | ||
| 138 | heap_memory->insert(begin(*heap_memory), heap_start - target, 0); | ||
| 139 | heap_start = target; | ||
| 140 | vm_manager.RefreshMemoryBlockMappings(heap_memory.get()); | ||
| 141 | } | ||
| 142 | if (target + size > heap_end) { | ||
| 143 | heap_memory->insert(end(*heap_memory), (target + size) - heap_end, 0); | ||
| 144 | heap_end = target + size; | ||
| 145 | vm_manager.RefreshMemoryBlockMappings(heap_memory.get()); | ||
| 146 | } | ||
| 147 | ASSERT(heap_end - heap_start == heap_memory->size()); | ||
| 148 | |||
| 149 | CASCADE_RESULT(auto vma, vm_manager.MapMemoryBlock(target, heap_memory, target - heap_start, size, MemoryState::Private)); | ||
| 150 | vm_manager.Reprotect(vma, perms); | ||
| 151 | |||
| 152 | return MakeResult<VAddr>(heap_end - size); | ||
| 153 | } | ||
| 154 | |||
| 155 | ResultCode Process::HeapFree(VAddr target, u32 size) { | ||
| 156 | if (target < Memory::HEAP_VADDR || target + size > Memory::HEAP_VADDR_END || target + size < target) { | ||
| 157 | return ERR_INVALID_ADDRESS; | ||
| 158 | } | ||
| 159 | |||
| 160 | ResultCode result = vm_manager.UnmapRange(target, size); | ||
| 161 | if (result.IsError()) return result; | ||
| 162 | |||
| 163 | return RESULT_SUCCESS; | ||
| 164 | } | ||
| 165 | |||
| 166 | ResultVal<VAddr> Process::LinearAllocate(VAddr target, u32 size, VMAPermission perms) { | ||
| 167 | if (linear_heap_memory == nullptr) { | ||
| 168 | // Initialize heap | ||
| 169 | linear_heap_memory = std::make_shared<std::vector<u8>>(); | ||
| 170 | } | ||
| 171 | |||
| 172 | VAddr heap_end = Memory::LINEAR_HEAP_VADDR + (u32)linear_heap_memory->size(); | ||
| 173 | // Games and homebrew only ever seem to pass 0 here (which lets the kernel decide the address), | ||
| 174 | // but explicit addresses are also accepted and respected. | ||
| 175 | if (target == 0) { | ||
| 176 | target = heap_end; | ||
| 177 | } | ||
| 178 | |||
| 179 | if (target < Memory::LINEAR_HEAP_VADDR || target + size > Memory::LINEAR_HEAP_VADDR_END || | ||
| 180 | target > heap_end || target + size < target) { | ||
| 181 | |||
| 182 | return ERR_INVALID_ADDRESS; | ||
| 183 | } | ||
| 184 | |||
| 185 | // Expansion of the linear heap is only allowed if you do an allocation immediatelly at its | ||
| 186 | // end. It's possible to free gaps in the middle of the heap and then reallocate them later, | ||
| 187 | // but expansions are only allowed at the end. | ||
| 188 | if (target == heap_end) { | ||
| 189 | linear_heap_memory->insert(linear_heap_memory->end(), size, 0); | ||
| 190 | vm_manager.RefreshMemoryBlockMappings(linear_heap_memory.get()); | ||
| 191 | } | ||
| 192 | |||
| 193 | size_t offset = target - Memory::LINEAR_HEAP_VADDR; | ||
| 194 | CASCADE_RESULT(auto vma, vm_manager.MapMemoryBlock(target, linear_heap_memory, offset, size, MemoryState::Continuous)); | ||
| 195 | vm_manager.Reprotect(vma, perms); | ||
| 196 | |||
| 197 | return MakeResult<VAddr>(target); | ||
| 198 | } | ||
| 199 | |||
| 200 | ResultCode Process::LinearFree(VAddr target, u32 size) { | ||
| 201 | if (linear_heap_memory == nullptr || target < Memory::LINEAR_HEAP_VADDR || | ||
| 202 | target + size > Memory::LINEAR_HEAP_VADDR_END || target + size < target) { | ||
| 203 | |||
| 204 | return ERR_INVALID_ADDRESS; | ||
| 205 | } | ||
| 206 | |||
| 207 | VAddr heap_end = Memory::LINEAR_HEAP_VADDR + (u32)linear_heap_memory->size(); | ||
| 208 | if (target + size > heap_end) { | ||
| 209 | return ERR_INVALID_ADDRESS_STATE; | ||
| 210 | } | ||
| 211 | |||
| 212 | ResultCode result = vm_manager.UnmapRange(target, size); | ||
| 213 | if (result.IsError()) return result; | ||
| 214 | |||
| 215 | if (target + size == heap_end) { | ||
| 216 | // End of linear heap has been freed, so check what's the last allocated block in it and | ||
| 217 | // reduce the size. | ||
| 218 | auto vma = vm_manager.FindVMA(target); | ||
| 219 | ASSERT(vma != vm_manager.vma_map.end()); | ||
| 220 | ASSERT(vma->second.type == VMAType::Free); | ||
| 221 | VAddr new_end = vma->second.base; | ||
| 222 | if (new_end >= Memory::LINEAR_HEAP_VADDR) { | ||
| 223 | linear_heap_memory->resize(new_end - Memory::LINEAR_HEAP_VADDR); | ||
| 224 | } | ||
| 225 | } | ||
| 226 | |||
| 227 | return RESULT_SUCCESS; | ||
| 228 | } | ||
| 229 | |||
| 120 | Kernel::Process::Process() {} | 230 | Kernel::Process::Process() {} |
| 121 | Kernel::Process::~Process() {} | 231 | Kernel::Process::~Process() {} |
| 122 | 232 | ||