summaryrefslogtreecommitdiff
path: root/src/core/hle/kernel/process.cpp
diff options
context:
space:
mode:
authorGravatar Yuri Kunde Schlesner2015-07-17 23:19:16 -0300
committerGravatar Yuri Kunde Schlesner2015-08-16 01:03:45 -0300
commitcdeeecf0807d0005356f30db0f7164c5891a9245 (patch)
treee32e6d673cdac358df0abd3d3ece13f37c1c28d5 /src/core/hle/kernel/process.cpp
parentMemory: Move PAGE_MASK and PAGE_BITS to memory.h (diff)
downloadyuzu-cdeeecf0807d0005356f30db0f7164c5891a9245.tar.gz
yuzu-cdeeecf0807d0005356f30db0f7164c5891a9245.tar.xz
yuzu-cdeeecf0807d0005356f30db0f7164c5891a9245.zip
Kernel: Properly implement ControlMemory FREE and COMMIT
Diffstat (limited to 'src/core/hle/kernel/process.cpp')
-rw-r--r--src/core/hle/kernel/process.cpp120
1 files changed, 115 insertions, 5 deletions
diff --git a/src/core/hle/kernel/process.cpp b/src/core/hle/kernel/process.cpp
index ad953cdbf..1db763999 100644
--- a/src/core/hle/kernel/process.cpp
+++ b/src/core/hle/kernel/process.cpp
@@ -36,8 +36,7 @@ SharedPtr<Process> Process::Create(SharedPtr<CodeSet> code_set) {
36 process->codeset = std::move(code_set); 36 process->codeset = std::move(code_set);
37 process->flags.raw = 0; 37 process->flags.raw = 0;
38 process->flags.memory_region = MemoryRegion::APPLICATION; 38 process->flags.memory_region = MemoryRegion::APPLICATION;
39 process->address_space = Common::make_unique<VMManager>(); 39 Memory::InitLegacyAddressSpace(process->vm_manager);
40 Memory::InitLegacyAddressSpace(*process->address_space);
41 40
42 return process; 41 return process;
43} 42}
@@ -104,19 +103,130 @@ void Process::ParseKernelCaps(const u32* kernel_caps, size_t len) {
104 103
105void Process::Run(s32 main_thread_priority, u32 stack_size) { 104void Process::Run(s32 main_thread_priority, u32 stack_size) {
106 auto MapSegment = [&](CodeSet::Segment& segment, VMAPermission permissions, MemoryState memory_state) { 105 auto MapSegment = [&](CodeSet::Segment& segment, VMAPermission permissions, MemoryState memory_state) {
107 auto vma = address_space->MapMemoryBlock(segment.addr, codeset->memory, 106 auto vma = vm_manager.MapMemoryBlock(segment.addr, codeset->memory,
108 segment.offset, segment.size, memory_state).Unwrap(); 107 segment.offset, segment.size, memory_state).Unwrap();
109 address_space->Reprotect(vma, permissions); 108 vm_manager.Reprotect(vma, permissions);
110 }; 109 };
111 110
111 // Map CodeSet segments
112 MapSegment(codeset->code, VMAPermission::ReadExecute, MemoryState::Code); 112 MapSegment(codeset->code, VMAPermission::ReadExecute, MemoryState::Code);
113 MapSegment(codeset->rodata, VMAPermission::Read, MemoryState::Code); 113 MapSegment(codeset->rodata, VMAPermission::Read, MemoryState::Code);
114 MapSegment(codeset->data, VMAPermission::ReadWrite, MemoryState::Private); 114 MapSegment(codeset->data, VMAPermission::ReadWrite, MemoryState::Private);
115 115
116 address_space->LogLayout(Log::Level::Debug); 116 // Allocate and map stack
117 vm_manager.MapMemoryBlock(Memory::HEAP_VADDR_END - stack_size,
118 std::make_shared<std::vector<u8>>(stack_size, 0), 0, stack_size, MemoryState::Locked
119 ).Unwrap();
120
121 vm_manager.LogLayout(Log::Level::Debug);
117 Kernel::SetupMainThread(codeset->entrypoint, main_thread_priority); 122 Kernel::SetupMainThread(codeset->entrypoint, main_thread_priority);
118} 123}
119 124
125ResultVal<VAddr> Process::HeapAllocate(VAddr target, u32 size, VMAPermission perms) {
126 if (target < Memory::HEAP_VADDR || target + size > Memory::HEAP_VADDR_END || target + size < target) {
127 return ERR_INVALID_ADDRESS;
128 }
129
130 if (heap_memory == nullptr) {
131 // Initialize heap
132 heap_memory = std::make_shared<std::vector<u8>>();
133 heap_start = heap_end = target;
134 }
135
136 // If necessary, expand backing vector to cover new heap extents.
137 if (target < heap_start) {
138 heap_memory->insert(begin(*heap_memory), heap_start - target, 0);
139 heap_start = target;
140 vm_manager.RefreshMemoryBlockMappings(heap_memory.get());
141 }
142 if (target + size > heap_end) {
143 heap_memory->insert(end(*heap_memory), (target + size) - heap_end, 0);
144 heap_end = target + size;
145 vm_manager.RefreshMemoryBlockMappings(heap_memory.get());
146 }
147 ASSERT(heap_end - heap_start == heap_memory->size());
148
149 CASCADE_RESULT(auto vma, vm_manager.MapMemoryBlock(target, heap_memory, target - heap_start, size, MemoryState::Private));
150 vm_manager.Reprotect(vma, perms);
151
152 return MakeResult<VAddr>(heap_end - size);
153}
154
155ResultCode Process::HeapFree(VAddr target, u32 size) {
156 if (target < Memory::HEAP_VADDR || target + size > Memory::HEAP_VADDR_END || target + size < target) {
157 return ERR_INVALID_ADDRESS;
158 }
159
160 ResultCode result = vm_manager.UnmapRange(target, size);
161 if (result.IsError()) return result;
162
163 return RESULT_SUCCESS;
164}
165
166ResultVal<VAddr> Process::LinearAllocate(VAddr target, u32 size, VMAPermission perms) {
167 if (linear_heap_memory == nullptr) {
168 // Initialize heap
169 linear_heap_memory = std::make_shared<std::vector<u8>>();
170 }
171
172 VAddr heap_end = Memory::LINEAR_HEAP_VADDR + (u32)linear_heap_memory->size();
173 // Games and homebrew only ever seem to pass 0 here (which lets the kernel decide the address),
174 // but explicit addresses are also accepted and respected.
175 if (target == 0) {
176 target = heap_end;
177 }
178
179 if (target < Memory::LINEAR_HEAP_VADDR || target + size > Memory::LINEAR_HEAP_VADDR_END ||
180 target > heap_end || target + size < target) {
181
182 return ERR_INVALID_ADDRESS;
183 }
184
185 // Expansion of the linear heap is only allowed if you do an allocation immediatelly at its
186 // end. It's possible to free gaps in the middle of the heap and then reallocate them later,
187 // but expansions are only allowed at the end.
188 if (target == heap_end) {
189 linear_heap_memory->insert(linear_heap_memory->end(), size, 0);
190 vm_manager.RefreshMemoryBlockMappings(linear_heap_memory.get());
191 }
192
193 size_t offset = target - Memory::LINEAR_HEAP_VADDR;
194 CASCADE_RESULT(auto vma, vm_manager.MapMemoryBlock(target, linear_heap_memory, offset, size, MemoryState::Continuous));
195 vm_manager.Reprotect(vma, perms);
196
197 return MakeResult<VAddr>(target);
198}
199
200ResultCode Process::LinearFree(VAddr target, u32 size) {
201 if (linear_heap_memory == nullptr || target < Memory::LINEAR_HEAP_VADDR ||
202 target + size > Memory::LINEAR_HEAP_VADDR_END || target + size < target) {
203
204 return ERR_INVALID_ADDRESS;
205 }
206
207 VAddr heap_end = Memory::LINEAR_HEAP_VADDR + (u32)linear_heap_memory->size();
208 if (target + size > heap_end) {
209 return ERR_INVALID_ADDRESS_STATE;
210 }
211
212 ResultCode result = vm_manager.UnmapRange(target, size);
213 if (result.IsError()) return result;
214
215 if (target + size == heap_end) {
216 // End of linear heap has been freed, so check what's the last allocated block in it and
217 // reduce the size.
218 auto vma = vm_manager.FindVMA(target);
219 ASSERT(vma != vm_manager.vma_map.end());
220 ASSERT(vma->second.type == VMAType::Free);
221 VAddr new_end = vma->second.base;
222 if (new_end >= Memory::LINEAR_HEAP_VADDR) {
223 linear_heap_memory->resize(new_end - Memory::LINEAR_HEAP_VADDR);
224 }
225 }
226
227 return RESULT_SUCCESS;
228}
229
120Kernel::Process::Process() {} 230Kernel::Process::Process() {}
121Kernel::Process::~Process() {} 231Kernel::Process::~Process() {}
122 232