rpcs3/rpcs3/util/vm.hpp
Nekotekina dba2baba9c Implement utils::memory_map_fd (partial)
Improve JIT profiling dump format (data + name, mmap)
Improve objdump interception util (better speed, fix bugs)
Rename spu_ubertrampoline to __ub+number
2022-01-26 15:46:16 +03:00

120 lines
2.8 KiB
C++

#pragma once
#include "util/types.hpp"
#include "util/atomic.hpp"
namespace utils
{
#ifdef _WIN32
using native_handle = void*;
#else
using native_handle = int;
#endif
// Memory protection type
enum class protection
{
rw, // Read + write (default)
ro, // Read only
no, // No access
wx, // Read + write + execute
rx, // Read + execute
};
/**
* Reserve `size` bytes of virtual memory and returns it.
* The memory should be committed before usage.
*/
void* memory_reserve(usz size, void* use_addr = nullptr);
/**
* Commit `size` bytes of virtual memory starting at pointer.
* That is, bake reserved memory with physical memory.
* pointer should belong to a range of reserved memory.
*/
void memory_commit(void* pointer, usz size, protection prot = protection::rw);
// Decommit all memory committed via commit_page_memory.
void memory_decommit(void* pointer, usz size);
// Decommit all memory and commit it again.
void memory_reset(void* pointer, usz size, protection prot = protection::rw);
// Free memory after reserved by memory_reserve, should specify original size
void memory_release(void* pointer, usz size);
// Set memory protection
void memory_protect(void* pointer, usz size, protection prot);
// Lock pages in memory
bool memory_lock(void* pointer, usz size);
// Map file descriptor
void* memory_map_fd(native_handle fd, usz size, protection prot);
// Shared memory handle
class shm
{
#ifdef _WIN32
void* m_handle{};
#else
int m_file{};
#endif
u32 m_flags{};
u64 m_size{};
atomic_t<void*> m_ptr{nullptr};
public:
explicit shm(u32 size, u32 flags = 0);
// Construct with specified path as sparse file storage
shm(u64 size, const std::string& storage);
shm(const shm&) = delete;
shm& operator=(const shm&) = delete;
~shm();
// Map shared memory
u8* map(void* ptr, protection prot = protection::rw, bool cow = false) const;
// Attempt to map shared memory fix fixed pointer
u8* try_map(void* ptr, protection prot = protection::rw, bool cow = false) const;
// Map shared memory over reserved memory region, which is unsafe (non-atomic) under Win32
u8* map_critical(void* ptr, protection prot = protection::rw, bool cow = false);
// Map shared memory into its own storage (not mapped by default)
u8* map_self(protection prot = protection::rw);
// Unmap shared memory
void unmap(void* ptr) const;
// Unmap shared memory, undoing map_critical
void unmap_critical(void* ptr);
// Unmap shared memory, undoing map_self()
void unmap_self();
// Get memory mapped by map_self()
u8* get() const
{
return static_cast<u8*>(+m_ptr);
}
u64 size() const
{
return m_size;
}
// Flags are unspecified, consider it userdata
u32 flags() const
{
return m_flags;
}
// Another userdata
u64 info = 0;
};
}