mirror of https://github.com/red-prig/fpPS4.git
957 lines
21 KiB
Plaintext
957 lines
21 KiB
Plaintext
unit dmem_map;
|
|
|
|
{$mode ObjFPC}{$H+}
|
|
|
|
interface
|
|
|
|
uses
|
|
vmparam,
|
|
kern_mtx;
|
|
|
|
Const
|
|
SCE_KERNEL_MAIN_DMEM_SIZE=$180000000; //6GB
|
|
|
|
SCE_KERNEL_WB_ONION = 0;
|
|
SCE_KERNEL_WC_GARLIC = 3;
|
|
SCE_KERNEL_WB_GARLIC =10;
|
|
|
|
//deprecated
|
|
SCE_KERNEL_WB_ONION_NONVOLATILE = 1;
|
|
SCE_KERNEL_WC_GARLIC_VOLATILE = 2;
|
|
SCE_KERNEL_WC_GARLIC_NONVOLATILE = 3;
|
|
SCE_KERNEL_WT_ONION_VOLATILE = 4;
|
|
SCE_KERNEL_WT_ONION_NONVOLATILE = 5;
|
|
SCE_KERNEL_WP_ONION_VOLATILE = 6;
|
|
SCE_KERNEL_WP_ONION_NONVOLATILE = 7;
|
|
SCE_KERNEL_UC_GARLIC_VOLATILE = 8;
|
|
SCE_KERNEL_UC_GARLIC_NONVOLATILE = 9;
|
|
|
|
type
|
|
pp_dmem_map_entry=^p_dmem_map_entry;
|
|
p_dmem_map_entry=^t_dmem_map_entry;
|
|
t_dmem_map_entry=packed record
|
|
prev :p_dmem_map_entry; // previous entry
|
|
next :p_dmem_map_entry; // next entry
|
|
left :p_dmem_map_entry; // left child in binary search tree
|
|
right :p_dmem_map_entry; // right child in binary search tree
|
|
start :DWORD; // start address
|
|
__end :DWORD; // end address
|
|
avail_ssize :DWORD; // amt can grow if this is a stack
|
|
adj_free :DWORD; // amount of adjacent free space
|
|
max_free :DWORD; // max free space in subtree
|
|
m_type :DWORD; // memory type
|
|
end;
|
|
|
|
p_dmem_map=^t_dmem_map;
|
|
t_dmem_map=packed object
|
|
header :t_dmem_map_entry; // List of entries
|
|
lock :mtx; // Lock for map data
|
|
nentries:DWORD; // Number of entries
|
|
size :DWORD; // size
|
|
root :p_dmem_map_entry; // Root of a binary search tree
|
|
property min_offset:DWORD read header.start write header.start;
|
|
property max_offset:DWORD read header.__end write header.__end;
|
|
end;
|
|
|
|
procedure dmem_map_entry_deallocate(entry:p_dmem_map_entry);
|
|
|
|
procedure dmem_map_lock(map:p_dmem_map);
|
|
function dmem_map_trylock(map:p_dmem_map):Boolean;
|
|
procedure dmem_map_unlock(map:p_dmem_map);
|
|
function dmem_map_locked(map:p_dmem_map):Boolean; inline;
|
|
|
|
procedure dmem_map_init(map:p_dmem_map;min,max:QWORD);
|
|
|
|
procedure dmem_map_entry_dispose(map:p_dmem_map;entry:p_dmem_map_entry); inline;
|
|
function dmem_map_entry_create(map:p_dmem_map):p_dmem_map_entry;
|
|
|
|
function dmem_map_lookup_entry(
|
|
map :p_dmem_map;
|
|
address :DWORD;
|
|
entry :pp_dmem_map_entry):Boolean;
|
|
|
|
function dmem_map_insert(
|
|
map :p_dmem_map;
|
|
start :DWORD;
|
|
__end :DWORD;
|
|
m_type :DWORD):Integer;
|
|
|
|
Function dmem_map_query_available(map:p_dmem_map;start,__end,align:QWORD;var oaddr,osize:QWORD):Integer;
|
|
|
|
function dmem_map_findspace(map :p_dmem_map;
|
|
start :DWORD;
|
|
length:DWORD;
|
|
addr :PDWORD):Integer;
|
|
|
|
function dmem_map_fixed(map :p_dmem_map;
|
|
start :DWORD;
|
|
length :DWORD;
|
|
m_type :DWORD;
|
|
overwr :Integer):Integer;
|
|
|
|
procedure dmem_map_simplify_entry(map:p_dmem_map;entry:p_dmem_map_entry);
|
|
|
|
procedure dmem_map_entry_delete(map:p_dmem_map;entry:p_dmem_map_entry);
|
|
|
|
function dmem_map_delete(map:p_dmem_map;start:DWORD;__end:DWORD):Integer;
|
|
|
|
implementation
|
|
|
|
uses
|
|
errno,
|
|
kern_thr;
|
|
|
|
function IDX_TO_OFF(x:DWORD):QWORD; inline;
|
|
begin
|
|
Result:=QWORD(x) shl PAGE_SHIFT;
|
|
end;
|
|
|
|
function OFF_TO_IDX(x:QWORD):DWORD; inline;
|
|
begin
|
|
Result:=QWORD(x) shr PAGE_SHIFT;
|
|
end;
|
|
|
|
function IsPowerOfTwo(x:QWORD):Boolean; inline;
|
|
begin
|
|
Result:=(x and (x - 1))=0;
|
|
end;
|
|
|
|
function fastIntLog2(i:QWORD):QWORD; inline;
|
|
begin
|
|
Result:=BsfQWORD(i);
|
|
end;
|
|
|
|
function AlignUp(addr:PtrUInt;alignment:PtrUInt):PtrUInt; inline;
|
|
var
|
|
tmp:PtrUInt;
|
|
begin
|
|
if (alignment=0) then Exit(addr);
|
|
tmp:=addr+PtrUInt(alignment-1);
|
|
Result:=tmp-(tmp mod alignment)
|
|
end;
|
|
|
|
function AlignDw(addr:PtrUInt;alignment:PtrUInt):PtrUInt; inline;
|
|
begin
|
|
Result:=addr-(addr mod alignment);
|
|
end;
|
|
|
|
procedure dmem_map_entry_deallocate(entry:p_dmem_map_entry);
|
|
begin
|
|
//if ((entry^.eflags and MAP_ENTRY_IS_SUB_MAP)=0) then
|
|
//begin
|
|
// vm_object_deallocate(entry^.vm_obj);
|
|
//end;
|
|
Freemem(entry);
|
|
end;
|
|
|
|
procedure DMEM_MAP_RANGE_CHECK(map:p_dmem_map;var start,__end:DWORD);
|
|
begin
|
|
if (start<map^.min_offset) then
|
|
begin
|
|
start:=map^.min_offset;
|
|
end;
|
|
if (__end>map^.max_offset) then
|
|
begin
|
|
__end:=map^.max_offset;
|
|
end;
|
|
if (start>__end) then
|
|
begin
|
|
start:=__end;
|
|
end;
|
|
end;
|
|
|
|
procedure dmem_map_lock(map:p_dmem_map);
|
|
begin
|
|
mtx_lock(map^.lock);
|
|
end;
|
|
|
|
function dmem_map_trylock(map:p_dmem_map):Boolean;
|
|
begin
|
|
Result:=mtx_trylock(map^.lock);
|
|
end;
|
|
|
|
procedure dmem_map_process_deferred;
|
|
var
|
|
td:p_kthread;
|
|
entry,next:p_dmem_map_entry;
|
|
begin
|
|
td:=curkthread;
|
|
if (td=nil) then Exit;
|
|
entry:=td^.td_map_def_user;
|
|
td^.td_map_def_user:=nil;
|
|
while (entry<>nil) do
|
|
begin
|
|
next:=entry^.next;
|
|
dmem_map_entry_deallocate(entry);
|
|
entry:=next;
|
|
end;
|
|
end;
|
|
|
|
procedure dmem_map_unlock(map:p_dmem_map);
|
|
begin
|
|
mtx_unlock(map^.lock);
|
|
dmem_map_process_deferred();
|
|
end;
|
|
|
|
function dmem_map_locked(map:p_dmem_map):Boolean; inline;
|
|
begin
|
|
Result:=mtx_owned(map^.lock);
|
|
end;
|
|
|
|
procedure DMEM_MAP_ASSERT_LOCKED(map:p_dmem_map); inline;
|
|
begin
|
|
Assert(dmem_map_locked(map));
|
|
end;
|
|
|
|
procedure _dmem_map_init(map:p_dmem_map;min,max:DWORD);
|
|
begin
|
|
map^.header.next:=@map^.header;
|
|
map^.header.prev:=@map^.header;
|
|
map^.min_offset :=min;
|
|
map^.max_offset :=max;
|
|
map^.header.adj_free:=(max-min);
|
|
map^.header.max_free:=(max-min);
|
|
map^.nentries:=0;
|
|
map^.size :=0;
|
|
map^.root:=nil;
|
|
end;
|
|
|
|
procedure dmem_map_init(map:p_dmem_map;min,max:QWORD);
|
|
begin
|
|
_dmem_map_init(map, OFF_TO_IDX(min), OFF_TO_IDX(max));
|
|
mtx_init(map^.lock,'dmem');
|
|
end;
|
|
|
|
procedure dmem_map_entry_dispose(map:p_dmem_map;entry:p_dmem_map_entry); inline;
|
|
begin
|
|
FreeMem(entry);
|
|
end;
|
|
|
|
function dmem_map_entry_create(map:p_dmem_map):p_dmem_map_entry;
|
|
var
|
|
new_entry:p_dmem_map_entry;
|
|
begin
|
|
new_entry:=AllocMem(SizeOf(t_dmem_map_entry));
|
|
Assert((new_entry<>nil),'dmem_map_entry_create: kernel resources exhausted');
|
|
Result:=new_entry;
|
|
end;
|
|
|
|
procedure dmem_map_entry_set_max_free(entry:p_dmem_map_entry);
|
|
begin
|
|
entry^.max_free:=entry^.adj_free;
|
|
if (entry^.left<>nil) then
|
|
if (entry^.left^.max_free>entry^.max_free) then
|
|
begin
|
|
entry^.max_free:=entry^.left^.max_free;
|
|
end;
|
|
if (entry^.right<>nil) then
|
|
if (entry^.right^.max_free>entry^.max_free) then
|
|
begin
|
|
entry^.max_free:=entry^.right^.max_free;
|
|
end;
|
|
end;
|
|
|
|
function dmem_map_entry_splay(addr:DWORD;root:p_dmem_map_entry):p_dmem_map_entry;
|
|
var
|
|
llist,rlist:p_dmem_map_entry;
|
|
ltree,rtree:p_dmem_map_entry;
|
|
y :p_dmem_map_entry;
|
|
begin
|
|
{ Special case of empty tree. }
|
|
if (root=nil) then Exit(root);
|
|
|
|
llist:=nil;
|
|
rlist:=nil;
|
|
repeat
|
|
{ root is never nil in here. }
|
|
if (addr<root^.start) then
|
|
begin
|
|
y:=root^.left;
|
|
if (y=nil) then break;
|
|
if (addr<y^.start) and (y^.left<>nil) then
|
|
begin
|
|
{ Rotate right and put y on rlist. }
|
|
root^.left:=y^.right;
|
|
y^.right:=root;
|
|
dmem_map_entry_set_max_free(root);
|
|
root:=y^.left;
|
|
y^.left:=rlist;
|
|
rlist:=y;
|
|
end else
|
|
begin
|
|
{ Put root on rlist. }
|
|
root^.left:=rlist;
|
|
rlist:=root;
|
|
root:=y;
|
|
end;
|
|
end else
|
|
if (addr>=root^.__end) then
|
|
begin
|
|
y:=root^.right;
|
|
if (y=nil) then break;
|
|
if (addr>=y^.__end) and (y^.right<>nil) then
|
|
begin
|
|
{ Rotate left and put y on llist. }
|
|
root^.right:=y^.left;
|
|
y^.left:=root;
|
|
dmem_map_entry_set_max_free(root);
|
|
root:=y^.right;
|
|
y^.right:=llist;
|
|
llist:=y;
|
|
end else
|
|
begin
|
|
{ Put root on llist. }
|
|
root^.right:=llist;
|
|
llist:=root;
|
|
root:=y;
|
|
end;
|
|
end else
|
|
begin
|
|
break;
|
|
end;
|
|
until false;
|
|
|
|
{
|
|
* Pass Two: Walk back up the two spines, flip the pointers
|
|
* and set max_free. The subtrees of the root go at the
|
|
* bottom of llist and rlist.
|
|
}
|
|
ltree:=root^.left;
|
|
while (llist<>nil) do
|
|
begin
|
|
y:=llist^.right;
|
|
llist^.right:=ltree;
|
|
dmem_map_entry_set_max_free(llist);
|
|
ltree:=llist;
|
|
llist:=y;
|
|
end;
|
|
rtree:=root^.right;
|
|
while (rlist<>nil) do
|
|
begin
|
|
y:=rlist^.left;
|
|
rlist^.left:=rtree;
|
|
dmem_map_entry_set_max_free(rlist);
|
|
rtree:=rlist;
|
|
rlist:=y;
|
|
end;
|
|
|
|
{
|
|
* Final assembly: add ltree and rtree as subtrees of root.
|
|
}
|
|
root^.left:=ltree;
|
|
root^.right:=rtree;
|
|
dmem_map_entry_set_max_free(root);
|
|
|
|
Result:=(root);
|
|
end;
|
|
|
|
procedure dmem_map_entry_link(
|
|
map :p_dmem_map;
|
|
after_where:p_dmem_map_entry;
|
|
entry :p_dmem_map_entry);
|
|
begin
|
|
DMEM_MAP_ASSERT_LOCKED(map);
|
|
|
|
Inc(map^.nentries);
|
|
entry^.prev:=after_where;
|
|
entry^.next:=after_where^.next;
|
|
entry^.next^.prev:=entry;
|
|
after_where^.next:=entry;
|
|
|
|
if (after_where<>@map^.header) then
|
|
begin
|
|
if (after_where<>map^.root) then
|
|
begin
|
|
dmem_map_entry_splay(after_where^.start, map^.root);
|
|
end;
|
|
entry^.right:=after_where^.right;
|
|
entry^.left:=after_where;
|
|
after_where^.right:=nil;
|
|
after_where^.adj_free:=entry^.start - after_where^.__end;
|
|
dmem_map_entry_set_max_free(after_where);
|
|
end else
|
|
begin
|
|
entry^.right:=map^.root;
|
|
entry^.left:=nil;
|
|
end;
|
|
if (entry^.next=@map^.header) then
|
|
begin
|
|
entry^.adj_free:=map^.max_offset-entry^.__end;
|
|
end else
|
|
begin
|
|
entry^.adj_free:=entry^.next^.start-entry^.__end;
|
|
end;
|
|
dmem_map_entry_set_max_free(entry);
|
|
map^.root:=entry;
|
|
end;
|
|
|
|
procedure dmem_map_entry_unlink(
|
|
map :p_dmem_map;
|
|
entry :p_dmem_map_entry);
|
|
var
|
|
next,prev,root:p_dmem_map_entry;
|
|
begin
|
|
DMEM_MAP_ASSERT_LOCKED(map);
|
|
|
|
if (entry<>map^.root) then
|
|
begin
|
|
dmem_map_entry_splay(entry^.start, map^.root);
|
|
end;
|
|
if (entry^.left=nil) then
|
|
begin
|
|
root:=entry^.right;
|
|
end else
|
|
begin
|
|
root:=dmem_map_entry_splay(entry^.start, entry^.left);
|
|
root^.right:=entry^.right;
|
|
if (root^.next=@map^.header) then
|
|
begin
|
|
root^.adj_free:=map^.max_offset-root^.__end;
|
|
end else
|
|
begin
|
|
root^.adj_free:=entry^.next^.start-root^.__end;
|
|
end;
|
|
dmem_map_entry_set_max_free(root);
|
|
end;
|
|
map^.root:=root;
|
|
|
|
prev:=entry^.prev;
|
|
next:=entry^.next;
|
|
next^.prev:=prev;
|
|
prev^.next:=next;
|
|
Dec(map^.nentries);
|
|
end;
|
|
|
|
procedure dmem_map_entry_resize_free(map:p_dmem_map;entry:p_dmem_map_entry);
|
|
begin
|
|
if (entry<>map^.root) then
|
|
begin
|
|
map^.root:=dmem_map_entry_splay(entry^.start, map^.root);
|
|
end;
|
|
|
|
if (entry^.next=@map^.header) then
|
|
begin
|
|
entry^.adj_free:=map^.max_offset-entry^.__end;
|
|
end else
|
|
begin
|
|
entry^.adj_free:=entry^.next^.start-entry^.__end;
|
|
end;
|
|
dmem_map_entry_set_max_free(entry);
|
|
end;
|
|
|
|
function dmem_map_lookup_entry(
|
|
map :p_dmem_map;
|
|
address :DWORD;
|
|
entry :pp_dmem_map_entry):Boolean;
|
|
var
|
|
cur:p_dmem_map_entry;
|
|
begin
|
|
DMEM_MAP_ASSERT_LOCKED(map);
|
|
|
|
{
|
|
* If the map is empty, then the map entry immediately preceding
|
|
* "address" is the map's header.
|
|
}
|
|
cur:=map^.root;
|
|
if (cur=nil) then
|
|
begin
|
|
entry^:=@map^.header;
|
|
end else
|
|
if (address>=cur^.start) and (cur^.__end>address) then
|
|
begin
|
|
entry^:=cur;
|
|
Exit(TRUE);
|
|
end else
|
|
begin
|
|
{
|
|
* Splay requires a write lock on the map. However, it only
|
|
* restructures the binary search tree; it does not otherwise
|
|
* change the map. Thus, the map's timestamp need not change
|
|
* on a temporary upgrade.
|
|
}
|
|
cur:=dmem_map_entry_splay(address,cur);
|
|
map^.root:=cur;
|
|
|
|
{
|
|
* If "address" is contained within a map entry, the new root
|
|
* is that map entry. Otherwise, the new root is a map entry
|
|
* immediately before or after "address".
|
|
}
|
|
if (address>=cur^.start) then
|
|
begin
|
|
entry^:=cur;
|
|
if (cur^.__end>address) then
|
|
begin
|
|
Exit(TRUE);
|
|
end;
|
|
end else
|
|
begin
|
|
entry^:=cur^.prev;
|
|
end;
|
|
end;
|
|
Result:=(FALSE);
|
|
end;
|
|
|
|
function dmem_map_insert(
|
|
map :p_dmem_map;
|
|
start :DWORD;
|
|
__end :DWORD;
|
|
m_type :DWORD):Integer;
|
|
var
|
|
new_entry :p_dmem_map_entry;
|
|
prev_entry:p_dmem_map_entry;
|
|
temp_entry:p_dmem_map_entry;
|
|
begin
|
|
DMEM_MAP_ASSERT_LOCKED(map);
|
|
|
|
{
|
|
* Check that the start and end points are not bogus.
|
|
}
|
|
if (start<map^.min_offset) or (__end>map^.max_offset) or (start>=__end) then
|
|
begin
|
|
Exit(EINVAL);
|
|
end;
|
|
|
|
{
|
|
* Find the entry prior to the proposed starting address; if it's part
|
|
* of an existing entry, this range is bogus.
|
|
}
|
|
if dmem_map_lookup_entry(map,start,@temp_entry) then
|
|
begin
|
|
Exit(EAGAIN);
|
|
end;
|
|
|
|
prev_entry:=temp_entry;
|
|
|
|
{
|
|
* Assert that the next entry doesn't overlap the end point.
|
|
}
|
|
if (prev_entry^.next<>@map^.header) and
|
|
(prev_entry^.next^.start<__end) then
|
|
begin
|
|
Exit(EAGAIN);
|
|
end;
|
|
|
|
if (prev_entry<>@map^.header) and
|
|
(prev_entry^.__end=start) then
|
|
begin
|
|
{
|
|
* We were able to extend the object. Determine if we
|
|
* can extend the previous map entry to include the
|
|
* new range as well.
|
|
}
|
|
if (prev_entry^.m_type=m_type) then
|
|
begin
|
|
map^.size:=map^.size+(__end - prev_entry^.__end);
|
|
prev_entry^.__end:=__end;
|
|
//change size
|
|
|
|
//pmap_enter_object
|
|
|
|
dmem_map_entry_resize_free(map, prev_entry);
|
|
dmem_map_simplify_entry(map, prev_entry);
|
|
Exit(0);
|
|
end;
|
|
|
|
end;
|
|
|
|
{
|
|
* NOTE: if conditionals fail, object can be nil here. This occurs
|
|
* in things like the buffer map where we manage kva but do not manage
|
|
* backing objects.
|
|
}
|
|
|
|
{
|
|
* Create a new entry
|
|
}
|
|
new_entry:=dmem_map_entry_create(map);
|
|
new_entry^.start:=start;
|
|
new_entry^.__end:=__end;
|
|
|
|
new_entry^.m_type:=m_type;
|
|
new_entry^.avail_ssize:=0;
|
|
|
|
{
|
|
* Insert the new entry into the list
|
|
}
|
|
dmem_map_entry_link(map, prev_entry, new_entry);
|
|
map^.size:=map^.size+(new_entry^.__end - new_entry^.start);
|
|
|
|
dmem_map_simplify_entry(map, new_entry);
|
|
|
|
//pmap_enter_object
|
|
|
|
Result:=0;
|
|
end;
|
|
|
|
Function dmem_map_query_available(map:p_dmem_map;start,__end,align:QWORD;var oaddr,osize:QWORD):Integer;
|
|
const
|
|
max_valid=QWORD($5000000000);
|
|
var
|
|
entry:p_dmem_map_entry;
|
|
|
|
r_addr,r_size:QWORD;
|
|
t_addr,t_size:QWORD;
|
|
t__end,t_free:QWORD;
|
|
begin
|
|
Result:=0;
|
|
|
|
if not IsPowerOfTwo(align) then
|
|
begin
|
|
Exit(EINVAL);
|
|
end;
|
|
|
|
if (align<PAGE_SIZE) then align:=PAGE_SIZE;
|
|
|
|
start:=(not (start shr 63)) and start;
|
|
if (start>max_valid) then start:=max_valid;
|
|
|
|
__end:=(not (__end shr 63)) and __end;
|
|
if (__end>max_valid) then __end:=max_valid;
|
|
|
|
start:=AlignUp(start,align);
|
|
|
|
dmem_map_lock(map);
|
|
|
|
if (map^.root=nil) then
|
|
begin
|
|
if (start>=IDX_TO_OFF(map^.max_offset)) then
|
|
begin
|
|
r_addr:=0;
|
|
r_size:=0;
|
|
Result:=ENOMEM;
|
|
end else
|
|
begin
|
|
r_addr:=start;
|
|
r_size:=IDX_TO_OFF(map^.max_offset)-start;
|
|
Result:=0;
|
|
end;
|
|
end else
|
|
begin
|
|
r_addr:=0;
|
|
r_size:=0;
|
|
Result:=ENOMEM;
|
|
|
|
map^.root:=dmem_map_entry_splay(OFF_TO_IDX(start), map^.root);
|
|
entry:=map^.root;
|
|
|
|
while (entry<>nil) do
|
|
begin
|
|
if (entry^.adj_free<>0) then
|
|
begin
|
|
t__end:=IDX_TO_OFF(entry^.__end);
|
|
t_free:=IDX_TO_OFF(entry^.adj_free);
|
|
t_addr:=AlignUp(t__end,align);
|
|
if (__end<t_addr) then Break;
|
|
t_size:=(t_addr-t__end);
|
|
if (t_size<t_free) then
|
|
begin
|
|
t_size:=t_free-t_size;
|
|
if (t_size>r_size) then
|
|
begin
|
|
r_addr:=t_addr;
|
|
r_size:=t_size;
|
|
end;
|
|
Result:=0;
|
|
end;
|
|
end;
|
|
entry:=entry^.next;
|
|
end;
|
|
end;
|
|
|
|
dmem_map_unlock(map);
|
|
|
|
oaddr:=r_addr;
|
|
osize:=r_size;
|
|
end;
|
|
|
|
function dmem_map_findspace(map :p_dmem_map;
|
|
start :DWORD;
|
|
length:DWORD;
|
|
addr :PDWORD):Integer;
|
|
label
|
|
_nxt;
|
|
var
|
|
entry:p_dmem_map_entry;
|
|
st:DWORD;
|
|
begin
|
|
{
|
|
* Request must fit within min/max VM address and must avoid
|
|
* address wrap.
|
|
}
|
|
if (start<map^.min_offset) then
|
|
begin
|
|
start:=map^.min_offset;
|
|
end;
|
|
if (start + length>map^.max_offset) or (start + length<start) then
|
|
begin
|
|
Exit(1);
|
|
end;
|
|
|
|
{ Empty tree means wide open address space. }
|
|
if (map^.root=nil) then
|
|
begin
|
|
addr^:=start;
|
|
Exit(0);
|
|
end;
|
|
|
|
{
|
|
* After splay, if start comes before root node, then there
|
|
* must be a gap from start to the root.
|
|
}
|
|
map^.root:=dmem_map_entry_splay(start, map^.root);
|
|
if (start + length<=map^.root^.start) then
|
|
begin
|
|
addr^:=start;
|
|
Exit(0);
|
|
end;
|
|
|
|
{
|
|
* Root is the last node that might begin its gap before
|
|
* start, and this is the last comparison where address
|
|
* wrap might be a problem.
|
|
}
|
|
|
|
if (start>map^.root^.__end) then
|
|
begin
|
|
st:=start;
|
|
end else
|
|
begin
|
|
st:=map^.root^.__end;
|
|
end;
|
|
|
|
if (length<=map^.root^.__end + map^.root^.adj_free - st) then
|
|
begin
|
|
addr^:=st;
|
|
Exit(0);
|
|
end;
|
|
|
|
{ With max_free, can immediately tell if no solution. }
|
|
entry:=map^.root^.right;
|
|
|
|
if (entry=nil) then
|
|
begin
|
|
Exit(1);
|
|
end;
|
|
|
|
if (length>entry^.max_free) then
|
|
begin
|
|
Exit(1);
|
|
end;
|
|
|
|
{
|
|
* Search the right subtree in the order: left subtree, root,
|
|
* right subtree (first fit). The previous splay implies that
|
|
* all regions in the right subtree have addresses>start.
|
|
}
|
|
while (entry<>nil) do
|
|
begin
|
|
if (entry^.left<>nil) then
|
|
begin
|
|
if not (entry^.left^.max_free>=length) then goto _nxt;
|
|
entry:=entry^.left;
|
|
end else
|
|
begin
|
|
_nxt:
|
|
if (entry^.adj_free>=length) then
|
|
begin
|
|
addr^:=entry^.__end;
|
|
Exit(0);
|
|
end else
|
|
begin
|
|
entry:=entry^.right;
|
|
end;
|
|
end;
|
|
end;
|
|
|
|
{ Can't get here, so panic if we do. }
|
|
Assert(false,'dmem_map_findspace: max_free corrupt');
|
|
end;
|
|
|
|
function dmem_map_fixed(map :p_dmem_map;
|
|
start :DWORD;
|
|
length :DWORD;
|
|
m_type :DWORD;
|
|
overwr :Integer):Integer;
|
|
var
|
|
__end:DWORD;
|
|
begin
|
|
__end:=start + length;
|
|
dmem_map_lock(map);
|
|
DMEM_MAP_RANGE_CHECK(map, start, __end);
|
|
if (overwr<>0) then
|
|
begin
|
|
dmem_map_delete(map, start, __end);
|
|
end;
|
|
Result:= dmem_map_insert(map, start, __end, m_type);
|
|
dmem_map_unlock(map);
|
|
end;
|
|
|
|
procedure dmem_map_simplify_entry(map:p_dmem_map;entry:p_dmem_map_entry);
|
|
var
|
|
next,prev:p_dmem_map_entry;
|
|
begin
|
|
//if ((entry^.eflags and (MAP_ENTRY_IS_SUB_MAP))<>0) then
|
|
//begin
|
|
// Exit;
|
|
//end;
|
|
|
|
prev:=entry^.prev;
|
|
if (prev<>@map^.header) then
|
|
begin
|
|
if (prev^.__end=entry^.start) and
|
|
(prev^.m_type=entry^.m_type) then
|
|
begin
|
|
dmem_map_entry_unlink(map, prev);
|
|
entry^.start:=prev^.start;
|
|
|
|
//change
|
|
if (entry^.prev<>@map^.header) then
|
|
begin
|
|
dmem_map_entry_resize_free(map, entry^.prev);
|
|
end;
|
|
|
|
dmem_map_entry_dispose(map, prev);
|
|
end;
|
|
end;
|
|
|
|
next:=entry^.next;
|
|
if (next<>@map^.header) then
|
|
begin
|
|
if (entry^.__end=next^.start) and
|
|
(next^.m_type=entry^.m_type) then
|
|
begin
|
|
dmem_map_entry_unlink(map, next);
|
|
entry^.__end:=next^.__end;
|
|
//change
|
|
dmem_map_entry_resize_free(map, entry);
|
|
|
|
dmem_map_entry_dispose(map, next);
|
|
end;
|
|
end;
|
|
end;
|
|
|
|
procedure _dmem_map_clip_start(map:p_dmem_map;entry:p_dmem_map_entry;start:DWORD);
|
|
var
|
|
new_entry:p_dmem_map_entry;
|
|
begin
|
|
DMEM_MAP_ASSERT_LOCKED(map);
|
|
|
|
dmem_map_simplify_entry(map, entry);
|
|
|
|
new_entry:=dmem_map_entry_create(map);
|
|
new_entry^:=entry^;
|
|
|
|
new_entry^.__end:=start;
|
|
entry^.start:=start;
|
|
|
|
dmem_map_entry_link(map, entry^.prev, new_entry);
|
|
|
|
//if ((entry^.eflags and MAP_ENTRY_IS_SUB_MAP)=0) then
|
|
//begin
|
|
// vm_object_reference(new_entry^.vm_obj);
|
|
//end;
|
|
end;
|
|
|
|
procedure dmem_map_clip_start(map:p_dmem_map;entry:p_dmem_map_entry;start:DWORD);
|
|
begin
|
|
if (start>entry^.start) then
|
|
begin
|
|
_dmem_map_clip_start(map,entry,start);
|
|
end;
|
|
end;
|
|
|
|
procedure _dmem_map_clip_end(map:p_dmem_map;entry:p_dmem_map_entry;__end:DWORD);
|
|
var
|
|
new_entry:p_dmem_map_entry;
|
|
begin
|
|
DMEM_MAP_ASSERT_LOCKED(map);
|
|
|
|
{
|
|
* Create a new entry and insert it AFTER the specified entry
|
|
}
|
|
new_entry:=dmem_map_entry_create(map);
|
|
new_entry^:=entry^;
|
|
|
|
new_entry^.start:=__end;
|
|
entry^.__end:=__end;
|
|
|
|
dmem_map_entry_link(map, entry, new_entry);
|
|
|
|
//if ((entry^.eflags and MAP_ENTRY_IS_SUB_MAP)=0) then
|
|
//begin
|
|
// vm_object_reference(new_entry^.vm_obj);
|
|
//end;
|
|
end;
|
|
|
|
procedure dmem_map_clip_end(map:p_dmem_map;entry:p_dmem_map_entry;__end:DWORD);
|
|
begin
|
|
if (__end<entry^.__end) then
|
|
begin
|
|
_dmem_map_clip_end(map,entry,__end);
|
|
end;
|
|
end;
|
|
|
|
procedure dmem_map_entry_delete(map:p_dmem_map;entry:p_dmem_map_entry);
|
|
var
|
|
size:DWORD;
|
|
begin
|
|
dmem_map_entry_unlink(map, entry);
|
|
size:=entry^.__end - entry^.start;
|
|
map^.size:=map^.size-size;
|
|
|
|
entry^.next:=curkthread^.td_map_def_user;
|
|
curkthread^.td_map_def_user:=entry;
|
|
end;
|
|
|
|
function dmem_map_delete(map:p_dmem_map;start:DWORD;__end:DWORD):Integer;
|
|
var
|
|
entry :p_dmem_map_entry;
|
|
first_entry:p_dmem_map_entry;
|
|
next :p_dmem_map_entry;
|
|
begin
|
|
DMEM_MAP_ASSERT_LOCKED(map);
|
|
if (start=__end) then
|
|
begin
|
|
Exit(0);
|
|
end;
|
|
|
|
{
|
|
* Find the start of the region, and clip it
|
|
}
|
|
if (not dmem_map_lookup_entry(map, start, @first_entry)) then
|
|
begin
|
|
entry:=first_entry^.next;
|
|
end else
|
|
begin
|
|
entry:=first_entry;
|
|
dmem_map_clip_start(map, entry, start);
|
|
end;
|
|
|
|
{
|
|
* Step through all entries in this region
|
|
}
|
|
while (entry<>@map^.header) and (entry^.start<__end) do
|
|
begin
|
|
|
|
dmem_map_clip_end(map, entry, __end);
|
|
|
|
next:=entry^.next;
|
|
|
|
//pmap_remove(map^.pmap,entry^.start,entry^.__end,entry^.protection);
|
|
|
|
{
|
|
* Delete the entry only after removing all pmap
|
|
* entries pointing to its pages. (Otherwise, its
|
|
* page frames may be reallocated, and any modify bits
|
|
* will be set in the wrong object!)
|
|
}
|
|
dmem_map_entry_delete(map, entry);
|
|
entry:=next;
|
|
end;
|
|
Result:=(0);
|
|
end;
|
|
|
|
end.
|
|
|