/*************************************************************************/ /* pool_allocator.cpp */ /*************************************************************************/ /* This file is part of: */ /* GODOT ENGINE */ /* http://www.godotengine.org */ /*************************************************************************/ /* Copyright (c) 2007-2016 Juan Linietsky, Ariel Manzur. */ /* */ /* Permission is hereby granted, free of charge, to any person obtaining */ /* a copy of this software and associated documentation files (the */ /* "Software"), to deal in the Software without restriction, including */ /* without limitation the rights to use, copy, modify, merge, publish, */ /* distribute, sublicense, and/or sell copies of the Software, and to */ /* permit persons to whom the Software is furnished to do so, subject to */ /* the following conditions: */ /* */ /* The above copyright notice and this permission notice shall be */ /* included in all copies or substantial portions of the Software. */ /* */ /* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, */ /* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF */ /* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.*/ /* IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY */ /* CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, */ /* TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE */ /* SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. */ /*************************************************************************/ #include "pool_allocator.h" #include "error_macros.h" #include "core/os/os.h" #include "os/memory.h" #include "os/copymem.h" #include "print_string.h" #include #define COMPACT_CHUNK( m_entry , m_to_pos ) \ do { \ void *_dst=&((unsigned char*)pool)[m_to_pos]; \ void *_src=&((unsigned char*)pool)[(m_entry).pos]; \ movemem(_dst,_src,aligned((m_entry).len)); \ (m_entry).pos=m_to_pos; \ } while (0); void PoolAllocator::mt_lock() const { } void PoolAllocator::mt_unlock() const { } bool PoolAllocator::get_free_entry(EntryArrayPos* p_pos) { if (entry_count==entry_max) return false; for (int i=0;i=p_for_size) { *p_pos=i; return true; } /* prepare for next one */ prev_entry_end_pos=entry_end(entry); } /* No holes between entrys, check at the end..*/ if ( (pool_size-prev_entry_end_pos)>=p_for_size ) { *p_pos=entry_count; return true; } return false; } void PoolAllocator::compact(int p_up_to) { uint32_t prev_entry_end_pos=0; if (p_up_to<0) p_up_to=entry_count; for (int i=0;i0 && !entry.lock) { COMPACT_CHUNK(entry,prev_entry_end_pos); } /* prepare for next one */ prev_entry_end_pos=entry_end(entry); } } void PoolAllocator::compact_up(int p_from) { uint32_t next_entry_end_pos=pool_size; // - static_area_size; for (int i=entry_count-1;i>=p_from;i--) { Entry &entry=entry_array[ entry_indices[ i ] ]; /* determine hole size to nextious entry */ int hole_size=next_entry_end_pos-(entry.pos+aligned(entry.len)); /* if we can compact, do it */ if (hole_size>0 && !entry.lock) { COMPACT_CHUNK(entry,(next_entry_end_pos-aligned(entry.len))); } /* prepare for next one */ next_entry_end_pos=entry.pos; } } bool PoolAllocator::find_entry_index(EntryIndicesPos *p_map_pos,Entry *p_entry) { EntryArrayPos entry_pos=entry_max; for (int i=0;i free_mem) OS::get_singleton()->debug_break(); #endif ERR_FAIL_COND_V(p_size>free_mem,POOL_ALLOCATOR_INVALID_ID); mt_lock(); if (entry_count==entry_max) { mt_unlock(); ERR_PRINT("entry_count==entry_max"); return POOL_ALLOCATOR_INVALID_ID; } int size_to_alloc=aligned(p_size); EntryIndicesPos new_entry_indices_pos; if (!find_hole(&new_entry_indices_pos, size_to_alloc)) { /* No hole could be found, try compacting mem */ compact(); /* Then search again */ if (!find_hole(&new_entry_indices_pos, size_to_alloc)) { mt_unlock(); ERR_PRINT("memory can't be compacted further"); return POOL_ALLOCATOR_INVALID_ID; } } EntryArrayPos new_entry_array_pos; bool found_free_entry=get_free_entry(&new_entry_array_pos); if (!found_free_entry) { mt_unlock(); ERR_FAIL_COND_V( !found_free_entry , POOL_ALLOCATOR_INVALID_ID ); } /* move all entry indices up, make room for this one */ for (int i=entry_count;i>new_entry_indices_pos;i-- ) { entry_indices[i]=entry_indices[i-1]; } entry_indices[new_entry_indices_pos]=new_entry_array_pos; entry_count++; Entry &entry=entry_array[ entry_indices[ new_entry_indices_pos ] ]; entry.len=p_size; entry.pos=(new_entry_indices_pos==0)?0:entry_end(entry_array[ entry_indices[ new_entry_indices_pos-1 ] ]); //alloc either at begining or end of previous entry.lock=0; entry.check=(check_count++)&CHECK_MASK; free_mem-=size_to_alloc; if (free_mem>CHECK_BITS; ERR_FAIL_INDEX_V(entry,entry_max,NULL); ERR_FAIL_COND_V(entry_array[entry].check!=check,NULL); ERR_FAIL_COND_V(entry_array[entry].len==0,NULL); return &entry_array[entry]; } const PoolAllocator::Entry * PoolAllocator::get_entry(ID p_mem) const { unsigned int check=p_mem&CHECK_MASK; int entry=p_mem>>CHECK_BITS; ERR_FAIL_INDEX_V(entry,entry_max,NULL); ERR_FAIL_COND_V(entry_array[entry].check!=check,NULL); ERR_FAIL_COND_V(entry_array[entry].len==0,NULL); return &entry_array[entry]; } void PoolAllocator::free(ID p_mem) { mt_lock(); Entry *e=get_entry(p_mem); if (!e) { mt_unlock(); ERR_PRINT("!e"); return; } if (e->lock) { mt_unlock(); ERR_PRINT("e->lock"); return; } EntryIndicesPos entry_indices_pos; bool index_found = find_entry_index(&entry_indices_pos,e); if (!index_found) { mt_unlock(); ERR_FAIL_COND(!index_found); } for (int i=entry_indices_pos;i<(entry_count-1);i++) { entry_indices[ i ] = entry_indices[ i+1 ]; } entry_count--; free_mem+=aligned(e->len); e->clear(); mt_unlock(); } int PoolAllocator::get_size(ID p_mem) const { int size; mt_lock(); const Entry *e=get_entry(p_mem); if (!e) { mt_unlock(); ERR_PRINT("!e"); return 0; } size=e->len; mt_unlock(); return size; } Error PoolAllocator::resize(ID p_mem,int p_new_size) { mt_lock(); Entry *e=get_entry(p_mem); if (!e) { mt_unlock(); ERR_FAIL_COND_V(!e,ERR_INVALID_PARAMETER); } if (needs_locking && e->lock) { mt_unlock(); ERR_FAIL_COND_V(e->lock,ERR_ALREADY_IN_USE); } int alloc_size = aligned(p_new_size); if (aligned(e->len)==alloc_size) { e->len=p_new_size; mt_unlock(); return OK; } else if (e->len>(uint32_t)p_new_size) { free_mem += aligned(e->len); free_mem -= alloc_size; e->len=p_new_size; mt_unlock(); return OK; } //p_new_size = align(p_new_size) int _total = pool_size; // - static_area_size; int _free = free_mem; // - static_area_size; if ((_free + aligned(e->len)) - alloc_size < 0) { mt_unlock(); ERR_FAIL_V( ERR_OUT_OF_MEMORY ); }; EntryIndicesPos entry_indices_pos; bool index_found = find_entry_index(&entry_indices_pos,e); if (!index_found) { mt_unlock(); ERR_FAIL_COND_V(!index_found,ERR_BUG); } //no need to move stuff around, it fits before the next block int next_pos; if (entry_indices_pos+1 == entry_count) { next_pos = pool_size; // - static_area_size; } else { next_pos = entry_array[entry_indices[entry_indices_pos+1]].pos; }; if ((next_pos - e->pos) > alloc_size) { free_mem+=aligned(e->len); e->len=p_new_size; free_mem-=alloc_size; mt_unlock(); return OK; } //it doesn't fit, compact around BEFORE current index (make room behind) compact(entry_indices_pos+1); if ((next_pos - e->pos) > alloc_size) { //now fits! hooray! free_mem+=aligned(e->len); e->len=p_new_size; free_mem-=alloc_size; mt_unlock(); if (free_mempos) > alloc_size) { //now fits! hooray! free_mem+=aligned(e->len); e->len=p_new_size; free_mem-=alloc_size; mt_unlock(); if (free_memlock++; mt_unlock(); return OK; } bool PoolAllocator::is_locked(ID p_mem) const { if (!needs_locking) return false; mt_lock(); const Entry *e=((PoolAllocator*)(this))->get_entry(p_mem); if (!e) { mt_unlock(); ERR_PRINT("!e"); return false; } bool locked = e->lock; mt_unlock(); return locked; } const void *PoolAllocator::get(ID p_mem) const { if (!needs_locking) { const Entry *e=get_entry(p_mem); ERR_FAIL_COND_V(!e,NULL); return &pool[e->pos]; } mt_lock(); const Entry *e=get_entry(p_mem); if (!e) { mt_unlock(); ERR_FAIL_COND_V(!e,NULL); } if (e->lock==0) { mt_unlock(); ERR_PRINT( "e->lock == 0" ); return NULL; } if (e->pos<0 || (int)e->pos>=pool_size) { mt_unlock(); ERR_PRINT("e->pos<0 || e->pos>=pool_size"); return NULL; } const void *ptr=&pool[e->pos]; mt_unlock(); return ptr; } void *PoolAllocator::get(ID p_mem) { if (!needs_locking) { Entry *e=get_entry(p_mem); if (!e) { ERR_FAIL_COND_V(!e,NULL); }; return &pool[e->pos]; } mt_lock(); Entry *e=get_entry(p_mem); if (!e) { mt_unlock(); ERR_FAIL_COND_V(!e,NULL); } if (e->lock==0) { //assert(0); mt_unlock(); ERR_PRINT( "e->lock == 0" ); return NULL; } if (e->pos<0 || (int)e->pos>=pool_size) { mt_unlock(); ERR_PRINT("e->pos<0 || e->pos>=pool_size"); return NULL; } void *ptr=&pool[e->pos]; mt_unlock(); return ptr; } void PoolAllocator::unlock(ID p_mem) { if (!needs_locking) return; mt_lock(); Entry *e=get_entry(p_mem); if (e->lock == 0 ) { mt_unlock(); ERR_PRINT( "e->lock == 0" ); return; } e->lock--; mt_unlock(); } int PoolAllocator::get_used_mem() const { return pool_size-free_mem; } int PoolAllocator::get_free_peak() { return free_mem_peak; } int PoolAllocator::get_free_mem() { return free_mem; } void PoolAllocator::create_pool(void * p_mem,int p_size,int p_max_entries) { pool=(uint8_t*)p_mem; pool_size=p_size; entry_array = memnew_arr( Entry, p_max_entries ); entry_indices = memnew_arr( int, p_max_entries ); entry_max = p_max_entries; entry_count=0; free_mem=p_size; free_mem_peak=p_size; check_count=0; } PoolAllocator::PoolAllocator(int p_size,bool p_needs_locking,int p_max_entries) { mem_ptr=Memory::alloc_static( p_size,"PoolAllocator()"); ERR_FAIL_COND(!mem_ptr); align=1; create_pool(mem_ptr,p_size,p_max_entries); needs_locking=p_needs_locking; } PoolAllocator::PoolAllocator(void * p_mem,int p_size, int p_align ,bool p_needs_locking,int p_max_entries) { if (p_align > 1) { uint8_t *mem8=(uint8_t*)p_mem; uint64_t ofs = (uint64_t)mem8; if (ofs%p_align) { int dif = p_align-(ofs%p_align); mem8+=p_align-(ofs%p_align); p_size -= dif; p_mem = (void*)mem8; }; }; create_pool( p_mem,p_size,p_max_entries); needs_locking=p_needs_locking; align=p_align; mem_ptr=NULL; } PoolAllocator::PoolAllocator(int p_align,int p_size,bool p_needs_locking,int p_max_entries) { ERR_FAIL_COND(p_align<1); mem_ptr=Memory::alloc_static( p_size+p_align,"PoolAllocator()"); uint8_t *mem8=(uint8_t*)mem_ptr; uint64_t ofs = (uint64_t)mem8; if (ofs%p_align) mem8+=p_align-(ofs%p_align); create_pool( mem8 ,p_size,p_max_entries); needs_locking=p_needs_locking; align=p_align; } PoolAllocator::~PoolAllocator() { if (mem_ptr) Memory::free_static( mem_ptr ); memdelete_arr( entry_array ); memdelete_arr( entry_indices ); }