diff options
Diffstat (limited to 'tdecore/kallocator.cpp')
-rw-r--r-- | tdecore/kallocator.cpp | 260 |
1 files changed, 260 insertions, 0 deletions
diff --git a/tdecore/kallocator.cpp b/tdecore/kallocator.cpp new file mode 100644 index 000000000..16c1b3625 --- /dev/null +++ b/tdecore/kallocator.cpp @@ -0,0 +1,260 @@ +/* + This file is part of the KDE libraries + + Copyright (C) 1999 Waldo Bastian (bastian@kde.org) + Copyright (C) 2002 Michael Matz (matz@kde.org) + + This library is free software; you can redistribute it and/or + modify it under the terms of the GNU Library General Public + License as published by the Free Software Foundation; either + version 2 of the License, or (at your option) any later version. + + This library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Library General Public License for more details. + + You should have received a copy of the GNU Library General Public License + along with this library; see the file COPYING.LIB. If not, write to + the Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor, + Boston, MA 02110-1301, USA. +*/ + +/* Fast zone memory allocator with deallocation support, for use as obstack + or as general purpose allocator. It does no compaction. If the usage + pattern is non-optimal it might waste some memory while running. E.g. + allocating many small things at once, and then deallocating only every + second one, there is a high chance, that actually no memory is freed. */ +// $Id$ + +#include "kallocator.h" +#include <kdebug.h> + +class KZoneAllocator::MemBlock +{ + public: + MemBlock(size_t s) : size(s), ref(0), older(0), newer(0) + { begin = new char[s]; } + ~MemBlock() { delete [] begin; } + bool is_in(void *ptr) const {return !(begin > (char *)ptr + || (begin + size) <= (char *)ptr); } + size_t size; + unsigned int ref; + char *begin; + MemBlock *older; + MemBlock *newer; +}; + +KZoneAllocator::KZoneAllocator(unsigned long _blockSize) +: currentBlock(0), blockSize(1), blockOffset(0), log2(0), num_blocks(0), + hashList(0), hashSize(0), hashDirty(true) +{ + while (blockSize < _blockSize) + blockSize <<= 1, log2++; + /* Make sure, that a block is allocated at the first time allocate() + is called (even with a 0 size). */ + blockOffset = blockSize + 1; +} + +KZoneAllocator::~KZoneAllocator() +{ + unsigned int count = 0; + if (hashList) { + /* No need to maintain the different lists in hashList[] anymore. + I.e. no need to use delBlock(). */ + for (unsigned int i = 0; i < hashSize; i++) + delete hashList[i]; + delete [] hashList; + hashList = 0; + } + MemBlock *next; + for (; currentBlock; currentBlock = next) { + next = currentBlock->older; + delete currentBlock; + count++; + } +#ifndef NDEBUG // as this is called quite late in the app, we don't care + // to use kdDebug + if (count > 1) + qDebug("zone still contained %d blocks", count); +#endif +} + +void KZoneAllocator::insertHash(MemBlock *b) +{ + unsigned long adr = ((unsigned long)b->begin) & (~(blockSize - 1)); + unsigned long end = ((unsigned long)b->begin) + blockSize; + while (adr < end) { + unsigned long key = adr >> log2; + key = key & (hashSize - 1); + if (!hashList[key]) + hashList[key] = new TQValueList<MemBlock *>; + hashList[key]->append(b); + adr += blockSize; + } +} + +/** Add a new memory block to the pool of blocks, + and reorganize the hash lists if needed. + @param b block to add + @internal +*/ +void KZoneAllocator::addBlock(MemBlock *b) +{ + b->newer = 0; + b->older = currentBlock; + if (currentBlock) + b->older->newer = b; + currentBlock = b; + num_blocks++; + /* If we either have no hashList at all, or since it's last construction + there are now many more blocks we reconstruct the list. But don't + make it larger than a certain maximum. */ + if (hashList && ((num_blocks / 4) > hashSize && hashSize < 64*1024)) + hashDirty = true; + /* Only insert this block into the hashlists, if we aren't going to + reconstruct them anyway. */ + if (hashList && !hashDirty) + insertHash (b); +} + +/** Reinitialize hash list. @internal */ +void KZoneAllocator::initHash() +{ + if (hashList) { + for (unsigned int i = 0; i < hashSize; i++) + delete hashList[i]; + delete [] hashList; + hashList = 0; + } + hashSize = 1; + while (hashSize < num_blocks) + hashSize <<= 1; + if (hashSize < 1024) + hashSize = 1024; + if (hashSize > 64*1024) + hashSize = 64*1024; + hashList = new TQValueList<MemBlock *> *[hashSize]; + memset (hashList, 0, sizeof(TQValueList<MemBlock*> *) * hashSize); + hashDirty = false; + for (MemBlock *b = currentBlock; b; b = b->older) + insertHash(b); +} + +/** Delete a memory block. This @em really returns the memory to the heap. + @param b block to delete + @internal +*/ +void KZoneAllocator::delBlock(MemBlock *b) +{ + /* Update also the hashlists if we aren't going to reconstruct them + soon. */ + if (hashList && !hashDirty) { + unsigned long adr = ((unsigned long)b->begin) & (~(blockSize - 1)); + unsigned long end = ((unsigned long)b->begin) + blockSize; + while (adr < end) { + unsigned long key = adr >> log2; + key = key & (hashSize - 1); + if (hashList[key]) { + TQValueList<MemBlock *> *list = hashList[key]; + TQValueList<MemBlock *>::Iterator it = list->begin(); + TQValueList<MemBlock *>::Iterator endit = list->end(); + for (; it != endit; ++it) + if (*it == b) { + list->remove(it); + break; + } + } + adr += blockSize; + } + } + if (b->older) + b->older->newer = b->newer; + if (b->newer) + b->newer->older = b->older; + if (b == currentBlock) { + currentBlock = 0; + blockOffset = blockSize; + } + delete b; + num_blocks--; +} + +void * +KZoneAllocator::allocate(size_t _size) +{ + // Use the size of (void *) as tqalignment + const size_t tqalignment = sizeof(void *) - 1; + _size = (_size + tqalignment) & ~tqalignment; + + if ((unsigned long) _size + blockOffset > blockSize) + { + if (_size > blockSize) { + qDebug("KZoneAllocator: allocating more than %lu bytes", blockSize); + return 0; + } + addBlock(new MemBlock(blockSize)); + blockOffset = 0; + //qDebug ("Allocating block #%d (%x)\n", num_blocks, currentBlock->begin); + } + void *result = (void *)(currentBlock->begin+blockOffset); + currentBlock->ref++; + blockOffset += _size; + return result; +} + +void +KZoneAllocator::deallocate(void *ptr) +{ + if (hashDirty) + initHash(); + + unsigned long key = (((unsigned long)ptr) >> log2) & (hashSize - 1); + TQValueList<MemBlock *> *list = hashList[key]; + if (!list) { + /* Can happen with certain usage pattern of intermixed free_since() + and deallocate(). */ + //qDebug("Uhoh"); + return; + } + TQValueList<MemBlock*>::ConstIterator it = list->begin(); + TQValueList<MemBlock*>::ConstIterator endit = list->end(); + for (; it != endit; ++it) { + MemBlock *cur = *it; + if (cur->is_in(ptr)) { + if (!--cur->ref) { + if (cur != currentBlock) + delBlock (cur); + else + blockOffset = 0; + } + return; + } + } + /* Can happen with certain usage pattern of intermixed free_since() + and deallocate(). */ + //qDebug("Uhoh2"); +} + +void +KZoneAllocator::free_since(void *ptr) +{ + /* If we have a hashList and it's not yet dirty, see, if we will dirty + it by removing too many blocks. This will make the below delBlock()s + faster. */ + if (hashList && !hashDirty) + { + const MemBlock *b; + unsigned int removed = 0; + for (b = currentBlock; b; b = b->older, removed++) + if (b->is_in (ptr)) + break; + if (hashSize >= 4 * (num_blocks - removed)) + hashDirty = true; + } + while (currentBlock && !currentBlock->is_in(ptr)) { + currentBlock = currentBlock->older; + delBlock (currentBlock->newer); + } + blockOffset = ((char*)ptr) - currentBlock->begin; +} |