This repository has been archived on 2023-10-09. You can view files and clone it, but cannot push or open issues or pull requests.
Files
blender-archive/source/blender/blenlib/intern/BLI_heap.c

421 lines
8.7 KiB
C

/*
* ***** BEGIN GPL LICENSE BLOCK *****
*
* This program is free software; you can redistribute it and/or
* modify it under the terms of the GNU General Public License
* as published by the Free Software Foundation; either version 2
* of the License, or (at your option) any later version.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License
* along with this program; if not, write to the Free Software Foundation,
* Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
*
* Contributor(s): Brecht Van Lommel
*
* ***** END GPL LICENSE BLOCK *****
*/
/** \file blender/blenlib/intern/BLI_heap.c
* \ingroup bli
*
* A min-heap / priority queue ADT.
*/
#include <stdlib.h>
#include <string.h>
#include "MEM_guardedalloc.h"
#include "BLI_utildefines.h"
#include "BLI_heap.h"
#include "BLI_strict_flags.h"
/***/
struct HeapNode {
void *ptr;
float value;
uint index;
};
struct HeapNode_Chunk {
struct HeapNode_Chunk *prev;
uint size;
uint bufsize;
struct HeapNode buf[0];
};
/**
* Number of nodes to include per #HeapNode_Chunk when no reserved size is passed,
* or we allocate past the reserved number.
*
* \note Optimize number for 64kb allocs.
* \note keep type in sync with tot_nodes in heap_node_alloc_chunk.
*/
#define HEAP_CHUNK_DEFAULT_NUM \
((uint)((MEM_SIZE_OPTIMAL((1 << 16) - sizeof(struct HeapNode_Chunk))) / sizeof(HeapNode)))
struct Heap {
uint size;
uint bufsize;
HeapNode **tree;
struct {
/* Always keep at least one chunk (never NULL) */
struct HeapNode_Chunk *chunk;
/* when NULL, allocate a new chunk */
HeapNode *free;
} nodes;
};
/** \name Internal Functions
* \{ */
#define HEAP_PARENT(i) (((i) - 1) >> 1)
#define HEAP_LEFT(i) (((i) << 1) + 1)
#define HEAP_RIGHT(i) (((i) << 1) + 2)
#define HEAP_COMPARE(a, b) ((a)->value < (b)->value)
#if 0 /* UNUSED */
#define HEAP_EQUALS(a, b) ((a)->value == (b)->value)
#endif
BLI_INLINE void heap_swap(Heap *heap, const uint i, const uint j)
{
#if 0
SWAP(uint, heap->tree[i]->index, heap->tree[j]->index);
SWAP(HeapNode *, heap->tree[i], heap->tree[j]);
#else
HeapNode **tree = heap->tree;
union {
uint index;
HeapNode *node;
} tmp;
SWAP_TVAL(tmp.index, tree[i]->index, tree[j]->index);
SWAP_TVAL(tmp.node, tree[i], tree[j]);
#endif
}
static void heap_down(Heap *heap, uint i)
{
/* size won't change in the loop */
const uint size = heap->size;
while (1) {
const uint l = HEAP_LEFT(i);
const uint r = HEAP_RIGHT(i);
uint smallest = i;
if ((l < size) && HEAP_COMPARE(heap->tree[l], heap->tree[smallest])) {
smallest = l;
}
if ((r < size) && HEAP_COMPARE(heap->tree[r], heap->tree[smallest])) {
smallest = r;
}
if (smallest == i) {
break;
}
heap_swap(heap, i, smallest);
i = smallest;
}
}
static void heap_up(Heap *heap, uint i)
{
while (i > 0) {
const uint p = HEAP_PARENT(i);
if (HEAP_COMPARE(heap->tree[p], heap->tree[i])) {
break;
}
heap_swap(heap, p, i);
i = p;
}
}
/** \} */
/** \name Internal Memory Management
* \{ */
static struct HeapNode_Chunk *heap_node_alloc_chunk(
uint tot_nodes, struct HeapNode_Chunk *chunk_prev)
{
struct HeapNode_Chunk *chunk = MEM_mallocN(
sizeof(struct HeapNode_Chunk) + (sizeof(HeapNode) * tot_nodes), __func__);
chunk->prev = chunk_prev;
chunk->bufsize = tot_nodes;
chunk->size = 0;
return chunk;
}
static struct HeapNode *heap_node_alloc(Heap *heap)
{
HeapNode *node;
if (heap->nodes.free) {
node = heap->nodes.free;
heap->nodes.free = heap->nodes.free->ptr;
}
else {
struct HeapNode_Chunk *chunk = heap->nodes.chunk;
if (UNLIKELY(chunk->size == chunk->bufsize)) {
chunk = heap->nodes.chunk = heap_node_alloc_chunk(HEAP_CHUNK_DEFAULT_NUM, chunk);
}
node = &chunk->buf[chunk->size++];
}
return node;
}
static void heap_node_free(Heap *heap, HeapNode *node)
{
node->ptr = heap->nodes.free;
heap->nodes.free = node;
}
/** \} */
/** \name Public Heap API
* \{ */
/**
* Creates a new heap. Removed nodes are recycled, so memory usage will not shrink.
*
* \note Use when the size of the heap is known in advance.
*/
Heap *BLI_heap_new_ex(uint tot_reserve)
{
Heap *heap = MEM_mallocN(sizeof(Heap), __func__);
/* ensure we have at least one so we can keep doubling it */
heap->size = 0;
heap->bufsize = MAX2(1u, tot_reserve);
heap->tree = MEM_mallocN(heap->bufsize * sizeof(HeapNode *), "BLIHeapTree");
heap->nodes.chunk = heap_node_alloc_chunk((tot_reserve > 1) ? tot_reserve : HEAP_CHUNK_DEFAULT_NUM, NULL);
heap->nodes.free = NULL;
return heap;
}
Heap *BLI_heap_new(void)
{
return BLI_heap_new_ex(1);
}
void BLI_heap_free(Heap *heap, HeapFreeFP ptrfreefp)
{
if (ptrfreefp) {
uint i;
for (i = 0; i < heap->size; i++) {
ptrfreefp(heap->tree[i]->ptr);
}
}
struct HeapNode_Chunk *chunk = heap->nodes.chunk;
do {
struct HeapNode_Chunk *chunk_prev;
chunk_prev = chunk->prev;
MEM_freeN(chunk);
chunk = chunk_prev;
} while (chunk);
MEM_freeN(heap->tree);
MEM_freeN(heap);
}
void BLI_heap_clear(Heap *heap, HeapFreeFP ptrfreefp)
{
if (ptrfreefp) {
uint i;
for (i = 0; i < heap->size; i++) {
ptrfreefp(heap->tree[i]->ptr);
}
}
heap->size = 0;
/* Remove all except the last chunk */
while (heap->nodes.chunk->prev) {
struct HeapNode_Chunk *chunk_prev = heap->nodes.chunk->prev;
MEM_freeN(heap->nodes.chunk);
heap->nodes.chunk = chunk_prev;
}
heap->nodes.chunk->size = 0;
heap->nodes.free = NULL;
}
/**
* Insert heap node with a value (often a 'cost') and pointer into the heap,
* duplicate values are allowed.
*/
HeapNode *BLI_heap_insert(Heap *heap, float value, void *ptr)
{
HeapNode *node;
if (UNLIKELY(heap->size >= heap->bufsize)) {
heap->bufsize *= 2;
heap->tree = MEM_reallocN(heap->tree, heap->bufsize * sizeof(*heap->tree));
}
node = heap_node_alloc(heap);
node->ptr = ptr;
node->value = value;
node->index = heap->size;
heap->tree[node->index] = node;
heap->size++;
heap_up(heap, node->index);
return node;
}
/**
* Convenience function since this is a common pattern.
*/
void BLI_heap_insert_or_update(Heap *heap, HeapNode **node_p, float value, void *ptr)
{
if (*node_p == NULL) {
*node_p = BLI_heap_insert(heap, value, ptr);
}
else {
BLI_heap_node_value_update_ptr(heap, *node_p, value, ptr);
}
}
bool BLI_heap_is_empty(const Heap *heap)
{
return (heap->size == 0);
}
uint BLI_heap_len(const Heap *heap)
{
return heap->size;
}
/**
* Return the top node of the heap.
* This is the node with the lowest value.
*/
HeapNode *BLI_heap_top(const Heap *heap)
{
return heap->tree[0];
}
/**
* Pop the top node off the heap and return it's pointer.
*/
void *BLI_heap_pop_min(Heap *heap)
{
BLI_assert(heap->size != 0);
void *ptr = heap->tree[0]->ptr;
heap_node_free(heap, heap->tree[0]);
if (--heap->size) {
heap_swap(heap, 0, heap->size);
heap_down(heap, 0);
}
return ptr;
}
void BLI_heap_remove(Heap *heap, HeapNode *node)
{
BLI_assert(heap->size != 0);
uint i = node->index;
while (i > 0) {
uint p = HEAP_PARENT(i);
heap_swap(heap, p, i);
i = p;
}
BLI_heap_pop_min(heap);
}
/**
* Can be used to avoid #BLI_heap_remove, #BLI_heap_insert calls,
* balancing the tree still has a performance cost,
* but is often much less than remove/insert, difference is most noticable with large heaps.
*/
void BLI_heap_node_value_update(Heap *heap, HeapNode *node, float value)
{
if (value < node->value) {
node->value = value;
heap_up(heap, node->index);
}
else if (value > node->value) {
node->value = value;
heap_down(heap, node->index);
}
}
void BLI_heap_node_value_update_ptr(Heap *heap, HeapNode *node, float value, void *ptr)
{
node->ptr = ptr; /* only difference */
if (value < node->value) {
node->value = value;
heap_up(heap, node->index);
}
else if (value > node->value) {
node->value = value;
heap_down(heap, node->index);
}
}
float BLI_heap_node_value(const HeapNode *node)
{
return node->value;
}
void *BLI_heap_node_ptr(const HeapNode *node)
{
return node->ptr;
}
static bool heap_is_minheap(const Heap *heap, uint root)
{
if (root < heap->size) {
const uint l = HEAP_LEFT(root);
if (l < heap->size) {
if (HEAP_COMPARE(heap->tree[l], heap->tree[root]) || !heap_is_minheap(heap, l)) {
return false;
}
}
const uint r = HEAP_RIGHT(root);
if (r < heap->size) {
if (HEAP_COMPARE(heap->tree[r], heap->tree[root]) || !heap_is_minheap(heap, r)) {
return false;
}
}
}
return true;
}
/**
* Only for checking internal errors (gtest).
*/
bool BLI_heap_is_valid(const Heap *heap)
{
return heap_is_minheap(heap, 0);
}
/** \} */