/* * SPDX-License-Identifier: BSD-2-Clause * * Copyright (c) 2023 Ventana Micro Systems Inc. * * Authors: * Anup Patel */ #include #include #include #include #include #include /* Minimum size and alignment of heap allocations */ #define HEAP_ALLOC_ALIGN 64 #define HEAP_HOUSEKEEPING_FACTOR 16 struct heap_node { struct sbi_dlist head; unsigned long addr; unsigned long size; }; struct sbi_heap_control { spinlock_t lock; unsigned long base; unsigned long size; unsigned long hkbase; unsigned long hksize; struct sbi_dlist free_node_list; struct sbi_dlist free_space_list; struct sbi_dlist used_space_list; }; struct sbi_heap_control global_hpctrl; void *sbi_malloc_from(struct sbi_heap_control *hpctrl, size_t size) { void *ret = NULL; struct heap_node *n, *np; if (!size) return NULL; size += HEAP_ALLOC_ALIGN - 1; size &= ~((unsigned long)HEAP_ALLOC_ALIGN - 1); spin_lock(&hpctrl->lock); np = NULL; sbi_list_for_each_entry(n, &hpctrl->free_space_list, head) { if (size <= n->size) { np = n; break; } } if (np) { if ((size < np->size) && !sbi_list_empty(&hpctrl->free_node_list)) { n = sbi_list_first_entry(&hpctrl->free_node_list, struct heap_node, head); sbi_list_del(&n->head); n->addr = np->addr + np->size - size; n->size = size; np->size -= size; sbi_list_add_tail(&n->head, &hpctrl->used_space_list); ret = (void *)n->addr; } else if (size == np->size) { sbi_list_del(&np->head); sbi_list_add_tail(&np->head, &hpctrl->used_space_list); ret = (void *)np->addr; } } spin_unlock(&hpctrl->lock); return ret; } void *sbi_zalloc_from(struct sbi_heap_control *hpctrl, size_t size) { void *ret = sbi_malloc_from(hpctrl, size); if (ret) sbi_memset(ret, 0, size); return ret; } void sbi_free_from(struct sbi_heap_control *hpctrl, void *ptr) { struct heap_node *n, *np; if (!ptr) return; spin_lock(&hpctrl->lock); np = NULL; sbi_list_for_each_entry(n, &hpctrl->used_space_list, head) { if ((n->addr <= (unsigned long)ptr) && ((unsigned long)ptr < (n->addr + n->size))) { np = n; break; } } if (!np) { spin_unlock(&hpctrl->lock); return; } sbi_list_del(&np->head); sbi_list_for_each_entry(n, &hpctrl->free_space_list, head) { if ((np->addr + np->size) == n->addr) { n->addr = np->addr; n->size += np->size; sbi_list_add_tail(&np->head, &hpctrl->free_node_list); np = NULL; break; } else if (np->addr == (n->addr + n->size)) { n->size += np->size; sbi_list_add_tail(&np->head, &hpctrl->free_node_list); np = NULL; break; } else if ((n->addr + n->size) < np->addr) { sbi_list_add(&np->head, &n->head); np = NULL; break; } } if (np) sbi_list_add_tail(&np->head, &hpctrl->free_space_list); spin_unlock(&hpctrl->lock); } unsigned long sbi_heap_free_space_from(struct sbi_heap_control *hpctrl) { struct heap_node *n; unsigned long ret = 0; spin_lock(&hpctrl->lock); sbi_list_for_each_entry(n, &hpctrl->free_space_list, head) ret += n->size; spin_unlock(&hpctrl->lock); return ret; } unsigned long sbi_heap_used_space_from(struct sbi_heap_control *hpctrl) { return hpctrl->size - hpctrl->hksize - sbi_heap_free_space(); } unsigned long sbi_heap_reserved_space_from(struct sbi_heap_control *hpctrl) { return hpctrl->hksize; } int sbi_heap_init_new(struct sbi_heap_control *hpctrl, unsigned long base, unsigned long size) { unsigned long i; struct heap_node *n; /* Initialize heap control */ SPIN_LOCK_INIT(hpctrl->lock); hpctrl->base = base; hpctrl->size = size; hpctrl->hkbase = hpctrl->base; hpctrl->hksize = hpctrl->size / HEAP_HOUSEKEEPING_FACTOR; hpctrl->hksize &= ~((unsigned long)HEAP_BASE_ALIGN - 1); SBI_INIT_LIST_HEAD(&hpctrl->free_node_list); SBI_INIT_LIST_HEAD(&hpctrl->free_space_list); SBI_INIT_LIST_HEAD(&hpctrl->used_space_list); /* Prepare free node list */ for (i = 0; i < (hpctrl->hksize / sizeof(*n)); i++) { n = (struct heap_node *)(hpctrl->hkbase + (sizeof(*n) * i)); SBI_INIT_LIST_HEAD(&n->head); n->addr = n->size = 0; sbi_list_add_tail(&n->head, &hpctrl->free_node_list); } /* Prepare free space list */ n = sbi_list_first_entry(&hpctrl->free_node_list, struct heap_node, head); sbi_list_del(&n->head); n->addr = hpctrl->hkbase + hpctrl->hksize; n->size = hpctrl->size - hpctrl->hksize; sbi_list_add_tail(&n->head, &hpctrl->free_space_list); return 0; } int sbi_heap_init(struct sbi_scratch *scratch) { /* Sanity checks on heap offset and size */ if (!scratch->fw_heap_size || (scratch->fw_heap_size & (HEAP_BASE_ALIGN - 1)) || (scratch->fw_heap_offset < scratch->fw_rw_offset) || (scratch->fw_size < (scratch->fw_heap_offset + scratch->fw_heap_size)) || (scratch->fw_heap_offset & (HEAP_BASE_ALIGN - 1))) return SBI_EINVAL; return sbi_heap_init_new(&global_hpctrl, scratch->fw_start + scratch->fw_heap_offset, scratch->fw_heap_size); } int sbi_heap_alloc_new(struct sbi_heap_control **hpctrl) { *hpctrl = sbi_calloc(1, sizeof(struct sbi_heap_control)); return 0; }