| /* |
| * Copyright © 2020, VideoLAN and dav1d authors |
| * Copyright © 2020, Two Orioles, LLC |
| * All rights reserved. |
| * |
| * Redistribution and use in source and binary forms, with or without |
| * modification, are permitted provided that the following conditions are met: |
| * |
| * 1. Redistributions of source code must retain the above copyright notice, this |
| * list of conditions and the following disclaimer. |
| * |
| * 2. Redistributions in binary form must reproduce the above copyright notice, |
| * this list of conditions and the following disclaimer in the documentation |
| * and/or other materials provided with the distribution. |
| * |
| * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND |
| * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED |
| * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE |
| * DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR |
| * ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES |
| * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; |
| * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND |
| * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
| * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS |
| * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
| */ |
| |
| #include "config.h" |
| |
| #include <stdint.h> |
| |
| #include "src/internal.h" |
| |
| #if TRACK_HEAP_ALLOCATIONS |
| #include <stdio.h> |
| |
| #include "src/log.h" |
| |
| #define DEFAULT_ALIGN 16 |
| |
| typedef struct { |
| size_t sz; |
| unsigned align; |
| enum AllocationType type; |
| } Dav1dAllocationData; |
| |
| typedef struct { |
| size_t curr_sz; |
| size_t peak_sz; |
| unsigned num_allocs; |
| unsigned num_reuses; |
| } AllocStats; |
| |
| static AllocStats tracked_allocs[N_ALLOC_TYPES]; |
| static size_t curr_total_sz; |
| static size_t peak_total_sz; |
| static pthread_mutex_t track_alloc_mutex = PTHREAD_MUTEX_INITIALIZER; |
| |
| static void *track_alloc(const enum AllocationType type, char *ptr, |
| const size_t sz, const size_t align) |
| { |
| assert(align >= sizeof(Dav1dAllocationData)); |
| if (ptr) { |
| ptr += align; |
| Dav1dAllocationData *const d = &((Dav1dAllocationData*)ptr)[-1]; |
| AllocStats *const s = &tracked_allocs[type]; |
| |
| d->sz = sz; |
| d->align = (unsigned)align; |
| d->type = type; |
| |
| pthread_mutex_lock(&track_alloc_mutex); |
| s->num_allocs++; |
| s->curr_sz += sz; |
| if (s->curr_sz > s->peak_sz) |
| s->peak_sz = s->curr_sz; |
| |
| curr_total_sz += sz; |
| if (curr_total_sz > peak_total_sz) |
| peak_total_sz = curr_total_sz; |
| pthread_mutex_unlock(&track_alloc_mutex); |
| } |
| return ptr; |
| } |
| |
| static void *track_free(char *const ptr) { |
| const Dav1dAllocationData *const d = &((Dav1dAllocationData*)ptr)[-1]; |
| const size_t sz = d->sz; |
| |
| pthread_mutex_lock(&track_alloc_mutex); |
| tracked_allocs[d->type].curr_sz -= sz; |
| curr_total_sz -= sz; |
| pthread_mutex_unlock(&track_alloc_mutex); |
| |
| return ptr - d->align; |
| } |
| |
| static void dav1d_track_reuse(const enum AllocationType type) { |
| pthread_mutex_lock(&track_alloc_mutex); |
| tracked_allocs[type].num_reuses++; |
| pthread_mutex_unlock(&track_alloc_mutex); |
| } |
| |
| void *dav1d_malloc(const enum AllocationType type, const size_t sz) { |
| void *const ptr = malloc(sz + DEFAULT_ALIGN); |
| return track_alloc(type, ptr, sz, DEFAULT_ALIGN); |
| } |
| |
| void *dav1d_alloc_aligned(const enum AllocationType type, |
| const size_t sz, const size_t align) |
| { |
| void *const ptr = dav1d_alloc_aligned_internal(sz + align, align); |
| return track_alloc(type, ptr, sz, align); |
| } |
| |
| void *dav1d_realloc(const enum AllocationType type, |
| void *ptr, const size_t sz) |
| { |
| if (!ptr) |
| return dav1d_malloc(type, sz); |
| ptr = realloc((char*)ptr - DEFAULT_ALIGN, sz + DEFAULT_ALIGN); |
| if (ptr) |
| ptr = track_free((char*)ptr + DEFAULT_ALIGN); |
| return track_alloc(type, ptr, sz, DEFAULT_ALIGN); |
| } |
| |
| void dav1d_free(void *ptr) { |
| if (ptr) |
| free(track_free(ptr)); |
| } |
| |
| void dav1d_free_aligned(void *ptr) { |
| if (ptr) { |
| dav1d_free_aligned_internal(track_free(ptr)); |
| } |
| } |
| |
| static COLD int cmp_stats(const void *const a, const void *const b) { |
| const size_t a_sz = ((const AllocStats*)a)->peak_sz; |
| const size_t b_sz = ((const AllocStats*)b)->peak_sz; |
| return a_sz < b_sz ? -1 : a_sz > b_sz; |
| } |
| |
| /* Insert spaces as thousands separators for better readability */ |
| static COLD int format_tsep(char *const s, const size_t n, const size_t value) { |
| if (value < 1000) |
| return snprintf(s, n, "%u", (unsigned)value); |
| |
| const int len = format_tsep(s, n, value / 1000); |
| assert((size_t)len < n); |
| return len + snprintf(s + len, n - len, " %03u", (unsigned)(value % 1000)); |
| } |
| |
| COLD void dav1d_log_alloc_stats(Dav1dContext *const c) { |
| static const char *const type_names[N_ALLOC_TYPES] = { |
| [ALLOC_BLOCK ] = "Block data", |
| [ALLOC_CDEF ] = "CDEF line buffers", |
| [ALLOC_CDF ] = "CDF contexts", |
| [ALLOC_COEF ] = "Coefficient data", |
| [ALLOC_COMMON_CTX] = "Common context data", |
| [ALLOC_DAV1DDATA ] = "Dav1dData", |
| [ALLOC_IPRED ] = "Intra pred edges", |
| [ALLOC_LF ] = "Loopfilter data", |
| [ALLOC_LR ] = "Looprestoration data", |
| [ALLOC_OBU_HDR ] = "OBU headers", |
| [ALLOC_OBU_META ] = "OBU metadata", |
| [ALLOC_PAL ] = "Palette data", |
| [ALLOC_PIC ] = "Picture buffers", |
| [ALLOC_PIC_CTX ] = "Picture context data", |
| [ALLOC_REFMVS ] = "Reference mv data", |
| [ALLOC_SEGMAP ] = "Segmentation maps", |
| [ALLOC_THREAD_CTX] = "Thread context data", |
| [ALLOC_TILE ] = "Tile data", |
| }; |
| |
| struct { |
| AllocStats stats; |
| enum AllocationType type; |
| } data[N_ALLOC_TYPES]; |
| unsigned total_allocs = 0; |
| unsigned total_reuses = 0; |
| |
| pthread_mutex_lock(&track_alloc_mutex); |
| for (int i = 0; i < N_ALLOC_TYPES; i++) { |
| AllocStats *const s = &data[i].stats; |
| *s = tracked_allocs[i]; |
| data[i].type = i; |
| total_allocs += s->num_allocs; |
| total_reuses += s->num_reuses; |
| } |
| size_t total_sz = peak_total_sz; |
| pthread_mutex_unlock(&track_alloc_mutex); |
| |
| /* Sort types by memory usage */ |
| qsort(&data, N_ALLOC_TYPES, sizeof(*data), cmp_stats); |
| |
| const double inv_total_share = 100.0 / total_sz; |
| char total_sz_buf[32]; |
| const int sz_len = 4 + format_tsep(total_sz_buf, sizeof(total_sz_buf), total_sz); |
| |
| dav1d_log(c, "\n Type Allocs Reuses Share Peak size\n" |
| "---------------------------------------------------------------------\n"); |
| for (int i = N_ALLOC_TYPES - 1; i >= 0; i--) { |
| const AllocStats *const s = &data[i].stats; |
| if (s->num_allocs) { |
| const double share = s->peak_sz * inv_total_share; |
| char sz_buf[32]; |
| format_tsep(sz_buf, sizeof(sz_buf), s->peak_sz); |
| dav1d_log(c, " %-20s%10u%10u%8.1f%%%*s\n", type_names[data[i].type], |
| s->num_allocs, s->num_reuses, share, sz_len, sz_buf); |
| } |
| } |
| dav1d_log(c, "---------------------------------------------------------------------\n" |
| "%31u%10u %s\n", |
| total_allocs, total_reuses, total_sz_buf); |
| } |
| #endif /* TRACK_HEAP_ALLOCATIONS */ |
| |
| static COLD void mem_pool_destroy(Dav1dMemPool *const pool) { |
| pthread_mutex_destroy(&pool->lock); |
| dav1d_free(pool); |
| } |
| |
| void dav1d_mem_pool_push(Dav1dMemPool *const pool, void *const ptr) { |
| pthread_mutex_lock(&pool->lock); |
| Dav1dMemPoolBuffer *const buf = (Dav1dMemPoolBuffer*)((uintptr_t)ptr - 64); |
| const int ref_cnt = --pool->ref_cnt; |
| if (!pool->end) { |
| buf->next = pool->buf; |
| pool->buf = buf; |
| pthread_mutex_unlock(&pool->lock); |
| assert(ref_cnt > 0); |
| } else { |
| pthread_mutex_unlock(&pool->lock); |
| dav1d_free_aligned(buf); |
| if (!ref_cnt) mem_pool_destroy(pool); |
| } |
| } |
| |
| void *dav1d_mem_pool_pop(Dav1dMemPool *const pool, const size_t size) { |
| pthread_mutex_lock(&pool->lock); |
| Dav1dMemPoolBuffer *buf = pool->buf; |
| pool->ref_cnt++; |
| |
| if (buf) { |
| pool->buf = buf->next; |
| pthread_mutex_unlock(&pool->lock); |
| if (buf->size != size) { |
| /* Reallocate if the size has changed */ |
| dav1d_free_aligned(buf); |
| goto alloc; |
| } |
| #if TRACK_HEAP_ALLOCATIONS |
| dav1d_track_reuse(pool->type); |
| #endif |
| } else { |
| pthread_mutex_unlock(&pool->lock); |
| alloc: |
| buf = dav1d_alloc_aligned(pool->type, size + 64, 64); |
| if (!buf) { |
| pthread_mutex_lock(&pool->lock); |
| const int ref_cnt = --pool->ref_cnt; |
| pthread_mutex_unlock(&pool->lock); |
| if (!ref_cnt) mem_pool_destroy(pool); |
| return NULL; |
| } |
| buf->size = size; |
| } |
| |
| return (void*)((uintptr_t)buf + 64); |
| } |
| |
| COLD int dav1d_mem_pool_init(const enum AllocationType type, |
| Dav1dMemPool **const ppool) |
| { |
| Dav1dMemPool *const pool = dav1d_malloc(ALLOC_COMMON_CTX, |
| sizeof(Dav1dMemPool)); |
| if (pool) { |
| if (!pthread_mutex_init(&pool->lock, NULL)) { |
| pool->buf = NULL; |
| pool->ref_cnt = 1; |
| pool->end = 0; |
| #if TRACK_HEAP_ALLOCATIONS |
| pool->type = type; |
| #endif |
| *ppool = pool; |
| return 0; |
| } |
| dav1d_free(pool); |
| } |
| *ppool = NULL; |
| return DAV1D_ERR(ENOMEM); |
| } |
| |
| COLD void dav1d_mem_pool_end(Dav1dMemPool *const pool) { |
| if (pool) { |
| pthread_mutex_lock(&pool->lock); |
| Dav1dMemPoolBuffer *buf = pool->buf; |
| const int ref_cnt = --pool->ref_cnt; |
| pool->buf = NULL; |
| pool->end = 1; |
| pthread_mutex_unlock(&pool->lock); |
| |
| while (buf) { |
| void *const ptr = buf; |
| buf = buf->next; |
| dav1d_free_aligned(ptr); |
| } |
| if (!ref_cnt) mem_pool_destroy(pool); |
| } |
| } |