mirror of
https://git.ffmpeg.org/ffmpeg.git
synced 2024-12-27 18:02:11 +00:00
72becd0627
It was always intended that the buffers returned by
RefStruct shall have the same alignment guarantees
as the buffers returned by av_malloc(); said alignment
depends upon the arch and the enabled instruction set
and the code used STRIDE_ALIGN as a proxy for this.
Yet since 7945d30e91
there is a better way to get av_malloc's alignment:
ALIGN_64 in mem_internal.h. So use this.
Signed-off-by: Andreas Rheinhardt <andreas.rheinhardt@outlook.com>
387 lines
11 KiB
C
387 lines
11 KiB
C
/*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
#include <stdatomic.h>
|
|
#include <stdint.h>
|
|
#include <string.h>
|
|
|
|
#include "refstruct.h"
|
|
|
|
#include "libavutil/avassert.h"
|
|
#include "libavutil/error.h"
|
|
#include "libavutil/macros.h"
|
|
#include "libavutil/mem.h"
|
|
#include "libavutil/mem_internal.h"
|
|
#include "libavutil/thread.h"
|
|
|
|
#ifndef REFSTRUCT_CHECKED
|
|
#ifndef ASSERT_LEVEL
|
|
#define ASSERT_LEVEL 0
|
|
#endif
|
|
#define REFSTRUCT_CHECKED (ASSERT_LEVEL >= 1)
|
|
#endif
|
|
|
|
#if REFSTRUCT_CHECKED
|
|
#define ff_assert(cond) av_assert0(cond)
|
|
#else
|
|
#define ff_assert(cond) ((void)0)
|
|
#endif
|
|
|
|
#define REFSTRUCT_COOKIE AV_NE((uint64_t)MKBETAG('R', 'e', 'f', 'S') << 32 | MKBETAG('t', 'r', 'u', 'c'), \
|
|
MKTAG('R', 'e', 'f', 'S') | (uint64_t)MKTAG('t', 'r', 'u', 'c') << 32)
|
|
|
|
#if __STDC_VERSION__ >= 201112L && !defined(_MSC_VER)
|
|
#define REFCOUNT_OFFSET FFALIGN(sizeof(RefCount), FFMAX(ALIGN_64, _Alignof(max_align_t)))
|
|
#else
|
|
#define REFCOUNT_OFFSET FFALIGN(sizeof(RefCount), ALIGN_64)
|
|
#endif
|
|
|
|
typedef struct RefCount {
|
|
/**
|
|
* An uintptr_t is big enough to hold the address of every reference,
|
|
* so no overflow can happen when incrementing the refcount as long as
|
|
* the user does not throw away references.
|
|
*/
|
|
atomic_uintptr_t refcount;
|
|
FFRefStructOpaque opaque;
|
|
void (*free_cb)(FFRefStructOpaque opaque, void *obj);
|
|
void (*free)(void *ref);
|
|
|
|
#if REFSTRUCT_CHECKED
|
|
uint64_t cookie;
|
|
#endif
|
|
} RefCount;
|
|
|
|
static RefCount *get_refcount(void *obj)
|
|
{
|
|
RefCount *ref = (RefCount*)((char*)obj - REFCOUNT_OFFSET);
|
|
ff_assert(ref->cookie == REFSTRUCT_COOKIE);
|
|
return ref;
|
|
}
|
|
|
|
static const RefCount *cget_refcount(const void *obj)
|
|
{
|
|
const RefCount *ref = (const RefCount*)((const char*)obj - REFCOUNT_OFFSET);
|
|
ff_assert(ref->cookie == REFSTRUCT_COOKIE);
|
|
return ref;
|
|
}
|
|
|
|
static void *get_userdata(void *buf)
|
|
{
|
|
return (char*)buf + REFCOUNT_OFFSET;
|
|
}
|
|
|
|
static void refcount_init(RefCount *ref, FFRefStructOpaque opaque,
|
|
void (*free_cb)(FFRefStructOpaque opaque, void *obj))
|
|
{
|
|
atomic_init(&ref->refcount, 1);
|
|
ref->opaque = opaque;
|
|
ref->free_cb = free_cb;
|
|
ref->free = av_free;
|
|
|
|
#if REFSTRUCT_CHECKED
|
|
ref->cookie = REFSTRUCT_COOKIE;
|
|
#endif
|
|
}
|
|
|
|
void *ff_refstruct_alloc_ext_c(size_t size, unsigned flags, FFRefStructOpaque opaque,
|
|
void (*free_cb)(FFRefStructOpaque opaque, void *obj))
|
|
{
|
|
void *buf, *obj;
|
|
|
|
if (size > SIZE_MAX - REFCOUNT_OFFSET)
|
|
return NULL;
|
|
buf = av_malloc(size + REFCOUNT_OFFSET);
|
|
if (!buf)
|
|
return NULL;
|
|
refcount_init(buf, opaque, free_cb);
|
|
obj = get_userdata(buf);
|
|
if (!(flags & FF_REFSTRUCT_FLAG_NO_ZEROING))
|
|
memset(obj, 0, size);
|
|
|
|
return obj;
|
|
}
|
|
|
|
void ff_refstruct_unref(void *objp)
|
|
{
|
|
void *obj;
|
|
RefCount *ref;
|
|
|
|
memcpy(&obj, objp, sizeof(obj));
|
|
if (!obj)
|
|
return;
|
|
memcpy(objp, &(void *){ NULL }, sizeof(obj));
|
|
|
|
ref = get_refcount(obj);
|
|
if (atomic_fetch_sub_explicit(&ref->refcount, 1, memory_order_acq_rel) == 1) {
|
|
if (ref->free_cb)
|
|
ref->free_cb(ref->opaque, obj);
|
|
ref->free(ref);
|
|
}
|
|
|
|
return;
|
|
}
|
|
|
|
void *ff_refstruct_ref(void *obj)
|
|
{
|
|
RefCount *ref = get_refcount(obj);
|
|
|
|
atomic_fetch_add_explicit(&ref->refcount, 1, memory_order_relaxed);
|
|
|
|
return obj;
|
|
}
|
|
|
|
const void *ff_refstruct_ref_c(const void *obj)
|
|
{
|
|
/* Casting const away here is fine, as it is only supposed
|
|
* to apply to the user's data and not our bookkeeping data. */
|
|
RefCount *ref = get_refcount((void*)obj);
|
|
|
|
atomic_fetch_add_explicit(&ref->refcount, 1, memory_order_relaxed);
|
|
|
|
return obj;
|
|
}
|
|
|
|
void ff_refstruct_replace(void *dstp, const void *src)
|
|
{
|
|
const void *dst;
|
|
memcpy(&dst, dstp, sizeof(dst));
|
|
|
|
if (src == dst)
|
|
return;
|
|
ff_refstruct_unref(dstp);
|
|
if (src) {
|
|
dst = ff_refstruct_ref_c(src);
|
|
memcpy(dstp, &dst, sizeof(dst));
|
|
}
|
|
}
|
|
|
|
int ff_refstruct_exclusive(const void *obj)
|
|
{
|
|
const RefCount *ref = cget_refcount(obj);
|
|
/* Casting const away here is safe, because it is a load.
|
|
* It is necessary because atomic_load_explicit() does not
|
|
* accept const atomics in C11 (see also N1807). */
|
|
return atomic_load_explicit((atomic_uintptr_t*)&ref->refcount, memory_order_acquire) == 1;
|
|
}
|
|
|
|
struct FFRefStructPool {
|
|
size_t size;
|
|
FFRefStructOpaque opaque;
|
|
int (*init_cb)(FFRefStructOpaque opaque, void *obj);
|
|
void (*reset_cb)(FFRefStructOpaque opaque, void *obj);
|
|
void (*free_entry_cb)(FFRefStructOpaque opaque, void *obj);
|
|
void (*free_cb)(FFRefStructOpaque opaque);
|
|
|
|
int uninited;
|
|
unsigned entry_flags;
|
|
unsigned pool_flags;
|
|
|
|
/** The number of outstanding entries not in available_entries. */
|
|
atomic_uintptr_t refcount;
|
|
/**
|
|
* This is a linked list of available entries;
|
|
* the RefCount's opaque pointer is used as next pointer
|
|
* for available entries.
|
|
* While the entries are in use, the opaque is a pointer
|
|
* to the corresponding FFRefStructPool.
|
|
*/
|
|
RefCount *available_entries;
|
|
AVMutex mutex;
|
|
};
|
|
|
|
static void pool_free(FFRefStructPool *pool)
|
|
{
|
|
ff_mutex_destroy(&pool->mutex);
|
|
if (pool->free_cb)
|
|
pool->free_cb(pool->opaque);
|
|
av_free(get_refcount(pool));
|
|
}
|
|
|
|
static void pool_free_entry(FFRefStructPool *pool, RefCount *ref)
|
|
{
|
|
if (pool->free_entry_cb)
|
|
pool->free_entry_cb(pool->opaque, get_userdata(ref));
|
|
av_free(ref);
|
|
}
|
|
|
|
static void pool_return_entry(void *ref_)
|
|
{
|
|
RefCount *ref = ref_;
|
|
FFRefStructPool *pool = ref->opaque.nc;
|
|
|
|
ff_mutex_lock(&pool->mutex);
|
|
if (!pool->uninited) {
|
|
ref->opaque.nc = pool->available_entries;
|
|
pool->available_entries = ref;
|
|
ref = NULL;
|
|
}
|
|
ff_mutex_unlock(&pool->mutex);
|
|
|
|
if (ref)
|
|
pool_free_entry(pool, ref);
|
|
|
|
if (atomic_fetch_sub_explicit(&pool->refcount, 1, memory_order_acq_rel) == 1)
|
|
pool_free(pool);
|
|
}
|
|
|
|
static void pool_reset_entry(FFRefStructOpaque opaque, void *entry)
|
|
{
|
|
FFRefStructPool *pool = opaque.nc;
|
|
|
|
pool->reset_cb(pool->opaque, entry);
|
|
}
|
|
|
|
static int refstruct_pool_get_ext(void *datap, FFRefStructPool *pool)
|
|
{
|
|
void *ret = NULL;
|
|
|
|
memcpy(datap, &(void *){ NULL }, sizeof(void*));
|
|
|
|
ff_mutex_lock(&pool->mutex);
|
|
ff_assert(!pool->uninited);
|
|
if (pool->available_entries) {
|
|
RefCount *ref = pool->available_entries;
|
|
ret = get_userdata(ref);
|
|
pool->available_entries = ref->opaque.nc;
|
|
ref->opaque.nc = pool;
|
|
atomic_init(&ref->refcount, 1);
|
|
}
|
|
ff_mutex_unlock(&pool->mutex);
|
|
|
|
if (!ret) {
|
|
RefCount *ref;
|
|
ret = ff_refstruct_alloc_ext(pool->size, pool->entry_flags, pool,
|
|
pool->reset_cb ? pool_reset_entry : NULL);
|
|
if (!ret)
|
|
return AVERROR(ENOMEM);
|
|
ref = get_refcount(ret);
|
|
ref->free = pool_return_entry;
|
|
if (pool->init_cb) {
|
|
int err = pool->init_cb(pool->opaque, ret);
|
|
if (err < 0) {
|
|
if (pool->pool_flags & FF_REFSTRUCT_POOL_FLAG_RESET_ON_INIT_ERROR)
|
|
pool->reset_cb(pool->opaque, ret);
|
|
if (pool->pool_flags & FF_REFSTRUCT_POOL_FLAG_FREE_ON_INIT_ERROR)
|
|
pool->free_entry_cb(pool->opaque, ret);
|
|
av_free(ref);
|
|
return err;
|
|
}
|
|
}
|
|
}
|
|
atomic_fetch_add_explicit(&pool->refcount, 1, memory_order_relaxed);
|
|
|
|
if (pool->pool_flags & FF_REFSTRUCT_POOL_FLAG_ZERO_EVERY_TIME)
|
|
memset(ret, 0, pool->size);
|
|
|
|
memcpy(datap, &ret, sizeof(ret));
|
|
|
|
return 0;
|
|
}
|
|
|
|
void *ff_refstruct_pool_get(FFRefStructPool *pool)
|
|
{
|
|
void *ret;
|
|
refstruct_pool_get_ext(&ret, pool);
|
|
return ret;
|
|
}
|
|
|
|
/**
|
|
* Hint: The content of pool_unref() and refstruct_pool_uninit()
|
|
* could currently be merged; they are only separate functions
|
|
* in case we would ever introduce weak references.
|
|
*/
|
|
static void pool_unref(void *ref)
|
|
{
|
|
FFRefStructPool *pool = get_userdata(ref);
|
|
if (atomic_fetch_sub_explicit(&pool->refcount, 1, memory_order_acq_rel) == 1)
|
|
pool_free(pool);
|
|
}
|
|
|
|
static void refstruct_pool_uninit(FFRefStructOpaque unused, void *obj)
|
|
{
|
|
FFRefStructPool *pool = obj;
|
|
RefCount *entry;
|
|
|
|
ff_mutex_lock(&pool->mutex);
|
|
ff_assert(!pool->uninited);
|
|
pool->uninited = 1;
|
|
entry = pool->available_entries;
|
|
pool->available_entries = NULL;
|
|
ff_mutex_unlock(&pool->mutex);
|
|
|
|
while (entry) {
|
|
void *next = entry->opaque.nc;
|
|
pool_free_entry(pool, entry);
|
|
entry = next;
|
|
}
|
|
}
|
|
|
|
FFRefStructPool *ff_refstruct_pool_alloc(size_t size, unsigned flags)
|
|
{
|
|
return ff_refstruct_pool_alloc_ext(size, flags, NULL, NULL, NULL, NULL, NULL);
|
|
}
|
|
|
|
FFRefStructPool *ff_refstruct_pool_alloc_ext_c(size_t size, unsigned flags,
|
|
FFRefStructOpaque opaque,
|
|
int (*init_cb)(FFRefStructOpaque opaque, void *obj),
|
|
void (*reset_cb)(FFRefStructOpaque opaque, void *obj),
|
|
void (*free_entry_cb)(FFRefStructOpaque opaque, void *obj),
|
|
void (*free_cb)(FFRefStructOpaque opaque))
|
|
{
|
|
FFRefStructPool *pool = ff_refstruct_alloc_ext(sizeof(*pool), 0, NULL,
|
|
refstruct_pool_uninit);
|
|
int err;
|
|
|
|
if (!pool)
|
|
return NULL;
|
|
get_refcount(pool)->free = pool_unref;
|
|
|
|
pool->size = size;
|
|
pool->opaque = opaque;
|
|
pool->init_cb = init_cb;
|
|
pool->reset_cb = reset_cb;
|
|
pool->free_entry_cb = free_entry_cb;
|
|
pool->free_cb = free_cb;
|
|
#define COMMON_FLAGS FF_REFSTRUCT_POOL_FLAG_NO_ZEROING
|
|
pool->entry_flags = flags & COMMON_FLAGS;
|
|
// Filter out nonsense combinations to avoid checks later.
|
|
if (!pool->reset_cb)
|
|
flags &= ~FF_REFSTRUCT_POOL_FLAG_RESET_ON_INIT_ERROR;
|
|
if (!pool->free_entry_cb)
|
|
flags &= ~FF_REFSTRUCT_POOL_FLAG_FREE_ON_INIT_ERROR;
|
|
pool->pool_flags = flags;
|
|
|
|
if (flags & FF_REFSTRUCT_POOL_FLAG_ZERO_EVERY_TIME) {
|
|
// We will zero the buffer before every use, so zeroing
|
|
// upon allocating the buffer is unnecessary.
|
|
pool->entry_flags |= FF_REFSTRUCT_FLAG_NO_ZEROING;
|
|
}
|
|
|
|
atomic_init(&pool->refcount, 1);
|
|
|
|
err = ff_mutex_init(&pool->mutex, NULL);
|
|
if (err) {
|
|
// Don't call ff_refstruct_uninit() on pool, as it hasn't been properly
|
|
// set up and is just a POD right now.
|
|
av_free(get_refcount(pool));
|
|
return NULL;
|
|
}
|
|
return pool;
|
|
}
|