ffmpeg/libavdevice/libndi_newtek_dec.c

341 lines
11 KiB
C
Raw Normal View History

/*
* Newtek NDI input
* Copyright (c) 2017 Maksym Veremeyenko
*
* This file is part of FFmpeg.
*
* FFmpeg is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* FFmpeg is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with FFmpeg; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
#include "libavformat/avformat.h"
#include "libavformat/internal.h"
#include "libavutil/opt.h"
#include "libavutil/imgutils.h"
#include "libndi_newtek_common.h"
struct NDIContext {
const AVClass *cclass;
/* Options */
int find_sources;
int64_t wait_sources;
int allow_video_fields;
/* Runtime */
NDIlib_recv_create_t *recv;
NDIlib_find_instance_t ndi_find;
/* Streams */
AVStream *video_st, *audio_st;
};
static int ndi_set_video_packet(AVFormatContext *avctx, NDIlib_video_frame_t *v, AVPacket *pkt)
{
int ret;
struct NDIContext *ctx = avctx->priv_data;
ret = av_new_packet(pkt, v->yres * v->line_stride_in_bytes);
if (ret < 0)
return ret;
pkt->dts = pkt->pts = av_rescale_q(v->timecode, NDI_TIME_BASE_Q, ctx->video_st->time_base);
pkt->duration = av_rescale_q(1, (AVRational){v->frame_rate_D, v->frame_rate_N}, ctx->video_st->time_base);
av_log(avctx, AV_LOG_DEBUG, "%s: pkt->dts = pkt->pts = %"PRId64", duration=%"PRId64", timecode=%"PRId64"\n",
__func__, pkt->dts, pkt->duration, v->timecode);
pkt->flags |= AV_PKT_FLAG_KEY;
pkt->stream_index = ctx->video_st->index;
memcpy(pkt->data, v->p_data, pkt->size);
return 0;
}
static int ndi_set_audio_packet(AVFormatContext *avctx, NDIlib_audio_frame_t *a, AVPacket *pkt)
{
int ret;
struct NDIContext *ctx = avctx->priv_data;
NDIlib_audio_frame_interleaved_16s_t dst;
ret = av_new_packet(pkt, 2 * a->no_samples * a->no_channels);
if (ret < 0)
return ret;
pkt->dts = pkt->pts = av_rescale_q(a->timecode, NDI_TIME_BASE_Q, ctx->audio_st->time_base);
pkt->duration = av_rescale_q(1, (AVRational){a->no_samples, a->sample_rate}, ctx->audio_st->time_base);
av_log(avctx, AV_LOG_DEBUG, "%s: pkt->dts = pkt->pts = %"PRId64", duration=%"PRId64", timecode=%"PRId64"\n",
__func__, pkt->dts, pkt->duration, a->timecode);
pkt->flags |= AV_PKT_FLAG_KEY;
pkt->stream_index = ctx->audio_st->index;
dst.reference_level = 0;
dst.p_data = (short *)pkt->data;
NDIlib_util_audio_to_interleaved_16s(a, &dst);
return 0;
}
static int ndi_find_sources(AVFormatContext *avctx, const char *name, NDIlib_source_t *source_to_connect_to)
{
int j = AVERROR(ENODEV);
unsigned int n, i;
struct NDIContext *ctx = avctx->priv_data;
const NDIlib_source_t *ndi_srcs = NULL;
const NDIlib_find_create_t find_create_desc = { .show_local_sources = true,
.p_groups = NULL, .p_extra_ips = NULL };
if (!ctx->ndi_find)
ctx->ndi_find = NDIlib_find_create2(&find_create_desc);
if (!ctx->ndi_find) {
av_log(avctx, AV_LOG_ERROR, "NDIlib_find_create failed.\n");
return AVERROR(EIO);
}
while (1)
{
int f, t = ctx->wait_sources / 1000;
av_log(avctx, AV_LOG_DEBUG, "Waiting for sources %d miliseconds\n", t);
f = NDIlib_find_wait_for_sources(ctx->ndi_find, t);
av_log(avctx, AV_LOG_DEBUG, "NDIlib_find_wait_for_sources returns %d\n", f);
if (!f)
break;
};
ndi_srcs = NDIlib_find_get_current_sources(ctx->ndi_find, &n);
if (ctx->find_sources)
av_log(avctx, AV_LOG_INFO, "Found %d NDI sources:\n", n);
for (i = 0; i < n; i++) {
if (ctx->find_sources)
av_log(avctx, AV_LOG_INFO, "\t'%s'\t'%s'\n", ndi_srcs[i].p_ndi_name, ndi_srcs[i].p_ip_address);
if (!strcmp(name, ndi_srcs[i].p_ndi_name)) {
*source_to_connect_to = ndi_srcs[i];
j = i;
}
}
return j;
}
static int ndi_read_header(AVFormatContext *avctx)
{
int ret;
NDIlib_recv_create_t recv_create_desc;
const NDIlib_tally_t tally_state = { .on_program = true, .on_preview = false };
struct NDIContext *ctx = avctx->priv_data;
if (!NDIlib_initialize()) {
av_log(avctx, AV_LOG_ERROR, "NDIlib_initialize failed.\n");
return AVERROR_EXTERNAL;
}
/* Find available sources. */
ret = ndi_find_sources(avctx, avctx->filename, &recv_create_desc.source_to_connect_to);
if (ctx->find_sources) {
return AVERROR_EXIT;
}
if (ret < 0)
return ret;
/* Create receiver description */
recv_create_desc.color_format = NDIlib_recv_color_format_e_UYVY_RGBA;
recv_create_desc.bandwidth = NDIlib_recv_bandwidth_highest;
recv_create_desc.allow_video_fields = ctx->allow_video_fields;
/* Create the receiver */
ctx->recv = NDIlib_recv_create(&recv_create_desc);
if (!ctx->recv) {
av_log(avctx, AV_LOG_ERROR, "NDIlib_recv_create2 failed.\n");
return AVERROR(EIO);
}
/* Set tally */
NDIlib_recv_set_tally(ctx->recv, &tally_state);
avctx->ctx_flags |= AVFMTCTX_NOHEADER;
return 0;
}
static int ndi_create_video_stream(AVFormatContext *avctx, NDIlib_video_frame_t *v)
{
AVStream *st;
AVRational tmp;
struct NDIContext *ctx = avctx->priv_data;
st = avformat_new_stream(avctx, NULL);
if (!st) {
av_log(avctx, AV_LOG_ERROR, "Cannot add video stream\n");
return AVERROR(ENOMEM);
}
st->time_base = NDI_TIME_BASE_Q;
st->r_frame_rate = av_make_q(v->frame_rate_N, v->frame_rate_D);
tmp = av_mul_q(av_d2q(v->picture_aspect_ratio, INT_MAX), (AVRational){v->yres, v->xres});
av_reduce(&st->sample_aspect_ratio.num, &st->sample_aspect_ratio.den, tmp.num, tmp.den, 1000);
st->codecpar->sample_aspect_ratio = st->sample_aspect_ratio;
st->codecpar->codec_type = AVMEDIA_TYPE_VIDEO;
st->codecpar->width = v->xres;
st->codecpar->height = v->yres;
st->codecpar->codec_id = AV_CODEC_ID_RAWVIDEO;
st->codecpar->bit_rate = av_rescale(v->xres * v->yres * 16, v->frame_rate_N, v->frame_rate_D);
st->codecpar->field_order = v->frame_format_type == NDIlib_frame_format_type_progressive
? AV_FIELD_PROGRESSIVE : AV_FIELD_TT;
if (NDIlib_FourCC_type_UYVY == v->FourCC || NDIlib_FourCC_type_UYVA == v->FourCC) {
st->codecpar->format = AV_PIX_FMT_UYVY422;
st->codecpar->codec_tag = MKTAG('U', 'Y', 'V', 'Y');
if (NDIlib_FourCC_type_UYVA == v->FourCC)
av_log(avctx, AV_LOG_WARNING, "Alpha channel ignored\n");
} else if (NDIlib_FourCC_type_BGRA == v->FourCC) {
st->codecpar->format = AV_PIX_FMT_BGRA;
st->codecpar->codec_tag = MKTAG('B', 'G', 'R', 'A');
} else if (NDIlib_FourCC_type_BGRX == v->FourCC) {
st->codecpar->format = AV_PIX_FMT_BGR0;
st->codecpar->codec_tag = MKTAG('B', 'G', 'R', '0');
} else if (NDIlib_FourCC_type_RGBA == v->FourCC) {
st->codecpar->format = AV_PIX_FMT_RGBA;
st->codecpar->codec_tag = MKTAG('R', 'G', 'B', 'A');
} else if (NDIlib_FourCC_type_RGBX == v->FourCC) {
st->codecpar->format = AV_PIX_FMT_RGB0;
st->codecpar->codec_tag = MKTAG('R', 'G', 'B', '0');
} else {
av_log(avctx, AV_LOG_ERROR, "Unsupported video stream format, v->FourCC=%d\n", v->FourCC);
return AVERROR(EINVAL);
}
avpriv_set_pts_info(st, 64, 1, NDI_TIME_BASE);
ctx->video_st = st;
return 0;
}
static int ndi_create_audio_stream(AVFormatContext *avctx, NDIlib_audio_frame_t *a)
{
AVStream *st;
struct NDIContext *ctx = avctx->priv_data;
st = avformat_new_stream(avctx, NULL);
if (!st) {
av_log(avctx, AV_LOG_ERROR, "Cannot add audio stream\n");
return AVERROR(ENOMEM);
}
st->codecpar->codec_type = AVMEDIA_TYPE_AUDIO;
st->codecpar->codec_id = AV_CODEC_ID_PCM_S16LE;
st->codecpar->sample_rate = a->sample_rate;
st->codecpar->channels = a->no_channels;
avpriv_set_pts_info(st, 64, 1, NDI_TIME_BASE);
ctx->audio_st = st;
return 0;
}
static int ndi_read_packet(AVFormatContext *avctx, AVPacket *pkt)
{
int ret = 0;
struct NDIContext *ctx = avctx->priv_data;
while (!ret) {
NDIlib_video_frame_t v;
NDIlib_audio_frame_t a;
NDIlib_metadata_frame_t m;
NDIlib_frame_type_e t;
av_log(avctx, AV_LOG_DEBUG, "NDIlib_recv_capture...\n");
t = NDIlib_recv_capture(ctx->recv, &v, &a, &m, 40);
av_log(avctx, AV_LOG_DEBUG, "NDIlib_recv_capture=%d\n", t);
if (t == NDIlib_frame_type_video) {
if (!ctx->video_st)
ret = ndi_create_video_stream(avctx, &v);
if (!ret)
ret = ndi_set_video_packet(avctx, &v, pkt);
NDIlib_recv_free_video(ctx->recv, &v);
break;
}
else if (t == NDIlib_frame_type_audio) {
if (!ctx->audio_st)
ret = ndi_create_audio_stream(avctx, &a);
if (!ret)
ret = ndi_set_audio_packet(avctx, &a, pkt);
NDIlib_recv_free_audio(ctx->recv, &a);
break;
}
else if (t == NDIlib_frame_type_metadata)
NDIlib_recv_free_metadata(ctx->recv, &m);
else if (t == NDIlib_frame_type_error){
av_log(avctx, AV_LOG_ERROR, "NDIlib_recv_capture failed with error\n");
ret = AVERROR(EIO);
}
};
return ret;
}
static int ndi_read_close(AVFormatContext *avctx)
{
struct NDIContext *ctx = (struct NDIContext *)avctx->priv_data;
if (ctx->recv)
NDIlib_recv_destroy(ctx->recv);
if (ctx->ndi_find)
NDIlib_find_destroy(ctx->ndi_find);
return 0;
}
#define OFFSET(x) offsetof(struct NDIContext, x)
#define DEC AV_OPT_FLAG_DECODING_PARAM
static const AVOption options[] = {
{ "find_sources", "Find available sources" , OFFSET(find_sources), AV_OPT_TYPE_BOOL, { .i64 = 0 }, 0, 1, DEC },
{ "wait_sources", "Time to wait until the number of online sources have changed" , OFFSET(wait_sources), AV_OPT_TYPE_DURATION, { .i64 = 1000000 }, 100000, 20000000, DEC },
{ "allow_video_fields", "When this flag is FALSE, all video that you receive will be progressive" , OFFSET(allow_video_fields), AV_OPT_TYPE_BOOL, { .i64 = 1 }, 0, 1, DEC },
{ NULL },
};
static const AVClass libndi_newtek_demuxer_class = {
.class_name = "NDI demuxer",
.item_name = av_default_item_name,
.option = options,
.version = LIBAVUTIL_VERSION_INT,
.category = AV_CLASS_CATEGORY_DEVICE_VIDEO_INPUT,
};
AVInputFormat ff_libndi_newtek_demuxer = {
.name = "libndi_newtek",
.long_name = NULL_IF_CONFIG_SMALL("Network Device Interface (NDI) input using NewTek library"),
.flags = AVFMT_NOFILE,
.priv_class = &libndi_newtek_demuxer_class,
.priv_data_size = sizeof(struct NDIContext),
.read_header = ndi_read_header,
.read_packet = ndi_read_packet,
.read_close = ndi_read_close,
};