/* * Blackmagic DeckLink input * Copyright (c) 2013-2014 Luca Barbato, Deti Fliegl * Copyright (c) 2014 Rafaël Carré * Copyright (c) 2017 Akamai Technologies, Inc. * * This file is part of FFmpeg. * * FFmpeg is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2.1 of the License, or (at your option) any later version. * * FFmpeg is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with FFmpeg; if not, write to the Free Software * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA */ #include using std::atomic; /* Include internal.h first to avoid conflict between winsock.h (used by * DeckLink headers) and winsock2.h (used by libavformat) in MSVC++ builds */ extern "C" { #include "libavformat/internal.h" } #include extern "C" { #include "config.h" #include "libavformat/avformat.h" #include "libavutil/avassert.h" #include "libavutil/avutil.h" #include "libavutil/common.h" #include "libavutil/imgutils.h" #include "libavutil/intreadwrite.h" #include "libavutil/time.h" #include "libavutil/mathematics.h" #include "libavutil/reverse.h" #include "avdevice.h" #if CONFIG_LIBZVBI #include #endif } #include "decklink_common.h" #include "decklink_dec.h" #define MAX_WIDTH_VANC 1920 const BMDDisplayMode AUTODETECT_DEFAULT_MODE = bmdModeNTSC; typedef struct VANCLineNumber { BMDDisplayMode mode; int vanc_start; int field0_vanc_end; int field1_vanc_start; int vanc_end; } VANCLineNumber; /* These VANC line numbers need not be very accurate. In any case * GetBufferForVerticalBlankingLine() will return an error when invalid * ancillary line number was requested. We just need to make sure that the * entire VANC region is covered, while making sure we don't decode VANC of * another source during switching*/ static VANCLineNumber vanc_line_numbers[] = { /* SD Modes */ {bmdModeNTSC, 11, 19, 274, 282}, {bmdModeNTSC2398, 11, 19, 274, 282}, {bmdModePAL, 7, 22, 320, 335}, {bmdModeNTSCp, 11, -1, -1, 39}, {bmdModePALp, 7, -1, -1, 45}, /* HD 1080 Modes */ {bmdModeHD1080p2398, 8, -1, -1, 42}, {bmdModeHD1080p24, 8, -1, -1, 42}, {bmdModeHD1080p25, 8, -1, -1, 42}, {bmdModeHD1080p2997, 8, -1, -1, 42}, {bmdModeHD1080p30, 8, -1, -1, 42}, {bmdModeHD1080i50, 8, 20, 570, 585}, {bmdModeHD1080i5994, 8, 20, 570, 585}, {bmdModeHD1080i6000, 8, 20, 570, 585}, {bmdModeHD1080p50, 8, -1, -1, 42}, {bmdModeHD1080p5994, 8, -1, -1, 42}, {bmdModeHD1080p6000, 8, -1, -1, 42}, /* HD 720 Modes */ {bmdModeHD720p50, 8, -1, -1, 26}, {bmdModeHD720p5994, 8, -1, -1, 26}, {bmdModeHD720p60, 8, -1, -1, 26}, /* For all other modes, for which we don't support VANC */ {bmdModeUnknown, 0, -1, -1, -1} }; class decklink_allocator : public IDeckLinkMemoryAllocator { public: decklink_allocator(): _refs(1) { } virtual ~decklink_allocator() { } // IDeckLinkMemoryAllocator methods virtual HRESULT STDMETHODCALLTYPE AllocateBuffer(unsigned int bufferSize, void* *allocatedBuffer) { void *buf = av_malloc(bufferSize + AV_INPUT_BUFFER_PADDING_SIZE); if (!buf) return E_OUTOFMEMORY; *allocatedBuffer = buf; return S_OK; } virtual HRESULT STDMETHODCALLTYPE ReleaseBuffer(void* buffer) { av_free(buffer); return S_OK; } virtual HRESULT STDMETHODCALLTYPE Commit() { return S_OK; } virtual HRESULT STDMETHODCALLTYPE Decommit() { return S_OK; } // IUnknown methods virtual HRESULT STDMETHODCALLTYPE QueryInterface(REFIID iid, LPVOID *ppv) { return E_NOINTERFACE; } virtual ULONG STDMETHODCALLTYPE AddRef(void) { return ++_refs; } virtual ULONG STDMETHODCALLTYPE Release(void) { int ret = --_refs; if (!ret) delete this; return ret; } private: std::atomic _refs; }; extern "C" { static void decklink_object_free(void *opaque, uint8_t *data) { IUnknown *obj = (class IUnknown *)opaque; obj->Release(); } } static int get_vanc_line_idx(BMDDisplayMode mode) { unsigned int i; for (i = 0; i < FF_ARRAY_ELEMS(vanc_line_numbers); i++) { if (mode == vanc_line_numbers[i].mode) return i; } /* Return the VANC idx for Unknown mode */ return i - 1; } static inline void clear_parity_bits(uint16_t *buf, int len) { int i; for (i = 0; i < len; i++) buf[i] &= 0xff; } static int check_vanc_parity_checksum(uint16_t *buf, int len, uint16_t checksum) { int i; uint16_t vanc_sum = 0; for (i = 3; i < len - 1; i++) { uint16_t v = buf[i]; int np = v >> 8; int p = av_parity(v & 0xff); if ((!!p ^ !!(v & 0x100)) || (np != 1 && np != 2)) { // Parity check failed return -1; } vanc_sum += v; } vanc_sum &= 0x1ff; vanc_sum |= ((~vanc_sum & 0x100) << 1); if (checksum != vanc_sum) { // Checksum verification failed return -1; } return 0; } /* The 10-bit VANC data is packed in V210, we only need the luma component. */ static void extract_luma_from_v210(uint16_t *dst, const uint8_t *src, int width) { int i; for (i = 0; i < width / 3; i++) { *dst++ = (src[1] >> 2) + ((src[2] & 15) << 6); *dst++ = src[4] + ((src[5] & 3) << 8); *dst++ = (src[6] >> 4) + ((src[7] & 63) << 4); src += 8; } } static void unpack_v210(uint16_t *dst, const uint8_t *src, int width) { int i; for (i = 0; i < width * 2 / 3; i++) { *dst++ = src[0] + ((src[1] & 3) << 8); *dst++ = (src[1] >> 2) + ((src[2] & 15) << 6); *dst++ = (src[2] >> 4) + ((src[3] & 63) << 4); src += 4; } } static uint8_t calc_parity_and_line_offset(int line) { uint8_t ret = (line < 313) << 5; if (line >= 7 && line <= 22) ret += line; if (line >= 320 && line <= 335) ret += (line - 313); return ret; } static void fill_data_unit_head(int line, uint8_t *tgt) { tgt[0] = 0x02; // data_unit_id tgt[1] = 0x2c; // data_unit_length tgt[2] = calc_parity_and_line_offset(line); // field_parity, line_offset tgt[3] = 0xe4; // framing code } #if CONFIG_LIBZVBI static uint8_t* teletext_data_unit_from_vbi_data(int line, uint8_t *src, uint8_t *tgt, vbi_pixfmt fmt) { vbi_bit_slicer slicer; vbi_bit_slicer_init(&slicer, 720, 13500000, 6937500, 6937500, 0x00aaaae4, 0xffff, 18, 6, 42 * 8, VBI_MODULATION_NRZ_MSB, fmt); if (vbi_bit_slice(&slicer, src, tgt + 4) == FALSE) return tgt; fill_data_unit_head(line, tgt); return tgt + 46; } static uint8_t* teletext_data_unit_from_vbi_data_10bit(int line, uint8_t *src, uint8_t *tgt) { uint8_t y[720]; uint8_t *py = y; uint8_t *pend = y + 720; /* The 10-bit VBI data is packed in V210, but libzvbi only supports 8-bit, * so we extract the 8 MSBs of the luma component, that is enough for * teletext bit slicing. */ while (py < pend) { *py++ = (src[1] >> 4) + ((src[2] & 15) << 4); *py++ = (src[4] >> 2) + ((src[5] & 3 ) << 6); *py++ = (src[6] >> 6) + ((src[7] & 63) << 2); src += 8; } return teletext_data_unit_from_vbi_data(line, y, tgt, VBI_PIXFMT_YUV420); } #endif static uint8_t* teletext_data_unit_from_op47_vbi_packet(int line, uint16_t *py, uint8_t *tgt) { int i; if (py[0] != 0x255 || py[1] != 0x255 || py[2] != 0x227) return tgt; fill_data_unit_head(line, tgt); py += 3; tgt += 4; for (i = 0; i < 42; i++) *tgt++ = ff_reverse[py[i] & 255]; return tgt; } static int linemask_matches(int line, int64_t mask) { int shift = -1; if (line >= 6 && line <= 22) shift = line - 6; if (line >= 318 && line <= 335) shift = line - 318 + 17; return shift >= 0 && ((1ULL << shift) & mask); } static uint8_t* teletext_data_unit_from_op47_data(uint16_t *py, uint16_t *pend, uint8_t *tgt, int64_t wanted_lines) { if (py < pend - 9) { if (py[0] == 0x151 && py[1] == 0x115 && py[3] == 0x102) { // identifier, identifier, format code for WST teletext uint16_t *descriptors = py + 4; int i; py += 9; for (i = 0; i < 5 && py < pend - 45; i++, py += 45) { int line = (descriptors[i] & 31) + (!(descriptors[i] & 128)) * 313; if (line && linemask_matches(line, wanted_lines)) tgt = teletext_data_unit_from_op47_vbi_packet(line, py, tgt); } } } return tgt; } static uint8_t* teletext_data_unit_from_ancillary_packet(uint16_t *py, uint16_t *pend, uint8_t *tgt, int64_t wanted_lines, int allow_multipacket) { uint16_t did = py[0]; // data id uint16_t sdid = py[1]; // secondary data id uint16_t dc = py[2] & 255; // data count py += 3; pend = FFMIN(pend, py + dc); if (did == 0x143 && sdid == 0x102) { // subtitle distribution packet tgt = teletext_data_unit_from_op47_data(py, pend, tgt, wanted_lines); } else if (allow_multipacket && did == 0x143 && sdid == 0x203) { // VANC multipacket py += 2; // priority, line/field while (py < pend - 3) { tgt = teletext_data_unit_from_ancillary_packet(py, pend, tgt, wanted_lines, 0); py += 4 + (py[2] & 255); // ndid, nsdid, ndc, line/field } } return tgt; } static uint8_t *vanc_to_cc(AVFormatContext *avctx, uint16_t *buf, size_t words, unsigned &cc_count) { size_t i, len = (buf[5] & 0xff) + 6 + 1; uint8_t cdp_sum, rate; uint16_t hdr, ftr; uint8_t *cc; uint16_t *cdp = &buf[6]; // CDP follows if (cdp[0] != 0x96 || cdp[1] != 0x69) { av_log(avctx, AV_LOG_WARNING, "Invalid CDP header 0x%.2x 0x%.2x\n", cdp[0], cdp[1]); return NULL; } len -= 7; // remove VANC header and checksum if (cdp[2] != len) { av_log(avctx, AV_LOG_WARNING, "CDP len %d != %zu\n", cdp[2], len); return NULL; } cdp_sum = 0; for (i = 0; i < len - 1; i++) cdp_sum += cdp[i]; cdp_sum = cdp_sum ? 256 - cdp_sum : 0; if (cdp[len - 1] != cdp_sum) { av_log(avctx, AV_LOG_WARNING, "CDP checksum invalid 0x%.4x != 0x%.4x\n", cdp_sum, cdp[len-1]); return NULL; } rate = cdp[3]; if (!(rate & 0x0f)) { av_log(avctx, AV_LOG_WARNING, "CDP frame rate invalid (0x%.2x)\n", rate); return NULL; } rate >>= 4; if (rate > 8) { av_log(avctx, AV_LOG_WARNING, "CDP frame rate invalid (0x%.2x)\n", rate); return NULL; } if (!(cdp[4] & 0x43)) /* ccdata_present | caption_service_active | reserved */ { av_log(avctx, AV_LOG_WARNING, "CDP flags invalid (0x%.2x)\n", cdp[4]); return NULL; } hdr = (cdp[5] << 8) | cdp[6]; if (cdp[7] != 0x72) /* ccdata_id */ { av_log(avctx, AV_LOG_WARNING, "Invalid ccdata_id 0x%.2x\n", cdp[7]); return NULL; } cc_count = cdp[8]; if (!(cc_count & 0xe0)) { av_log(avctx, AV_LOG_WARNING, "Invalid cc_count 0x%.2x\n", cc_count); return NULL; } cc_count &= 0x1f; if ((len - 13) < cc_count * 3) { av_log(avctx, AV_LOG_WARNING, "Invalid cc_count %d (> %zu)\n", cc_count * 3, len - 13); return NULL; } if (cdp[len - 4] != 0x74) /* footer id */ { av_log(avctx, AV_LOG_WARNING, "Invalid footer id 0x%.2x\n", cdp[len-4]); return NULL; } ftr = (cdp[len - 3] << 8) | cdp[len - 2]; if (ftr != hdr) { av_log(avctx, AV_LOG_WARNING, "Header 0x%.4x != Footer 0x%.4x\n", hdr, ftr); return NULL; } cc = (uint8_t *)av_malloc(cc_count * 3); if (cc == NULL) { av_log(avctx, AV_LOG_WARNING, "CC - av_malloc failed for cc_count = %d\n", cc_count); return NULL; } for (size_t i = 0; i < cc_count; i++) { cc[3*i + 0] = cdp[9 + 3*i+0] /* & 3 */; cc[3*i + 1] = cdp[9 + 3*i+1]; cc[3*i + 2] = cdp[9 + 3*i+2]; } cc_count *= 3; return cc; } static uint8_t *get_metadata(AVFormatContext *avctx, uint16_t *buf, size_t width, uint8_t *tgt, size_t tgt_size, AVPacket *pkt) { decklink_cctx *cctx = (struct decklink_cctx *) avctx->priv_data; uint16_t *max_buf = buf + width; while (buf < max_buf - 6) { int len; uint16_t did = buf[3] & 0xFF; // data id uint16_t sdid = buf[4] & 0xFF; // secondary data id /* Check for VANC header */ if (buf[0] != 0 || buf[1] != 0x3ff || buf[2] != 0x3ff) { return tgt; } len = (buf[5] & 0xff) + 6 + 1; if (len > max_buf - buf) { av_log(avctx, AV_LOG_WARNING, "Data Count (%d) > data left (%zu)\n", len, max_buf - buf); return tgt; } if (did == 0x43 && (sdid == 0x02 || sdid == 0x03) && cctx->teletext_lines && width == 1920 && tgt_size >= 1920) { if (check_vanc_parity_checksum(buf, len, buf[len - 1]) < 0) { av_log(avctx, AV_LOG_WARNING, "VANC parity or checksum incorrect\n"); goto skip_packet; } tgt = teletext_data_unit_from_ancillary_packet(buf + 3, buf + len, tgt, cctx->teletext_lines, 1); } else if (did == 0x61 && sdid == 0x01) { unsigned int data_len; uint8_t *data; if (check_vanc_parity_checksum(buf, len, buf[len - 1]) < 0) { av_log(avctx, AV_LOG_WARNING, "VANC parity or checksum incorrect\n"); goto skip_packet; } clear_parity_bits(buf, len); data = vanc_to_cc(avctx, buf, width, data_len); if (data) { if (av_packet_add_side_data(pkt, AV_PKT_DATA_A53_CC, data, data_len) < 0) av_free(data); } } else { av_log(avctx, AV_LOG_DEBUG, "Unknown meta data DID = 0x%.2x SDID = 0x%.2x\n", did, sdid); } skip_packet: buf += len; } return tgt; } static void avpacket_queue_init(AVFormatContext *avctx, AVPacketQueue *q) { struct decklink_cctx *ctx = (struct decklink_cctx *)avctx->priv_data; memset(q, 0, sizeof(AVPacketQueue)); pthread_mutex_init(&q->mutex, NULL); pthread_cond_init(&q->cond, NULL); q->avctx = avctx; q->max_q_size = ctx->queue_size; } static void avpacket_queue_flush(AVPacketQueue *q) { AVPacketList *pkt, *pkt1; pthread_mutex_lock(&q->mutex); for (pkt = q->first_pkt; pkt != NULL; pkt = pkt1) { pkt1 = pkt->next; av_packet_unref(&pkt->pkt); av_freep(&pkt); } q->last_pkt = NULL; q->first_pkt = NULL; q->nb_packets = 0; q->size = 0; pthread_mutex_unlock(&q->mutex); } static void avpacket_queue_end(AVPacketQueue *q) { avpacket_queue_flush(q); pthread_mutex_destroy(&q->mutex); pthread_cond_destroy(&q->cond); } static unsigned long long avpacket_queue_size(AVPacketQueue *q) { unsigned long long size; pthread_mutex_lock(&q->mutex); size = q->size; pthread_mutex_unlock(&q->mutex); return size; } static int avpacket_queue_put(AVPacketQueue *q, AVPacket *pkt) { AVPacketList *pkt1; // Drop Packet if queue size is > maximum queue size if (avpacket_queue_size(q) > (uint64_t)q->max_q_size) { av_packet_unref(pkt); av_log(q->avctx, AV_LOG_WARNING, "Decklink input buffer overrun!\n"); return -1; } /* ensure the packet is reference counted */ if (av_packet_make_refcounted(pkt) < 0) { av_packet_unref(pkt); return -1; } pkt1 = (AVPacketList *)av_malloc(sizeof(AVPacketList)); if (!pkt1) { av_packet_unref(pkt); return -1; } av_packet_move_ref(&pkt1->pkt, pkt); pkt1->next = NULL; pthread_mutex_lock(&q->mutex); if (!q->last_pkt) { q->first_pkt = pkt1; } else { q->last_pkt->next = pkt1; } q->last_pkt = pkt1; q->nb_packets++; q->size += pkt1->pkt.size + sizeof(*pkt1); pthread_cond_signal(&q->cond); pthread_mutex_unlock(&q->mutex); return 0; } static int avpacket_queue_get(AVPacketQueue *q, AVPacket *pkt, int block) { AVPacketList *pkt1; int ret; pthread_mutex_lock(&q->mutex); for (;; ) { pkt1 = q->first_pkt; if (pkt1) { q->first_pkt = pkt1->next; if (!q->first_pkt) { q->last_pkt = NULL; } q->nb_packets--; q->size -= pkt1->pkt.size + sizeof(*pkt1); *pkt = pkt1->pkt; av_free(pkt1); ret = 1; break; } else if (!block) { ret = 0; break; } else { pthread_cond_wait(&q->cond, &q->mutex); } } pthread_mutex_unlock(&q->mutex); return ret; } class decklink_input_callback : public IDeckLinkInputCallback { public: decklink_input_callback(AVFormatContext *_avctx); ~decklink_input_callback(); virtual HRESULT STDMETHODCALLTYPE QueryInterface(REFIID iid, LPVOID *ppv) { return E_NOINTERFACE; } virtual ULONG STDMETHODCALLTYPE AddRef(void); virtual ULONG STDMETHODCALLTYPE Release(void); virtual HRESULT STDMETHODCALLTYPE VideoInputFormatChanged(BMDVideoInputFormatChangedEvents, IDeckLinkDisplayMode*, BMDDetectedVideoInputFormatFlags); virtual HRESULT STDMETHODCALLTYPE VideoInputFrameArrived(IDeckLinkVideoInputFrame*, IDeckLinkAudioInputPacket*); private: std::atomic _refs; AVFormatContext *avctx; decklink_ctx *ctx; int no_video; int64_t initial_video_pts; int64_t initial_audio_pts; }; decklink_input_callback::decklink_input_callback(AVFormatContext *_avctx) : _refs(1) { avctx = _avctx; decklink_cctx *cctx = (struct decklink_cctx *)avctx->priv_data; ctx = (struct decklink_ctx *)cctx->ctx; no_video = 0; initial_audio_pts = initial_video_pts = AV_NOPTS_VALUE; } decklink_input_callback::~decklink_input_callback() { } ULONG decklink_input_callback::AddRef(void) { return ++_refs; } ULONG decklink_input_callback::Release(void) { int ret = --_refs; if (!ret) delete this; return ret; } static int64_t get_pkt_pts(IDeckLinkVideoInputFrame *videoFrame, IDeckLinkAudioInputPacket *audioFrame, int64_t wallclock, int64_t abs_wallclock, DecklinkPtsSource pts_src, AVRational time_base, int64_t *initial_pts, int copyts) { int64_t pts = AV_NOPTS_VALUE; BMDTimeValue bmd_pts; BMDTimeValue bmd_duration; HRESULT res = E_INVALIDARG; switch (pts_src) { case PTS_SRC_AUDIO: if (audioFrame) res = audioFrame->GetPacketTime(&bmd_pts, time_base.den); break; case PTS_SRC_VIDEO: if (videoFrame) res = videoFrame->GetStreamTime(&bmd_pts, &bmd_duration, time_base.den); break; case PTS_SRC_REFERENCE: if (videoFrame) res = videoFrame->GetHardwareReferenceTimestamp(time_base.den, &bmd_pts, &bmd_duration); break; case PTS_SRC_WALLCLOCK: /* fall through */ case PTS_SRC_ABS_WALLCLOCK: { /* MSVC does not support compound literals like AV_TIME_BASE_Q * in C++ code (compiler error C4576) */ AVRational timebase; timebase.num = 1; timebase.den = AV_TIME_BASE; if (pts_src == PTS_SRC_WALLCLOCK) pts = av_rescale_q(wallclock, timebase, time_base); else pts = av_rescale_q(abs_wallclock, timebase, time_base); break; } } if (res == S_OK) pts = bmd_pts / time_base.num; if (!copyts) { if (pts != AV_NOPTS_VALUE && *initial_pts == AV_NOPTS_VALUE) *initial_pts = pts; if (*initial_pts != AV_NOPTS_VALUE) pts -= *initial_pts; } return pts; } HRESULT decklink_input_callback::VideoInputFrameArrived( IDeckLinkVideoInputFrame *videoFrame, IDeckLinkAudioInputPacket *audioFrame) { void *frameBytes; void *audioFrameBytes; BMDTimeValue frameTime; BMDTimeValue frameDuration; int64_t wallclock = 0, abs_wallclock = 0; struct decklink_cctx *cctx = (struct decklink_cctx *) avctx->priv_data; if (ctx->autodetect) { if (videoFrame && !(videoFrame->GetFlags() & bmdFrameHasNoInputSource) && ctx->bmd_mode == bmdModeUnknown) { ctx->bmd_mode = AUTODETECT_DEFAULT_MODE; } return S_OK; } // Drop the frames till system's timestamp aligns with the configured value. if (0 == ctx->frameCount && cctx->timestamp_align) { AVRational remainder = av_make_q(av_gettime() % cctx->timestamp_align, 1000000); AVRational frame_duration = av_inv_q(ctx->video_st->r_frame_rate); if (av_cmp_q(remainder, frame_duration) > 0) { ++ctx->dropped; return S_OK; } } ctx->frameCount++; if (ctx->audio_pts_source == PTS_SRC_WALLCLOCK || ctx->video_pts_source == PTS_SRC_WALLCLOCK) wallclock = av_gettime_relative(); if (ctx->audio_pts_source == PTS_SRC_ABS_WALLCLOCK || ctx->video_pts_source == PTS_SRC_ABS_WALLCLOCK) abs_wallclock = av_gettime(); // Handle Video Frame if (videoFrame) { AVPacket pkt; av_init_packet(&pkt); if (ctx->frameCount % 25 == 0) { unsigned long long qsize = avpacket_queue_size(&ctx->queue); av_log(avctx, AV_LOG_DEBUG, "Frame received (#%lu) - Valid (%liB) - QSize %fMB\n", ctx->frameCount, videoFrame->GetRowBytes() * videoFrame->GetHeight(), (double)qsize / 1024 / 1024); } videoFrame->GetBytes(&frameBytes); videoFrame->GetStreamTime(&frameTime, &frameDuration, ctx->video_st->time_base.den); if (videoFrame->GetFlags() & bmdFrameHasNoInputSource) { if (ctx->draw_bars && videoFrame->GetPixelFormat() == bmdFormat8BitYUV) { unsigned bars[8] = { 0xEA80EA80, 0xD292D210, 0xA910A9A5, 0x90229035, 0x6ADD6ACA, 0x51EF515A, 0x286D28EF, 0x10801080 }; int width = videoFrame->GetWidth(); int height = videoFrame->GetHeight(); unsigned *p = (unsigned *)frameBytes; for (int y = 0; y < height; y++) { for (int x = 0; x < width; x += 2) *p++ = bars[(x * 8) / width]; } } if (!no_video) { av_log(avctx, AV_LOG_WARNING, "Frame received (#%lu) - No input signal detected " "- Frames dropped %u\n", ctx->frameCount, ++ctx->dropped); } no_video = 1; } else { if (no_video) { av_log(avctx, AV_LOG_WARNING, "Frame received (#%lu) - Input returned " "- Frames dropped %u\n", ctx->frameCount, ++ctx->dropped); } no_video = 0; // Handle Timecode (if requested) if (ctx->tc_format) { IDeckLinkTimecode *timecode; if (videoFrame->GetTimecode(ctx->tc_format, &timecode) == S_OK) { const char *tc = NULL; DECKLINK_STR decklink_tc; if (timecode->GetString(&decklink_tc) == S_OK) { tc = DECKLINK_STRDUP(decklink_tc); DECKLINK_FREE(decklink_tc); } timecode->Release(); if (tc) { AVDictionary* metadata_dict = NULL; int metadata_len; uint8_t* packed_metadata; if (av_dict_set(&metadata_dict, "timecode", tc, AV_DICT_DONT_STRDUP_VAL) >= 0) { packed_metadata = av_packet_pack_dictionary(metadata_dict, &metadata_len); av_dict_free(&metadata_dict); if (packed_metadata) { if (av_packet_add_side_data(&pkt, AV_PKT_DATA_STRINGS_METADATA, packed_metadata, metadata_len) < 0) av_freep(&packed_metadata); else if (!ctx->tc_seen) ctx->tc_seen = ctx->frameCount; } } } } else { av_log(avctx, AV_LOG_DEBUG, "Unable to find timecode.\n"); } } } if (ctx->tc_format && cctx->wait_for_tc && !ctx->tc_seen) { av_log(avctx, AV_LOG_WARNING, "No TC detected yet. wait_for_tc set. Dropping. \n"); av_log(avctx, AV_LOG_WARNING, "Frame received (#%lu) - " "- Frames dropped %u\n", ctx->frameCount, ++ctx->dropped); return S_OK; } pkt.pts = get_pkt_pts(videoFrame, audioFrame, wallclock, abs_wallclock, ctx->video_pts_source, ctx->video_st->time_base, &initial_video_pts, cctx->copyts); pkt.dts = pkt.pts; pkt.duration = frameDuration; //To be made sure it still applies pkt.flags |= AV_PKT_FLAG_KEY; pkt.stream_index = ctx->video_st->index; pkt.data = (uint8_t *)frameBytes; pkt.size = videoFrame->GetRowBytes() * videoFrame->GetHeight(); //fprintf(stderr,"Video Frame size %d ts %d\n", pkt.size, pkt.pts); if (!no_video) { IDeckLinkVideoFrameAncillary *vanc; AVPacket txt_pkt; uint8_t txt_buf0[3531]; // 35 * 46 bytes decoded teletext lines + 1 byte data_identifier + 1920 bytes OP47 decode buffer uint8_t *txt_buf = txt_buf0; if (videoFrame->GetAncillaryData(&vanc) == S_OK) { int i; int64_t line_mask = 1; BMDPixelFormat vanc_format = vanc->GetPixelFormat(); txt_buf[0] = 0x10; // data_identifier - EBU_data txt_buf++; #if CONFIG_LIBZVBI if (ctx->bmd_mode == bmdModePAL && ctx->teletext_lines && (vanc_format == bmdFormat8BitYUV || vanc_format == bmdFormat10BitYUV)) { av_assert0(videoFrame->GetWidth() == 720); for (i = 6; i < 336; i++, line_mask <<= 1) { uint8_t *buf; if ((ctx->teletext_lines & line_mask) && vanc->GetBufferForVerticalBlankingLine(i, (void**)&buf) == S_OK) { if (vanc_format == bmdFormat8BitYUV) txt_buf = teletext_data_unit_from_vbi_data(i, buf, txt_buf, VBI_PIXFMT_UYVY); else txt_buf = teletext_data_unit_from_vbi_data_10bit(i, buf, txt_buf); } if (i == 22) i = 317; } } #endif if (vanc_format == bmdFormat10BitYUV && videoFrame->GetWidth() <= MAX_WIDTH_VANC) { int idx = get_vanc_line_idx(ctx->bmd_mode); for (i = vanc_line_numbers[idx].vanc_start; i <= vanc_line_numbers[idx].vanc_end; i++) { uint8_t *buf; if (vanc->GetBufferForVerticalBlankingLine(i, (void**)&buf) == S_OK) { uint16_t vanc[MAX_WIDTH_VANC]; size_t vanc_size = videoFrame->GetWidth(); if (ctx->bmd_mode == bmdModeNTSC && videoFrame->GetWidth() * 2 <= MAX_WIDTH_VANC) { vanc_size = vanc_size * 2; unpack_v210(vanc, buf, videoFrame->GetWidth()); } else { extract_luma_from_v210(vanc, buf, videoFrame->GetWidth()); } txt_buf = get_metadata(avctx, vanc, vanc_size, txt_buf, sizeof(txt_buf0) - (txt_buf - txt_buf0), &pkt); } if (i == vanc_line_numbers[idx].field0_vanc_end) i = vanc_line_numbers[idx].field1_vanc_start - 1; } } vanc->Release(); if (txt_buf - txt_buf0 > 1) { int stuffing_units = (4 - ((45 + txt_buf - txt_buf0) / 46) % 4) % 4; while (stuffing_units--) { memset(txt_buf, 0xff, 46); txt_buf[1] = 0x2c; // data_unit_length txt_buf += 46; } av_init_packet(&txt_pkt); txt_pkt.pts = pkt.pts; txt_pkt.dts = pkt.dts; txt_pkt.stream_index = ctx->teletext_st->index; txt_pkt.data = txt_buf0; txt_pkt.size = txt_buf - txt_buf0; if (avpacket_queue_put(&ctx->queue, &txt_pkt) < 0) { ++ctx->dropped; } } } } pkt.buf = av_buffer_create(pkt.data, pkt.size, decklink_object_free, videoFrame, 0); if (pkt.buf) videoFrame->AddRef(); if (avpacket_queue_put(&ctx->queue, &pkt) < 0) { ++ctx->dropped; } } // Handle Audio Frame if (audioFrame) { AVPacket pkt; BMDTimeValue audio_pts; av_init_packet(&pkt); //hack among hacks pkt.size = audioFrame->GetSampleFrameCount() * ctx->audio_st->codecpar->channels * (ctx->audio_depth / 8); audioFrame->GetBytes(&audioFrameBytes); audioFrame->GetPacketTime(&audio_pts, ctx->audio_st->time_base.den); pkt.pts = get_pkt_pts(videoFrame, audioFrame, wallclock, abs_wallclock, ctx->audio_pts_source, ctx->audio_st->time_base, &initial_audio_pts, cctx->copyts); pkt.dts = pkt.pts; //fprintf(stderr,"Audio Frame size %d ts %d\n", pkt.size, pkt.pts); pkt.flags |= AV_PKT_FLAG_KEY; pkt.stream_index = ctx->audio_st->index; pkt.data = (uint8_t *)audioFrameBytes; if (avpacket_queue_put(&ctx->queue, &pkt) < 0) { ++ctx->dropped; } } return S_OK; } HRESULT decklink_input_callback::VideoInputFormatChanged( BMDVideoInputFormatChangedEvents events, IDeckLinkDisplayMode *mode, BMDDetectedVideoInputFormatFlags) { ctx->bmd_mode = mode->GetDisplayMode(); return S_OK; } static int decklink_autodetect(struct decklink_cctx *cctx) { struct decklink_ctx *ctx = (struct decklink_ctx *)cctx->ctx; DECKLINK_BOOL autodetect_supported = false; int i; if (ctx->attr->GetFlag(BMDDeckLinkSupportsInputFormatDetection, &autodetect_supported) != S_OK) return -1; if (autodetect_supported == false) return -1; ctx->autodetect = 1; ctx->bmd_mode = bmdModeUnknown; if (ctx->dli->EnableVideoInput(AUTODETECT_DEFAULT_MODE, bmdFormat8BitYUV, bmdVideoInputEnableFormatDetection) != S_OK) { return -1; } if (ctx->dli->StartStreams() != S_OK) { return -1; } // 1 second timeout for (i = 0; i < 10; i++) { av_usleep(100000); /* Sometimes VideoInputFrameArrived is called without the * bmdFrameHasNoInputSource flag before VideoInputFormatChanged. * So don't break for bmd_mode == AUTODETECT_DEFAULT_MODE. */ if (ctx->bmd_mode != bmdModeUnknown && ctx->bmd_mode != AUTODETECT_DEFAULT_MODE) break; } ctx->dli->PauseStreams(); ctx->dli->FlushStreams(); ctx->autodetect = 0; if (ctx->bmd_mode != bmdModeUnknown) { cctx->format_code = (char *)av_mallocz(5); if (!cctx->format_code) return -1; AV_WB32(cctx->format_code, ctx->bmd_mode); return 0; } else { return -1; } } extern "C" { av_cold int ff_decklink_read_close(AVFormatContext *avctx) { struct decklink_cctx *cctx = (struct decklink_cctx *)avctx->priv_data; struct decklink_ctx *ctx = (struct decklink_ctx *)cctx->ctx; if (ctx->dli) { ctx->dli->StopStreams(); ctx->dli->DisableVideoInput(); ctx->dli->DisableAudioInput(); } ff_decklink_cleanup(avctx); avpacket_queue_end(&ctx->queue); av_freep(&cctx->ctx); return 0; } av_cold int ff_decklink_read_header(AVFormatContext *avctx) { struct decklink_cctx *cctx = (struct decklink_cctx *)avctx->priv_data; struct decklink_ctx *ctx; class decklink_allocator *allocator; class decklink_input_callback *input_callback; AVStream *st; HRESULT result; int ret; ctx = (struct decklink_ctx *) av_mallocz(sizeof(struct decklink_ctx)); if (!ctx) return AVERROR(ENOMEM); ctx->list_devices = cctx->list_devices; ctx->list_formats = cctx->list_formats; ctx->teletext_lines = cctx->teletext_lines; ctx->preroll = cctx->preroll; ctx->duplex_mode = cctx->duplex_mode; if (cctx->tc_format > 0 && (unsigned int)cctx->tc_format < FF_ARRAY_ELEMS(decklink_timecode_format_map)) ctx->tc_format = decklink_timecode_format_map[cctx->tc_format]; if (cctx->video_input > 0 && (unsigned int)cctx->video_input < FF_ARRAY_ELEMS(decklink_video_connection_map)) ctx->video_input = decklink_video_connection_map[cctx->video_input]; if (cctx->audio_input > 0 && (unsigned int)cctx->audio_input < FF_ARRAY_ELEMS(decklink_audio_connection_map)) ctx->audio_input = decklink_audio_connection_map[cctx->audio_input]; ctx->audio_pts_source = cctx->audio_pts_source; ctx->video_pts_source = cctx->video_pts_source; ctx->draw_bars = cctx->draw_bars; ctx->audio_depth = cctx->audio_depth; cctx->ctx = ctx; /* Check audio channel option for valid values: 2, 8 or 16 */ switch (cctx->audio_channels) { case 2: case 8: case 16: break; default: av_log(avctx, AV_LOG_ERROR, "Value of channels option must be one of 2, 8 or 16\n"); return AVERROR(EINVAL); } /* Check audio bit depth option for valid values: 16 or 32 */ switch (cctx->audio_depth) { case 16: case 32: break; default: av_log(avctx, AV_LOG_ERROR, "Value for audio bit depth option must be either 16 or 32\n"); return AVERROR(EINVAL); } /* List available devices. */ if (ctx->list_devices) { av_log(avctx, AV_LOG_WARNING, "The -list_devices option is deprecated and will be removed. Please use ffmpeg -sources decklink instead.\n"); ff_decklink_list_devices_legacy(avctx, 1, 0); return AVERROR_EXIT; } ret = ff_decklink_init_device(avctx, avctx->url); if (ret < 0) return ret; /* Get input device. */ if (ctx->dl->QueryInterface(IID_IDeckLinkInput, (void **) &ctx->dli) != S_OK) { av_log(avctx, AV_LOG_ERROR, "Could not open input device from '%s'\n", avctx->url); ret = AVERROR(EIO); goto error; } if (ff_decklink_set_configs(avctx, DIRECTION_IN) < 0) { av_log(avctx, AV_LOG_ERROR, "Could not set input configuration\n"); ret = AVERROR(EIO); goto error; } /* List supported formats. */ if (ctx->list_formats) { ff_decklink_list_formats(avctx, DIRECTION_IN); ret = AVERROR_EXIT; goto error; } input_callback = new decklink_input_callback(avctx); ret = (ctx->dli->SetCallback(input_callback) == S_OK ? 0 : AVERROR_EXTERNAL); input_callback->Release(); if (ret < 0) { av_log(avctx, AV_LOG_ERROR, "Cannot set input callback\n"); goto error; } allocator = new decklink_allocator(); ret = (ctx->dli->SetVideoInputFrameMemoryAllocator(allocator) == S_OK ? 0 : AVERROR_EXTERNAL); allocator->Release(); if (ret < 0) { av_log(avctx, AV_LOG_ERROR, "Cannot set custom memory allocator\n"); goto error; } if (!cctx->format_code) { if (decklink_autodetect(cctx) < 0) { av_log(avctx, AV_LOG_ERROR, "Cannot Autodetect input stream or No signal\n"); ret = AVERROR(EIO); goto error; } av_log(avctx, AV_LOG_INFO, "Autodetected the input mode\n"); } if (ff_decklink_set_format(avctx, DIRECTION_IN) < 0) { av_log(avctx, AV_LOG_ERROR, "Could not set format code %s for %s\n", cctx->format_code ? cctx->format_code : "(unset)", avctx->url); ret = AVERROR(EIO); goto error; } #if !CONFIG_LIBZVBI if (ctx->teletext_lines && ctx->bmd_mode == bmdModePAL) { av_log(avctx, AV_LOG_ERROR, "Libzvbi support is needed for capturing SD PAL teletext, please recompile FFmpeg.\n"); ret = AVERROR(ENOSYS); goto error; } #endif /* Setup streams. */ st = avformat_new_stream(avctx, NULL); if (!st) { av_log(avctx, AV_LOG_ERROR, "Cannot add stream\n"); ret = AVERROR(ENOMEM); goto error; } st->codecpar->codec_type = AVMEDIA_TYPE_AUDIO; st->codecpar->codec_id = cctx->audio_depth == 32 ? AV_CODEC_ID_PCM_S32LE : AV_CODEC_ID_PCM_S16LE; st->codecpar->sample_rate = bmdAudioSampleRate48kHz; st->codecpar->channels = cctx->audio_channels; avpriv_set_pts_info(st, 64, 1, 1000000); /* 64 bits pts in us */ ctx->audio_st=st; st = avformat_new_stream(avctx, NULL); if (!st) { av_log(avctx, AV_LOG_ERROR, "Cannot add stream\n"); ret = AVERROR(ENOMEM); goto error; } st->codecpar->codec_type = AVMEDIA_TYPE_VIDEO; st->codecpar->width = ctx->bmd_width; st->codecpar->height = ctx->bmd_height; st->time_base.den = ctx->bmd_tb_den; st->time_base.num = ctx->bmd_tb_num; st->r_frame_rate = av_make_q(st->time_base.den, st->time_base.num); switch((BMDPixelFormat)cctx->raw_format) { case bmdFormat8BitYUV: st->codecpar->codec_id = AV_CODEC_ID_RAWVIDEO; st->codecpar->codec_tag = MKTAG('U', 'Y', 'V', 'Y'); st->codecpar->format = AV_PIX_FMT_UYVY422; st->codecpar->bit_rate = av_rescale(ctx->bmd_width * ctx->bmd_height * 16, st->time_base.den, st->time_base.num); break; case bmdFormat10BitYUV: st->codecpar->codec_id = AV_CODEC_ID_V210; st->codecpar->codec_tag = MKTAG('V','2','1','0'); st->codecpar->bit_rate = av_rescale(ctx->bmd_width * ctx->bmd_height * 64, st->time_base.den, st->time_base.num * 3); st->codecpar->bits_per_coded_sample = 10; break; case bmdFormat8BitARGB: st->codecpar->codec_id = AV_CODEC_ID_RAWVIDEO; st->codecpar->format = AV_PIX_FMT_0RGB; st->codecpar->codec_tag = avcodec_pix_fmt_to_codec_tag((enum AVPixelFormat)st->codecpar->format); st->codecpar->bit_rate = av_rescale(ctx->bmd_width * ctx->bmd_height * 32, st->time_base.den, st->time_base.num); break; case bmdFormat8BitBGRA: st->codecpar->codec_id = AV_CODEC_ID_RAWVIDEO; st->codecpar->format = AV_PIX_FMT_BGR0; st->codecpar->codec_tag = avcodec_pix_fmt_to_codec_tag((enum AVPixelFormat)st->codecpar->format); st->codecpar->bit_rate = av_rescale(ctx->bmd_width * ctx->bmd_height * 32, st->time_base.den, st->time_base.num); break; case bmdFormat10BitRGB: st->codecpar->codec_id = AV_CODEC_ID_R210; st->codecpar->codec_tag = MKTAG('R','2','1','0'); st->codecpar->format = AV_PIX_FMT_RGB48LE; st->codecpar->bit_rate = av_rescale(ctx->bmd_width * ctx->bmd_height * 30, st->time_base.den, st->time_base.num); st->codecpar->bits_per_coded_sample = 10; break; default: av_log(avctx, AV_LOG_ERROR, "Raw Format %.4s not supported\n", (char*) &cctx->raw_format); ret = AVERROR(EINVAL); goto error; } switch (ctx->bmd_field_dominance) { case bmdUpperFieldFirst: st->codecpar->field_order = AV_FIELD_TT; break; case bmdLowerFieldFirst: st->codecpar->field_order = AV_FIELD_BB; break; case bmdProgressiveFrame: case bmdProgressiveSegmentedFrame: st->codecpar->field_order = AV_FIELD_PROGRESSIVE; break; } avpriv_set_pts_info(st, 64, 1, 1000000); /* 64 bits pts in us */ ctx->video_st=st; if (ctx->teletext_lines) { st = avformat_new_stream(avctx, NULL); if (!st) { av_log(avctx, AV_LOG_ERROR, "Cannot add stream\n"); ret = AVERROR(ENOMEM); goto error; } st->codecpar->codec_type = AVMEDIA_TYPE_SUBTITLE; st->time_base.den = ctx->bmd_tb_den; st->time_base.num = ctx->bmd_tb_num; st->codecpar->codec_id = AV_CODEC_ID_DVB_TELETEXT; avpriv_set_pts_info(st, 64, 1, 1000000); /* 64 bits pts in us */ ctx->teletext_st = st; } av_log(avctx, AV_LOG_VERBOSE, "Using %d input audio channels\n", ctx->audio_st->codecpar->channels); result = ctx->dli->EnableAudioInput(bmdAudioSampleRate48kHz, cctx->audio_depth == 32 ? bmdAudioSampleType32bitInteger : bmdAudioSampleType16bitInteger, ctx->audio_st->codecpar->channels); if (result != S_OK) { av_log(avctx, AV_LOG_ERROR, "Cannot enable audio input\n"); ret = AVERROR(EIO); goto error; } result = ctx->dli->EnableVideoInput(ctx->bmd_mode, (BMDPixelFormat) cctx->raw_format, bmdVideoInputFlagDefault); if (result != S_OK) { av_log(avctx, AV_LOG_ERROR, "Cannot enable video input\n"); ret = AVERROR(EIO); goto error; } avpacket_queue_init (avctx, &ctx->queue); if (ctx->dli->StartStreams() != S_OK) { av_log(avctx, AV_LOG_ERROR, "Cannot start input stream\n"); ret = AVERROR(EIO); goto error; } return 0; error: ff_decklink_cleanup(avctx); return ret; } int ff_decklink_read_packet(AVFormatContext *avctx, AVPacket *pkt) { struct decklink_cctx *cctx = (struct decklink_cctx *)avctx->priv_data; struct decklink_ctx *ctx = (struct decklink_ctx *)cctx->ctx; avpacket_queue_get(&ctx->queue, pkt, 1); if (ctx->tc_format && !(av_dict_get(ctx->video_st->metadata, "timecode", NULL, 0))) { int size; const uint8_t *side_metadata = av_packet_get_side_data(pkt, AV_PKT_DATA_STRINGS_METADATA, &size); if (side_metadata) { if (av_packet_unpack_dictionary(side_metadata, size, &ctx->video_st->metadata) < 0) av_log(avctx, AV_LOG_ERROR, "Unable to set timecode\n"); } } return 0; } int ff_decklink_list_input_devices(AVFormatContext *avctx, struct AVDeviceInfoList *device_list) { return ff_decklink_list_devices(avctx, device_list, 1, 0); } } /* extern "C" */