diff options
Diffstat (limited to 'subprojects/gst-plugins-bad/sys/decklink/gstdecklinkvideosrc.cpp')
-rw-r--r-- | subprojects/gst-plugins-bad/sys/decklink/gstdecklinkvideosrc.cpp | 1712 |
1 files changed, 1712 insertions, 0 deletions
diff --git a/subprojects/gst-plugins-bad/sys/decklink/gstdecklinkvideosrc.cpp b/subprojects/gst-plugins-bad/sys/decklink/gstdecklinkvideosrc.cpp new file mode 100644 index 0000000000..2385a2339a --- /dev/null +++ b/subprojects/gst-plugins-bad/sys/decklink/gstdecklinkvideosrc.cpp @@ -0,0 +1,1712 @@ +/* GStreamer + * Copyright (C) 2011 David Schleef <ds@entropywave.com> + * Copyright (C) 2014 Sebastian Dröge <sebastian@centricular.com> + * Copyright (C) 2015 Florian Langlois <florian.langlois@fr.thalesgroup.com> + * + * This library is free software; you can redistribute it and/or + * modify it under the terms of the GNU Library General Public + * License as published by the Free Software Foundation; either + * version 2 of the License, or (at your option) any later version. + * + * This library is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + * Library General Public License for more details. + * + * You should have received a copy of the GNU Library General Public + * License along with this library; if not, write to the + * Free Software Foundation, Inc., 51 Franklin Street, Suite 500, + * Boston, MA 02110-1335, USA. + */ +/** + * SECTION:element-decklinkvideosrc + * @short_description: Inputs Video from a BlackMagic DeckLink Device + * + * Capture Video from a BlackMagic DeckLink Device. + * + * ## Sample pipeline + * |[ + * gst-launch-1.0 \ + * decklinkvideosrc device-number=0 connection=sdi mode=1080p25 ! \ + * autovideosink + * ]| + * Capturing 1080p25 video from the SDI-In of Card 0. Devices are numbered + * starting with 0. + * + * # Duplex-Mode: + * Certain DechLink Cards like the Duo2 or the Quad2 contain two or four + * independent SDI units with two connectors each. These units can operate either + * in half- or in full-duplex mode. + * + * The Duplex-Mode of a Card can be configured using the `duplex-mode`-Property. + * Cards that to not support Duplex-Modes are not influenced by the property. + * + * ## Half-Duplex-Mode (default): + * By default decklinkvideosrc will configure them into half-duplex mode, so that + * each connector acts as if it were an independent DeckLink Card which can either + * be used as an Input or as an Output. In this mode the Duo2 can be used as as 4 SDI + * In-/Outputs and the Quad2 as 8 SDI In-/Outputs. + * + * |[ + * gst-launch-1.0 \ + * decklinkvideosrc device-number=0 mode=1080p25 ! c. \ + * decklinkvideosrc device-number=1 mode=1080p25 ! c. \ + * decklinkvideosrc device-number=2 mode=1080p25 ! c. \ + * decklinkvideosrc device-number=3 mode=1080p25 ! c. \ + * compositor name=c \ + * sink_0::xpos=0 sink_0::ypos=0 sink_0::width=960 sink_0::height=540 \ + * sink_1::xpos=960 sink_1::ypos=0 sink_1::width=960 sink_1::height=540 \ + * sink_2::xpos=0 sink_2::ypos=540 sink_2::width=960 sink_2::height=540 \ + * sink_3::xpos=960 sink_3::ypos=540 sink_3::width=960 sink_3::height=540 ! \ + * video/x-raw,width=1920,height=1080 ! \ + * autovideosink + * ]| + * Capture 1080p25 from the first 4 units in the System (ie. the 4 Connectors of + * a Duo2 Card) and compose them into a 2x2 grid. + * + * |[ + * gst-launch-1.0 \ + * videotestsrc foreground-color=0x0000ff00 ! decklinkvideosink device-number=0 mode=1080p25 \ + * decklinkvideosrc device-number=1 mode=1080p25 ! autovideosink \ + * decklinkvideosrc device-number=2 mode=1080p25 ! autovideosink \ + * videotestsrc foreground-color=0x00ff0000 ! decklinkvideosink device-number=3 mode=1080p25 + * ]| + * Capture 1080p25 from the second and third unit in the System, + * Playout a Test-Screen with colored Snow on the first and fourth unit + * (ie. the Connectors 1-4 of a Duo2 unit). + * + * ## Device-Number-Mapping in Half-Duplex-Mode + * The device-number to connector-mapping in half-duplex-mode is as follows for the Duo2 + * - `device-number=0` SDI1 + * - `device-number=1` SDI3 + * - `device-number=2` SDI2 + * - `device-number=3` SDI4 + * + * And for the Quad2 + * - `device-number=0` SDI1 + * - `device-number=1` SDI3 + * - `device-number=2` SDI5 + * - `device-number=3` SDI7 + * - `device-number=4` SDI2 + * - `device-number=5` SDI4 + * - `device-number=6` SDI6 + * - `device-number=7` SDI8 + * + * ## Full-Duplex-Mode: + * When operating in full-duplex mode, two connectors of a unit are combined to + * a single device, performing extra processing with the second connection. + * + * This mode is most useful for Playout. See @decklinkvideosink. + * For Capturing the options are as follows: + * + * When capturing from a duplex-unit, the secondary port outputs the captured image + * unchanged. + * |[ + * gst-launch-1.0 \ + * decklinkvideosrc device-number=0 mode=1080p25 duplex-mode=full ! \ + * autovideosink + * ]| + * + * When simultaneously capturing and playing out onto the same device, the + * secondary port outputs the played out video. Note, that this can also be + * achieved using half-duplex mode. + * |[ + * gst-launch-1.0 \ + * decklinkvideosrc device-number=0 mode=1080p25 duplex-mode=full ! \ + * videoflip video-direction=vert ! \ + * decklinkvideosink device-number=0 mode=1080p25 duplex-mode=full + * ]| + * Capturing Video on the primary port of device 0, output flipped version of the + * video on secondary port of the same device. + * + * ## Device-Number-Mapping in Full-Duplex-Mode + * The device-number to connector-mapping in full-duplex-mode is as follows for the Duo2 + * - `device-number=0` SDI1 primary, SDI2 secondary + * - `device-number=1` SDI3 primaty, SDI4 secondary + * + * And for the Quad2 + * - `device-number=0` SDI1 primary, SDI2 secondary + * - `device-number=1` SDI3 primaty, SDI4 secondary + * - `device-number=2` SDI5 primary, SDI6 secondary + * - `device-number=3` SDI7 primary, SDI8 secondary + */ + +#ifdef HAVE_CONFIG_H +#include "config.h" +#endif + +#include "gstdecklinkvideosrc.h" +#include <string.h> + +GST_DEBUG_CATEGORY_STATIC (gst_decklink_video_src_debug); +#define GST_CAT_DEFAULT gst_decklink_video_src_debug + +#define DEFAULT_MODE (GST_DECKLINK_MODE_AUTO) +#define DEFAULT_CONNECTION (GST_DECKLINK_CONNECTION_AUTO) +#define DEFAULT_BUFFER_SIZE (5) +#define DEFAULT_OUTPUT_STREAM_TIME (FALSE) +#define DEFAULT_SKIP_FIRST_TIME (0) +#define DEFAULT_DROP_NO_SIGNAL_FRAMES (FALSE) +#define DEFAULT_OUTPUT_CC (FALSE) +#define DEFAULT_OUTPUT_AFD_BAR (FALSE) + +#ifndef ABSDIFF +#define ABSDIFF(x, y) ( (x) > (y) ? ((x) - (y)) : ((y) - (x)) ) +#endif + +enum +{ + PROP_0, + PROP_MODE, + PROP_CONNECTION, + PROP_DEVICE_NUMBER, + PROP_BUFFER_SIZE, + PROP_VIDEO_FORMAT, + PROP_PROFILE_ID, + PROP_TIMECODE_FORMAT, + PROP_OUTPUT_STREAM_TIME, + PROP_SKIP_FIRST_TIME, + PROP_DROP_NO_SIGNAL_FRAMES, + PROP_SIGNAL, + PROP_HW_SERIAL_NUMBER, + PROP_OUTPUT_CC, + PROP_OUTPUT_AFD_BAR, +}; + +typedef struct +{ + IDeckLinkVideoInputFrame *frame; + GstClockTime timestamp, duration; + GstClockTime stream_timestamp; + GstClockTime stream_duration; + GstClockTime hardware_timestamp; + GstClockTime hardware_duration; + GstDecklinkModeEnum mode; + BMDPixelFormat format; + GstVideoTimeCode *tc; + gboolean no_signal; +} CaptureFrame; + +static void +capture_frame_clear (CaptureFrame * frame) +{ + if (frame->frame) + frame->frame->Release (); + if (frame->tc) + gst_video_time_code_free (frame->tc); + memset (frame, 0, sizeof (*frame)); +} + +typedef struct +{ + IDeckLinkVideoInputFrame *frame; + IDeckLinkInput *input; +} VideoFrame; + +static void +video_frame_free (void *data) +{ + VideoFrame *frame = (VideoFrame *) data; + + frame->frame->Release (); + frame->input->Release (); + g_free (frame); +} + +static void gst_decklink_video_src_set_property (GObject * object, + guint property_id, const GValue * value, GParamSpec * pspec); +static void gst_decklink_video_src_get_property (GObject * object, + guint property_id, GValue * value, GParamSpec * pspec); +static void gst_decklink_video_src_finalize (GObject * object); + +static GstStateChangeReturn +gst_decklink_video_src_change_state (GstElement * element, + GstStateChange transition); + +static GstCaps *gst_decklink_video_src_get_caps (GstBaseSrc * bsrc, + GstCaps * filter); +static gboolean gst_decklink_video_src_query (GstBaseSrc * bsrc, + GstQuery * query); +static gboolean gst_decklink_video_src_unlock (GstBaseSrc * bsrc); +static gboolean gst_decklink_video_src_unlock_stop (GstBaseSrc * bsrc); + +static GstFlowReturn gst_decklink_video_src_create (GstPushSrc * psrc, + GstBuffer ** buffer); + +static gboolean gst_decklink_video_src_open (GstDecklinkVideoSrc * self); +static gboolean gst_decklink_video_src_close (GstDecklinkVideoSrc * self); + +static gboolean gst_decklink_video_src_stop (GstDecklinkVideoSrc * self); + +static void gst_decklink_video_src_start_streams (GstElement * element); + +#define parent_class gst_decklink_video_src_parent_class +G_DEFINE_TYPE (GstDecklinkVideoSrc, gst_decklink_video_src, GST_TYPE_PUSH_SRC); +GST_ELEMENT_REGISTER_DEFINE_WITH_CODE (decklinkvideosrc, "decklinkvideosrc", GST_RANK_NONE, + GST_TYPE_DECKLINK_VIDEO_SRC, decklink_element_init (plugin)); + +static void +gst_decklink_video_src_class_init (GstDecklinkVideoSrcClass * klass) +{ + GObjectClass *gobject_class = G_OBJECT_CLASS (klass); + GstElementClass *element_class = GST_ELEMENT_CLASS (klass); + GstBaseSrcClass *basesrc_class = GST_BASE_SRC_CLASS (klass); + GstPushSrcClass *pushsrc_class = GST_PUSH_SRC_CLASS (klass); + GstCaps *templ_caps; + + gobject_class->set_property = gst_decklink_video_src_set_property; + gobject_class->get_property = gst_decklink_video_src_get_property; + gobject_class->finalize = gst_decklink_video_src_finalize; + + element_class->change_state = + GST_DEBUG_FUNCPTR (gst_decklink_video_src_change_state); + + basesrc_class->query = GST_DEBUG_FUNCPTR (gst_decklink_video_src_query); + basesrc_class->negotiate = NULL; + basesrc_class->get_caps = GST_DEBUG_FUNCPTR (gst_decklink_video_src_get_caps); + basesrc_class->unlock = GST_DEBUG_FUNCPTR (gst_decklink_video_src_unlock); + basesrc_class->unlock_stop = + GST_DEBUG_FUNCPTR (gst_decklink_video_src_unlock_stop); + + pushsrc_class->create = GST_DEBUG_FUNCPTR (gst_decklink_video_src_create); + + g_object_class_install_property (gobject_class, PROP_MODE, + g_param_spec_enum ("mode", "Playback Mode", + "Video Mode to use for playback", + GST_TYPE_DECKLINK_MODE, DEFAULT_MODE, + (GParamFlags) (G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS | + G_PARAM_CONSTRUCT))); + + g_object_class_install_property (gobject_class, PROP_CONNECTION, + g_param_spec_enum ("connection", "Connection", + "Video input connection to use", + GST_TYPE_DECKLINK_CONNECTION, DEFAULT_CONNECTION, + (GParamFlags) (G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS | + G_PARAM_CONSTRUCT))); + + g_object_class_install_property (gobject_class, PROP_DEVICE_NUMBER, + g_param_spec_int ("device-number", "Device number", + "Output device instance to use", 0, G_MAXINT, 0, + (GParamFlags) (G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS | + G_PARAM_CONSTRUCT))); + + g_object_class_install_property (gobject_class, PROP_BUFFER_SIZE, + g_param_spec_uint ("buffer-size", "Buffer Size", + "Size of internal buffer in number of video frames", 1, + G_MAXINT, DEFAULT_BUFFER_SIZE, + (GParamFlags) (G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS))); + + g_object_class_install_property (gobject_class, PROP_VIDEO_FORMAT, + g_param_spec_enum ("video-format", "Video format", + "Video format type to use for input (Only use auto for mode=auto)", + GST_TYPE_DECKLINK_VIDEO_FORMAT, GST_DECKLINK_VIDEO_FORMAT_AUTO, + (GParamFlags) (G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS | + G_PARAM_CONSTRUCT))); + + /** + * GstDecklinkVideoSrc:profile + * + * Specifies decklink profile to use. + * + * Since: 1.20 + */ + g_object_class_install_property (gobject_class, PROP_PROFILE_ID, + g_param_spec_enum ("profile", "Profile", + "Certain DeckLink devices such as the DeckLink 8K Pro, the DeckLink " + "Quad 2 and the DeckLink Duo 2 support multiple profiles to " + "configure the capture and playback behavior of its sub-devices." + "For the DeckLink Duo 2 and DeckLink Quad 2, a profile is shared " + "between any 2 sub-devices that utilize the same connectors. For the " + "DeckLink 8K Pro, a profile is shared between all 4 sub-devices. Any " + "sub-devices that share a profile are considered to be part of the " + "same profile group." + "DeckLink Duo 2 support configuration of the duplex mode of " + "individual sub-devices.", + GST_TYPE_DECKLINK_PROFILE_ID, GST_DECKLINK_PROFILE_ID_DEFAULT, + (GParamFlags) (G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS | + G_PARAM_CONSTRUCT))); + + g_object_class_install_property (gobject_class, PROP_TIMECODE_FORMAT, + g_param_spec_enum ("timecode-format", "Timecode format", + "Timecode format type to use for input", + GST_TYPE_DECKLINK_TIMECODE_FORMAT, + GST_DECKLINK_TIMECODE_FORMAT_RP188ANY, + (GParamFlags) (G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS | + G_PARAM_CONSTRUCT))); + + g_object_class_install_property (gobject_class, PROP_OUTPUT_STREAM_TIME, + g_param_spec_boolean ("output-stream-time", "Output Stream Time", + "Output stream time directly instead of translating to pipeline clock", + DEFAULT_OUTPUT_STREAM_TIME, + (GParamFlags) (G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS))); + + g_object_class_install_property (gobject_class, PROP_SKIP_FIRST_TIME, + g_param_spec_uint64 ("skip-first-time", "Skip First Time", + "Skip that much time of initial frames after starting", 0, + G_MAXUINT64, DEFAULT_SKIP_FIRST_TIME, + (GParamFlags) (G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS))); + + g_object_class_install_property (gobject_class, PROP_DROP_NO_SIGNAL_FRAMES, + g_param_spec_boolean ("drop-no-signal-frames", "Drop No Signal Frames", + "Drop frames that are marked as having no input signal", + DEFAULT_DROP_NO_SIGNAL_FRAMES, + (GParamFlags) (G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS))); + + g_object_class_install_property (gobject_class, PROP_SIGNAL, + g_param_spec_boolean ("signal", "Input signal available", + "True if there is a valid input signal available", + FALSE, (GParamFlags) (G_PARAM_READABLE | G_PARAM_STATIC_STRINGS))); + + g_object_class_install_property (gobject_class, PROP_HW_SERIAL_NUMBER, + g_param_spec_string ("hw-serial-number", "Hardware serial number", + "The serial number (hardware ID) of the Decklink card", + NULL, (GParamFlags) (G_PARAM_READABLE | G_PARAM_STATIC_STRINGS))); + + g_object_class_install_property (gobject_class, PROP_OUTPUT_CC, + g_param_spec_boolean ("output-cc", "Output Closed Caption", + "Extract and output CC as GstMeta (if present)", + DEFAULT_OUTPUT_CC, + (GParamFlags) (G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS))); + + g_object_class_install_property (gobject_class, PROP_OUTPUT_AFD_BAR, + g_param_spec_boolean ("output-afd-bar", "Output AFD/Bar data", + "Extract and output AFD/Bar as GstMeta (if present)", + DEFAULT_OUTPUT_AFD_BAR, + (GParamFlags) (G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS))); + + templ_caps = gst_decklink_mode_get_template_caps (TRUE); + gst_element_class_add_pad_template (element_class, + gst_pad_template_new ("src", GST_PAD_SRC, GST_PAD_ALWAYS, templ_caps)); + gst_caps_unref (templ_caps); + + gst_element_class_set_static_metadata (element_class, "Decklink Video Source", + "Video/Source/Hardware", "Decklink Source", + "David Schleef <ds@entropywave.com>, " + "Sebastian Dröge <sebastian@centricular.com>"); + + GST_DEBUG_CATEGORY_INIT (gst_decklink_video_src_debug, "decklinkvideosrc", + 0, "debug category for decklinkvideosrc element"); +} + +static void +gst_decklink_video_src_init (GstDecklinkVideoSrc * self) +{ + self->mode = DEFAULT_MODE; + self->caps_mode = GST_DECKLINK_MODE_AUTO; + self->caps_format = bmdFormat8BitYUV; + self->connection = DEFAULT_CONNECTION; + self->device_number = 0; + self->buffer_size = DEFAULT_BUFFER_SIZE; + self->video_format = GST_DECKLINK_VIDEO_FORMAT_AUTO; + self->profile_id = GST_DECKLINK_PROFILE_ID_DEFAULT; + self->timecode_format = bmdTimecodeRP188Any; + self->signal_state = SIGNAL_STATE_UNKNOWN; + self->output_stream_time = DEFAULT_OUTPUT_STREAM_TIME; + self->skip_first_time = DEFAULT_SKIP_FIRST_TIME; + self->drop_no_signal_frames = DEFAULT_DROP_NO_SIGNAL_FRAMES; + self->output_cc = DEFAULT_OUTPUT_CC; + self->output_afd_bar = DEFAULT_OUTPUT_AFD_BAR; + + self->window_size = 64; + self->times = g_new (GstClockTime, 4 * self->window_size); + self->times_temp = self->times + 2 * self->window_size; + self->window_fill = 0; + self->window_skip = 1; + self->window_skip_count = 0; + self->skipped_last = 0; + self->skip_from_timestamp = GST_CLOCK_TIME_NONE; + self->skip_to_timestamp = GST_CLOCK_TIME_NONE; + + gst_base_src_set_live (GST_BASE_SRC (self), TRUE); + gst_base_src_set_format (GST_BASE_SRC (self), GST_FORMAT_TIME); + + gst_pad_use_fixed_caps (GST_BASE_SRC_PAD (self)); + + g_mutex_init (&self->lock); + g_cond_init (&self->cond); + + self->current_frames = + gst_queue_array_new_for_struct (sizeof (CaptureFrame), + DEFAULT_BUFFER_SIZE); +} + +void +gst_decklink_video_src_set_property (GObject * object, guint property_id, + const GValue * value, GParamSpec * pspec) +{ + GstDecklinkVideoSrc *self = GST_DECKLINK_VIDEO_SRC_CAST (object); + + switch (property_id) { + case PROP_MODE: + self->mode = (GstDecklinkModeEnum) g_value_get_enum (value); + /* setting the default value for caps_mode here: if mode==auto then we + * configure caps_mode from the caps, if mode!=auto we set caps_mode to + * the same value as the mode. so self->caps_mode is essentially + * self->mode with mode=auto filtered into whatever we got from the + * negotiation */ + if (self->mode != GST_DECKLINK_MODE_AUTO) + self->caps_mode = self->mode; + break; + case PROP_CONNECTION: + self->connection = (GstDecklinkConnectionEnum) g_value_get_enum (value); + break; + case PROP_DEVICE_NUMBER: + self->device_number = g_value_get_int (value); + break; + case PROP_BUFFER_SIZE: + self->buffer_size = g_value_get_uint (value); + break; + case PROP_VIDEO_FORMAT: + self->video_format = (GstDecklinkVideoFormat) g_value_get_enum (value); + switch (self->video_format) { + case GST_DECKLINK_VIDEO_FORMAT_8BIT_YUV: + case GST_DECKLINK_VIDEO_FORMAT_10BIT_YUV: + case GST_DECKLINK_VIDEO_FORMAT_8BIT_ARGB: + case GST_DECKLINK_VIDEO_FORMAT_8BIT_BGRA: + self->caps_format = + gst_decklink_pixel_format_from_type (self->video_format); + case GST_DECKLINK_VIDEO_FORMAT_AUTO: + break; + default: + GST_ELEMENT_WARNING (GST_ELEMENT (self), CORE, NOT_IMPLEMENTED, + ("Format %d not supported", self->video_format), (NULL)); + break; + } + break; + case PROP_PROFILE_ID: + self->profile_id = (GstDecklinkProfileId) g_value_get_enum (value); + break; + case PROP_TIMECODE_FORMAT: + self->timecode_format = + gst_decklink_timecode_format_from_enum ((GstDecklinkTimecodeFormat) + g_value_get_enum (value)); + break; + case PROP_OUTPUT_STREAM_TIME: + self->output_stream_time = g_value_get_boolean (value); + break; + case PROP_SKIP_FIRST_TIME: + self->skip_first_time = g_value_get_uint64 (value); + break; + case PROP_DROP_NO_SIGNAL_FRAMES: + self->drop_no_signal_frames = g_value_get_boolean (value); + break; + case PROP_OUTPUT_CC: + self->output_cc = g_value_get_boolean (value); + break; + case PROP_OUTPUT_AFD_BAR: + self->output_afd_bar = g_value_get_boolean (value); + break; + default: + G_OBJECT_WARN_INVALID_PROPERTY_ID (object, property_id, pspec); + break; + } +} + +void +gst_decklink_video_src_get_property (GObject * object, guint property_id, + GValue * value, GParamSpec * pspec) +{ + GstDecklinkVideoSrc *self = GST_DECKLINK_VIDEO_SRC_CAST (object); + + switch (property_id) { + case PROP_MODE: + g_value_set_enum (value, self->mode); + break; + case PROP_CONNECTION: + g_value_set_enum (value, self->connection); + break; + case PROP_DEVICE_NUMBER: + g_value_set_int (value, self->device_number); + break; + case PROP_BUFFER_SIZE: + g_value_set_uint (value, self->buffer_size); + break; + case PROP_VIDEO_FORMAT: + g_value_set_enum (value, self->video_format); + break; + case PROP_PROFILE_ID: + g_value_set_enum (value, self->profile_id); + break; + case PROP_TIMECODE_FORMAT: + g_value_set_enum (value, + gst_decklink_timecode_format_to_enum (self->timecode_format)); + break; + case PROP_OUTPUT_STREAM_TIME: + g_value_set_boolean (value, self->output_stream_time); + break; + case PROP_SKIP_FIRST_TIME: + g_value_set_uint64 (value, self->skip_first_time); + break; + case PROP_DROP_NO_SIGNAL_FRAMES: + g_value_set_boolean (value, self->drop_no_signal_frames); + break; + case PROP_SIGNAL: + g_value_set_boolean (value, self->signal_state == SIGNAL_STATE_AVAILABLE); + break; + case PROP_HW_SERIAL_NUMBER: + if (self->input) + g_value_set_string (value, self->input->hw_serial_number); + else + g_value_set_string (value, NULL); + break; + case PROP_OUTPUT_CC: + g_value_set_boolean (value, self->output_cc); + break; + case PROP_OUTPUT_AFD_BAR: + g_value_set_boolean (value, self->output_afd_bar); + break; + default: + G_OBJECT_WARN_INVALID_PROPERTY_ID (object, property_id, pspec); + break; + } +} + +void +gst_decklink_video_src_finalize (GObject * object) +{ + GstDecklinkVideoSrc *self = GST_DECKLINK_VIDEO_SRC_CAST (object); + + g_free (self->times); + self->times = NULL; + g_mutex_clear (&self->lock); + g_cond_clear (&self->cond); + + if (self->current_frames) { + while (gst_queue_array_get_length (self->current_frames) > 0) { + CaptureFrame *tmp = (CaptureFrame *) + gst_queue_array_pop_head_struct (self->current_frames); + capture_frame_clear (tmp); + } + gst_queue_array_free (self->current_frames); + self->current_frames = NULL; + } + + G_OBJECT_CLASS (parent_class)->finalize (object); +} + +static gboolean +gst_decklink_video_src_start (GstDecklinkVideoSrc * self) +{ + const GstDecklinkMode *mode; + BMDVideoInputFlags flags; + HRESULT ret; + BMDPixelFormat format; + + g_mutex_lock (&self->input->lock); + if (self->input->video_enabled) { + g_mutex_unlock (&self->input->lock); + return TRUE; + } + g_mutex_unlock (&self->input->lock); + + if (self->input->config && self->connection != GST_DECKLINK_CONNECTION_AUTO) { + ret = self->input->config->SetInt (bmdDeckLinkConfigVideoInputConnection, + gst_decklink_get_connection (self->connection)); + if (ret != S_OK) { + GST_ERROR_OBJECT (self, + "Failed to set configuration (input source): 0x%08lx", + (unsigned long) ret); + return FALSE; + } + + if (self->connection == GST_DECKLINK_CONNECTION_COMPOSITE) { + ret = self->input->config->SetInt (bmdDeckLinkConfigAnalogVideoInputFlags, + bmdAnalogVideoFlagCompositeSetup75); + if (ret != S_OK) { + GST_ERROR_OBJECT (self, + "Failed to set configuration (composite setup): 0x%08lx", + (unsigned long) ret); + return FALSE; + } + } + } + + flags = bmdVideoInputFlagDefault; + if (self->mode == GST_DECKLINK_MODE_AUTO) { + bool autoDetection = false; + + if (self->input->attributes) { + ret = + self->input-> + attributes->GetFlag (BMDDeckLinkSupportsInputFormatDetection, + &autoDetection); + if (ret != S_OK) { + GST_ERROR_OBJECT (self, + "Failed to get attribute (autodetection): 0x%08lx", + (unsigned long) ret); + return FALSE; + } + if (autoDetection) + flags |= bmdVideoInputEnableFormatDetection; + } + if (!autoDetection) { + GST_ERROR_OBJECT (self, "Failed to activate auto-detection"); + return FALSE; + } + } + + mode = gst_decklink_get_mode (self->mode); + g_assert (mode != NULL); + + format = self->caps_format; + ret = self->input->input->EnableVideoInput (mode->mode, format, flags); + if (ret != S_OK) { + GST_WARNING_OBJECT (self, "Failed to enable video input: 0x%08lx", + (unsigned long) ret); + return FALSE; + } + + g_mutex_lock (&self->input->lock); + self->input->mode = mode; + self->input->video_enabled = TRUE; + if (self->input->start_streams) + self->input->start_streams (self->input->videosrc); + g_mutex_unlock (&self->input->lock); + + self->skipped_last = 0; + self->skip_from_timestamp = GST_CLOCK_TIME_NONE; + self->skip_to_timestamp = GST_CLOCK_TIME_NONE; + self->aspect_ratio_flag = -1; + + return TRUE; +} + +static void +gst_decklink_video_src_update_time_mapping (GstDecklinkVideoSrc * self, + GstClockTime capture_time, GstClockTime stream_time) +{ + if (self->window_skip_count == 0) { + GstClockTime num, den, b, xbase; + gdouble r_squared; + + self->times[2 * self->window_fill] = stream_time; + self->times[2 * self->window_fill + 1] = capture_time; + + self->window_fill++; + self->window_skip_count++; + if (self->window_skip_count >= self->window_skip) + self->window_skip_count = 0; + + if (self->window_fill >= self->window_size) { + guint fps = + ((gdouble) self->info.fps_n + self->info.fps_d - + 1) / ((gdouble) self->info.fps_d); + + /* Start by updating first every frame, once full every second frame, + * etc. until we update once every 4 seconds */ + if (self->window_skip < 4 * fps) + self->window_skip *= 2; + if (self->window_skip >= 4 * fps) + self->window_skip = 4 * fps; + + self->window_fill = 0; + self->window_filled = TRUE; + } + + /* First sample ever, create some basic mapping to start */ + if (!self->window_filled && self->window_fill == 1) { + self->current_time_mapping.xbase = stream_time; + self->current_time_mapping.b = capture_time; + self->current_time_mapping.num = 1; + self->current_time_mapping.den = 1; + self->next_time_mapping_pending = FALSE; + } + + /* Only bother calculating anything here once we had enough measurements, + * i.e. let's take the window size as a start */ + if (self->window_filled && + gst_calculate_linear_regression (self->times, self->times_temp, + self->window_size, &num, &den, &b, &xbase, &r_squared)) { + + GST_DEBUG_OBJECT (self, + "Calculated new time mapping: pipeline time = %lf * (stream time - %" + G_GUINT64_FORMAT ") + %" G_GUINT64_FORMAT " (%lf)", + ((gdouble) num) / ((gdouble) den), xbase, b, r_squared); + + self->next_time_mapping.xbase = xbase; + self->next_time_mapping.b = b; + self->next_time_mapping.num = num; + self->next_time_mapping.den = den; + self->next_time_mapping_pending = TRUE; + } + } else { + self->window_skip_count++; + if (self->window_skip_count >= self->window_skip) + self->window_skip_count = 0; + } + + if (self->next_time_mapping_pending) { + GstClockTime expected, new_calculated, diff, max_diff; + + expected = + gst_clock_adjust_with_calibration (NULL, stream_time, + self->current_time_mapping.xbase, self->current_time_mapping.b, + self->current_time_mapping.num, self->current_time_mapping.den); + new_calculated = + gst_clock_adjust_with_calibration (NULL, stream_time, + self->next_time_mapping.xbase, self->next_time_mapping.b, + self->next_time_mapping.num, self->next_time_mapping.den); + + if (new_calculated > expected) + diff = new_calculated - expected; + else + diff = expected - new_calculated; + + /* At most 5% frame duration change per update */ + max_diff = + gst_util_uint64_scale (GST_SECOND / 20, self->info.fps_d, + self->info.fps_n); + + GST_DEBUG_OBJECT (self, + "New time mapping causes difference of %" GST_TIME_FORMAT, + GST_TIME_ARGS (diff)); + GST_DEBUG_OBJECT (self, "Maximum allowed per frame %" GST_TIME_FORMAT, + GST_TIME_ARGS (max_diff)); + + if (diff > max_diff) { + /* adjust so that we move that much closer */ + if (new_calculated > expected) { + self->current_time_mapping.b = expected + max_diff; + self->current_time_mapping.xbase = stream_time; + } else { + self->current_time_mapping.b = expected - max_diff; + self->current_time_mapping.xbase = stream_time; + } + } else { + self->current_time_mapping.xbase = self->next_time_mapping.xbase; + self->current_time_mapping.b = self->next_time_mapping.b; + self->current_time_mapping.num = self->next_time_mapping.num; + self->current_time_mapping.den = self->next_time_mapping.den; + self->next_time_mapping_pending = FALSE; + } + } +} + +static void +gst_decklink_video_src_got_frame (GstElement * element, + IDeckLinkVideoInputFrame * frame, GstDecklinkModeEnum mode, + GstClockTime capture_time, GstClockTime stream_time, + GstClockTime stream_duration, GstClockTime hardware_time, + GstClockTime hardware_duration, IDeckLinkTimecode * dtc, gboolean no_signal) +{ + GstDecklinkVideoSrc *self = GST_DECKLINK_VIDEO_SRC_CAST (element); + GstClockTime timestamp, duration; + + GST_LOG_OBJECT (self, + "Got video frame at %" GST_TIME_FORMAT " / %" GST_TIME_FORMAT " (%" + GST_TIME_FORMAT "), no signal: %d", GST_TIME_ARGS (capture_time), + GST_TIME_ARGS (stream_time), GST_TIME_ARGS (stream_duration), no_signal); + + g_mutex_lock (&self->lock); + if (self->first_time == GST_CLOCK_TIME_NONE) + self->first_time = stream_time; + + if (self->skip_first_time > 0 + && stream_time - self->first_time < self->skip_first_time) { + g_mutex_unlock (&self->lock); + GST_DEBUG_OBJECT (self, + "Skipping frame as requested: %" GST_TIME_FORMAT " < %" GST_TIME_FORMAT, + GST_TIME_ARGS (stream_time), + GST_TIME_ARGS (self->skip_first_time + self->first_time)); + return; + } + + if (self->drop_no_signal_frames && no_signal) { + CaptureFrame f; + memset (&f, 0, sizeof (f)); + + /* Notify the streaming thread about the signal loss */ + gst_queue_array_push_tail_struct (self->current_frames, &f); + g_cond_signal (&self->cond); + g_mutex_unlock (&self->lock); + + return; + } + + gst_decklink_video_src_update_time_mapping (self, capture_time, stream_time); + if (self->output_stream_time) { + timestamp = stream_time; + duration = stream_duration; + } else { + timestamp = + gst_clock_adjust_with_calibration (NULL, stream_time, + self->current_time_mapping.xbase, self->current_time_mapping.b, + self->current_time_mapping.num, self->current_time_mapping.den); + duration = + gst_util_uint64_scale (stream_duration, self->current_time_mapping.num, + self->current_time_mapping.den); + } + + GST_LOG_OBJECT (self, + "Converted times to %" GST_TIME_FORMAT " (%" + GST_TIME_FORMAT ")", GST_TIME_ARGS (timestamp), GST_TIME_ARGS (duration)); + + if (!self->flushing) { + CaptureFrame f; + const GstDecklinkMode *bmode; + GstVideoTimeCodeFlags flags = GST_VIDEO_TIME_CODE_FLAGS_NONE; + guint field_count = 0; + guint skipped_frames = 0; + + while (gst_queue_array_get_length (self->current_frames) >= + self->buffer_size) { + CaptureFrame *tmp = (CaptureFrame *) + gst_queue_array_pop_head_struct (self->current_frames); + if (tmp->frame) { + if (skipped_frames == 0 && self->skipped_last == 0) + self->skip_from_timestamp = tmp->timestamp; + skipped_frames++; + self->skip_to_timestamp = tmp->timestamp; + } + capture_frame_clear (tmp); + } + + if (self->skipped_last == 0 && skipped_frames > 0) { + GST_WARNING_OBJECT (self, "Starting to drop frames"); + } + + if (skipped_frames == 0 && self->skipped_last > 0) { + GST_ELEMENT_WARNING_WITH_DETAILS (self, + STREAM, FAILED, + ("Dropped %u old frames from %" GST_TIME_FORMAT " to %" + GST_TIME_FORMAT, self->skipped_last, + GST_TIME_ARGS (self->skip_from_timestamp), + GST_TIME_ARGS (self->skip_to_timestamp)), + (NULL), + ("dropped", G_TYPE_UINT, self->skipped_last, + "from", G_TYPE_UINT64, self->skip_from_timestamp, + "to", G_TYPE_UINT64, self->skip_to_timestamp, NULL)); + self->skipped_last = 0; + } + + self->skipped_last += skipped_frames; + + memset (&f, 0, sizeof (f)); + f.frame = frame; + f.timestamp = timestamp; + f.duration = duration; + f.stream_timestamp = stream_time; + f.stream_duration = stream_duration; + f.hardware_timestamp = hardware_time; + f.hardware_duration = hardware_duration; + f.mode = mode; + f.format = frame->GetPixelFormat (); + f.no_signal = no_signal; + if (dtc != NULL) { + uint8_t hours, minutes, seconds, frames; + HRESULT res; + + res = dtc->GetComponents (&hours, &minutes, &seconds, &frames); + if (res != S_OK) { + GST_ERROR ("Could not get components for timecode %p: 0x%08lx", dtc, + (unsigned long) res); + f.tc = NULL; + } else { + GST_DEBUG_OBJECT (self, "Got timecode %02d:%02d:%02d:%02d", + hours, minutes, seconds, frames); + bmode = gst_decklink_get_mode (mode); + if (bmode->interlaced) + flags = + (GstVideoTimeCodeFlags) (flags | + GST_VIDEO_TIME_CODE_FLAGS_INTERLACED); + if (bmode->fps_d == 1001) { + if (bmode->fps_n == 30000 || bmode->fps_n == 60000) { + /* Some occurrences have been spotted where the driver mistakenly + * fails to set the drop-frame flag for drop-frame timecodes. + * Assume always drop-frame for 29.97 and 59.94 FPS */ + flags = + (GstVideoTimeCodeFlags) (flags | + GST_VIDEO_TIME_CODE_FLAGS_DROP_FRAME); + } else { + /* Drop-frame isn't defined for any other framerates (e.g. 23.976) + * */ + flags = + (GstVideoTimeCodeFlags) (flags & + ~GST_VIDEO_TIME_CODE_FLAGS_DROP_FRAME); + } + } + f.tc = + gst_video_time_code_new (bmode->fps_n, bmode->fps_d, NULL, flags, + hours, minutes, seconds, frames, field_count); + } + dtc->Release (); + } else { + f.tc = NULL; + } + + frame->AddRef (); + gst_queue_array_push_tail_struct (self->current_frames, &f); + g_cond_signal (&self->cond); + } + g_mutex_unlock (&self->lock); +} + +static void +extract_vbi_line (GstDecklinkVideoSrc * self, GstBuffer ** buffer, + IDeckLinkVideoFrameAncillary * vanc_frame, guint field2_offset, guint line, + gboolean * found_cc_out, gboolean * found_afd_bar_out) +{ + GstVideoAncillary gstanc; + const guint8 *vancdata; + gboolean found_cc = FALSE, found_afd_bar = FALSE; + + if (vanc_frame->GetBufferForVerticalBlankingLine (field2_offset + line, + (void **) &vancdata) != S_OK) + return; + + GST_DEBUG_OBJECT (self, "Checking for VBI data on field line %u (field %u)", + field2_offset + line, field2_offset ? 2 : 1); + gst_video_vbi_parser_add_line (self->vbiparser, vancdata); + + /* Check if CC or AFD/Bar is on this line if we didn't find any on a + * previous line. Remember the line where we found them */ + + while (gst_video_vbi_parser_get_ancillary (self->vbiparser, + &gstanc) == GST_VIDEO_VBI_PARSER_RESULT_OK) { + switch (GST_VIDEO_ANCILLARY_DID16 (&gstanc)) { + case GST_VIDEO_ANCILLARY_DID16_S334_EIA_708: + if (*found_cc_out || !self->output_cc) + continue; + + GST_DEBUG_OBJECT (self, + "Adding CEA-708 CDP meta to buffer for line %u", + field2_offset + line); + GST_MEMDUMP_OBJECT (self, "CDP", gstanc.data, gstanc.data_count); + gst_buffer_add_video_caption_meta (*buffer, + GST_VIDEO_CAPTION_TYPE_CEA708_CDP, gstanc.data, gstanc.data_count); + + found_cc = TRUE; + if (field2_offset) + self->last_cc_vbi_line_field2 = line; + else + self->last_cc_vbi_line = line; + break; + case GST_VIDEO_ANCILLARY_DID16_S334_EIA_608: + if (*found_cc_out || !self->output_cc) + continue; + + GST_DEBUG_OBJECT (self, + "Adding CEA-608 meta to buffer for line %u", field2_offset + line); + GST_MEMDUMP_OBJECT (self, "CEA608", gstanc.data, gstanc.data_count); + gst_buffer_add_video_caption_meta (*buffer, + GST_VIDEO_CAPTION_TYPE_CEA608_S334_1A, gstanc.data, + gstanc.data_count); + + found_cc = TRUE; + if (field2_offset) + self->last_cc_vbi_line_field2 = line; + else + self->last_cc_vbi_line = line; + break; + case GST_VIDEO_ANCILLARY_DID16_S2016_3_AFD_BAR:{ + GstVideoAFDValue afd; + gboolean is_letterbox; + guint16 bar1, bar2; + + if (*found_afd_bar_out || !self->output_afd_bar) + continue; + + GST_DEBUG_OBJECT (self, + "Adding AFD/Bar meta to buffer for line %u", field2_offset + line); + GST_MEMDUMP_OBJECT (self, "AFD/Bar", gstanc.data, gstanc.data_count); + + if (gstanc.data_count < 8) { + GST_WARNING_OBJECT (self, "AFD/Bar data too small"); + continue; + } + + self->aspect_ratio_flag = (gstanc.data[0] >> 2) & 0x1; + + afd = (GstVideoAFDValue) ((gstanc.data[0] >> 3) & 0xf); + is_letterbox = ((gstanc.data[3] >> 4) & 0x3) == 0; + bar1 = GST_READ_UINT16_BE (&gstanc.data[4]); + bar2 = GST_READ_UINT16_BE (&gstanc.data[6]); + + gst_buffer_add_video_afd_meta (*buffer, field2_offset ? 1 : 0, + GST_VIDEO_AFD_SPEC_SMPTE_ST2016_1, afd); + gst_buffer_add_video_bar_meta (*buffer, field2_offset ? 1 : 0, + is_letterbox, bar1, bar2); + + found_afd_bar = TRUE; + if (field2_offset) + self->last_afd_bar_vbi_line_field2 = line; + else + self->last_afd_bar_vbi_line = line; + break; + } + default: + /* otherwise continue looking */ + continue; + } + } + + if (found_cc) + *found_cc_out = TRUE; + if (found_afd_bar) + *found_afd_bar_out = TRUE; +} + +static void +extract_vbi (GstDecklinkVideoSrc * self, GstBuffer ** buffer, VideoFrame * vf) +{ + IDeckLinkVideoFrameAncillary *vanc_frame = NULL; + gint line; + GstVideoFormat videoformat; + GstDecklinkModeEnum mode_enum; + const GstDecklinkMode *mode; + gboolean found_cc = FALSE, found_afd_bar = FALSE; + + if (vf->frame->GetAncillaryData (&vanc_frame) != S_OK) + return; + + videoformat = + gst_decklink_video_format_from_type (vanc_frame->GetPixelFormat ()); + mode_enum = + gst_decklink_get_mode_enum_from_bmd (vanc_frame->GetDisplayMode ()); + mode = gst_decklink_get_mode (mode_enum); + + if (videoformat == GST_VIDEO_FORMAT_UNKNOWN) { + GST_DEBUG_OBJECT (self, "Unknown video format for Ancillary data"); + vanc_frame->Release (); + return; + } + + if ((videoformat != self->anc_vformat || mode->width != self->anc_width) + && self->vbiparser) { + gst_video_vbi_parser_free (self->vbiparser); + self->vbiparser = NULL; + } + + if (self->vbiparser == NULL) { + self->vbiparser = gst_video_vbi_parser_new (videoformat, mode->width); + self->anc_vformat = videoformat; + self->anc_width = mode->width; + } + + GST_DEBUG_OBJECT (self, "Checking for ancillary data in VBI"); + + /* First check last known lines, if any */ + if (self->last_cc_vbi_line > 0) { + extract_vbi_line (self, buffer, vanc_frame, 0, self->last_cc_vbi_line, + &found_cc, &found_afd_bar); + } + if (self->last_afd_bar_vbi_line > 0 + && self->last_cc_vbi_line != self->last_afd_bar_vbi_line) { + extract_vbi_line (self, buffer, vanc_frame, 0, self->last_afd_bar_vbi_line, + &found_cc, &found_afd_bar); + } + + if (!found_cc) + self->last_cc_vbi_line = -1; + if (!found_afd_bar) + self->last_afd_bar_vbi_line = -1; + + if ((self->output_cc && !found_cc) || (self->output_afd_bar + && !found_afd_bar)) { + /* Otherwise loop through the first 21 lines and hope to find the data */ + /* FIXME: For the different formats the number of lines that can contain + * VANC are different */ + for (line = 1; line < 22; line++) { + extract_vbi_line (self, buffer, vanc_frame, 0, line, &found_cc, + &found_afd_bar); + + /* If we found everything we wanted to extract, stop here */ + if ((!self->output_cc || found_cc) && + (!self->output_afd_bar || found_afd_bar)) + break; + } + } + + /* Do the same for field 2 in case of interlaced content */ + if (GST_VIDEO_INFO_IS_INTERLACED (&self->info)) { + gboolean found_cc_field2 = FALSE, found_afd_bar_field2 = FALSE; + guint field2_offset = 0; + + /* The VANC lines for the second field are at an offset, depending on + * the format in use + */ + switch (self->info.height) { + case 486: + /* NTSC: 525 / 2 + 1 */ + field2_offset = 263; + break; + case 576: + /* PAL: 625 / 2 + 1 */ + field2_offset = 313; + break; + case 1080: + /* 1080i: 1125 / 2 + 1 */ + field2_offset = 563; + break; + default: + g_assert_not_reached (); + } + + /* First try the same lines as for field 1 if we don't know yet */ + if (self->last_cc_vbi_line_field2 <= 0) + self->last_cc_vbi_line_field2 = self->last_cc_vbi_line; + if (self->last_afd_bar_vbi_line_field2 <= 0) + self->last_afd_bar_vbi_line_field2 = self->last_afd_bar_vbi_line; + + if (self->last_cc_vbi_line_field2 > 0) { + extract_vbi_line (self, buffer, vanc_frame, field2_offset, + self->last_cc_vbi_line_field2, &found_cc_field2, + &found_afd_bar_field2); + } + if (self->last_afd_bar_vbi_line_field2 > 0 + && self->last_cc_vbi_line_field2 != + self->last_afd_bar_vbi_line_field2) { + extract_vbi_line (self, buffer, vanc_frame, field2_offset, + self->last_afd_bar_vbi_line_field2, &found_cc_field2, + &found_afd_bar_field2); + } + + if (!found_cc_field2) + self->last_cc_vbi_line_field2 = -1; + if (!found_afd_bar_field2) + self->last_afd_bar_vbi_line_field2 = -1; + + if (((self->output_cc && !found_cc_field2) || (self->output_afd_bar + && !found_afd_bar_field2))) { + for (line = 1; line < 22; line++) { + extract_vbi_line (self, buffer, vanc_frame, field2_offset, line, + &found_cc_field2, &found_afd_bar_field2); + + /* If we found everything we wanted to extract, stop here */ + if ((!self->output_cc || found_cc_field2) && + (!self->output_afd_bar || found_afd_bar_field2)) + break; + } + } + } + + vanc_frame->Release (); +} + +static GstFlowReturn +gst_decklink_video_src_create (GstPushSrc * bsrc, GstBuffer ** buffer) +{ + GstDecklinkVideoSrc *self = GST_DECKLINK_VIDEO_SRC_CAST (bsrc); + GstFlowReturn flow_ret = GST_FLOW_OK; + const guint8 *data; + gsize data_size; + VideoFrame *vf; + CaptureFrame f; + GstCaps *caps; + gboolean caps_changed = FALSE; + const GstDecklinkMode *mode; + static GstStaticCaps stream_reference = + GST_STATIC_CAPS ("timestamp/x-decklink-stream"); + static GstStaticCaps hardware_reference = + GST_STATIC_CAPS ("timestamp/x-decklink-hardware"); + + if (!gst_decklink_video_src_start (self)) { + return GST_FLOW_NOT_NEGOTIATED; + } + + g_mutex_lock (&self->lock); +retry: + while (gst_queue_array_is_empty (self->current_frames) && !self->flushing) { + g_cond_wait (&self->cond, &self->lock); + } + + if (self->flushing) { + GST_DEBUG_OBJECT (self, "Flushing"); + g_mutex_unlock (&self->lock); + return GST_FLOW_FLUSHING; + } + + f = *(CaptureFrame *) gst_queue_array_pop_head_struct (self->current_frames); + + // We will have no frame if frames without signal are dropped immediately + // but we still have to signal that it's lost here. + if (f.no_signal || !f.frame) { + if (self->signal_state != SIGNAL_STATE_LOST) { + self->signal_state = SIGNAL_STATE_LOST; + g_object_notify (G_OBJECT (self), "signal"); + GST_ELEMENT_WARNING (GST_ELEMENT (self), RESOURCE, READ, ("Signal lost"), + ("No input source was detected - video frames invalid")); + } + // If we have no frame here, simply retry until we got one + if (!f.frame) { + capture_frame_clear (&f); + goto retry; + } + } else { + GstDecklinkSignalState previous_signal_state = self->signal_state; + + if (previous_signal_state != SIGNAL_STATE_AVAILABLE) { + self->signal_state = SIGNAL_STATE_AVAILABLE; + g_object_notify (G_OBJECT (self), "signal"); + } + + if (previous_signal_state == SIGNAL_STATE_LOST) { + GST_ELEMENT_INFO (GST_ELEMENT (self), RESOURCE, READ, + ("Signal recovered"), ("Input source detected")); + } + } + + // If we're not flushing, we should have a valid frame from the queue + g_assert (f.frame != NULL); + + // Create output buffer + f.frame->GetBytes ((gpointer *) & data); + data_size = f.frame->GetHeight() * f.frame->GetRowBytes(); + + vf = (VideoFrame *) g_malloc0 (sizeof (VideoFrame)); + + *buffer = + gst_buffer_new_wrapped_full ((GstMemoryFlags) GST_MEMORY_FLAG_READONLY, + (gpointer) data, data_size, 0, data_size, vf, + (GDestroyNotify) video_frame_free); + + vf->frame = f.frame; + f.frame->AddRef (); + vf->input = self->input->input; + vf->input->AddRef (); + + // Reset aspect ratio flag if the mode has changed. The new mode might not + // have AFD/Bar VANC. + if (self->caps_mode != f.mode) { + self->aspect_ratio_flag = -1; + } + // If we have a format that supports VANC and we are asked to extract CC, + // then do it here. + if ((self->output_cc || self->output_afd_bar) + && self->signal_state != SIGNAL_STATE_LOST) + extract_vbi (self, buffer, vf); + + if (!gst_pad_has_current_caps (GST_BASE_SRC_PAD (self))) { + caps_changed = TRUE; + } + // If there was AFD information with the aspect ratio flag set and the mode + // is auto then we have to switch from normal NTSC/PAL to the widescreen + // variants + if (self->aspect_ratio_flag != -1 && self->mode == GST_DECKLINK_MODE_AUTO) { + switch (f.mode) { + case GST_DECKLINK_MODE_NTSC: + f.mode = + self->aspect_ratio_flag == + 1 ? GST_DECKLINK_MODE_NTSC_WIDESCREEN : GST_DECKLINK_MODE_NTSC; + break; + case GST_DECKLINK_MODE_NTSC_P: + f.mode = + self->aspect_ratio_flag == + 1 ? GST_DECKLINK_MODE_NTSC_P_WIDESCREEN : GST_DECKLINK_MODE_NTSC_P; + break; + case GST_DECKLINK_MODE_NTSC2398: + f.mode = + self->aspect_ratio_flag == + 1 ? GST_DECKLINK_MODE_NTSC2398_WIDESCREEN : + GST_DECKLINK_MODE_NTSC2398; + break; + case GST_DECKLINK_MODE_PAL: + f.mode = + self->aspect_ratio_flag == + 1 ? GST_DECKLINK_MODE_PAL_WIDESCREEN : GST_DECKLINK_MODE_PAL; + break; + case GST_DECKLINK_MODE_PAL_P: + f.mode = + self->aspect_ratio_flag == + 1 ? GST_DECKLINK_MODE_PAL_P_WIDESCREEN : GST_DECKLINK_MODE_PAL_P; + break; + default: + break; + } + } + + if (self->caps_mode != f.mode) { + if (self->mode == GST_DECKLINK_MODE_AUTO + || !gst_pad_has_current_caps (GST_BASE_SRC_PAD (self))) { + GST_DEBUG_OBJECT (self, "Mode changed from %d to %d", self->caps_mode, + f.mode); + caps_changed = TRUE; + self->caps_mode = f.mode; + } else { + g_mutex_unlock (&self->lock); + GST_ELEMENT_ERROR (self, CORE, NEGOTIATION, + ("Invalid mode in captured frame"), + ("Mode set to %d but captured %d", self->caps_mode, f.mode)); + capture_frame_clear (&f); + gst_clear_buffer (buffer); + return GST_FLOW_NOT_NEGOTIATED; + } + } + if (self->caps_format != f.format) { + if (self->video_format == GST_DECKLINK_VIDEO_FORMAT_AUTO + || !gst_pad_has_current_caps (GST_BASE_SRC_PAD (self))) { + GST_DEBUG_OBJECT (self, "Format changed from %d to %d", self->caps_format, + f.format); + caps_changed = TRUE; + self->caps_format = f.format; + } else { + g_mutex_unlock (&self->lock); + GST_ELEMENT_ERROR (self, CORE, NEGOTIATION, + ("Invalid pixel format in captured frame"), + ("Format set to %d but captured %d", self->caps_format, f.format)); + capture_frame_clear (&f); + gst_clear_buffer (buffer); + return GST_FLOW_NOT_NEGOTIATED; + } + } + + /* 1 ns error can be just a rounding error, so that's OK. The Decklink + * drivers give us a really steady stream time, so anything above 1 ns can't + * be a rounding error and is therefore something to worry about */ + if (self->expected_stream_time != GST_CLOCK_TIME_NONE && + ABSDIFF (self->expected_stream_time, f.stream_timestamp) > 1) { + GstMessage *msg; + GstClockTime running_time; + + self->dropped += f.stream_timestamp - self->expected_stream_time; + running_time = gst_segment_to_running_time (&GST_BASE_SRC (self)->segment, + GST_FORMAT_TIME, f.timestamp); + + msg = + gst_message_new_qos (GST_OBJECT (self), TRUE, running_time, + f.stream_timestamp, f.timestamp, f.duration); + gst_message_set_qos_stats (msg, GST_FORMAT_TIME, self->processed, + self->dropped); + gst_element_post_message (GST_ELEMENT (self), msg); + } + if (self->first_stream_time == GST_CLOCK_TIME_NONE) + self->first_stream_time = f.stream_timestamp; + self->processed = + f.stream_timestamp - self->dropped - self->first_stream_time; + self->expected_stream_time = f.stream_timestamp + f.stream_duration; + + g_mutex_unlock (&self->lock); + if (caps_changed) { + self->last_cc_vbi_line = -1; + self->last_afd_bar_vbi_line = -1; + self->last_cc_vbi_line_field2 = -1; + self->last_afd_bar_vbi_line_field2 = -1; + caps = gst_decklink_mode_get_caps (f.mode, f.format, TRUE); + gst_video_info_from_caps (&self->info, caps); + gst_base_src_set_caps (GST_BASE_SRC_CAST (bsrc), caps); + gst_element_post_message (GST_ELEMENT_CAST (self), + gst_message_new_latency (GST_OBJECT_CAST (self))); + gst_caps_unref (caps); + if (self->vbiparser) { + gst_video_vbi_parser_free (self->vbiparser); + self->vbiparser = NULL; + self->anc_vformat = GST_VIDEO_FORMAT_UNKNOWN; + self->anc_width = 0; + } + } + + if (f.no_signal) + GST_BUFFER_FLAG_SET (*buffer, GST_BUFFER_FLAG_GAP); + GST_BUFFER_TIMESTAMP (*buffer) = f.timestamp; + GST_BUFFER_DURATION (*buffer) = f.duration; + if (f.tc != NULL) + gst_buffer_add_video_time_code_meta (*buffer, f.tc); + gst_buffer_add_reference_timestamp_meta (*buffer, + gst_static_caps_get (&stream_reference), f.stream_timestamp, + f.stream_duration); + gst_buffer_add_reference_timestamp_meta (*buffer, + gst_static_caps_get (&hardware_reference), f.hardware_timestamp, + f.hardware_duration); + + mode = gst_decklink_get_mode (self->caps_mode); + if (mode->interlaced && mode->tff) + GST_BUFFER_FLAG_SET (*buffer, + GST_VIDEO_BUFFER_FLAG_TFF | GST_VIDEO_BUFFER_FLAG_INTERLACED); + else if (mode->interlaced) + GST_BUFFER_FLAG_SET (*buffer, GST_VIDEO_BUFFER_FLAG_INTERLACED); + + GST_DEBUG_OBJECT (self, + "Outputting buffer %p with timestamp %" GST_TIME_FORMAT " and duration %" + GST_TIME_FORMAT, *buffer, GST_TIME_ARGS (GST_BUFFER_TIMESTAMP (*buffer)), + GST_TIME_ARGS (GST_BUFFER_DURATION (*buffer))); + + capture_frame_clear (&f); + + return flow_ret; +} + +static GstCaps * +gst_decklink_video_src_get_caps (GstBaseSrc * bsrc, GstCaps * filter) +{ + GstDecklinkVideoSrc *self = GST_DECKLINK_VIDEO_SRC_CAST (bsrc); + GstCaps *caps; + + if (self->mode != GST_DECKLINK_MODE_AUTO) { + caps = gst_decklink_mode_get_caps (self->mode, self->caps_format, TRUE); + } else if (self->caps_mode != GST_DECKLINK_MODE_AUTO) { + caps = + gst_decklink_mode_get_caps (self->caps_mode, self->caps_format, TRUE); + } else { + caps = gst_pad_get_pad_template_caps (GST_BASE_SRC_PAD (bsrc)); + } + + if (filter) { + GstCaps *tmp = + gst_caps_intersect_full (filter, caps, GST_CAPS_INTERSECT_FIRST); + gst_caps_unref (caps); + caps = tmp; + } + + return caps; +} + +static gboolean +gst_decklink_video_src_query (GstBaseSrc * bsrc, GstQuery * query) +{ + GstDecklinkVideoSrc *self = GST_DECKLINK_VIDEO_SRC_CAST (bsrc); + gboolean ret = TRUE; + + switch (GST_QUERY_TYPE (query)) { + case GST_QUERY_LATENCY:{ + if (self->input) { + GstClockTime min, max; + const GstDecklinkMode *mode; + + g_mutex_lock (&self->lock); + mode = gst_decklink_get_mode (self->caps_mode); + g_mutex_unlock (&self->lock); + + min = gst_util_uint64_scale_ceil (GST_SECOND, mode->fps_d, mode->fps_n); + max = self->buffer_size * min; + + gst_query_set_latency (query, TRUE, min, max); + ret = TRUE; + } else { + ret = FALSE; + } + + break; + } + default: + ret = GST_BASE_SRC_CLASS (parent_class)->query (bsrc, query); + break; + } + + return ret; +} + +static gboolean +gst_decklink_video_src_unlock (GstBaseSrc * bsrc) +{ + GstDecklinkVideoSrc *self = GST_DECKLINK_VIDEO_SRC_CAST (bsrc); + + g_mutex_lock (&self->lock); + self->flushing = TRUE; + g_cond_signal (&self->cond); + g_mutex_unlock (&self->lock); + + return TRUE; +} + +static gboolean +gst_decklink_video_src_unlock_stop (GstBaseSrc * bsrc) +{ + GstDecklinkVideoSrc *self = GST_DECKLINK_VIDEO_SRC_CAST (bsrc); + + g_mutex_lock (&self->lock); + self->flushing = FALSE; + while (gst_queue_array_get_length (self->current_frames) > 0) { + CaptureFrame *tmp = + (CaptureFrame *) gst_queue_array_pop_head_struct (self->current_frames); + capture_frame_clear (tmp); + } + g_mutex_unlock (&self->lock); + + return TRUE; +} + +static gboolean +gst_decklink_video_src_open (GstDecklinkVideoSrc * self) +{ + const GstDecklinkMode *mode; + + GST_DEBUG_OBJECT (self, "Opening"); + + self->input = + gst_decklink_acquire_nth_input (self->device_number, + GST_ELEMENT_CAST (self), FALSE); + if (!self->input) { + GST_ERROR_OBJECT (self, "Failed to acquire input"); + return FALSE; + } + + g_object_notify (G_OBJECT (self), "hw-serial-number"); + + mode = gst_decklink_get_mode (self->mode); + g_assert (mode != NULL); + g_mutex_lock (&self->input->lock); + self->input->mode = mode; + self->input->format = self->caps_format; + self->input->got_video_frame = gst_decklink_video_src_got_frame; + self->input->start_streams = gst_decklink_video_src_start_streams; + g_mutex_unlock (&self->input->lock); + + return TRUE; +} + +static gboolean +gst_decklink_video_src_close (GstDecklinkVideoSrc * self) +{ + + GST_DEBUG_OBJECT (self, "Closing"); + + if (self->input) { + g_mutex_lock (&self->input->lock); + self->input->got_video_frame = NULL; + self->input->mode = NULL; + self->input->video_enabled = FALSE; + self->input->start_streams = NULL; + g_mutex_unlock (&self->input->lock); + + gst_decklink_release_nth_input (self->device_number, + GST_ELEMENT_CAST (self), FALSE); + self->input = NULL; + } + + return TRUE; +} + +static gboolean +gst_decklink_video_src_stop (GstDecklinkVideoSrc * self) +{ + GST_DEBUG_OBJECT (self, "Stopping"); + + while (gst_queue_array_get_length (self->current_frames) > 0) { + CaptureFrame *tmp = + (CaptureFrame *) gst_queue_array_pop_head_struct (self->current_frames); + capture_frame_clear (tmp); + } + self->caps_mode = GST_DECKLINK_MODE_AUTO; + + if (self->input && self->input->video_enabled) { + g_mutex_lock (&self->input->lock); + self->input->video_enabled = FALSE; + g_mutex_unlock (&self->input->lock); + + self->input->input->DisableVideoInput (); + } + + if (self->vbiparser) { + gst_video_vbi_parser_free (self->vbiparser); + self->vbiparser = NULL; + self->anc_vformat = GST_VIDEO_FORMAT_UNKNOWN; + self->anc_width = 0; + } + + return TRUE; +} + +static void +gst_decklink_video_src_start_streams (GstElement * element) +{ + GstDecklinkVideoSrc *self = GST_DECKLINK_VIDEO_SRC_CAST (element); + HRESULT res; + + if (self->input->video_enabled && (!self->input->audiosrc + || self->input->audio_enabled) + && (GST_STATE (self) == GST_STATE_PLAYING + || GST_STATE_PENDING (self) == GST_STATE_PLAYING)) { + GST_DEBUG_OBJECT (self, "Starting streams"); + + g_mutex_lock (&self->lock); + self->first_time = GST_CLOCK_TIME_NONE; + self->window_fill = 0; + self->window_filled = FALSE; + self->window_skip = 1; + self->window_skip_count = 0; + self->current_time_mapping.xbase = 0; + self->current_time_mapping.b = 0; + self->current_time_mapping.num = 1; + self->current_time_mapping.den = 1; + self->next_time_mapping.xbase = 0; + self->next_time_mapping.b = 0; + self->next_time_mapping.num = 1; + self->next_time_mapping.den = 1; + g_mutex_unlock (&self->lock); + res = self->input->input->StartStreams (); + if (res != S_OK) { + GST_ELEMENT_ERROR (self, STREAM, FAILED, + (NULL), ("Failed to start streams: 0x%08lx", (unsigned long) res)); + return; + } + } else { + GST_DEBUG_OBJECT (self, "Not starting streams yet"); + } +} + +static GstStateChangeReturn +gst_decklink_video_src_change_state (GstElement * element, + GstStateChange transition) +{ + GstDecklinkVideoSrc *self = GST_DECKLINK_VIDEO_SRC_CAST (element); + GstStateChangeReturn ret = GST_STATE_CHANGE_SUCCESS; + + switch (transition) { + case GST_STATE_CHANGE_NULL_TO_READY: + self->processed = 0; + self->dropped = 0; + self->expected_stream_time = GST_CLOCK_TIME_NONE; + self->first_stream_time = GST_CLOCK_TIME_NONE; + if (!gst_decklink_video_src_open (self)) { + ret = GST_STATE_CHANGE_FAILURE; + goto out; + } + if (self->mode == GST_DECKLINK_MODE_AUTO && + self->video_format != GST_DECKLINK_VIDEO_FORMAT_AUTO) { + GST_WARNING_OBJECT (self, "Warning: mode=auto and format!=auto may \ + not work"); + } + self->vbiparser = NULL; + self->anc_vformat = GST_VIDEO_FORMAT_UNKNOWN; + self->anc_width = 0; + break; + case GST_STATE_CHANGE_READY_TO_PAUSED: + self->flushing = FALSE; + break; + default: + break; + } + + if (ret == GST_STATE_CHANGE_FAILURE) + return ret; + ret = GST_ELEMENT_CLASS (parent_class)->change_state (element, transition); + if (ret == GST_STATE_CHANGE_FAILURE) + return ret; + + switch (transition) { + case GST_STATE_CHANGE_PAUSED_TO_READY: + self->signal_state = SIGNAL_STATE_UNKNOWN; + + gst_decklink_video_src_stop (self); + break; + case GST_STATE_CHANGE_PLAYING_TO_PAUSED:{ + HRESULT res; + + GST_DEBUG_OBJECT (self, "Stopping streams"); + + res = self->input->input->StopStreams (); + if (res != S_OK) { + GST_ELEMENT_ERROR (self, STREAM, FAILED, + (NULL), ("Failed to stop streams: 0x%08lx", (unsigned long) res)); + ret = GST_STATE_CHANGE_FAILURE; + } + break; + } + case GST_STATE_CHANGE_PAUSED_TO_PLAYING:{ + g_mutex_lock (&self->input->lock); + if (self->input->start_streams) + self->input->start_streams (self->input->videosrc); + g_mutex_unlock (&self->input->lock); + + break; + } + case GST_STATE_CHANGE_READY_TO_NULL: + gst_decklink_video_src_close (self); + break; + default: + break; + } +out: + + return ret; +} |