mirror of
https://gitlab.freedesktop.org/gstreamer/gst-plugins-rs.git
synced 2024-06-10 02:09:22 +00:00
1359 lines
46 KiB
Rust
1359 lines
46 KiB
Rust
|
use glib;
|
||
|
use glib::prelude::*;
|
||
|
use gst;
|
||
|
use gst::prelude::*;
|
||
|
use gst_video;
|
||
|
use gst_video::prelude::*;
|
||
|
|
||
|
use byte_slice_cast::AsMutSliceOf;
|
||
|
|
||
|
use std::collections::VecDeque;
|
||
|
use std::sync::atomic::{AtomicUsize, Ordering};
|
||
|
use std::sync::{Arc, Condvar, Mutex, Weak};
|
||
|
use std::thread;
|
||
|
|
||
|
use super::*;
|
||
|
|
||
|
enum ReceiverInfo {
|
||
|
Connecting {
|
||
|
id: usize,
|
||
|
ndi_name: Option<String>,
|
||
|
ip_address: Option<String>,
|
||
|
video: Option<Weak<ReceiverInner>>,
|
||
|
audio: Option<Weak<ReceiverInner>>,
|
||
|
},
|
||
|
Connected {
|
||
|
id: usize,
|
||
|
ndi_name: String,
|
||
|
ip_address: String,
|
||
|
recv: RecvInstance,
|
||
|
video: Option<Weak<ReceiverInner>>,
|
||
|
audio: Option<Weak<ReceiverInner>>,
|
||
|
},
|
||
|
}
|
||
|
|
||
|
lazy_static! {
|
||
|
static ref HASHMAP_RECEIVERS: Mutex<HashMap<usize, ReceiverInfo>> = {
|
||
|
let m = HashMap::new();
|
||
|
Mutex::new(m)
|
||
|
};
|
||
|
}
|
||
|
|
||
|
static ID_RECEIVER: AtomicUsize = AtomicUsize::new(0);
|
||
|
|
||
|
#[derive(Clone)]
|
||
|
pub struct Receiver(Arc<ReceiverInner>);
|
||
|
|
||
|
#[derive(Debug)]
|
||
|
pub enum ReceiverItem {
|
||
|
AudioBuffer(gst::Buffer, gst_audio::AudioInfo),
|
||
|
VideoBuffer(gst::Buffer, gst_video::VideoInfo),
|
||
|
Flushing,
|
||
|
Timeout,
|
||
|
Error(gst::FlowError),
|
||
|
}
|
||
|
|
||
|
struct ReceiverInner {
|
||
|
id: usize,
|
||
|
|
||
|
queue: ReceiverQueue,
|
||
|
|
||
|
video: bool,
|
||
|
|
||
|
recv: Mutex<Option<RecvInstance>>,
|
||
|
recv_cond: Condvar,
|
||
|
|
||
|
cat: gst::DebugCategory,
|
||
|
element: glib::WeakRef<gst_base::BaseSrc>,
|
||
|
timestamp_mode: TimestampMode,
|
||
|
timeout: u32,
|
||
|
|
||
|
thread: Mutex<Option<std::thread::JoinHandle<()>>>,
|
||
|
}
|
||
|
|
||
|
#[derive(Clone)]
|
||
|
struct ReceiverQueue(Arc<(Mutex<ReceiverQueueInner>, Condvar)>);
|
||
|
|
||
|
struct ReceiverQueueInner {
|
||
|
// If we should be capturing at all or go out of our capture loop
|
||
|
//
|
||
|
// This is true as long as the source element is in Paused/Playing
|
||
|
capturing: bool,
|
||
|
|
||
|
// If we're flushing right now and all buffers should simply be discarded
|
||
|
// and capture() directly returns Flushing
|
||
|
flushing: bool,
|
||
|
|
||
|
// If we're playing right now or not: if not we simply discard everything captured
|
||
|
playing: bool,
|
||
|
// Queue containing our buffers. This holds at most 5 buffers at a time.
|
||
|
//
|
||
|
// On timeout/error will contain a single item and then never be filled again
|
||
|
buffer_queue: VecDeque<ReceiverItem>,
|
||
|
|
||
|
error: Option<gst::FlowError>,
|
||
|
timeout: bool,
|
||
|
}
|
||
|
|
||
|
#[derive(Clone)]
|
||
|
pub struct ReceiverControlHandle {
|
||
|
queue: ReceiverQueue,
|
||
|
}
|
||
|
|
||
|
impl ReceiverControlHandle {
|
||
|
pub fn set_flushing(&self, flushing: bool) {
|
||
|
let mut queue = (self.queue.0).0.lock().unwrap();
|
||
|
queue.flushing = flushing;
|
||
|
(self.queue.0).1.notify_all();
|
||
|
}
|
||
|
|
||
|
pub fn set_playing(&self, playing: bool) {
|
||
|
let mut queue = (self.queue.0).0.lock().unwrap();
|
||
|
queue.playing = playing;
|
||
|
}
|
||
|
|
||
|
pub fn shutdown(&self) {
|
||
|
let mut queue = (self.queue.0).0.lock().unwrap();
|
||
|
queue.capturing = false;
|
||
|
(self.queue.0).1.notify_all();
|
||
|
}
|
||
|
}
|
||
|
|
||
|
impl Receiver {
|
||
|
fn new(
|
||
|
info: &mut ReceiverInfo,
|
||
|
video: bool,
|
||
|
timestamp_mode: TimestampMode,
|
||
|
timeout: u32,
|
||
|
element: &gst_base::BaseSrc,
|
||
|
cat: gst::DebugCategory,
|
||
|
) -> Self {
|
||
|
let (id, storage, recv) = if video {
|
||
|
match info {
|
||
|
ReceiverInfo::Connecting {
|
||
|
id, ref mut video, ..
|
||
|
} => (*id, video, None),
|
||
|
ReceiverInfo::Connected {
|
||
|
id,
|
||
|
ref mut video,
|
||
|
ref mut recv,
|
||
|
..
|
||
|
} => (*id, video, Some(recv.clone())),
|
||
|
}
|
||
|
} else {
|
||
|
match info {
|
||
|
ReceiverInfo::Connecting {
|
||
|
id, ref mut audio, ..
|
||
|
} => (*id, audio, None),
|
||
|
ReceiverInfo::Connected {
|
||
|
id,
|
||
|
ref mut audio,
|
||
|
ref mut recv,
|
||
|
..
|
||
|
} => (*id, audio, Some(recv.clone())),
|
||
|
}
|
||
|
};
|
||
|
assert!(storage.is_none());
|
||
|
|
||
|
let receiver = Receiver(Arc::new(ReceiverInner {
|
||
|
id,
|
||
|
queue: ReceiverQueue(Arc::new((
|
||
|
Mutex::new(ReceiverQueueInner {
|
||
|
capturing: true,
|
||
|
playing: false,
|
||
|
flushing: false,
|
||
|
buffer_queue: VecDeque::with_capacity(5),
|
||
|
error: None,
|
||
|
timeout: false,
|
||
|
}),
|
||
|
Condvar::new(),
|
||
|
))),
|
||
|
video,
|
||
|
recv: Mutex::new(recv),
|
||
|
recv_cond: Condvar::new(),
|
||
|
cat,
|
||
|
element: element.downgrade(),
|
||
|
timestamp_mode,
|
||
|
timeout,
|
||
|
thread: Mutex::new(None),
|
||
|
}));
|
||
|
|
||
|
let weak = Arc::downgrade(&receiver.0);
|
||
|
let thread = thread::spawn(move || {
|
||
|
use std::panic;
|
||
|
|
||
|
let weak_clone = weak.clone();
|
||
|
match panic::catch_unwind(panic::AssertUnwindSafe(move || receive_thread(&weak_clone)))
|
||
|
{
|
||
|
Ok(_) => (),
|
||
|
Err(_) => {
|
||
|
if let Some(receiver) = weak.upgrade().map(Receiver) {
|
||
|
if let Some(element) = receiver.0.element.upgrade() {
|
||
|
gst_element_error!(
|
||
|
element,
|
||
|
gst::LibraryError::Failed,
|
||
|
["Panic while connecting to NDI source"]
|
||
|
);
|
||
|
}
|
||
|
|
||
|
let mut queue = (receiver.0.queue.0).0.lock().unwrap();
|
||
|
queue.error = Some(gst::FlowError::Error);
|
||
|
(receiver.0.queue.0).1.notify_one();
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
});
|
||
|
|
||
|
let weak = Arc::downgrade(&receiver.0);
|
||
|
*storage = Some(weak);
|
||
|
|
||
|
*receiver.0.thread.lock().unwrap() = Some(thread);
|
||
|
|
||
|
receiver
|
||
|
}
|
||
|
|
||
|
pub fn receiver_control_handle(&self) -> ReceiverControlHandle {
|
||
|
ReceiverControlHandle {
|
||
|
queue: self.0.queue.clone(),
|
||
|
}
|
||
|
}
|
||
|
|
||
|
pub fn set_flushing(&self, flushing: bool) {
|
||
|
let mut queue = (self.0.queue.0).0.lock().unwrap();
|
||
|
queue.flushing = flushing;
|
||
|
(self.0.queue.0).1.notify_all();
|
||
|
}
|
||
|
|
||
|
pub fn set_playing(&self, playing: bool) {
|
||
|
let mut queue = (self.0.queue.0).0.lock().unwrap();
|
||
|
queue.playing = playing;
|
||
|
}
|
||
|
|
||
|
pub fn shutdown(&self) {
|
||
|
let mut queue = (self.0.queue.0).0.lock().unwrap();
|
||
|
queue.capturing = false;
|
||
|
(self.0.queue.0).1.notify_all();
|
||
|
}
|
||
|
|
||
|
pub fn capture(&self) -> ReceiverItem {
|
||
|
let mut queue = (self.0.queue.0).0.lock().unwrap();
|
||
|
loop {
|
||
|
if let Some(err) = queue.error {
|
||
|
return ReceiverItem::Error(err);
|
||
|
} else if queue.buffer_queue.is_empty() && queue.timeout {
|
||
|
return ReceiverItem::Timeout;
|
||
|
} else if queue.flushing || !queue.capturing {
|
||
|
return ReceiverItem::Flushing;
|
||
|
} else if let Some(item) = queue.buffer_queue.pop_front() {
|
||
|
return item;
|
||
|
}
|
||
|
|
||
|
queue = (self.0.queue.0).1.wait(queue).unwrap();
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
|
||
|
impl Drop for ReceiverInner {
|
||
|
fn drop(&mut self) {
|
||
|
// Will shut down the receiver thread on the next iteration
|
||
|
let mut queue = (self.queue.0).0.lock().unwrap();
|
||
|
queue.capturing = false;
|
||
|
drop(queue);
|
||
|
|
||
|
let element = self.element.upgrade();
|
||
|
|
||
|
if let Some(ref element) = element {
|
||
|
gst_debug!(self.cat, obj: element, "Closing NDI connection...");
|
||
|
}
|
||
|
|
||
|
let mut receivers = HASHMAP_RECEIVERS.lock().unwrap();
|
||
|
{
|
||
|
let val = receivers.get_mut(&self.id).unwrap();
|
||
|
let (audio, video) = match val {
|
||
|
ReceiverInfo::Connecting {
|
||
|
ref mut audio,
|
||
|
ref mut video,
|
||
|
..
|
||
|
} => (audio, video),
|
||
|
ReceiverInfo::Connected {
|
||
|
ref mut audio,
|
||
|
ref mut video,
|
||
|
..
|
||
|
} => (audio, video),
|
||
|
};
|
||
|
if video.is_some() && audio.is_some() {
|
||
|
if self.video {
|
||
|
*video = None;
|
||
|
} else {
|
||
|
*audio = None;
|
||
|
}
|
||
|
return;
|
||
|
}
|
||
|
}
|
||
|
receivers.remove(&self.id);
|
||
|
|
||
|
if let Some(ref element) = element {
|
||
|
gst_debug!(self.cat, obj: element, "Closed NDI connection");
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
|
||
|
pub fn connect_ndi(
|
||
|
cat: gst::DebugCategory,
|
||
|
element: &gst_base::BaseSrc,
|
||
|
ip_address: Option<&str>,
|
||
|
ndi_name: Option<&str>,
|
||
|
receiver_ndi_name: &str,
|
||
|
connect_timeout: u32,
|
||
|
bandwidth: NDIlib_recv_bandwidth_e,
|
||
|
timestamp_mode: TimestampMode,
|
||
|
timeout: u32,
|
||
|
) -> Option<Receiver> {
|
||
|
gst_debug!(cat, obj: element, "Starting NDI connection...");
|
||
|
|
||
|
let mut receivers = HASHMAP_RECEIVERS.lock().unwrap();
|
||
|
|
||
|
let video = element.get_type() == ndivideosrc::NdiVideoSrc::get_type();
|
||
|
|
||
|
// Check if we already have a receiver for this very stream
|
||
|
for val in receivers.values_mut() {
|
||
|
let (val_audio, val_video, val_ip_address, val_ndi_name) = match val {
|
||
|
ReceiverInfo::Connecting {
|
||
|
ref mut audio,
|
||
|
ref mut video,
|
||
|
ref ip_address,
|
||
|
ref ndi_name,
|
||
|
..
|
||
|
} => (
|
||
|
audio,
|
||
|
video,
|
||
|
ip_address.as_ref().map(String::as_ref),
|
||
|
ndi_name.as_ref().map(String::as_ref),
|
||
|
),
|
||
|
ReceiverInfo::Connected {
|
||
|
ref mut audio,
|
||
|
ref mut video,
|
||
|
ref ip_address,
|
||
|
ref ndi_name,
|
||
|
..
|
||
|
} => (
|
||
|
audio,
|
||
|
video,
|
||
|
Some(ip_address.as_str()),
|
||
|
Some(ndi_name.as_str()),
|
||
|
),
|
||
|
};
|
||
|
|
||
|
if val_ip_address == ip_address || val_ndi_name == ndi_name {
|
||
|
if (val_video.is_some() || !video) && (val_audio.is_some() || video) {
|
||
|
gst_error!(
|
||
|
cat,
|
||
|
obj: element,
|
||
|
"Source with ndi-name '{:?}' and ip-address '{:?}' already in use for {}",
|
||
|
val_ndi_name,
|
||
|
val_ip_address,
|
||
|
if video { "video" } else { "audio" },
|
||
|
);
|
||
|
|
||
|
return None;
|
||
|
} else {
|
||
|
return Some(Receiver::new(
|
||
|
val,
|
||
|
video,
|
||
|
timestamp_mode,
|
||
|
timeout,
|
||
|
element,
|
||
|
cat,
|
||
|
));
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// Otherwise asynchronously search for it and return the receiver to the caller
|
||
|
let id_receiver = ID_RECEIVER.fetch_add(1, Ordering::SeqCst);
|
||
|
let mut info = ReceiverInfo::Connecting {
|
||
|
id: id_receiver,
|
||
|
ndi_name: ndi_name.map(String::from),
|
||
|
ip_address: ip_address.map(String::from),
|
||
|
video: None,
|
||
|
audio: None,
|
||
|
};
|
||
|
|
||
|
let receiver = Receiver::new(&mut info, video, timestamp_mode, timeout, element, cat);
|
||
|
|
||
|
receivers.insert(id_receiver, info);
|
||
|
|
||
|
let receiver_ndi_name = String::from(receiver_ndi_name);
|
||
|
let element = element.clone();
|
||
|
thread::spawn(move || {
|
||
|
use std::panic;
|
||
|
|
||
|
let res = match panic::catch_unwind(move || {
|
||
|
connect_ndi_async(
|
||
|
cat,
|
||
|
&element,
|
||
|
id_receiver,
|
||
|
receiver_ndi_name,
|
||
|
connect_timeout,
|
||
|
bandwidth,
|
||
|
)
|
||
|
}) {
|
||
|
Ok(res) => res,
|
||
|
Err(_) => Err(Some(gst_error_msg!(
|
||
|
gst::LibraryError::Failed,
|
||
|
["Panic while connecting to NDI source"]
|
||
|
))),
|
||
|
};
|
||
|
|
||
|
match res {
|
||
|
Ok(_) => (),
|
||
|
Err(None) => {
|
||
|
gst_debug!(cat, "Shutting down while connecting");
|
||
|
}
|
||
|
Err(Some(err)) => {
|
||
|
gst_error!(cat, "Error while connecting: {:?}", err);
|
||
|
let mut receivers = HASHMAP_RECEIVERS.lock().unwrap();
|
||
|
let info = match receivers.get_mut(&id_receiver) {
|
||
|
None => return,
|
||
|
Some(val) => val,
|
||
|
};
|
||
|
|
||
|
let (audio, video) = match info {
|
||
|
ReceiverInfo::Connecting {
|
||
|
ref audio,
|
||
|
ref video,
|
||
|
..
|
||
|
} => (audio, video),
|
||
|
ReceiverInfo::Connected { .. } => unreachable!(),
|
||
|
};
|
||
|
|
||
|
assert!(audio.is_some() || video.is_some());
|
||
|
|
||
|
if let Some(audio) = audio.as_ref().and_then(|v| v.upgrade()).map(Receiver) {
|
||
|
if let Some(element) = audio.0.element.upgrade() {
|
||
|
element.post_error_message(&err);
|
||
|
}
|
||
|
let audio_recv = audio.0.recv.lock().unwrap();
|
||
|
let mut queue = (audio.0.queue.0).0.lock().unwrap();
|
||
|
assert!(audio_recv.is_none());
|
||
|
queue.error = Some(gst::FlowError::Error);
|
||
|
audio.0.recv_cond.notify_one();
|
||
|
(audio.0.queue.0).1.notify_one();
|
||
|
}
|
||
|
|
||
|
if let Some(video) = video.as_ref().and_then(|v| v.upgrade()).map(Receiver) {
|
||
|
if let Some(element) = video.0.element.upgrade() {
|
||
|
element.post_error_message(&err);
|
||
|
}
|
||
|
let video_recv = video.0.recv.lock().unwrap();
|
||
|
let mut queue = (video.0.queue.0).0.lock().unwrap();
|
||
|
assert!(video_recv.is_none());
|
||
|
queue.error = Some(gst::FlowError::Error);
|
||
|
video.0.recv_cond.notify_one();
|
||
|
(video.0.queue.0).1.notify_one();
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
});
|
||
|
|
||
|
Some(receiver)
|
||
|
}
|
||
|
|
||
|
fn connect_ndi_async(
|
||
|
cat: gst::DebugCategory,
|
||
|
element: &gst_base::BaseSrc,
|
||
|
id_receiver: usize,
|
||
|
receiver_ndi_name: String,
|
||
|
connect_timeout: u32,
|
||
|
bandwidth: NDIlib_recv_bandwidth_e,
|
||
|
) -> Result<(), Option<gst::ErrorMessage>> {
|
||
|
let mut find = match FindInstance::builder().build() {
|
||
|
None => {
|
||
|
return Err(Some(gst_error_msg!(
|
||
|
gst::CoreError::Negotiation,
|
||
|
["Cannot run NDI: NDIlib_find_create_v2 error"]
|
||
|
)));
|
||
|
}
|
||
|
Some(find) => find,
|
||
|
};
|
||
|
|
||
|
let timer = time::Instant::now();
|
||
|
let source = loop {
|
||
|
find.wait_for_sources(100);
|
||
|
let sources = find.get_current_sources();
|
||
|
|
||
|
gst_debug!(
|
||
|
cat,
|
||
|
obj: element,
|
||
|
"Total sources found in network {}",
|
||
|
sources.len(),
|
||
|
);
|
||
|
|
||
|
{
|
||
|
let receivers = HASHMAP_RECEIVERS.lock().unwrap();
|
||
|
let info = match receivers.get(&id_receiver) {
|
||
|
None => return Err(None),
|
||
|
Some(val) => val,
|
||
|
};
|
||
|
|
||
|
let (ndi_name, ip_address) = match info {
|
||
|
ReceiverInfo::Connecting {
|
||
|
ref ndi_name,
|
||
|
ref ip_address,
|
||
|
ref audio,
|
||
|
ref video,
|
||
|
..
|
||
|
} => {
|
||
|
assert!(audio.is_some() || video.is_some());
|
||
|
(ndi_name, ip_address)
|
||
|
}
|
||
|
ReceiverInfo::Connected { .. } => unreachable!(),
|
||
|
};
|
||
|
|
||
|
let source = sources.iter().find(|s| {
|
||
|
Some(s.ndi_name()) == ndi_name.as_ref().map(String::as_str)
|
||
|
|| Some(s.ip_address()) == ip_address.as_ref().map(String::as_str)
|
||
|
});
|
||
|
|
||
|
if let Some(source) = source {
|
||
|
break source.to_owned();
|
||
|
}
|
||
|
}
|
||
|
|
||
|
if timer.elapsed().as_millis() >= connect_timeout as u128 {
|
||
|
return Err(Some(gst_error_msg!(
|
||
|
gst::ResourceError::NotFound,
|
||
|
["Stream not found"]
|
||
|
)));
|
||
|
}
|
||
|
};
|
||
|
|
||
|
gst_debug!(
|
||
|
cat,
|
||
|
obj: element,
|
||
|
"Connecting to NDI source with ndi-name '{}' and ip-address '{}'",
|
||
|
source.ndi_name(),
|
||
|
source.ip_address(),
|
||
|
);
|
||
|
|
||
|
// FIXME: Ideally we would use NDIlib_recv_color_format_fastest here but that seems to be
|
||
|
// broken with interlaced content currently
|
||
|
let recv = RecvInstance::builder(&source, &receiver_ndi_name)
|
||
|
.bandwidth(bandwidth)
|
||
|
.color_format(NDIlib_recv_color_format_e::NDIlib_recv_color_format_UYVY_BGRA)
|
||
|
.allow_video_fields(true)
|
||
|
.build();
|
||
|
let recv = match recv {
|
||
|
None => {
|
||
|
return Err(Some(gst_error_msg!(
|
||
|
gst::CoreError::Negotiation,
|
||
|
["Cannot run NDI: NDIlib_recv_create_v3 error"]
|
||
|
)));
|
||
|
}
|
||
|
Some(recv) => recv,
|
||
|
};
|
||
|
|
||
|
recv.set_tally(&Tally::default());
|
||
|
|
||
|
let enable_hw_accel = MetadataFrame::new(0, Some("<ndi_hwaccel enabled=\"true\"/>"));
|
||
|
recv.send_metadata(&enable_hw_accel);
|
||
|
|
||
|
let mut receivers = HASHMAP_RECEIVERS.lock().unwrap();
|
||
|
let info = match receivers.get_mut(&id_receiver) {
|
||
|
None => return Err(None),
|
||
|
Some(val) => val,
|
||
|
};
|
||
|
|
||
|
let (audio, video) = match info {
|
||
|
ReceiverInfo::Connecting {
|
||
|
ref audio,
|
||
|
ref video,
|
||
|
..
|
||
|
} => (audio.clone(), video.clone()),
|
||
|
ReceiverInfo::Connected { .. } => unreachable!(),
|
||
|
};
|
||
|
|
||
|
assert!(audio.is_some() || video.is_some());
|
||
|
|
||
|
*info = ReceiverInfo::Connected {
|
||
|
id: id_receiver,
|
||
|
ndi_name: source.ndi_name().to_owned(),
|
||
|
ip_address: source.ip_address().to_owned(),
|
||
|
recv: recv.clone(),
|
||
|
video: video.clone(),
|
||
|
audio: audio.clone(),
|
||
|
};
|
||
|
|
||
|
gst_debug!(cat, obj: element, "Started NDI connection");
|
||
|
|
||
|
if let Some(audio) = audio.and_then(|v| v.upgrade()).map(Receiver) {
|
||
|
let mut audio_recv = audio.0.recv.lock().unwrap();
|
||
|
assert!(audio_recv.is_none());
|
||
|
*audio_recv = Some(recv.clone());
|
||
|
audio.0.recv_cond.notify_one();
|
||
|
}
|
||
|
|
||
|
if let Some(video) = video.and_then(|v| v.upgrade()).map(Receiver) {
|
||
|
let mut video_recv = video.0.recv.lock().unwrap();
|
||
|
assert!(video_recv.is_none());
|
||
|
*video_recv = Some(recv.clone());
|
||
|
video.0.recv_cond.notify_one();
|
||
|
}
|
||
|
|
||
|
Ok(())
|
||
|
}
|
||
|
|
||
|
fn receive_thread(receiver: &Weak<ReceiverInner>) {
|
||
|
// First loop until we actually are connected, or an error happened
|
||
|
let recv = {
|
||
|
let receiver = match receiver.upgrade().map(Receiver) {
|
||
|
None => return,
|
||
|
Some(receiver) => receiver,
|
||
|
};
|
||
|
|
||
|
let element = receiver.0.element.upgrade().unwrap();
|
||
|
|
||
|
let mut recv = receiver.0.recv.lock().unwrap();
|
||
|
loop {
|
||
|
{
|
||
|
let queue = (receiver.0.queue.0).0.lock().unwrap();
|
||
|
if !queue.capturing {
|
||
|
gst_debug!(receiver.0.cat, obj: &element, "Shutting down");
|
||
|
return;
|
||
|
}
|
||
|
|
||
|
// If an error happened in the meantime, just go out of here
|
||
|
if let Some(_) = queue.error {
|
||
|
gst_error!(
|
||
|
receiver.0.cat,
|
||
|
obj: &element,
|
||
|
"Error while waiting for connection"
|
||
|
);
|
||
|
return;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
if let Some(ref recv) = *recv {
|
||
|
break recv.clone();
|
||
|
}
|
||
|
|
||
|
recv = receiver.0.recv_cond.wait(recv).unwrap();
|
||
|
}
|
||
|
};
|
||
|
|
||
|
// Now first capture frames until the queues are empty so that we're sure that we output only
|
||
|
// the very latest frame that is available now
|
||
|
loop {
|
||
|
let receiver = match receiver.upgrade().map(Receiver) {
|
||
|
None => return,
|
||
|
Some(receiver) => receiver,
|
||
|
};
|
||
|
|
||
|
let element = receiver.0.element.upgrade().unwrap();
|
||
|
|
||
|
{
|
||
|
let queue = (receiver.0.queue.0).0.lock().unwrap();
|
||
|
if !queue.capturing {
|
||
|
gst_debug!(receiver.0.cat, obj: &element, "Shutting down");
|
||
|
return;
|
||
|
}
|
||
|
|
||
|
// If an error happened in the meantime, just go out of here
|
||
|
if let Some(_) = queue.error {
|
||
|
gst_error!(
|
||
|
receiver.0.cat,
|
||
|
obj: &element,
|
||
|
"Error while waiting for connection"
|
||
|
);
|
||
|
return;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
let queue = recv.get_queue();
|
||
|
if (!receiver.0.video && queue.audio_frames() <= 1)
|
||
|
|| (receiver.0.video && queue.video_frames() <= 1)
|
||
|
{
|
||
|
break;
|
||
|
}
|
||
|
|
||
|
let _ = recv.capture(receiver.0.video, !receiver.0.video, false, 0);
|
||
|
}
|
||
|
|
||
|
// And if that went fine, capture until we're done
|
||
|
loop {
|
||
|
let receiver = match receiver.upgrade().map(Receiver) {
|
||
|
None => break,
|
||
|
Some(receiver) => receiver,
|
||
|
};
|
||
|
|
||
|
let element = receiver.0.element.upgrade().unwrap();
|
||
|
|
||
|
{
|
||
|
let queue = (receiver.0.queue.0).0.lock().unwrap();
|
||
|
if !queue.capturing {
|
||
|
gst_debug!(receiver.0.cat, obj: &element, "Shutting down");
|
||
|
break;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
let res = if receiver.0.video {
|
||
|
receiver
|
||
|
.capture_video(&element, &recv)
|
||
|
.map(|(buffer, info)| ReceiverItem::VideoBuffer(buffer, info))
|
||
|
} else {
|
||
|
receiver
|
||
|
.capture_audio(&element, &recv)
|
||
|
.map(|(buffer, info)| ReceiverItem::AudioBuffer(buffer, info))
|
||
|
};
|
||
|
|
||
|
match res {
|
||
|
Ok(item) => {
|
||
|
let mut queue = (receiver.0.queue.0).0.lock().unwrap();
|
||
|
while queue.buffer_queue.len() > 5 {
|
||
|
gst_warning!(
|
||
|
receiver.0.cat,
|
||
|
obj: &element,
|
||
|
"Dropping old buffer -- queue has {} items",
|
||
|
queue.buffer_queue.len()
|
||
|
);
|
||
|
queue.buffer_queue.pop_front();
|
||
|
}
|
||
|
queue.buffer_queue.push_back(item);
|
||
|
(receiver.0.queue.0).1.notify_one();
|
||
|
}
|
||
|
Err(gst::FlowError::Eos) => {
|
||
|
gst_debug!(receiver.0.cat, obj: &element, "Signalling EOS");
|
||
|
let mut queue = (receiver.0.queue.0).0.lock().unwrap();
|
||
|
queue.timeout = true;
|
||
|
(receiver.0.queue.0).1.notify_one();
|
||
|
}
|
||
|
Err(gst::FlowError::CustomError) => {
|
||
|
// Flushing, nothing to be done here except for emptying our queue
|
||
|
let mut queue = (receiver.0.queue.0).0.lock().unwrap();
|
||
|
queue.buffer_queue.clear();
|
||
|
(receiver.0.queue.0).1.notify_one();
|
||
|
}
|
||
|
Err(err) => {
|
||
|
gst_error!(receiver.0.cat, obj: &element, "Signalling error");
|
||
|
let mut queue = (receiver.0.queue.0).0.lock().unwrap();
|
||
|
if queue.error.is_none() {
|
||
|
queue.error = Some(err);
|
||
|
}
|
||
|
(receiver.0.queue.0).1.notify_one();
|
||
|
break;
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
|
||
|
impl Receiver {
|
||
|
fn capture_video(
|
||
|
&self,
|
||
|
element: &gst_base::BaseSrc,
|
||
|
recv: &RecvInstance,
|
||
|
) -> Result<(gst::Buffer, gst_video::VideoInfo), gst::FlowError> {
|
||
|
let timeout = time::Instant::now();
|
||
|
let mut flushing;
|
||
|
let mut playing;
|
||
|
|
||
|
let video_frame = loop {
|
||
|
{
|
||
|
let queue = (self.0.queue.0).0.lock().unwrap();
|
||
|
playing = queue.playing;
|
||
|
flushing = queue.flushing;
|
||
|
if !queue.capturing {
|
||
|
gst_debug!(self.0.cat, obj: element, "Shutting down");
|
||
|
return Err(gst::FlowError::Flushing);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
let res = match recv.capture(true, false, false, 50) {
|
||
|
Err(_) => Err(()),
|
||
|
Ok(None) => Ok(None),
|
||
|
Ok(Some(Frame::Video(frame))) => Ok(Some(frame)),
|
||
|
_ => unreachable!(),
|
||
|
};
|
||
|
|
||
|
let video_frame = match res {
|
||
|
Err(_) => {
|
||
|
gst_element_error!(element, gst::ResourceError::Read, ["NDI frame type error received, assuming that the source closed the stream...."]);
|
||
|
return Err(gst::FlowError::Error);
|
||
|
}
|
||
|
Ok(None) if timeout.elapsed().as_millis() >= self.0.timeout as u128 => {
|
||
|
return Err(gst::FlowError::Eos);
|
||
|
}
|
||
|
Ok(None) => {
|
||
|
gst_debug!(
|
||
|
self.0.cat,
|
||
|
obj: element,
|
||
|
"No video frame received yet, retry"
|
||
|
);
|
||
|
continue;
|
||
|
}
|
||
|
Ok(Some(frame)) => frame,
|
||
|
};
|
||
|
|
||
|
break video_frame;
|
||
|
};
|
||
|
|
||
|
// Simply read all video frames while flushing but don't copy them or anything to
|
||
|
// make sure that we're not accumulating anything here
|
||
|
if !playing || flushing {
|
||
|
gst_debug!(self.0.cat, obj: element, "Flushing, dropping buffer");
|
||
|
return Err(gst::FlowError::CustomError);
|
||
|
}
|
||
|
|
||
|
let pts = self.calculate_video_timestamp(element, &video_frame);
|
||
|
let info = self.create_video_info(element, &video_frame)?;
|
||
|
|
||
|
let buffer = self.create_video_buffer(element, pts, &info, &video_frame)?;
|
||
|
|
||
|
gst_log!(self.0.cat, obj: element, "Produced buffer {:?}", buffer);
|
||
|
|
||
|
Ok((buffer, info))
|
||
|
}
|
||
|
|
||
|
fn calculate_video_timestamp(
|
||
|
&self,
|
||
|
element: &gst_base::BaseSrc,
|
||
|
video_frame: &VideoFrame,
|
||
|
) -> gst::ClockTime {
|
||
|
let clock = element.get_clock().unwrap();
|
||
|
|
||
|
// For now take the current running time as PTS. At a later time we
|
||
|
// will want to work with the timestamp given by the NDI SDK if available
|
||
|
let now = clock.get_time();
|
||
|
let base_time = element.get_base_time();
|
||
|
let receive_time = now - base_time;
|
||
|
|
||
|
let real_time_now = gst::ClockTime::from(glib::get_real_time() as u64 * 1000);
|
||
|
let timestamp = if video_frame.timestamp() == ndisys::NDIlib_recv_timestamp_undefined {
|
||
|
gst::CLOCK_TIME_NONE
|
||
|
} else {
|
||
|
gst::ClockTime::from(video_frame.timestamp() as u64 * 100)
|
||
|
};
|
||
|
let timecode = gst::ClockTime::from(video_frame.timecode() as u64 * 100);
|
||
|
|
||
|
gst_log!(
|
||
|
self.0.cat,
|
||
|
obj: element,
|
||
|
"NDI video frame received: {:?} with timecode {} and timestamp {}, receive time {}, local time now {}",
|
||
|
video_frame,
|
||
|
timecode,
|
||
|
timestamp,
|
||
|
receive_time,
|
||
|
real_time_now,
|
||
|
);
|
||
|
|
||
|
let pts = match self.0.timestamp_mode {
|
||
|
TimestampMode::ReceiveTime => receive_time,
|
||
|
TimestampMode::Timecode => timecode,
|
||
|
TimestampMode::Timestamp if timestamp.is_none() => receive_time,
|
||
|
TimestampMode::Timestamp => {
|
||
|
// Timestamps are relative to the UNIX epoch
|
||
|
if real_time_now > timestamp {
|
||
|
let diff = real_time_now - timestamp;
|
||
|
if diff > receive_time {
|
||
|
0.into()
|
||
|
} else {
|
||
|
receive_time - diff
|
||
|
}
|
||
|
} else {
|
||
|
let diff = timestamp - real_time_now;
|
||
|
receive_time + diff
|
||
|
}
|
||
|
}
|
||
|
};
|
||
|
|
||
|
gst_log!(
|
||
|
self.0.cat,
|
||
|
obj: element,
|
||
|
"Calculated pts for video frame: {:?}",
|
||
|
pts
|
||
|
);
|
||
|
|
||
|
pts
|
||
|
}
|
||
|
|
||
|
fn create_video_info(
|
||
|
&self,
|
||
|
_element: &gst_base::BaseSrc,
|
||
|
video_frame: &VideoFrame,
|
||
|
) -> Result<gst_video::VideoInfo, gst::FlowError> {
|
||
|
// YV12 and I420 are swapped in the NDI SDK compared to GStreamer
|
||
|
let format = match video_frame.fourcc() {
|
||
|
ndisys::NDIlib_FourCC_type_e::NDIlib_FourCC_type_UYVY => gst_video::VideoFormat::Uyvy,
|
||
|
ndisys::NDIlib_FourCC_type_e::NDIlib_FourCC_type_YV12 => gst_video::VideoFormat::I420,
|
||
|
ndisys::NDIlib_FourCC_type_e::NDIlib_FourCC_type_NV12 => gst_video::VideoFormat::Nv12,
|
||
|
ndisys::NDIlib_FourCC_type_e::NDIlib_FourCC_type_I420 => gst_video::VideoFormat::Yv12,
|
||
|
ndisys::NDIlib_FourCC_type_e::NDIlib_FourCC_type_BGRA => gst_video::VideoFormat::Bgra,
|
||
|
ndisys::NDIlib_FourCC_type_e::NDIlib_FourCC_type_BGRX => gst_video::VideoFormat::Bgrx,
|
||
|
ndisys::NDIlib_FourCC_type_e::NDIlib_FourCC_type_RGBA => gst_video::VideoFormat::Rgba,
|
||
|
ndisys::NDIlib_FourCC_type_e::NDIlib_FourCC_type_RGBX => gst_video::VideoFormat::Rgbx,
|
||
|
ndisys::NDIlib_FourCC_type_e::NDIlib_FourCC_type_UYVA => gst_video::VideoFormat::Uyvy,
|
||
|
};
|
||
|
|
||
|
let par = gst::Fraction::approximate_f32(video_frame.picture_aspect_ratio()).unwrap()
|
||
|
* gst::Fraction::new(video_frame.yres(), video_frame.xres());
|
||
|
|
||
|
#[cfg(feature = "interlaced-fields")]
|
||
|
{
|
||
|
let mut builder = gst_video::VideoInfo::new(
|
||
|
format,
|
||
|
video_frame.xres() as u32,
|
||
|
video_frame.yres() as u32,
|
||
|
)
|
||
|
.fps(gst::Fraction::from(video_frame.frame_rate()))
|
||
|
.par(par)
|
||
|
.interlace_mode(match video_frame.frame_format_type() {
|
||
|
ndisys::NDIlib_frame_format_type_e::NDIlib_frame_format_type_progressive => {
|
||
|
gst_video::VideoInterlaceMode::Progressive
|
||
|
}
|
||
|
ndisys::NDIlib_frame_format_type_e::NDIlib_frame_format_type_interleaved => {
|
||
|
gst_video::VideoInterlaceMode::Interleaved
|
||
|
}
|
||
|
_ => gst_video::VideoInterlaceMode::Alternate,
|
||
|
});
|
||
|
|
||
|
/* Requires GStreamer 1.12 at least */
|
||
|
if video_frame.frame_format_type()
|
||
|
== ndisys::NDIlib_frame_format_type_e::NDIlib_frame_format_type_interleaved
|
||
|
{
|
||
|
builder = builder.field_order(gst_video::VideoFieldOrder::TopFieldFirst);
|
||
|
}
|
||
|
|
||
|
Ok(builder.build().unwrap())
|
||
|
}
|
||
|
|
||
|
#[cfg(not(feature = "interlaced-fields"))]
|
||
|
{
|
||
|
if video_frame.frame_format_type()
|
||
|
!= ndisys::NDIlib_frame_format_type_e::NDIlib_frame_format_type_progressive
|
||
|
&& video_frame.frame_format_type()
|
||
|
!= ndisys::NDIlib_frame_format_type_e::NDIlib_frame_format_type_interleaved
|
||
|
{
|
||
|
gst_element_error!(
|
||
|
element,
|
||
|
gst::StreamError::Format,
|
||
|
["Separate field interlacing not supported"]
|
||
|
);
|
||
|
return Err(gst::FlowError::NotNegotiated);
|
||
|
}
|
||
|
|
||
|
let builder = gst_video::VideoInfo::new(
|
||
|
format,
|
||
|
video_frame.xres() as u32,
|
||
|
video_frame.yres() as u32,
|
||
|
)
|
||
|
.fps(gst::Fraction::from(video_frame.frame_rate()))
|
||
|
.par(par)
|
||
|
.interlace_mode(
|
||
|
if video_frame.frame_format_type()
|
||
|
== ndisys::NDIlib_frame_format_type_e::NDIlib_frame_format_type_progressive
|
||
|
{
|
||
|
gst_video::VideoInterlaceMode::Progressive
|
||
|
} else {
|
||
|
gst_video::VideoInterlaceMode::Interleaved
|
||
|
},
|
||
|
);
|
||
|
Ok(builder.build().unwrap());
|
||
|
}
|
||
|
}
|
||
|
|
||
|
fn create_video_buffer(
|
||
|
&self,
|
||
|
element: &gst_base::BaseSrc,
|
||
|
pts: gst::ClockTime,
|
||
|
info: &gst_video::VideoInfo,
|
||
|
video_frame: &VideoFrame,
|
||
|
) -> Result<gst::Buffer, gst::FlowError> {
|
||
|
let mut buffer = gst::Buffer::with_size(info.size()).unwrap();
|
||
|
{
|
||
|
let duration = gst::SECOND
|
||
|
.mul_div_floor(
|
||
|
video_frame.frame_rate().1 as u64,
|
||
|
video_frame.frame_rate().0 as u64,
|
||
|
)
|
||
|
.unwrap_or(gst::CLOCK_TIME_NONE);
|
||
|
let buffer = buffer.get_mut().unwrap();
|
||
|
buffer.set_pts(pts);
|
||
|
buffer.set_duration(duration);
|
||
|
|
||
|
#[cfg(feature = "reference-timestamps")]
|
||
|
{
|
||
|
gst::ReferenceTimestampMeta::add(
|
||
|
buffer,
|
||
|
&*TIMECODE_CAPS,
|
||
|
gst::ClockTime::from(video_frame.timecode() as u64 * 100),
|
||
|
gst::CLOCK_TIME_NONE,
|
||
|
);
|
||
|
if video_frame.timestamp() != ndisys::NDIlib_recv_timestamp_undefined {
|
||
|
gst::ReferenceTimestampMeta::add(
|
||
|
buffer,
|
||
|
&*TIMESTAMP_CAPS,
|
||
|
gst::ClockTime::from(video_frame.timestamp() as u64 * 100),
|
||
|
gst::CLOCK_TIME_NONE,
|
||
|
);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
#[cfg(feature = "interlaced-fields")]
|
||
|
{
|
||
|
match video_frame.frame_format_type() {
|
||
|
ndisys::NDIlib_frame_format_type_e::NDIlib_frame_format_type_interleaved => {
|
||
|
buffer.set_video_flags(
|
||
|
gst_video::VideoBufferFlags::INTERLACED
|
||
|
| gst_video::VideoBufferFlags::TFF,
|
||
|
);
|
||
|
}
|
||
|
ndisys::NDIlib_frame_format_type_e::NDIlib_frame_format_type_field_0 => {
|
||
|
buffer.set_video_flags(
|
||
|
gst_video::VideoBufferFlags::INTERLACED
|
||
|
| gst_video::VideoBufferFlags::TOP_FIELD,
|
||
|
);
|
||
|
}
|
||
|
ndisys::NDIlib_frame_format_type_e::NDIlib_frame_format_type_field_1 => {
|
||
|
buffer.set_video_flags(
|
||
|
gst_video::VideoBufferFlags::INTERLACED
|
||
|
| gst_video::VideoBufferFlags::BOTTOM_FIELD,
|
||
|
);
|
||
|
}
|
||
|
_ => (),
|
||
|
};
|
||
|
}
|
||
|
|
||
|
#[cfg(not(feature = "interlaced-fields"))]
|
||
|
{
|
||
|
if video_frame.frame_format_type()
|
||
|
== ndisys::NDIlib_frame_format_type_e::NDIlib_frame_format_type_interleaved
|
||
|
{
|
||
|
buffer.set_video_flags(
|
||
|
gst_video::VideoBufferFlags::INTERLACED | gst_video::VideoBufferFlags::TFF,
|
||
|
);
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
|
||
|
self.copy_video_frame(element, info, buffer, video_frame)
|
||
|
}
|
||
|
|
||
|
fn copy_video_frame(
|
||
|
&self,
|
||
|
_element: &gst_base::BaseSrc,
|
||
|
info: &gst_video::VideoInfo,
|
||
|
buffer: gst::Buffer,
|
||
|
video_frame: &VideoFrame,
|
||
|
) -> Result<gst::Buffer, gst::FlowError> {
|
||
|
// FIXME: Error handling if frame dimensions don't match
|
||
|
let mut vframe = gst_video::VideoFrame::from_buffer_writable(buffer, info).unwrap();
|
||
|
|
||
|
match info.format() {
|
||
|
gst_video::VideoFormat::Uyvy
|
||
|
| gst_video::VideoFormat::Bgra
|
||
|
| gst_video::VideoFormat::Bgrx
|
||
|
| gst_video::VideoFormat::Rgba
|
||
|
| gst_video::VideoFormat::Rgbx => {
|
||
|
let line_bytes = if info.format() == gst_video::VideoFormat::Uyvy {
|
||
|
2 * vframe.width() as usize
|
||
|
} else {
|
||
|
4 * vframe.width() as usize
|
||
|
};
|
||
|
let dest_stride = vframe.plane_stride()[0] as usize;
|
||
|
let dest = vframe.plane_data_mut(0).unwrap();
|
||
|
let src_stride = video_frame.line_stride_in_bytes() as usize;
|
||
|
let src = video_frame.data();
|
||
|
|
||
|
for (dest, src) in dest
|
||
|
.chunks_exact_mut(dest_stride)
|
||
|
.zip(src.chunks_exact(src_stride))
|
||
|
{
|
||
|
dest.copy_from_slice(src);
|
||
|
dest.copy_from_slice(&src[..line_bytes]);
|
||
|
}
|
||
|
}
|
||
|
gst_video::VideoFormat::Nv12 => {
|
||
|
// First plane
|
||
|
{
|
||
|
let line_bytes = vframe.width() as usize;
|
||
|
let dest_stride = vframe.plane_stride()[0] as usize;
|
||
|
let dest = vframe.plane_data_mut(0).unwrap();
|
||
|
let src_stride = video_frame.line_stride_in_bytes() as usize;
|
||
|
let src = video_frame.data();
|
||
|
|
||
|
for (dest, src) in dest
|
||
|
.chunks_exact_mut(dest_stride)
|
||
|
.zip(src.chunks_exact(src_stride))
|
||
|
{
|
||
|
dest.copy_from_slice(&src[..line_bytes]);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// Second plane
|
||
|
{
|
||
|
let line_bytes = vframe.width() as usize;
|
||
|
let dest_stride = vframe.plane_stride()[1] as usize;
|
||
|
let dest = vframe.plane_data_mut(1).unwrap();
|
||
|
let src_stride = video_frame.line_stride_in_bytes() as usize;
|
||
|
let src = &video_frame.data()[(video_frame.yres() as usize * src_stride)..];
|
||
|
|
||
|
for (dest, src) in dest
|
||
|
.chunks_exact_mut(dest_stride)
|
||
|
.zip(src.chunks_exact(src_stride))
|
||
|
{
|
||
|
dest.copy_from_slice(&src[..line_bytes]);
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
gst_video::VideoFormat::Yv12 | gst_video::VideoFormat::I420 => {
|
||
|
// First plane
|
||
|
{
|
||
|
let line_bytes = vframe.width() as usize;
|
||
|
let dest_stride = vframe.plane_stride()[0] as usize;
|
||
|
let dest = vframe.plane_data_mut(0).unwrap();
|
||
|
let src_stride = video_frame.line_stride_in_bytes() as usize;
|
||
|
let src = video_frame.data();
|
||
|
|
||
|
for (dest, src) in dest
|
||
|
.chunks_exact_mut(dest_stride)
|
||
|
.zip(src.chunks_exact(src_stride))
|
||
|
{
|
||
|
dest.copy_from_slice(&src[..line_bytes]);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// Second plane
|
||
|
{
|
||
|
let line_bytes = (vframe.width() as usize + 1) / 2;
|
||
|
let dest_stride = vframe.plane_stride()[1] as usize;
|
||
|
let dest = vframe.plane_data_mut(1).unwrap();
|
||
|
let src_stride = video_frame.line_stride_in_bytes() as usize;
|
||
|
let src_stride1 = video_frame.line_stride_in_bytes() as usize / 2;
|
||
|
let src = &video_frame.data()[(video_frame.yres() as usize * src_stride)..];
|
||
|
|
||
|
for (dest, src) in dest
|
||
|
.chunks_exact_mut(dest_stride)
|
||
|
.zip(src.chunks_exact(src_stride1))
|
||
|
{
|
||
|
dest.copy_from_slice(&src[..line_bytes]);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// Third plane
|
||
|
{
|
||
|
let line_bytes = (vframe.width() as usize + 1) / 2;
|
||
|
let dest_stride = vframe.plane_stride()[2] as usize;
|
||
|
let dest = vframe.plane_data_mut(2).unwrap();
|
||
|
let src_stride = video_frame.line_stride_in_bytes() as usize;
|
||
|
let src_stride1 = video_frame.line_stride_in_bytes() as usize / 2;
|
||
|
let src = &video_frame.data()[(video_frame.yres() as usize * src_stride
|
||
|
+ (video_frame.yres() as usize + 1) / 2 * src_stride1)..];
|
||
|
|
||
|
for (dest, src) in dest
|
||
|
.chunks_exact_mut(dest_stride)
|
||
|
.zip(src.chunks_exact(src_stride1))
|
||
|
{
|
||
|
dest.copy_from_slice(&src[..line_bytes]);
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
_ => unreachable!(),
|
||
|
}
|
||
|
|
||
|
Ok(vframe.into_buffer())
|
||
|
}
|
||
|
|
||
|
fn capture_audio(
|
||
|
&self,
|
||
|
element: &gst_base::BaseSrc,
|
||
|
recv: &RecvInstance,
|
||
|
) -> Result<(gst::Buffer, gst_audio::AudioInfo), gst::FlowError> {
|
||
|
let timeout = time::Instant::now();
|
||
|
let mut flushing;
|
||
|
let mut playing;
|
||
|
|
||
|
let audio_frame = loop {
|
||
|
{
|
||
|
let queue = (self.0.queue.0).0.lock().unwrap();
|
||
|
flushing = queue.flushing;
|
||
|
playing = queue.playing;
|
||
|
if !queue.capturing {
|
||
|
gst_debug!(self.0.cat, obj: element, "Shutting down");
|
||
|
return Err(gst::FlowError::Flushing);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
let res = match recv.capture(false, true, false, 50) {
|
||
|
Err(_) => Err(()),
|
||
|
Ok(None) => Ok(None),
|
||
|
Ok(Some(Frame::Audio(frame))) => Ok(Some(frame)),
|
||
|
_ => unreachable!(),
|
||
|
};
|
||
|
|
||
|
let audio_frame = match res {
|
||
|
Err(_) => {
|
||
|
gst_element_error!(element, gst::ResourceError::Read, ["NDI frame type error received, assuming that the source closed the stream...."]);
|
||
|
return Err(gst::FlowError::Error);
|
||
|
}
|
||
|
Ok(None) if timeout.elapsed().as_millis() >= self.0.timeout as u128 => {
|
||
|
return Err(gst::FlowError::Eos);
|
||
|
}
|
||
|
Ok(None) => {
|
||
|
gst_debug!(
|
||
|
self.0.cat,
|
||
|
obj: element,
|
||
|
"No audio frame received yet, retry"
|
||
|
);
|
||
|
continue;
|
||
|
}
|
||
|
Ok(Some(frame)) => frame,
|
||
|
};
|
||
|
|
||
|
break audio_frame;
|
||
|
};
|
||
|
|
||
|
// Simply read all video frames while flushing but don't copy them or anything to
|
||
|
// make sure that we're not accumulating anything here
|
||
|
if !playing || flushing {
|
||
|
gst_debug!(self.0.cat, obj: element, "Flushing, dropping buffer");
|
||
|
return Err(gst::FlowError::CustomError);
|
||
|
}
|
||
|
|
||
|
let pts = self.calculate_audio_timestamp(element, &audio_frame);
|
||
|
let info = self.create_audio_info(element, &audio_frame)?;
|
||
|
|
||
|
let buffer = self.create_audio_buffer(element, pts, &info, &audio_frame)?;
|
||
|
|
||
|
gst_log!(self.0.cat, obj: element, "Produced buffer {:?}", buffer);
|
||
|
|
||
|
Ok((buffer, info))
|
||
|
}
|
||
|
|
||
|
fn calculate_audio_timestamp(
|
||
|
&self,
|
||
|
element: &gst_base::BaseSrc,
|
||
|
audio_frame: &AudioFrame,
|
||
|
) -> gst::ClockTime {
|
||
|
let clock = element.get_clock().unwrap();
|
||
|
|
||
|
// For now take the current running time as PTS. At a later time we
|
||
|
// will want to work with the timestamp given by the NDI SDK if available
|
||
|
let now = clock.get_time();
|
||
|
let base_time = element.get_base_time();
|
||
|
let receive_time = now - base_time;
|
||
|
|
||
|
let real_time_now = gst::ClockTime::from(glib::get_real_time() as u64 * 1000);
|
||
|
let timestamp = if audio_frame.timestamp() == ndisys::NDIlib_recv_timestamp_undefined {
|
||
|
gst::CLOCK_TIME_NONE
|
||
|
} else {
|
||
|
gst::ClockTime::from(audio_frame.timestamp() as u64 * 100)
|
||
|
};
|
||
|
let timecode = gst::ClockTime::from(audio_frame.timecode() as u64 * 100);
|
||
|
|
||
|
gst_log!(
|
||
|
self.0.cat,
|
||
|
obj: element,
|
||
|
"NDI audio frame received: {:?} with timecode {} and timestamp {}, receive time {}, local time now {}",
|
||
|
audio_frame,
|
||
|
timecode,
|
||
|
timestamp,
|
||
|
receive_time,
|
||
|
real_time_now,
|
||
|
);
|
||
|
|
||
|
let pts = match self.0.timestamp_mode {
|
||
|
TimestampMode::ReceiveTime => receive_time,
|
||
|
TimestampMode::Timecode => timecode,
|
||
|
TimestampMode::Timestamp if timestamp.is_none() => receive_time,
|
||
|
TimestampMode::Timestamp => {
|
||
|
// Timestamps are relative to the UNIX epoch
|
||
|
if real_time_now > timestamp {
|
||
|
let diff = real_time_now - timestamp;
|
||
|
if diff > receive_time {
|
||
|
0.into()
|
||
|
} else {
|
||
|
receive_time - diff
|
||
|
}
|
||
|
} else {
|
||
|
let diff = timestamp - real_time_now;
|
||
|
receive_time + diff
|
||
|
}
|
||
|
}
|
||
|
};
|
||
|
|
||
|
gst_log!(
|
||
|
self.0.cat,
|
||
|
obj: element,
|
||
|
"Calculated pts for audio frame: {:?}",
|
||
|
pts
|
||
|
);
|
||
|
|
||
|
pts
|
||
|
}
|
||
|
|
||
|
fn create_audio_info(
|
||
|
&self,
|
||
|
_element: &gst_base::BaseSrc,
|
||
|
audio_frame: &AudioFrame,
|
||
|
) -> Result<gst_audio::AudioInfo, gst::FlowError> {
|
||
|
let builder = gst_audio::AudioInfo::new(
|
||
|
gst_audio::AUDIO_FORMAT_S16,
|
||
|
audio_frame.sample_rate() as u32,
|
||
|
audio_frame.no_channels() as u32,
|
||
|
);
|
||
|
|
||
|
Ok(builder.build().unwrap())
|
||
|
}
|
||
|
|
||
|
fn create_audio_buffer(
|
||
|
&self,
|
||
|
_element: &gst_base::BaseSrc,
|
||
|
pts: gst::ClockTime,
|
||
|
info: &gst_audio::AudioInfo,
|
||
|
audio_frame: &AudioFrame,
|
||
|
) -> Result<gst::Buffer, gst::FlowError> {
|
||
|
// We multiply by 2 because is the size in bytes of an i16 variable
|
||
|
let buff_size = (audio_frame.no_samples() as u32 * info.bpf()) as usize;
|
||
|
let mut buffer = gst::Buffer::with_size(buff_size).unwrap();
|
||
|
{
|
||
|
let duration = gst::SECOND
|
||
|
.mul_div_floor(
|
||
|
audio_frame.no_samples() as u64,
|
||
|
audio_frame.sample_rate() as u64,
|
||
|
)
|
||
|
.unwrap_or(gst::CLOCK_TIME_NONE);
|
||
|
let buffer = buffer.get_mut().unwrap();
|
||
|
|
||
|
buffer.set_pts(pts);
|
||
|
buffer.set_duration(duration);
|
||
|
|
||
|
#[cfg(feature = "reference-timestamps")]
|
||
|
{
|
||
|
gst::ReferenceTimestampMeta::add(
|
||
|
buffer,
|
||
|
&*TIMECODE_CAPS,
|
||
|
gst::ClockTime::from(audio_frame.timecode() as u64 * 100),
|
||
|
gst::CLOCK_TIME_NONE,
|
||
|
);
|
||
|
if audio_frame.timestamp() != ndisys::NDIlib_recv_timestamp_undefined {
|
||
|
gst::ReferenceTimestampMeta::add(
|
||
|
buffer,
|
||
|
&*TIMESTAMP_CAPS,
|
||
|
gst::ClockTime::from(audio_frame.timestamp() as u64 * 100),
|
||
|
gst::CLOCK_TIME_NONE,
|
||
|
);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
audio_frame.copy_to_interleaved_16s(
|
||
|
buffer
|
||
|
.map_writable()
|
||
|
.unwrap()
|
||
|
.as_mut_slice_of::<i16>()
|
||
|
.unwrap(),
|
||
|
);
|
||
|
}
|
||
|
|
||
|
Ok(buffer)
|
||
|
}
|
||
|
}
|