| 1 | /* |
| 2 | * Copyright (C) 2012, 2015, 2016, 2018 Igalia S.L |
| 3 | * Copyright (C) 2015, 2016, 2018 Metrological Group B.V. |
| 4 | * |
| 5 | * This library is free software; you can redistribute it and/or |
| 6 | * modify it under the terms of the GNU Lesser General Public |
| 7 | * License as published by the Free Software Foundation; either |
| 8 | * version 2 of the License, or (at your option) any later version. |
| 9 | * |
| 10 | * This library is distributed in the hope that it will be useful, |
| 11 | * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| 12 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| 13 | * Lesser General Public License for more details. |
| 14 | * |
| 15 | * You should have received a copy of the GNU Lesser General Public |
| 16 | * License along with this library; if not, write to the Free Software |
| 17 | * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
| 18 | */ |
| 19 | #include "config.h" |
| 20 | |
| 21 | #if USE(GSTREAMER) && USE(LIBWEBRTC) |
| 22 | #include "GStreamerVideoFrameLibWebRTC.h" |
| 23 | |
| 24 | #include <thread> |
| 25 | |
| 26 | namespace WebCore { |
| 27 | |
| 28 | const GRefPtr<GstSample> GStreamerSampleFromLibWebRTCVideoFrame(const webrtc::VideoFrame& frame) |
| 29 | { |
| 30 | if (frame.video_frame_buffer()->type() == webrtc::VideoFrameBuffer::Type::kNative) { |
| 31 | auto framebuffer = static_cast<GStreamerVideoFrameLibWebRTC*>(frame.video_frame_buffer().get()); |
| 32 | auto gstsample = framebuffer->getSample(); |
| 33 | |
| 34 | GST_LOG("Reusing native GStreamer sample: %p" , gstsample.get()); |
| 35 | |
| 36 | return gstsample; |
| 37 | } |
| 38 | |
| 39 | auto webrtcbuffer = frame.video_frame_buffer().get()->ToI420(); |
| 40 | // FIXME - Check lifetime of those buffers. |
| 41 | const uint8_t* comps[3] = { |
| 42 | webrtcbuffer->DataY(), |
| 43 | webrtcbuffer->DataU(), |
| 44 | webrtcbuffer->DataV() |
| 45 | }; |
| 46 | |
| 47 | GstVideoInfo info; |
| 48 | gst_video_info_set_format(&info, GST_VIDEO_FORMAT_I420, frame.width(), frame.height()); |
| 49 | auto buffer = adoptGRef(gst_buffer_new()); |
| 50 | for (gint i = 0; i < 3; i++) { |
| 51 | gsize compsize = GST_VIDEO_INFO_COMP_STRIDE(&info, i) * GST_VIDEO_INFO_COMP_HEIGHT(&info, i); |
| 52 | |
| 53 | GstMemory* comp = gst_memory_new_wrapped( |
| 54 | static_cast<GstMemoryFlags>(GST_MEMORY_FLAG_PHYSICALLY_CONTIGUOUS | GST_MEMORY_FLAG_READONLY), |
| 55 | const_cast<gpointer>(reinterpret_cast<const void*>(comps[i])), compsize, 0, compsize, webrtcbuffer, nullptr); |
| 56 | gst_buffer_append_memory(buffer.get(), comp); |
| 57 | } |
| 58 | |
| 59 | auto caps = adoptGRef(gst_video_info_to_caps(&info)); |
| 60 | auto sample = adoptGRef(gst_sample_new(buffer.get(), caps.get(), nullptr, nullptr)); |
| 61 | return sample; |
| 62 | } |
| 63 | |
| 64 | rtc::scoped_refptr<webrtc::VideoFrameBuffer> GStreamerVideoFrameLibWebRTC::create(GstSample * sample) |
| 65 | { |
| 66 | GstVideoInfo info; |
| 67 | |
| 68 | if (!gst_video_info_from_caps(&info, gst_sample_get_caps(sample))) |
| 69 | ASSERT_NOT_REACHED(); |
| 70 | |
| 71 | return rtc::scoped_refptr<webrtc::VideoFrameBuffer>(new GStreamerVideoFrameLibWebRTC(sample, info)); |
| 72 | } |
| 73 | |
| 74 | std::unique_ptr<webrtc::VideoFrame> LibWebRTCVideoFrameFromGStreamerSample(GstSample* sample, webrtc::VideoRotation rotation, |
| 75 | int64_t timestamp, int64_t renderTimeMs) |
| 76 | { |
| 77 | auto frameBuffer(GStreamerVideoFrameLibWebRTC::create(sample)); |
| 78 | |
| 79 | return std::unique_ptr<webrtc::VideoFrame>( |
| 80 | new webrtc::VideoFrame(frameBuffer, timestamp, renderTimeMs, rotation)); |
| 81 | } |
| 82 | |
| 83 | webrtc::VideoFrameBuffer::Type GStreamerVideoFrameLibWebRTC::type() const |
| 84 | { |
| 85 | return Type::kNative; |
| 86 | } |
| 87 | |
| 88 | GRefPtr<GstSample> GStreamerVideoFrameLibWebRTC::getSample() |
| 89 | { |
| 90 | return m_sample.get(); |
| 91 | } |
| 92 | |
| 93 | rtc::scoped_refptr<webrtc::I420BufferInterface> GStreamerVideoFrameLibWebRTC::ToI420() |
| 94 | { |
| 95 | GstMappedFrame inFrame(m_sample, GST_MAP_READ); |
| 96 | |
| 97 | if (!inFrame) { |
| 98 | GST_WARNING("Could not map frame" ); |
| 99 | |
| 100 | return nullptr; |
| 101 | } |
| 102 | |
| 103 | auto newBuffer = m_bufferPool.CreateBuffer(inFrame.width(), inFrame.height()); |
| 104 | ASSERT(newBuffer); |
| 105 | if (!newBuffer) { |
| 106 | GST_WARNING("RealtimeOutgoingVideoSourceGStreamer::videoSampleAvailable unable to allocate buffer for conversion to YUV" ); |
| 107 | return nullptr; |
| 108 | } |
| 109 | |
| 110 | if (inFrame.format() != GST_VIDEO_FORMAT_I420) { |
| 111 | GstVideoInfo outInfo; |
| 112 | |
| 113 | gst_video_info_set_format(&outInfo, GST_VIDEO_FORMAT_I420, inFrame.width(), |
| 114 | inFrame.height()); |
| 115 | auto info = inFrame.info(); |
| 116 | outInfo.fps_n = info->fps_n; |
| 117 | outInfo.fps_d = info->fps_d; |
| 118 | |
| 119 | GRefPtr<GstBuffer> buffer = adoptGRef(gst_buffer_new_wrapped_full(GST_MEMORY_FLAG_NO_SHARE, newBuffer->MutableDataY(), |
| 120 | outInfo.size, 0, outInfo.size, nullptr, nullptr)); |
| 121 | |
| 122 | GstMappedFrame outFrame(buffer.get(), outInfo, GST_MAP_WRITE); |
| 123 | |
| 124 | GUniquePtr<GstVideoConverter> videoConverter(gst_video_converter_new(inFrame.info(), |
| 125 | &outInfo, gst_structure_new("GstVideoConvertConfig" , |
| 126 | GST_VIDEO_CONVERTER_OPT_THREADS, G_TYPE_UINT, std::thread::hardware_concurrency() || 1 , nullptr))); |
| 127 | |
| 128 | ASSERT(videoConverter); |
| 129 | |
| 130 | gst_video_converter_frame(videoConverter.get(), inFrame.get(), outFrame.get()); |
| 131 | |
| 132 | return newBuffer; |
| 133 | } |
| 134 | |
| 135 | newBuffer->Copy( |
| 136 | inFrame.width(), |
| 137 | inFrame.height(), |
| 138 | inFrame.ComponentData(0), |
| 139 | inFrame.ComponentStride(0), |
| 140 | inFrame.ComponentData(1), |
| 141 | inFrame.ComponentStride(1), |
| 142 | inFrame.ComponentData(2), |
| 143 | inFrame.ComponentStride(2)); |
| 144 | |
| 145 | return newBuffer; |
| 146 | } |
| 147 | } |
| 148 | #endif // USE(LIBWEBRTC) |
| 149 | |