/* GStreamer * * Copyright (C) 2001-2002 Ronald Bultje * 2006 Edgard Lima * 2009 Texas Instruments, Inc - http://www.ti.com/ * * SPDX-FileCopyrightText: Copyright (c) 2018-2025 NVIDIA CORPORATION & AFFILIATES. All rights reserved. * SPDX-License-Identifier: LGPL-2.0-only * * gstv4l2bufferpool.c V4L2 buffer pool class * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Library General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Library General Public License for more details. * * You should have received a copy of the GNU Library General Public * License along with this library; if not, write to the * Free Software Foundation, Inc., 51 Franklin St, Fifth Floor, * Boston, MA 02110-1301, USA. */ #ifdef HAVE_CONFIG_H # include #endif #ifndef _GNU_SOURCE # define _GNU_SOURCE /* O_CLOEXEC */ #endif #include #include #include #include #include "gst/video/video.h" #include "gst/video/gstvideometa.h" #include "gst/video/gstvideopool.h" #include "gst/allocators/gstdmabuf.h" #include #include "gstv4l2object.h" #include "nvbufsurftransform.h" #include "gst/gst-i18n-plugin.h" #include #ifdef USE_V4L2_TARGET_NV #include #endif GST_DEBUG_CATEGORY_STATIC (v4l2bufferpool_debug); GST_DEBUG_CATEGORY_STATIC (CAT_PERFORMANCE); #define GST_CAT_DEFAULT v4l2bufferpool_debug #define GST_V4L2_IMPORT_QUARK gst_v4l2_buffer_pool_import_quark () /* * GstV4l2BufferPool: */ #define gst_v4l2_buffer_pool_parent_class parent_class G_DEFINE_TYPE (GstV4l2BufferPool, gst_v4l2_buffer_pool, GST_TYPE_BUFFER_POOL); enum _GstV4l2BufferPoolAcquireFlags { GST_V4L2_BUFFER_POOL_ACQUIRE_FLAG_RESURRECT = GST_BUFFER_POOL_ACQUIRE_FLAG_LAST, GST_V4L2_BUFFER_POOL_ACQUIRE_FLAG_LAST }; static void gst_v4l2_buffer_pool_release_buffer (GstBufferPool * bpool, GstBuffer * buffer); #ifdef USE_V4L2_TARGET_NV #define VPx_FRAME_HEADER_SIZE 12 static void report_metadata (GstV4l2Object * obj, guint32 buffer_index, v4l2_ctrl_videodec_outputbuf_metadata * metadata); static void v4l2_video_dec_get_enable_frame_type_reporting (GstV4l2Object * obj, guint32 buffer_index, v4l2_ctrl_videodec_outputbuf_metadata * dec_metadata); static void v4l2_video_dec_clear_poll_interrupt (GstV4l2Object * obj); #endif static gboolean #ifdef USE_V4L2_TARGET_NV gst_v4l2_is_buffer_valid (GstBuffer * buffer, GstV4l2MemoryGroup ** out_group, gboolean is_encode) #else gst_v4l2_is_buffer_valid (GstBuffer * buffer, GstV4l2MemoryGroup ** out_group) #endif { GstMemory *mem = gst_buffer_peek_memory (buffer, 0); gboolean valid = FALSE; if (GST_BUFFER_FLAG_IS_SET (buffer, GST_BUFFER_FLAG_TAG_MEMORY)) goto done; if (gst_is_dmabuf_memory (mem)) mem = gst_mini_object_get_qdata (GST_MINI_OBJECT (mem), GST_V4L2_MEMORY_QUARK); if (mem && gst_is_v4l2_memory (mem)) { GstV4l2Memory *vmem = (GstV4l2Memory *) mem; GstV4l2MemoryGroup *group = vmem->group; gint i, n_mem; #ifndef USE_V4L2_TARGET_NV n_mem = group->n_mem; #else /* We appended just one memory in the alloc_buffer function. */ if (is_encode) n_mem = group->n_mem; else n_mem = 1; #endif if (n_mem != (gint)gst_buffer_n_memory (buffer)) goto done; for (i = 0; i < n_mem; i++) { if (group->mem[i] != gst_buffer_peek_memory (buffer, i)) goto done; if (!gst_memory_is_writable (group->mem[i])) goto done; } valid = TRUE; if (out_group) *out_group = group; } done: return valid; } static NvBufSurfTransform_Error CopySurfTransform(NvBufSurface* src, NvBufSurface* dest) { NvBufSurfTransform_Error status; NvBufSurfTransformParams transformParams; NvBufSurfTransformRect srcRect; NvBufSurfTransformRect destRect; srcRect.top = srcRect.left = 0; destRect.top = destRect.left = 0; srcRect.width = src->surfaceList[0].width; srcRect.height = src->surfaceList[0].height; destRect.width = dest->surfaceList[0].width; destRect.height = dest->surfaceList[0].height; transformParams.src_rect = &srcRect; transformParams.dst_rect = &destRect; transformParams.transform_flag = NVBUFSURF_TRANSFORM_FILTER; transformParams.transform_flip = NvBufSurfTransform_None; transformParams.transform_filter = NvBufSurfTransformInter_Nearest; status = NvBufSurfTransform(src, dest, &transformParams); return status; } static GstFlowReturn gst_v4l2_buffer_pool_copy_buffer (GstV4l2BufferPool * pool, GstBuffer * dest, GstBuffer * src) { const GstVideoFormatInfo *finfo = pool->caps_info.finfo; GST_LOG_OBJECT (pool, "copying buffer"); #ifdef USE_V4L2_TARGET_NV gboolean ret; gint retn = 0; #endif if (finfo && (finfo->format != GST_VIDEO_FORMAT_UNKNOWN && finfo->format != GST_VIDEO_FORMAT_ENCODED)) { #ifndef USE_V4L2_TARGET_NV GstVideoFrame src_frame, dest_frame; GST_DEBUG_OBJECT (pool, "copy video frame"); /* we have raw video, use videoframe copy to get strides right */ if (!gst_video_frame_map (&src_frame, &pool->caps_info, src, GST_MAP_READ)) goto invalid_buffer; if (!gst_video_frame_map (&dest_frame, &pool->caps_info, dest, GST_MAP_WRITE)) { gst_video_frame_unmap (&src_frame); goto invalid_buffer; } gst_video_frame_copy (&dest_frame, &src_frame); gst_video_frame_unmap (&src_frame); gst_video_frame_unmap (&dest_frame); #endif } else { GstMapInfo map; GST_DEBUG_OBJECT (pool, "copy raw bytes"); if (!gst_buffer_map (src, &map, GST_MAP_READ)) goto invalid_buffer; gst_buffer_fill (dest, 0, map.data, gst_buffer_get_size (src)); gst_buffer_unmap (src, &map); gst_buffer_resize (dest, 0, gst_buffer_get_size (src)); } #ifndef USE_V4L2_TARGET_NV gst_buffer_copy_into (dest, src, GST_BUFFER_COPY_FLAGS | GST_BUFFER_COPY_TIMESTAMPS, 0, -1); #else if (((!strcmp (pool->obj->videodev, V4L2_DEVICE_PATH_NVDEC)) && (is_cuvid == FALSE)) || ((!strcmp (pool->obj->videodev, V4L2_DEVICE_PATH_NVDEC_ALT)) && (is_cuvid == FALSE)) || ((!strcmp (pool->obj->videodev, V4L2_DEVICE_PATH_NVDEC_MCCOY)) && (is_cuvid == TRUE)) || ((!strcmp (pool->obj->videodev, V4L2_DEVICE_PATH_NVENC) || !strcmp (pool->obj->videodev, V4L2_DEVICE_PATH_NVENC_ALT)) && (!V4L2_TYPE_IS_OUTPUT (pool->obj->type)))) { ret = gst_buffer_copy_into (dest, src, GST_BUFFER_COPY_FLAGS | GST_BUFFER_COPY_TIMESTAMPS, 0, -1); if (ret == FALSE) GST_ERROR_OBJECT (src,"Copy Failed"); } if ((!strcmp (pool->obj->videodev, V4L2_DEVICE_PATH_NVENC) || !strcmp (pool->obj->videodev, V4L2_DEVICE_PATH_NVENC_ALT)) && !V4L2_TYPE_IS_OUTPUT (pool->obj->type)) { GstMapInfo outmap = { NULL, (GstMapFlags) 0, NULL, 0, 0, }; void *sBaseAddr = NULL; GstV4l2Memory *outmemory = NULL; outmemory = (GstV4l2Memory *)gst_buffer_peek_memory (src, 0); NvBufSurface *nvbuf_surf = NULL; gboolean already_mapped = 0; retn = NvBufSurfaceFromFd(outmemory->dmafd, (void**)(&nvbuf_surf)); if (retn != 0) { GST_ERROR_OBJECT (src,"NvBufSurfaceFromFd Failed for fd = %d", outmemory->dmafd); return FALSE; } if (!nvbuf_surf->surfaceList[0].mappedAddr.addr[0]) retn = NvBufSurfaceMap(nvbuf_surf, 0, 0, NVBUF_MAP_READ_WRITE); else //Dont do unmapping already_mapped = TRUE; if (retn != 0) { GST_ERROR_OBJECT (src,"NvBufSurfaceMap Failed for fd = %d", outmemory->dmafd); return FALSE; } sBaseAddr = (void*)nvbuf_surf->surfaceList[0].mappedAddr.addr[0]; if (!gst_buffer_map (dest, &outmap, GST_MAP_WRITE)) goto invalid_buffer; memcpy (outmap.data, sBaseAddr, gst_buffer_get_size (src)); //Unmap only if we have mapped it if (already_mapped == FALSE) retn = NvBufSurfaceUnMap(nvbuf_surf, 0, 0); if (retn != 0) { GST_ERROR_OBJECT (src,"NvBufSurfaceUnMap Failed for fd = %d", outmemory->dmafd); gst_buffer_unmap (dest, &outmap); return FALSE; } gst_buffer_unmap (dest, &outmap); } if ((!strcmp (pool->obj->videodev, V4L2_DEVICE_PATH_NVENC) || !strcmp (pool->obj->videodev, V4L2_DEVICE_PATH_NVENC_ALT)) && V4L2_TYPE_IS_OUTPUT (pool->obj->type)) { GstV4l2Memory *inmemory = NULL; GstMapInfo inmap = { NULL, (GstMapFlags) 0, NULL, 0, 0, }; if (is_cuvid == FALSE) { #ifndef USE_V4L2_TARGET_NV_X86 NvBufSurfTransformParams transform_params; memset(&transform_params, 0, sizeof(NvBufSurfTransformParams)); if (!gst_buffer_map (src, &inmap, GST_MAP_READ)) goto invalid_buffer; NvBufSurface *src_buf = (NvBufSurface *)inmap.data; inmemory = (GstV4l2Memory *)gst_buffer_peek_memory (dest, 0); NvBufSurface *nvbuf_surf = 0; NvBufSurfaceFromFd(inmemory->dmafd, (void**)(&nvbuf_surf)); retn = NvBufSurfTransform (src_buf, nvbuf_surf, &transform_params); if (retn != 0) { GST_ERROR_OBJECT(src, "NvBufSurfTransform Failed"); gst_buffer_unmap(src, &inmap); return GST_FLOW_ERROR; } gst_buffer_unmap(src, &inmap); #endif } if (is_cuvid == TRUE){ NvBufSurface *src_bufsurf = NULL; NvBufSurface *dst_bufsurf = NULL; if (!gst_buffer_map (src, &inmap, GST_MAP_READ)) goto invalid_buffer; src_bufsurf = (NvBufSurface*)inmap.data; inmemory = (GstV4l2Memory *)gst_buffer_peek_memory (dest, 0); retn = NvBufSurfaceFromFd(inmemory->dmafd, (void**)(&dst_bufsurf)); if (retn != 0) { GST_ERROR_OBJECT(src, "NvBufSurfaceFromFd Failed"); gst_buffer_unmap(src, &inmap); return GST_FLOW_ERROR; } if (CopySurfTransform(src_bufsurf, dst_bufsurf) != NvBufSurfTransformError_Success) { GST_ERROR_OBJECT(src, "ERROR in BufSurfacecopy \n"); gst_buffer_unmap(src, &inmap); return GST_FLOW_ERROR; } gst_buffer_unmap(src, &inmap); } } #endif GST_CAT_LOG_OBJECT (CAT_PERFORMANCE, pool, "slow copy into buffer %p", dest); return GST_FLOW_OK; invalid_buffer: { GST_ERROR_OBJECT (pool, "could not map buffer"); return GST_FLOW_ERROR; } } struct UserPtrData { GstBuffer *buffer; gboolean is_frame; GstVideoFrame frame; GstMapInfo map; }; static GQuark gst_v4l2_buffer_pool_import_quark (void) { static GQuark quark = 0; if (quark == 0) quark = g_quark_from_string ("GstV4l2BufferPoolUsePtrData"); return quark; } static void _unmap_userptr_frame (struct UserPtrData *data) { if (data->is_frame) gst_video_frame_unmap (&data->frame); else gst_buffer_unmap (data->buffer, &data->map); if (data->buffer) gst_buffer_unref (data->buffer); g_slice_free (struct UserPtrData, data); } static GstFlowReturn gst_v4l2_buffer_pool_import_userptr (GstV4l2BufferPool * pool, GstBuffer * dest, GstBuffer * src) { GstFlowReturn ret = GST_FLOW_OK; GstV4l2MemoryGroup *group = NULL; GstMapFlags flags; const GstVideoFormatInfo *finfo = pool->caps_info.finfo; struct UserPtrData *data = NULL; GST_LOG_OBJECT (pool, "importing userptr"); /* get the group */ #ifdef USE_V4L2_TARGET_NV if (!gst_v4l2_is_buffer_valid (dest, &group, pool->obj->is_encode)) goto not_our_buffer; #else if (!gst_v4l2_is_buffer_valid (dest, &group)) goto not_our_buffer; #endif if (V4L2_TYPE_IS_OUTPUT (pool->obj->type)) flags = GST_MAP_READ; else flags = GST_MAP_WRITE; data = g_slice_new0 (struct UserPtrData); if (finfo && (finfo->format != GST_VIDEO_FORMAT_UNKNOWN && finfo->format != GST_VIDEO_FORMAT_ENCODED)) { gsize size[GST_VIDEO_MAX_PLANES] = { 0, }; #ifndef USE_V4L2_TARGET_NV gint i; #else guint i; #endif data->is_frame = TRUE; if (!gst_video_frame_map (&data->frame, &pool->caps_info, src, flags)) goto invalid_buffer; for (i = 0; i < GST_VIDEO_FORMAT_INFO_N_PLANES (finfo); i++) { if (GST_VIDEO_FORMAT_INFO_IS_TILED (finfo)) { gint tinfo = GST_VIDEO_FRAME_PLANE_STRIDE (&data->frame, i); size[i] = GST_VIDEO_TILE_X_TILES (tinfo) * GST_VIDEO_TILE_Y_TILES (tinfo) * GST_VIDEO_FORMAT_INFO_TILE_SIZE (finfo, i); } else { size[i] = GST_VIDEO_FRAME_PLANE_STRIDE (&data->frame, i) * GST_VIDEO_FRAME_COMP_HEIGHT (&data->frame, i); } } /* In the single planar API, planes must be contiguous in memory and * therefore they must have expected size. ie: no padding. * To check these conditions, we check that plane 'i' start address * + plane 'i' size equals to plane 'i+1' start address */ if (!V4L2_TYPE_IS_MULTIPLANAR (pool->obj->type)) { for (i = 0; i < (GST_VIDEO_FORMAT_INFO_N_PLANES (finfo) - 1); i++) { const struct v4l2_pix_format *pix_fmt = &pool->obj->format.fmt.pix; gpointer tmp; gint estride = gst_v4l2_object_extrapolate_stride (finfo, i, pix_fmt->bytesperline); guint eheight = GST_VIDEO_FORMAT_INFO_SCALE_HEIGHT (finfo, i, pix_fmt->height); tmp = ((guint8 *) data->frame.data[i]) + estride * eheight; if (tmp != data->frame.data[i + 1]) goto non_contiguous_mem; } } if (!gst_v4l2_allocator_import_userptr (pool->vallocator, group, data->frame.info.size, finfo->n_planes, data->frame.data, size)) goto import_failed; } else { gpointer ptr[1]; gsize size[1]; data->is_frame = FALSE; if (!gst_buffer_map (src, &data->map, flags)) goto invalid_buffer; ptr[0] = data->map.data; size[0] = data->map.size; if (!gst_v4l2_allocator_import_userptr (pool->vallocator, group, data->map.size, 1, ptr, size)) goto import_failed; } data->buffer = gst_buffer_ref (src); gst_mini_object_set_qdata (GST_MINI_OBJECT (dest), GST_V4L2_IMPORT_QUARK, data, (GDestroyNotify) _unmap_userptr_frame); gst_buffer_copy_into (dest, src, GST_BUFFER_COPY_FLAGS | GST_BUFFER_COPY_TIMESTAMPS, 0, -1); return ret; not_our_buffer: { GST_ERROR_OBJECT (pool, "destination buffer invalid or not from our pool"); return GST_FLOW_ERROR; } invalid_buffer: { GST_ERROR_OBJECT (pool, "could not map buffer"); g_slice_free (struct UserPtrData, data); return GST_FLOW_ERROR; } non_contiguous_mem: { GST_ERROR_OBJECT (pool, "memory is not contiguous or plane size mismatch"); _unmap_userptr_frame (data); return GST_FLOW_ERROR; } import_failed: { GST_ERROR_OBJECT (pool, "failed to import data"); _unmap_userptr_frame (data); return GST_FLOW_ERROR; } } static GstFlowReturn gst_v4l2_buffer_pool_import_dmabuf (GstV4l2BufferPool * pool, GstBuffer * dest, GstBuffer * src) { GstV4l2MemoryGroup *group = NULL; guint n_mem = gst_buffer_n_memory (src); #ifndef USE_V4L2_TARGET_NV gint i; GstMemory *dma_mem[GST_VIDEO_MAX_PLANES] = { 0 }; #else guint i; #endif GST_LOG_OBJECT (pool, "importing dmabuf"); #ifdef USE_V4L2_TARGET_NV if (!gst_v4l2_is_buffer_valid (dest, &group, pool->obj->is_encode)) goto not_our_buffer; #else if (!gst_v4l2_is_buffer_valid (dest, &group)) goto not_our_buffer; #endif if (n_mem > GST_VIDEO_MAX_PLANES) goto too_many_mems; #ifndef USE_V4L2_TARGET_NV for (i = 0; i < n_mem; i++) dma_mem[i] = gst_buffer_peek_memory (src, i); if (!gst_v4l2_allocator_import_dmabuf (pool->vallocator, group, n_mem, dma_mem)) goto import_failed; #else g_return_val_if_fail (pool->vallocator->memory == V4L2_MEMORY_DMABUF, FALSE); if ((!strcmp (pool->obj->videodev, V4L2_DEVICE_PATH_NVENC) || !strcmp (pool->obj->videodev, V4L2_DEVICE_PATH_NVENC_ALT)) && V4L2_TYPE_IS_OUTPUT (pool->obj->type)) { gint dmafd = -1; GstV4l2Memory *mem = NULL; GstMemory *inmemory = NULL; GstMapInfo inmap = { NULL, (GstMapFlags) 0, NULL, 0, 0, }; if (!gst_buffer_map (src, &inmap, GST_MAP_READ)) { GST_ERROR_OBJECT (pool, "could not map input buffer"); return GST_FLOW_ERROR; } NvBufSurface *src_bufsurf = (NvBufSurface*)inmap.data; if ((is_cuvid == FALSE) && ((src_bufsurf->memType == NVBUF_MEM_CUDA_PINNED) || (src_bufsurf->memType == NVBUF_MEM_CUDA_DEVICE) || (src_bufsurf->memType == NVBUF_MEM_CUDA_UNIFIED))) { GST_ERROR_OBJECT (pool, "Input CUDA Memory not supported on Jeston for output-io-mode=dmabuf-import," "element = %s", GST_ELEMENT_NAME(src)); gst_buffer_unmap (src, &inmap); goto invalid_buffer; } dmafd = src_bufsurf->surfaceList->bufferDesc; /* NOTE: gst-memory with input buffer for nvidia proprietary plugins mostly will be 1, though this may not always be the case as can have per plane separate gst-memory */ inmemory = (GstMemory *)gst_buffer_peek_memory (src, 0); for (i = 0; i < (guint)group->n_mem; i++) { gsize size, offset, maxsize; size = gst_memory_get_sizes (inmemory, &offset, &maxsize); mem = (GstV4l2Memory *) group->mem[i]; /* Update memory */ mem->mem.maxsize = maxsize; mem->mem.offset = offset; mem->mem.size = size; mem->dmafd = dmafd; /* Update v4l2 structure */ group->planes[i].length = maxsize; group->planes[i].bytesused = size + offset; group->planes[i].m.fd = dmafd; group->planes[i].data_offset = offset; } /* Copy into buffer structure if not using planes */ if (!V4L2_TYPE_IS_MULTIPLANAR (pool->obj->type)) { group->buffer.bytesused = group->planes[0].bytesused; group->buffer.length = group->planes[0].length; group->buffer.m.fd = group->planes[0].m.userptr; /* Check if data_offset > 0 and fail for non-multi-planar */ g_assert (group->planes[0].data_offset == 0); } else { group->buffer.length = group->n_mem; } gst_buffer_unmap(src, &inmap); } else { GST_INFO_OBJECT (pool, "DMABUF_IMPORT io mode not supported for device %s ", pool->obj->videodev); return GST_FLOW_ERROR; } #endif gst_mini_object_set_qdata (GST_MINI_OBJECT (dest), GST_V4L2_IMPORT_QUARK, gst_buffer_ref (src), (GDestroyNotify) gst_buffer_unref); gst_buffer_copy_into (dest, src, GST_BUFFER_COPY_FLAGS | GST_BUFFER_COPY_TIMESTAMPS, 0, -1); return GST_FLOW_OK; invalid_buffer: { return GST_FLOW_ERROR; } not_our_buffer: { GST_ERROR_OBJECT (pool, "destination buffer invalid or not from our pool"); return GST_FLOW_ERROR; } too_many_mems: { GST_ERROR_OBJECT (pool, "could not map buffer"); return GST_FLOW_ERROR; } #ifndef USE_V4L2_TARGET_NV import_failed: { GST_ERROR_OBJECT (pool, "failed to import dmabuf"); return GST_FLOW_ERROR; } #endif } static GstFlowReturn gst_v4l2_buffer_pool_prepare_buffer (GstV4l2BufferPool * pool, GstBuffer * dest, GstBuffer * src) { GstFlowReturn ret = GST_FLOW_OK; gboolean own_src = FALSE; if (src == NULL) { if (pool->other_pool == NULL) { GST_ERROR_OBJECT (pool, "can't prepare buffer, source buffer missing"); return GST_FLOW_ERROR; } ret = gst_buffer_pool_acquire_buffer (pool->other_pool, &src, NULL); if (ret != GST_FLOW_OK) { GST_ERROR_OBJECT (pool, "failed to acquire buffer from downstream pool"); goto done; } own_src = TRUE; } switch (pool->obj->mode) { case GST_V4L2_IO_MMAP: case GST_V4L2_IO_DMABUF: ret = gst_v4l2_buffer_pool_copy_buffer (pool, dest, src); break; case GST_V4L2_IO_USERPTR: ret = gst_v4l2_buffer_pool_import_userptr (pool, dest, src); break; case GST_V4L2_IO_DMABUF_IMPORT: ret = gst_v4l2_buffer_pool_import_dmabuf (pool, dest, src); break; default: break; } if (own_src) gst_buffer_unref (src); done: return ret; } static GstFlowReturn gst_v4l2_buffer_pool_alloc_buffer (GstBufferPool * bpool, GstBuffer ** buffer, GstBufferPoolAcquireParams * params) { GstV4l2BufferPool *pool = GST_V4L2_BUFFER_POOL (bpool); GstV4l2MemoryGroup *group = NULL; GstBuffer *newbuf = NULL; GstV4l2Object *obj; GstVideoInfo *info; obj = pool->obj; info = &obj->info; switch (obj->mode) { case GST_V4L2_IO_RW: newbuf = gst_buffer_new_allocate (pool->allocator, pool->size, &pool->params); break; case GST_V4L2_IO_MMAP: group = gst_v4l2_allocator_alloc_mmap (pool->vallocator); break; case GST_V4L2_IO_DMABUF: group = gst_v4l2_allocator_alloc_dmabuf (pool->vallocator, pool->allocator); break; case GST_V4L2_IO_USERPTR: group = gst_v4l2_allocator_alloc_userptr (pool->vallocator); break; case GST_V4L2_IO_DMABUF_IMPORT: group = gst_v4l2_allocator_alloc_dmabufin (pool->vallocator); break; default: newbuf = NULL; g_assert_not_reached (); break; } if (group != NULL) { gint i; newbuf = gst_buffer_new (); for (i = 0; i < group->n_mem; i++) gst_buffer_append_memory (newbuf, group->mem[i]); #ifdef USE_V4L2_TARGET_NV if (!V4L2_TYPE_IS_OUTPUT(obj->type) && is_cuvid == FALSE) { GstMapInfo map = GST_MAP_INFO_INIT; NvBufSurface *nvbuf_surf = NULL; gst_buffer_map (newbuf, &map, GST_MAP_READ); nvbuf_surf = (NvBufSurface *) map.data; if (g_queue_find(pool->allocated_surfaces_queue, nvbuf_surf) == NULL) { g_queue_push_tail (pool->allocated_surfaces_queue, nvbuf_surf); } gst_buffer_unmap (newbuf, &map); } #endif } else if (newbuf == NULL) { goto allocation_failed; } /* add metadata to raw video buffers */ if (pool->add_videometa) gst_buffer_add_video_meta_full (newbuf, GST_VIDEO_FRAME_FLAG_NONE, GST_VIDEO_INFO_FORMAT (info), GST_VIDEO_INFO_WIDTH (info), GST_VIDEO_INFO_HEIGHT (info), GST_VIDEO_INFO_N_PLANES (info), info->offset, info->stride); *buffer = newbuf; return GST_FLOW_OK; /* ERRORS */ allocation_failed: { GST_ERROR_OBJECT (pool, "failed to allocate buffer"); return GST_FLOW_ERROR; } } static gboolean gst_v4l2_buffer_pool_set_config (GstBufferPool * bpool, GstStructure * config) { GstV4l2BufferPool *pool = GST_V4L2_BUFFER_POOL (bpool); GstV4l2Object *obj = pool->obj; GstCaps *caps; guint size, min_buffers = 0, max_buffers = 0; GstAllocator *allocator = NULL; GstAllocationParams params = { 0, }; gboolean can_allocate = FALSE; gboolean updated = FALSE; gboolean ret; gst_allocation_params_init (¶ms); pool->add_videometa = gst_buffer_pool_config_has_option (config, GST_BUFFER_POOL_OPTION_VIDEO_META); /* parse the config and keep around */ if (!gst_buffer_pool_config_get_params (config, &caps, &size, &min_buffers, &max_buffers)) goto wrong_config; if (!gst_buffer_pool_config_get_allocator (config, &allocator, ¶ms)) goto wrong_config; GST_DEBUG_OBJECT (pool, "config %" GST_PTR_FORMAT, config); if (pool->allocator) gst_object_unref (pool->allocator); pool->allocator = NULL; switch (obj->mode) { case GST_V4L2_IO_DMABUF: pool->allocator = gst_dmabuf_allocator_new (); can_allocate = GST_V4L2_ALLOCATOR_CAN_ALLOCATE (pool->vallocator, MMAP); break; case GST_V4L2_IO_MMAP: can_allocate = GST_V4L2_ALLOCATOR_CAN_ALLOCATE (pool->vallocator, MMAP); break; case GST_V4L2_IO_USERPTR: can_allocate = GST_V4L2_ALLOCATOR_CAN_ALLOCATE (pool->vallocator, USERPTR); break; case GST_V4L2_IO_DMABUF_IMPORT: can_allocate = GST_V4L2_ALLOCATOR_CAN_ALLOCATE (pool->vallocator, DMABUF); break; case GST_V4L2_IO_RW: if (allocator) pool->allocator = g_object_ref (allocator); pool->params = params; /* No need to change the configuration */ goto done; break; default: g_assert_not_reached (); break; } /* libv4l2 conversion code does not handle CREATE_BUFS, and may lead to * instability and crash, disable it for now */ if (can_allocate && obj->fmtdesc->flags & V4L2_FMT_FLAG_EMULATED) { GST_WARNING_OBJECT (pool, "libv4l2 converter detected, disabling CREATE_BUFS"); can_allocate = FALSE; GST_OBJECT_FLAG_UNSET (pool->vallocator, GST_V4L2_ALLOCATOR_FLAG_MMAP_CREATE_BUFS | GST_V4L2_ALLOCATOR_FLAG_USERPTR_CREATE_BUFS | GST_V4L2_ALLOCATOR_FLAG_DMABUF_CREATE_BUFS); } if (min_buffers < GST_V4L2_MIN_BUFFERS) { updated = TRUE; min_buffers = GST_V4L2_MIN_BUFFERS; GST_INFO_OBJECT (pool, "increasing minimum buffers to %u", min_buffers); } /* respect driver requirements */ if (min_buffers < obj->min_buffers) { updated = TRUE; min_buffers = obj->min_buffers; GST_INFO_OBJECT (pool, "increasing minimum buffers to %u", min_buffers); } #ifdef USE_V4L2_TARGET_NV if (max_buffers > NV_VIDEO_MAX_FRAME || max_buffers == 0) { updated = TRUE; max_buffers = NV_VIDEO_MAX_FRAME; GST_INFO_OBJECT (pool, "reducing maximum buffers to %u", max_buffers); } #else if (max_buffers > VIDEO_MAX_FRAME || max_buffers == 0) { updated = TRUE; max_buffers = VIDEO_MAX_FRAME; GST_INFO_OBJECT (pool, "reducing maximum buffers to %u", max_buffers); } #endif if (min_buffers > max_buffers) { updated = TRUE; min_buffers = max_buffers; GST_INFO_OBJECT (pool, "reducing minimum buffers to %u", min_buffers); } else if (min_buffers != max_buffers) { if (!can_allocate) { updated = TRUE; max_buffers = min_buffers; GST_INFO_OBJECT (pool, "can't allocate, setting maximum to minimum"); } } if (!pool->add_videometa && obj->need_video_meta) { GST_INFO_OBJECT (pool, "adding needed video meta"); updated = TRUE; gst_buffer_pool_config_add_option (config, GST_BUFFER_POOL_OPTION_VIDEO_META); } /* Always update the config to ensure the configured size matches */ if ((!strcmp (obj->videodev, V4L2_DEVICE_PATH_NVENC) || !strcmp (obj->videodev, V4L2_DEVICE_PATH_NVENC_ALT)) && (obj->mode == GST_V4L2_IO_DMABUF_IMPORT)) { /*For DMABUF Import, queue size should be kept to max to avoid frame drops coming from decoder*/ min_buffers = max_buffers = 24; /* NvMM Encoder has MAX Q size as 24*/ gst_buffer_pool_config_set_params(config, caps, sizeof(NvBufSurface), min_buffers, max_buffers); } else gst_buffer_pool_config_set_params (config, caps, obj->info.size, min_buffers, max_buffers); #ifdef USE_V4L2_TARGET_NV /* Need to adjust the size to 0th plane's size since we will only output v4l2 memory associated with 0th plane. */ if (!V4L2_TYPE_IS_OUTPUT(obj->type)) gst_buffer_pool_config_set_params (config, caps, obj->info.width * obj->info.height, min_buffers, max_buffers); #endif /* keep a GstVideoInfo with defaults for the when we need to copy */ gst_video_info_from_caps (&pool->caps_info, caps); done: ret = GST_BUFFER_POOL_CLASS (parent_class)->set_config (bpool, config); /* If anything was changed documentation recommand to return FALSE */ return !updated && ret; /* ERRORS */ wrong_config: { GST_ERROR_OBJECT (pool, "invalid config %" GST_PTR_FORMAT, config); return FALSE; } } static GstFlowReturn gst_v4l2_buffer_pool_resurect_buffer (GstV4l2BufferPool * pool) { GstBufferPoolAcquireParams params = { 0 }; GstBuffer *buffer = NULL; GstFlowReturn ret; GST_DEBUG_OBJECT (pool, "A buffer was lost, reallocating it"); /* block recursive calls to this function */ g_signal_handler_block (pool->vallocator, pool->group_released_handler); params.flags = (GstBufferPoolAcquireFlags) GST_V4L2_BUFFER_POOL_ACQUIRE_FLAG_RESURRECT | GST_BUFFER_POOL_ACQUIRE_FLAG_DONTWAIT; ret = gst_buffer_pool_acquire_buffer (GST_BUFFER_POOL (pool), &buffer, ¶ms); if (ret == GST_FLOW_OK) gst_buffer_unref (buffer); g_signal_handler_unblock (pool->vallocator, pool->group_released_handler); return ret; } static gboolean gst_v4l2_buffer_pool_streamon (GstV4l2BufferPool * pool) { GstV4l2Object *obj = pool->obj; if (pool->streaming) return TRUE; switch (obj->mode) { case GST_V4L2_IO_MMAP: case GST_V4L2_IO_USERPTR: case GST_V4L2_IO_DMABUF: case GST_V4L2_IO_DMABUF_IMPORT: if (!V4L2_TYPE_IS_OUTPUT (pool->obj->type)) { guint num_queued; guint i, n = 0; num_queued = g_atomic_int_get (&pool->num_queued); #ifdef USE_V4L2_TARGET_NV GST_DEBUG_OBJECT (pool, "num_allocated %d num_queued %d allocator count %d \ dynamic_allocation %d\n", pool->num_allocated, num_queued, pool->vallocator->count, pool->enable_dynamic_allocation); #endif if (num_queued < pool->num_allocated) n = pool->num_allocated - num_queued; /* For captures, we need to enqueue buffers before we start streaming, * so the driver don't underflow immediatly. As we have put then back * into the base class queue, resurect them, then releasing will queue * them back. */ for (i = 0; i < n; i++) gst_v4l2_buffer_pool_resurect_buffer (pool); } if (obj->ioctl (pool->video_fd, VIDIOC_STREAMON, &obj->type) < 0) goto streamon_failed; pool->streaming = TRUE; GST_DEBUG_OBJECT (pool, "Started streaming"); break; default: break; } return TRUE; streamon_failed: { GST_ERROR_OBJECT (pool, "error with STREAMON %d (%s)", errno, g_strerror (errno)); return FALSE; } } /* Call with streamlock held, or when streaming threads are down */ static void gst_v4l2_buffer_pool_streamoff (GstV4l2BufferPool * pool) { GstBufferPoolClass *pclass = GST_BUFFER_POOL_CLASS (parent_class); GstV4l2Object *obj = pool->obj; #ifdef USE_V4L2_TARGET_NV GstBuffer *buffers[NV_VIDEO_MAX_FRAME] = {NULL}; #else GstBuffer *buffers[VIDEO_MAX_FRAME] = {NULL}; #endif gint i; GST_OBJECT_LOCK (pool); if (!pool->streaming) { GST_OBJECT_UNLOCK (pool); return; } switch (obj->mode) { case GST_V4L2_IO_MMAP: case GST_V4L2_IO_USERPTR: case GST_V4L2_IO_DMABUF: case GST_V4L2_IO_DMABUF_IMPORT: if (obj->ioctl (pool->video_fd, VIDIOC_STREAMOFF, &obj->type) < 0) GST_WARNING_OBJECT (pool, "STREAMOFF failed with errno %d (%s)", errno, g_strerror (errno)); pool->streaming = FALSE; GST_DEBUG_OBJECT (pool, "Stopped streaming"); if (pool->vallocator) gst_v4l2_allocator_flush (pool->vallocator); break; default: break; } #ifdef USE_V4L2_TARGET_NV for (i = 0; i < NV_VIDEO_MAX_FRAME; i++) { #else for (i = 0; i < VIDEO_MAX_FRAME; i++) { #endif if (pool->buffers[i]) { buffers[i] = pool->buffers[i]; pool->buffers[i] = NULL; g_atomic_int_add (&pool->num_queued, -1); } } GST_OBJECT_UNLOCK (pool); #ifdef USE_V4L2_TARGET_NV for (i = 0; i < NV_VIDEO_MAX_FRAME; i++) { #else for (i = 0; i < VIDEO_MAX_FRAME; i++) { #endif if (buffers[i]) { GstBuffer *buffer = buffers[i]; GstBufferPool *bpool = GST_BUFFER_POOL (pool); if (V4L2_TYPE_IS_OUTPUT (pool->obj->type)) gst_v4l2_buffer_pool_release_buffer (bpool, buffer); else /* Don't re-enqueue capture buffer on stop */ pclass->release_buffer (bpool, buffer); } } } static gboolean gst_v4l2_buffer_pool_start (GstBufferPool * bpool) { GstV4l2BufferPool *pool = GST_V4L2_BUFFER_POOL (bpool); GstBufferPoolClass *pclass = GST_BUFFER_POOL_CLASS (parent_class); GstV4l2Object *obj = pool->obj; GstStructure *config; GstCaps *caps; guint size = 0, min_buffers = 0, max_buffers = 0; guint max_latency, min_latency, copy_threshold = 0; gboolean can_allocate = FALSE, ret = TRUE; GST_DEBUG_OBJECT (pool, "activating pool"); config = gst_buffer_pool_get_config (bpool); if (!gst_buffer_pool_config_get_params (config, &caps, &size, &min_buffers, &max_buffers)) goto wrong_config; min_latency = MAX (GST_V4L2_MIN_BUFFERS, obj->min_buffers); switch (obj->mode) { case GST_V4L2_IO_RW: can_allocate = TRUE; #ifdef HAVE_LIBV4L2 /* This workaround a unfixable bug in libv4l2 when RW is emulated on top * of MMAP. In this case, the first read initialize the queues, but the * poll before that will always fail. Doing an empty read, forces the * queue to be initialized now. We only do this if we have a streaming * driver. */ if (obj->device_caps & V4L2_CAP_STREAMING) obj->read (obj->video_fd, NULL, 0); #endif break; case GST_V4L2_IO_DMABUF: case GST_V4L2_IO_MMAP: { guint count; can_allocate = GST_V4L2_ALLOCATOR_CAN_ALLOCATE (pool->vallocator, MMAP); /* first, lets request buffers, and see how many we can get: */ GST_DEBUG_OBJECT (pool, "requesting %d MMAP buffers", min_buffers); count = gst_v4l2_allocator_start (pool->vallocator, min_buffers, V4L2_MEMORY_MMAP); pool->num_allocated = count; if (count < GST_V4L2_MIN_BUFFERS) { min_buffers = count; goto no_buffers; } /* V4L2 buffer pool are often very limited in the amount of buffers it * can offer. The copy_threshold will workaround this limitation by * falling back to copy if the pipeline needed more buffers. This also * prevent having to do REQBUFS(N)/REQBUFS(0) everytime configure is * called. */ if (count != min_buffers || pool->enable_copy_threshold) { GST_WARNING_OBJECT (pool, "Uncertain or not enough buffers, enabling copy threshold"); min_buffers = count; copy_threshold = min_latency; } break; } case GST_V4L2_IO_USERPTR: { guint count; can_allocate = GST_V4L2_ALLOCATOR_CAN_ALLOCATE (pool->vallocator, USERPTR); GST_DEBUG_OBJECT (pool, "requesting %d USERPTR buffers", min_buffers); count = gst_v4l2_allocator_start (pool->vallocator, min_buffers, V4L2_MEMORY_USERPTR); pool->num_allocated = count; /* There is no rational to not get what we asked */ if (count < min_buffers) { min_buffers = count; goto no_buffers; } min_buffers = count; break; } case GST_V4L2_IO_DMABUF_IMPORT: { guint count; can_allocate = GST_V4L2_ALLOCATOR_CAN_ALLOCATE (pool->vallocator, DMABUF); GST_DEBUG_OBJECT (pool, "requesting %d DMABUF buffers", min_buffers); count = gst_v4l2_allocator_start (pool->vallocator, min_buffers, V4L2_MEMORY_DMABUF); pool->num_allocated = count; /* There is no rational to not get what we asked */ if (count < min_buffers) { min_buffers = count; goto no_buffers; } min_buffers = count; break; } default: min_buffers = 0; copy_threshold = 0; g_assert_not_reached (); break; } if (can_allocate) max_latency = max_buffers; else max_latency = min_buffers; pool->size = size; pool->copy_threshold = copy_threshold; pool->max_latency = max_latency; pool->min_latency = min_latency; pool->num_queued = 0; if (max_buffers != 0 && max_buffers < min_buffers) max_buffers = min_buffers; gst_buffer_pool_config_set_params (config, caps, size, min_buffers, max_buffers); pclass->set_config (bpool, config); gst_structure_free (config); if (pool->other_pool) if (!gst_buffer_pool_set_active (pool->other_pool, TRUE)) goto other_pool_failed; /* now, allocate the buffers: */ if (!pclass->start (bpool)) goto start_failed; if (!V4L2_TYPE_IS_OUTPUT (obj->type)) { pool->group_released_handler = g_signal_connect_swapped (pool->vallocator, "group-released", G_CALLBACK (gst_v4l2_buffer_pool_resurect_buffer), pool); ret = gst_v4l2_buffer_pool_streamon (pool); } return ret; /* ERRORS */ wrong_config: { GST_ERROR_OBJECT (pool, "invalid config %" GST_PTR_FORMAT, config); gst_structure_free (config); return FALSE; } no_buffers: { GST_ERROR_OBJECT (pool, "we received %d buffer from device '%s', we want at least %d", min_buffers, obj->videodev, GST_V4L2_MIN_BUFFERS); gst_structure_free (config); return FALSE; } start_failed: { GST_ERROR_OBJECT (pool, "allocate failed"); return FALSE; } other_pool_failed: { GST_ERROR_OBJECT (pool, "failed to active the other pool %" GST_PTR_FORMAT, pool->other_pool); return FALSE; } } static gboolean gst_v4l2_buffer_pool_stop (GstBufferPool * bpool) { GstV4l2BufferPool *pool = GST_V4L2_BUFFER_POOL (bpool); gboolean ret; GST_DEBUG_OBJECT (pool, "stopping pool"); #ifdef USE_V4L2_TARGET_NV /* * On EOS, video_fd becomes -1, this makes VIDIOC_REQBUFS API fail which * internally releases hardware buffers causing memory leak. * In below code NvBufSurfaces are destroyed explicitly. */ if (pool->vallocator) { if (pool->vallocator->obj->video_fd == -1 && !V4L2_TYPE_IS_OUTPUT (pool->vallocator->obj->type) && is_cuvid == FALSE) { while(g_queue_get_length(pool->allocated_surfaces_queue) > 0) { int retval = 0; NvBufSurface* allocated_surface = NULL; allocated_surface = g_queue_pop_head (pool->allocated_surfaces_queue); if (allocated_surface) { retval = NvBufSurfaceDestroy(allocated_surface); if(retval < 0) { GST_ERROR_OBJECT (pool, "failed to destroy nvbufsurface"); return GST_V4L2_ERROR; } } } } } #endif if (pool->group_released_handler > 0) { g_signal_handler_disconnect (pool->vallocator, pool->group_released_handler); pool->group_released_handler = 0; } if (pool->other_pool) { gst_buffer_pool_set_active (pool->other_pool, FALSE); gst_object_unref (pool->other_pool); pool->other_pool = NULL; } gst_v4l2_buffer_pool_streamoff (pool); ret = GST_BUFFER_POOL_CLASS (parent_class)->stop (bpool); if (ret && pool->vallocator) { GstV4l2Return vret; vret = gst_v4l2_allocator_stop (pool->vallocator); if (vret == GST_V4L2_BUSY) GST_WARNING_OBJECT (pool, "some buffers are still outstanding"); ret = (vret == GST_V4L2_OK); } #ifdef USE_V4L2_TARGET_NV /* Make sure, there are no outstanding buffers to be unreffed by * downstream component. In case of DRC, wait before reseting decoder * until last buffer on capture plane is unreffed. */ if (!V4L2_TYPE_IS_OUTPUT(pool->obj->type)) { g_mutex_lock (&pool->obj->cplane_stopped_lock); pool->obj->capture_plane_stopped = TRUE; g_cond_signal (&pool->obj->cplane_stopped_cond); g_mutex_unlock (&pool->obj->cplane_stopped_lock); } #endif return ret; } static void gst_v4l2_buffer_pool_flush_start (GstBufferPool * bpool) { GstV4l2BufferPool *pool = GST_V4L2_BUFFER_POOL (bpool); GST_DEBUG_OBJECT (pool, "start flushing"); #ifndef USE_V4L2_TARGET_NV gst_poll_set_flushing (pool->poll, TRUE); #else if (is_cuvid == TRUE) { v4l2_video_dec_clear_poll_interrupt (pool->obj); } #endif GST_OBJECT_LOCK (pool); pool->empty = FALSE; g_cond_broadcast (&pool->empty_cond); GST_OBJECT_UNLOCK (pool); if (pool->other_pool) gst_buffer_pool_set_flushing (pool->other_pool, TRUE); GST_DEBUG_OBJECT (pool, "End flushing"); } static void gst_v4l2_buffer_pool_flush_stop (GstBufferPool * bpool) { GstV4l2BufferPool *pool = GST_V4L2_BUFFER_POOL (bpool); GST_DEBUG_OBJECT (pool, "stop flushing"); if (pool->other_pool) gst_buffer_pool_set_flushing (pool->other_pool, FALSE); #ifndef USE_V4L2_TARGET_NV gst_poll_set_flushing (pool->poll, FALSE); #endif } static GstFlowReturn gst_v4l2_buffer_pool_poll (GstV4l2BufferPool * pool) { gint ret; /* In RW mode there is no queue, hence no need to wait while the queue is * empty */ if (pool->obj->mode != GST_V4L2_IO_RW) { GST_OBJECT_LOCK (pool); while (pool->empty) g_cond_wait (&pool->empty_cond, GST_OBJECT_GET_LOCK (pool)); GST_OBJECT_UNLOCK (pool); } if (!pool->can_poll_device) goto done; GST_LOG_OBJECT (pool, "polling device"); again: ret = gst_poll_wait (pool->poll, GST_CLOCK_TIME_NONE); if (G_UNLIKELY (ret < 0)) { switch (errno) { case EBUSY: goto stopped; case EAGAIN: case EINTR: goto again; case ENXIO: GST_WARNING_OBJECT (pool, "v4l2 device doesn't support polling. Disabling" " using libv4l2 in this case may cause deadlocks"); pool->can_poll_device = FALSE; goto done; default: goto select_error; } } if (gst_poll_fd_has_error (pool->poll, &pool->pollfd)) goto select_error; done: return GST_FLOW_OK; /* ERRORS */ stopped: { GST_DEBUG_OBJECT (pool, "stop called"); return GST_FLOW_FLUSHING; } select_error: { GST_ELEMENT_ERROR (pool->obj->element, RESOURCE, READ, (NULL), ("poll error %d: %s (%d)", ret, g_strerror (errno), errno)); return GST_FLOW_ERROR; } } static GstFlowReturn gst_v4l2_buffer_pool_qbuf (GstV4l2BufferPool * pool, GstBuffer * buf) { GstV4l2MemoryGroup *group = NULL; const GstV4l2Object *obj = pool->obj; GstClockTime timestamp; gint index; #ifdef USE_V4L2_TARGET_NV if (!gst_v4l2_is_buffer_valid (buf, &group, pool->obj->is_encode)) { GST_ERROR_OBJECT (pool, "invalid buffer %p", buf); return GST_FLOW_ERROR; } #else if (!gst_v4l2_is_buffer_valid (buf, &group)) { GST_ERROR_OBJECT (pool, "invalid buffer %p", buf); return GST_FLOW_ERROR; } #endif index = group->buffer.index; if (pool->buffers[index] != NULL) goto already_queued; GST_LOG_OBJECT (pool, "queuing buffer %i", index); if (V4L2_TYPE_IS_OUTPUT (obj->type)) { enum v4l2_field field; /* Except when field is set to alternate, buffer field is the same as * the one defined in format */ if (V4L2_TYPE_IS_MULTIPLANAR (obj->type)) field = obj->format.fmt.pix_mp.field; else field = obj->format.fmt.pix.field; /* NB: At this moment, we can't have alternate mode because it not handled * yet */ if (field == V4L2_FIELD_ALTERNATE) { if (GST_BUFFER_FLAG_IS_SET (buf, GST_VIDEO_FRAME_FLAG_TFF)) field = V4L2_FIELD_TOP; else field = V4L2_FIELD_BOTTOM; } group->buffer.field = field; } if (GST_BUFFER_TIMESTAMP_IS_VALID (buf)) { timestamp = GST_BUFFER_TIMESTAMP (buf); GST_TIME_TO_TIMEVAL (timestamp, group->buffer.timestamp); } GST_OBJECT_LOCK (pool); g_atomic_int_inc (&pool->num_queued); pool->buffers[index] = buf; if (!gst_v4l2_allocator_qbuf (pool->vallocator, group)) goto queue_failed; pool->empty = FALSE; g_cond_signal (&pool->empty_cond); GST_OBJECT_UNLOCK (pool); return GST_FLOW_OK; already_queued: { GST_ERROR_OBJECT (pool, "the buffer %i was already queued", index); return GST_FLOW_ERROR; } queue_failed: { GST_ERROR_OBJECT (pool, "could not queue a buffer %i", index); /* Mark broken buffer to the allocator */ GST_BUFFER_FLAG_SET (buf, GST_BUFFER_FLAG_TAG_MEMORY); g_atomic_int_add (&pool->num_queued, -1); pool->buffers[index] = NULL; GST_OBJECT_UNLOCK (pool); return GST_FLOW_ERROR; } } static GstFlowReturn gst_v4l2_buffer_pool_dqbuf (GstV4l2BufferPool * pool, GstBuffer ** buffer) { GstFlowReturn res; GstBuffer *outbuf; GstV4l2Object *obj = pool->obj; GstClockTime timestamp; GstV4l2MemoryGroup *group; GstVideoMeta *vmeta; gsize size; gint i; if ((res = gst_v4l2_buffer_pool_poll (pool)) != GST_FLOW_OK) goto poll_failed; GST_LOG_OBJECT (pool, "dequeueing a buffer"); res = gst_v4l2_allocator_dqbuf (pool->vallocator, &group); if (res == GST_FLOW_EOS) goto eos; if (res != GST_FLOW_OK) goto dqbuf_failed; /* get our GstBuffer with that index from the pool, if the buffer was * outstanding we have a serious problem. */ outbuf = pool->buffers[group->buffer.index]; if (outbuf == NULL) goto no_buffer; /* mark the buffer outstanding */ pool->buffers[group->buffer.index] = NULL; if (g_atomic_int_dec_and_test (&pool->num_queued)) { GST_OBJECT_LOCK (pool); pool->empty = TRUE; GST_OBJECT_UNLOCK (pool); } #ifdef USE_V4L2_TARGET_NV if (pool->obj->type == V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE && obj->enableMVBufferMeta && (!strcmp (obj->videodev, V4L2_DEVICE_PATH_NVENC) || !strcmp (obj->videodev, V4L2_DEVICE_PATH_NVENC_ALT))) { v4l2_ctrl_videoenc_outputbuf_metadata_MV enc_mv_metadata; memset ((void *) &enc_mv_metadata, 0, sizeof (enc_mv_metadata)); if (get_motion_vectors (obj, group->buffer.index, &enc_mv_metadata) == 0) { guint32 numMVs = enc_mv_metadata.bufSize / sizeof (MVInfo), i; MVInfo *pInfo = enc_mv_metadata.pMVInfo; g_print ("Num MVs = %d \n", numMVs); for (i = 0; i < numMVs; i++, pInfo++) g_print ("%d: mv_x=%d mv_y=%d weight=%d\n ", i, pInfo->mv_x, pInfo->mv_y, pInfo->weight); } } if (pool->obj->type == V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE && (((!strcmp(obj->videodev, V4L2_DEVICE_PATH_NVDEC)) && (is_cuvid == FALSE)) || ((!strcmp(obj->videodev, V4L2_DEVICE_PATH_NVDEC_ALT)) && (is_cuvid == FALSE)) || ((!strcmp(obj->videodev, V4L2_DEVICE_PATH_NVDEC_MCCOY)) && (is_cuvid == TRUE))) && (obj->Enable_frame_type_reporting || obj->Enable_error_check)) { v4l2_ctrl_videodec_outputbuf_metadata dec_metadata; memset ((void *) &dec_metadata, 0, sizeof (dec_metadata)); v4l2_video_dec_get_enable_frame_type_reporting (obj, group->buffer.index, &dec_metadata); report_metadata (obj, group->buffer.index, &dec_metadata); } #endif timestamp = GST_TIMEVAL_TO_TIME (group->buffer.timestamp); size = 0; vmeta = gst_buffer_get_video_meta (outbuf); for (i = 0; i < group->n_mem; i++) { GST_LOG_OBJECT (pool, "dequeued buffer %p seq:%d (ix=%d), mem %p used %d, plane=%d, flags %08x, ts %" GST_TIME_FORMAT ", pool-queued=%d, buffer=%p", outbuf, group->buffer.sequence, group->buffer.index, group->mem[i], group->planes[i].bytesused, i, group->buffer.flags, GST_TIME_ARGS (timestamp), pool->num_queued, outbuf); if (vmeta) { vmeta->offset[i] = size; size += gst_memory_get_sizes (group->mem[i], NULL, NULL); } } /* Ignore timestamp and field for OUTPUT device */ if (V4L2_TYPE_IS_OUTPUT (obj->type)) goto done; /* Check for driver bug in reporting feild */ if (group->buffer.field == V4L2_FIELD_ANY) { /* Only warn once to avoid the spamming */ #ifndef GST_DISABLE_GST_DEBUG if (!pool->has_warned_on_buggy_field) { pool->has_warned_on_buggy_field = TRUE; GST_WARNING_OBJECT (pool, "Driver should never set v4l2_buffer.field to ANY"); } #endif /* Use the value from the format (works for UVC bug) */ group->buffer.field = obj->format.fmt.pix.field; /* If driver also has buggy S_FMT, assume progressive */ if (group->buffer.field == V4L2_FIELD_ANY) { #ifndef GST_DISABLE_GST_DEBUG if (!pool->has_warned_on_buggy_field) { pool->has_warned_on_buggy_field = TRUE; GST_WARNING_OBJECT (pool, "Driver should never set v4l2_format.pix.field to ANY"); } #endif group->buffer.field = V4L2_FIELD_NONE; } } /* set top/bottom field first if v4l2_buffer has the information */ switch (group->buffer.field) { case V4L2_FIELD_NONE: GST_BUFFER_FLAG_UNSET (outbuf, GST_VIDEO_BUFFER_FLAG_INTERLACED); GST_BUFFER_FLAG_UNSET (outbuf, GST_VIDEO_BUFFER_FLAG_TFF); break; case V4L2_FIELD_INTERLACED_TB: GST_BUFFER_FLAG_SET (outbuf, GST_VIDEO_BUFFER_FLAG_INTERLACED); GST_BUFFER_FLAG_SET (outbuf, GST_VIDEO_BUFFER_FLAG_TFF); break; case V4L2_FIELD_INTERLACED_BT: GST_BUFFER_FLAG_SET (outbuf, GST_VIDEO_BUFFER_FLAG_INTERLACED); GST_BUFFER_FLAG_UNSET (outbuf, GST_VIDEO_BUFFER_FLAG_TFF); break; case V4L2_FIELD_INTERLACED: GST_BUFFER_FLAG_SET (outbuf, GST_VIDEO_BUFFER_FLAG_INTERLACED); if (obj->tv_norm == V4L2_STD_NTSC_M || obj->tv_norm == V4L2_STD_NTSC_M_JP || obj->tv_norm == V4L2_STD_NTSC_M_KR) { GST_BUFFER_FLAG_UNSET (outbuf, GST_VIDEO_BUFFER_FLAG_TFF); } else { GST_BUFFER_FLAG_SET (outbuf, GST_VIDEO_BUFFER_FLAG_TFF); } break; default: GST_BUFFER_FLAG_UNSET (outbuf, GST_VIDEO_BUFFER_FLAG_INTERLACED); GST_BUFFER_FLAG_UNSET (outbuf, GST_VIDEO_BUFFER_FLAG_TFF); GST_FIXME_OBJECT (pool, "Unhandled enum v4l2_field %d - treating as progressive", group->buffer.field); break; } if (GST_VIDEO_INFO_FORMAT (&obj->info) == GST_VIDEO_FORMAT_ENCODED) { if ((group->buffer.flags & V4L2_BUF_FLAG_KEYFRAME) || GST_V4L2_PIXELFORMAT (obj) == V4L2_PIX_FMT_MJPEG || GST_V4L2_PIXELFORMAT (obj) == V4L2_PIX_FMT_JPEG || GST_V4L2_PIXELFORMAT (obj) == V4L2_PIX_FMT_PJPG) GST_BUFFER_FLAG_UNSET (outbuf, GST_BUFFER_FLAG_DELTA_UNIT); else GST_BUFFER_FLAG_SET (outbuf, GST_BUFFER_FLAG_DELTA_UNIT); } if (group->buffer.flags & V4L2_BUF_FLAG_ERROR) GST_BUFFER_FLAG_SET (outbuf, GST_BUFFER_FLAG_CORRUPTED); GST_BUFFER_TIMESTAMP (outbuf) = timestamp; GST_BUFFER_OFFSET (outbuf) = group->buffer.sequence; GST_BUFFER_OFFSET_END (outbuf) = group->buffer.sequence + 1; done: *buffer = outbuf; return GST_FLOW_OK; /* ERRORS */ poll_failed: { GST_DEBUG_OBJECT (pool, "poll error %s", gst_flow_get_name (res)); return res; } eos: { return GST_FLOW_EOS; } dqbuf_failed: { return GST_FLOW_ERROR; } no_buffer: { GST_ERROR_OBJECT (pool, "No free buffer found in the pool at index %d.", group->buffer.index); return GST_FLOW_ERROR; } } static GstFlowReturn gst_v4l2_buffer_pool_acquire_buffer (GstBufferPool * bpool, GstBuffer ** buffer, GstBufferPoolAcquireParams * params) { GstFlowReturn ret; GstV4l2BufferPool *pool = GST_V4L2_BUFFER_POOL (bpool); GstBufferPoolClass *pclass = GST_BUFFER_POOL_CLASS (parent_class); GstV4l2Object *obj = pool->obj; GST_DEBUG_OBJECT (pool, "acquire"); /* If this is being called to resurect a lost buffer */ if (params && params->flags & GST_V4L2_BUFFER_POOL_ACQUIRE_FLAG_RESURRECT) { ret = pclass->acquire_buffer (bpool, buffer, params); goto done; } switch (obj->type) { case V4L2_BUF_TYPE_VIDEO_CAPTURE: case V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE: /* capture, This function should return a buffer with new captured data */ switch (obj->mode) { case GST_V4L2_IO_RW: { /* take empty buffer from the pool */ ret = pclass->acquire_buffer (bpool, buffer, params); break; } case GST_V4L2_IO_DMABUF: case GST_V4L2_IO_MMAP: case GST_V4L2_IO_USERPTR: case GST_V4L2_IO_DMABUF_IMPORT: { /* just dequeue a buffer, we basically use the queue of v4l2 as the * storage for our buffers. This function does poll first so we can * interrupt it fine. */ ret = gst_v4l2_buffer_pool_dqbuf (pool, buffer); break; } default: ret = GST_FLOW_ERROR; g_assert_not_reached (); break; } break; case V4L2_BUF_TYPE_VIDEO_OUTPUT: case V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE: /* playback, This function should return an empty buffer */ switch (obj->mode) { case GST_V4L2_IO_RW: /* get an empty buffer */ ret = pclass->acquire_buffer (bpool, buffer, params); break; case GST_V4L2_IO_MMAP: case GST_V4L2_IO_DMABUF: case GST_V4L2_IO_USERPTR: case GST_V4L2_IO_DMABUF_IMPORT: /* get a free unqueued buffer */ ret = pclass->acquire_buffer (bpool, buffer, params); break; default: ret = GST_FLOW_ERROR; g_assert_not_reached (); break; } break; default: ret = GST_FLOW_ERROR; g_assert_not_reached (); break; } done: return ret; } static void gst_v4l2_buffer_pool_release_buffer (GstBufferPool * bpool, GstBuffer * buffer) { GstV4l2BufferPool *pool = GST_V4L2_BUFFER_POOL (bpool); GstBufferPoolClass *pclass = GST_BUFFER_POOL_CLASS (parent_class); GstV4l2Object *obj = pool->obj; GST_DEBUG_OBJECT (pool, "release buffer %p", buffer); switch (obj->type) { case V4L2_BUF_TYPE_VIDEO_CAPTURE: case V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE: /* capture, put the buffer back in the queue so that we can refill it * later. */ switch (obj->mode) { case GST_V4L2_IO_RW: /* release back in the pool */ pclass->release_buffer (bpool, buffer); break; case GST_V4L2_IO_DMABUF: case GST_V4L2_IO_MMAP: case GST_V4L2_IO_USERPTR: case GST_V4L2_IO_DMABUF_IMPORT: { GstV4l2MemoryGroup *group; #ifdef USE_V4L2_TARGET_NV if (gst_v4l2_is_buffer_valid (buffer, &group, pool->obj->is_encode)) { #else if (gst_v4l2_is_buffer_valid (buffer, &group)) { #endif gst_v4l2_allocator_reset_group (pool->vallocator, group); /* queue back in the device */ if (pool->other_pool) gst_v4l2_buffer_pool_prepare_buffer (pool, buffer, NULL); if (gst_v4l2_buffer_pool_qbuf (pool, buffer) != GST_FLOW_OK) pclass->release_buffer (bpool, buffer); } else { /* Simply release invalide/modified buffer, the allocator will * give it back later */ GST_BUFFER_FLAG_SET (buffer, GST_BUFFER_FLAG_TAG_MEMORY); pclass->release_buffer (bpool, buffer); } break; } default: g_assert_not_reached (); break; } break; case V4L2_BUF_TYPE_VIDEO_OUTPUT: case V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE: switch (obj->mode) { case GST_V4L2_IO_RW: /* release back in the pool */ pclass->release_buffer (bpool, buffer); break; case GST_V4L2_IO_MMAP: case GST_V4L2_IO_DMABUF: case GST_V4L2_IO_USERPTR: case GST_V4L2_IO_DMABUF_IMPORT: { GstV4l2MemoryGroup *group; guint index; #ifdef USE_V4L2_TARGET_NV if (!gst_v4l2_is_buffer_valid (buffer, &group, pool->obj->is_encode)) { #else if (!gst_v4l2_is_buffer_valid (buffer, &group)) { #endif /* Simply release invalide/modified buffer, the allocator will * give it back later */ GST_BUFFER_FLAG_SET (buffer, GST_BUFFER_FLAG_TAG_MEMORY); pclass->release_buffer (bpool, buffer); break; } index = group->buffer.index; if (pool->buffers[index] == NULL) { GST_LOG_OBJECT (pool, "buffer %u not queued, putting on free list", index); /* Remove qdata, this will unmap any map data in userptr */ gst_mini_object_set_qdata (GST_MINI_OBJECT (buffer), GST_V4L2_IMPORT_QUARK, NULL, NULL); /* reset to default size */ gst_v4l2_allocator_reset_group (pool->vallocator, group); /* playback, put the buffer back in the queue to refill later. */ pclass->release_buffer (bpool, buffer); } else { /* the buffer is queued in the device but maybe not played yet. We just * leave it there and not make it available for future calls to acquire * for now. The buffer will be dequeued and reused later. */ GST_LOG_OBJECT (pool, "buffer %u is queued", index); } break; } default: g_assert_not_reached (); break; } break; default: g_assert_not_reached (); break; } } static void gst_v4l2_buffer_pool_dispose (GObject * object) { GstV4l2BufferPool *pool = GST_V4L2_BUFFER_POOL (object); if (pool->vallocator) gst_object_unref (pool->vallocator); pool->vallocator = NULL; if (pool->allocator) gst_object_unref (pool->allocator); pool->allocator = NULL; if (pool->other_pool) gst_object_unref (pool->other_pool); pool->other_pool = NULL; G_OBJECT_CLASS (parent_class)->dispose (object); } static void gst_v4l2_buffer_pool_finalize (GObject * object) { GstV4l2BufferPool *pool = GST_V4L2_BUFFER_POOL (object); if (pool->video_fd >= 0) pool->obj->close (pool->video_fd); #ifdef USE_V4L2_TARGET_NV g_queue_free(pool->allocated_surfaces_queue); #else gst_poll_free (pool->poll); #endif /* This can't be done in dispose method because we must not set pointer * to NULL as it is part of the v4l2object and dispose could be called * multiple times */ gst_object_unref (pool->obj->element); g_cond_clear (&pool->empty_cond); /* FIXME have we done enough here ? */ G_OBJECT_CLASS (parent_class)->finalize (object); } static void gst_v4l2_buffer_pool_init (GstV4l2BufferPool * pool) { #ifndef USE_V4L2_TARGET_NV pool->poll = gst_poll_new (TRUE); #endif g_cond_init (&pool->empty_cond); pool->empty = TRUE; #ifdef USE_V4L2_TARGET_NV pool->allocated_surfaces_queue = g_queue_new(); #endif } static void gst_v4l2_buffer_pool_class_init (GstV4l2BufferPoolClass * klass) { GObjectClass *object_class = G_OBJECT_CLASS (klass); GstBufferPoolClass *bufferpool_class = GST_BUFFER_POOL_CLASS (klass); object_class->dispose = gst_v4l2_buffer_pool_dispose; object_class->finalize = gst_v4l2_buffer_pool_finalize; bufferpool_class->start = gst_v4l2_buffer_pool_start; bufferpool_class->stop = gst_v4l2_buffer_pool_stop; bufferpool_class->set_config = gst_v4l2_buffer_pool_set_config; bufferpool_class->alloc_buffer = gst_v4l2_buffer_pool_alloc_buffer; bufferpool_class->acquire_buffer = gst_v4l2_buffer_pool_acquire_buffer; bufferpool_class->release_buffer = gst_v4l2_buffer_pool_release_buffer; bufferpool_class->flush_start = gst_v4l2_buffer_pool_flush_start; bufferpool_class->flush_stop = gst_v4l2_buffer_pool_flush_stop; GST_DEBUG_CATEGORY_INIT (v4l2bufferpool_debug, "v4l2bufferpool", 0, "V4L2 Buffer Pool"); GST_DEBUG_CATEGORY_GET (CAT_PERFORMANCE, "GST_PERFORMANCE"); } /** * gst_v4l2_buffer_pool_new: * @obj: the v4l2 object owning the pool * * Construct a new buffer pool. * * Returns: the new pool, use gst_object_unref() to free resources */ GstBufferPool * gst_v4l2_buffer_pool_new (GstV4l2Object * obj, GstCaps * caps) { GstV4l2BufferPool *pool; GstStructure *config; gchar *name, *parent_name; gint fd; fd = obj->dup (obj->video_fd); if (fd < 0) goto dup_failed; /* setting a significant unique name */ parent_name = gst_object_get_name (GST_OBJECT (obj->element)); name = g_strconcat (parent_name, ":", "pool:", V4L2_TYPE_IS_OUTPUT (obj->type) ? "sink" : "src", NULL); g_free (parent_name); pool = (GstV4l2BufferPool *) g_object_new (GST_TYPE_V4L2_BUFFER_POOL, "name", name, NULL); g_object_ref_sink (pool); g_free (name); #ifndef USE_V4L2_TARGET_NV gst_poll_fd_init (&pool->pollfd); pool->pollfd.fd = fd; gst_poll_add_fd (pool->poll, &pool->pollfd); if (V4L2_TYPE_IS_OUTPUT (obj->type)) gst_poll_fd_ctl_write (pool->poll, &pool->pollfd, TRUE); else gst_poll_fd_ctl_read (pool->poll, &pool->pollfd, TRUE); #endif pool->video_fd = fd; pool->obj = obj; /* TODO: Check with poll_device set to FALSE */ #ifdef USE_V4L2_TARGET_NV pool->can_poll_device = FALSE; #endif pool->vallocator = gst_v4l2_allocator_new (GST_OBJECT (pool), obj); if (pool->vallocator == NULL) goto allocator_failed; gst_object_ref (obj->element); config = gst_buffer_pool_get_config (GST_BUFFER_POOL_CAST (pool)); #ifndef USE_V4L2_TARGET_NV gst_buffer_pool_config_set_params (config, caps, obj->info.size, 0, 0); #else /* TODO: Fix below once have a single source for Jetson TX1, TX2 and Xavier */ if (((!strcmp(obj->videodev, V4L2_DEVICE_PATH_NVDEC)) && (is_cuvid == FALSE)) || ((!strcmp(obj->videodev, V4L2_DEVICE_PATH_NVDEC_ALT)) && (is_cuvid == FALSE)) || ((!strcmp(obj->videodev, V4L2_DEVICE_PATH_NVDEC_MCCOY)) && (is_cuvid == TRUE))) { gst_buffer_pool_config_set_params (config, caps, obj->info.size, 0, 0); /* Need to adjust the size to 0th plane's size since we will only output v4l2 memory associated with 0th plane. */ if (!V4L2_TYPE_IS_OUTPUT(obj->type)) gst_buffer_pool_config_set_params (config, caps, obj->info.width * obj->info.height, 0, 0); } if (!strcmp(obj->videodev, V4L2_DEVICE_PATH_NVENC) || !strcmp(obj->videodev, V4L2_DEVICE_PATH_NVENC_ALT)) gst_buffer_pool_config_set_params (config, caps, obj->info.size, 0, 0); #endif /* This will simply set a default config, but will not configure the pool * because min and max are not valid */ gst_buffer_pool_set_config (GST_BUFFER_POOL_CAST (pool), config); return GST_BUFFER_POOL (pool); /* ERRORS */ dup_failed: { GST_ERROR ("failed to dup fd %d (%s)", errno, g_strerror (errno)); return NULL; } allocator_failed: { GST_ERROR_OBJECT (pool, "Failed to create V4L2 allocator"); gst_object_unref (pool); return NULL; } } static GstFlowReturn gst_v4l2_do_read (GstV4l2BufferPool * pool, GstBuffer * buf) { GstFlowReturn res; GstV4l2Object *obj = pool->obj; gint amount; GstMapInfo map = GST_MAP_INFO_INIT; gint toread; toread = obj->info.size; GST_LOG_OBJECT (pool, "reading %d bytes into buffer %p", toread, buf); gst_buffer_map (buf, &map, GST_MAP_WRITE); do { if ((res = gst_v4l2_buffer_pool_poll (pool)) != GST_FLOW_OK) goto poll_error; amount = obj->read (obj->video_fd, map.data, toread); if (amount == toread) { break; } else if (amount == -1) { if (errno == EAGAIN || errno == EINTR) { continue; } else goto read_error; } else { /* short reads can happen if a signal interrupts the read */ continue; } } while (TRUE); GST_LOG_OBJECT (pool, "read %d bytes", amount); gst_buffer_unmap (buf, &map); gst_buffer_resize (buf, 0, amount); return GST_FLOW_OK; /* ERRORS */ poll_error: { GST_DEBUG ("poll error %s", gst_flow_get_name (res)); goto cleanup; } read_error: { GST_ELEMENT_ERROR (obj->element, RESOURCE, READ, (_("Error reading %d bytes from device '%s'."), toread, obj->videodev), GST_ERROR_SYSTEM); res = GST_FLOW_ERROR; goto cleanup; } cleanup: { gst_buffer_unmap (buf, &map); gst_buffer_resize (buf, 0, 0); return res; } } /** * gst_v4l2_buffer_pool_process: * @bpool: a #GstBufferPool * @buf: a #GstBuffer, maybe be replaced * * Process @buf in @bpool. For capture devices, this functions fills @buf with * data from the device. For output devices, this functions send the contents of * @buf to the device for playback. * * Returns: %GST_FLOW_OK on success. */ GstFlowReturn gst_v4l2_buffer_pool_process (GstV4l2BufferPool * pool, GstBuffer ** buf) { GstFlowReturn ret = GST_FLOW_OK; GstBufferPool *bpool = GST_BUFFER_POOL_CAST (pool); GstV4l2Object *obj = pool->obj; GST_DEBUG_OBJECT (pool, "process buffer %p", buf); if (GST_BUFFER_POOL_IS_FLUSHING (pool)) return GST_FLOW_FLUSHING; switch (obj->type) { case V4L2_BUF_TYPE_VIDEO_CAPTURE: case V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE: /* capture */ switch (obj->mode) { case GST_V4L2_IO_RW: /* capture into the buffer */ ret = gst_v4l2_do_read (pool, *buf); break; case GST_V4L2_IO_MMAP: case GST_V4L2_IO_DMABUF: { GstBuffer *tmp; if ((*buf)->pool == bpool) { guint num_queued; gsize size = gst_buffer_get_size (*buf); if (size == 0) { if (GST_BUFFER_FLAG_IS_SET (*buf, GST_BUFFER_FLAG_CORRUPTED)) goto buffer_corrupted; else goto eos; } num_queued = g_atomic_int_get (&pool->num_queued); GST_TRACE_OBJECT (pool, "Only %i buffer left in the capture queue.", num_queued); /* If we have no more buffer, and can allocate it time to do so */ #ifdef USE_V4L2_TARGET_NV if (num_queued == 0 && pool->enable_dynamic_allocation) { #else if (num_queued == 0) { #endif if (GST_V4L2_ALLOCATOR_CAN_ALLOCATE (pool->vallocator, MMAP)) { ret = gst_v4l2_buffer_pool_resurect_buffer (pool); if (ret == GST_FLOW_OK) goto done; } } /* start copying buffers when we are running low on buffers */ #ifdef USE_V4L2_TARGET_NV GST_DEBUG_OBJECT (pool, "num_queued %d copy_threshold %d dynamic_allocation %d\n", num_queued, pool->copy_threshold, pool->enable_dynamic_allocation); if (num_queued < pool->copy_threshold && pool->enable_dynamic_allocation) { if (GST_V4L2_ALLOCATOR_CAN_ALLOCATE (pool->vallocator, MMAP)) { ret = gst_v4l2_buffer_pool_resurect_buffer (pool); if (ret == GST_FLOW_OK) goto done; } } #else if (num_queued < pool->copy_threshold) { GstBuffer *copy; if (GST_V4L2_ALLOCATOR_CAN_ALLOCATE (pool->vallocator, MMAP)) { ret = gst_v4l2_buffer_pool_resurect_buffer (pool); if (ret == GST_FLOW_OK) goto done; } /* copy the buffer */ copy = gst_buffer_copy_region (*buf, GST_BUFFER_COPY_ALL | GST_BUFFER_COPY_DEEP, 0, -1); GST_LOG_OBJECT (pool, "copy buffer %p->%p", *buf, copy); /* and requeue so that we can continue capturing */ gst_buffer_unref (*buf); *buf = copy; } #endif ret = GST_FLOW_OK; /* nothing, data was inside the buffer when we did _acquire() */ goto done; } /* buffer not from our pool, grab a frame and copy it into the target */ if ((ret = gst_v4l2_buffer_pool_dqbuf (pool, &tmp)) != GST_FLOW_OK) goto done; /* An empty buffer on capture indicates the end of stream */ if (gst_buffer_get_size (tmp) == 0) { gboolean corrupted = GST_BUFFER_FLAG_IS_SET (tmp, GST_BUFFER_FLAG_CORRUPTED); gst_v4l2_buffer_pool_release_buffer (bpool, tmp); if (corrupted) goto buffer_corrupted; else goto eos; } ret = gst_v4l2_buffer_pool_copy_buffer (pool, *buf, tmp); /* an queue the buffer again after the copy */ gst_v4l2_buffer_pool_release_buffer (bpool, tmp); if (ret != GST_FLOW_OK) goto copy_failed; break; } case GST_V4L2_IO_USERPTR: { struct UserPtrData *data; GstBuffer *tmp; /* Replace our buffer with downstream allocated buffer */ data = gst_mini_object_steal_qdata (GST_MINI_OBJECT (*buf), GST_V4L2_IMPORT_QUARK); tmp = gst_buffer_ref (data->buffer); _unmap_userptr_frame (data); /* Now tmp is writable, copy the flags and timestamp */ gst_buffer_copy_into (tmp, *buf, GST_BUFFER_COPY_FLAGS | GST_BUFFER_COPY_TIMESTAMPS, 0, -1); gst_buffer_replace (buf, tmp); gst_buffer_unref (tmp); break; } case GST_V4L2_IO_DMABUF_IMPORT: { GstBuffer *tmp; /* Replace our buffer with downstream allocated buffer */ tmp = gst_mini_object_steal_qdata (GST_MINI_OBJECT (*buf), GST_V4L2_IMPORT_QUARK); gst_buffer_copy_into (tmp, *buf, GST_BUFFER_COPY_FLAGS | GST_BUFFER_COPY_TIMESTAMPS, 0, -1); gst_buffer_replace (buf, tmp); gst_buffer_unref (tmp); break; } default: g_assert_not_reached (); break; } break; case V4L2_BUF_TYPE_VIDEO_OUTPUT: case V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE: /* playback */ switch (obj->mode) { case GST_V4L2_IO_RW: /* FIXME, do write() */ GST_WARNING_OBJECT (pool, "implement write()"); break; case GST_V4L2_IO_USERPTR: case GST_V4L2_IO_DMABUF_IMPORT: case GST_V4L2_IO_DMABUF: case GST_V4L2_IO_MMAP: { GstBuffer *to_queue = NULL; #ifndef USE_V4L2_TARGET_NV GstV4l2MemoryGroup *group; #else GstV4l2MemoryGroup *group = NULL; #endif gint index; if ((*buf)->pool != bpool) goto copying; #ifdef USE_V4L2_TARGET_NV if (!gst_v4l2_is_buffer_valid (*buf, &group, pool->obj->is_encode)) goto copying; #else if (!gst_v4l2_is_buffer_valid (*buf, &group)) goto copying; #endif index = group->buffer.index; GST_LOG_OBJECT (pool, "processing buffer %i from our pool", index); if (pool->buffers[index] != NULL) { GST_LOG_OBJECT (pool, "buffer %i already queued, copying", index); goto copying; } /* we can queue directly */ to_queue = gst_buffer_ref (*buf); copying: if (to_queue == NULL) { GstBufferPoolAcquireParams params = { 0 }; GST_LOG_OBJECT (pool, "alloc buffer from our pool"); /* this can return EOS if all buffers are outstanding which would * be strange because we would expect the upstream element to have * allocated them and returned to us.. */ params.flags = GST_BUFFER_POOL_ACQUIRE_FLAG_DONTWAIT; ret = gst_buffer_pool_acquire_buffer (bpool, &to_queue, ¶ms); if (ret != GST_FLOW_OK) goto acquire_failed; ret = gst_v4l2_buffer_pool_prepare_buffer (pool, to_queue, *buf); if (ret != GST_FLOW_OK) { gst_buffer_unref (to_queue); goto prepare_failed; } } if ((ret = gst_v4l2_buffer_pool_qbuf (pool, to_queue)) != GST_FLOW_OK) goto queue_failed; /* if we are not streaming yet (this is the first buffer, start * streaming now */ if (!gst_v4l2_buffer_pool_streamon (pool)) { /* don't check return value because qbuf would have failed */ #ifdef USE_V4L2_TARGET_NV gst_v4l2_is_buffer_valid (to_queue, &group, pool->obj->is_encode); #else gst_v4l2_is_buffer_valid (to_queue, &group); #endif /* qbuf has stored to_queue buffer but we are not in * streaming state, so the flush logic won't be performed. * To avoid leaks, flush the allocator and restore the queued * buffer as non-queued */ gst_v4l2_allocator_flush (pool->vallocator); pool->buffers[group->buffer.index] = NULL; gst_mini_object_set_qdata (GST_MINI_OBJECT (to_queue), GST_V4L2_IMPORT_QUARK, NULL, NULL); gst_buffer_unref (to_queue); g_atomic_int_add (&pool->num_queued, -1); goto start_failed; } /* Remove our ref, we will still hold this buffer in acquire as needed, * otherwise the pool will think it is outstanding and will refuse to stop. */ gst_buffer_unref (to_queue); #ifndef USE_V4L2_TARGET_NV if (g_atomic_int_get (&pool->num_queued) >= pool->min_latency) { #else if (g_atomic_int_get (&pool->num_queued) >= (gint) pool->min_latency) { #endif GstBuffer *out; /* all buffers are queued, try to dequeue one and release it back * into the pool so that _acquire can get to it again. */ ret = gst_v4l2_buffer_pool_dqbuf (pool, &out); if (ret == GST_FLOW_OK && out->pool == NULL) /* release the rendered buffer back into the pool. This wakes up any * thread waiting for a buffer in _acquire(). */ gst_v4l2_buffer_pool_release_buffer (bpool, out); } break; } default: g_assert_not_reached (); break; } break; default: g_assert_not_reached (); break; } done: return ret; /* ERRORS */ copy_failed: { GST_ERROR_OBJECT (pool, "failed to copy buffer"); return ret; } buffer_corrupted: { GST_WARNING_OBJECT (pool, "Dropping corrupted buffer without payload"); gst_buffer_unref (*buf); *buf = NULL; return GST_V4L2_FLOW_CORRUPTED_BUFFER; } eos: { GST_DEBUG_OBJECT (pool, "end of stream reached"); gst_buffer_unref (*buf); *buf = NULL; return GST_V4L2_FLOW_LAST_BUFFER; } acquire_failed: { if (ret == GST_FLOW_FLUSHING) GST_DEBUG_OBJECT (pool, "flushing"); else GST_WARNING_OBJECT (pool, "failed to acquire a buffer: %s", gst_flow_get_name (ret)); return ret; } prepare_failed: { GST_ERROR_OBJECT (pool, "failed to prepare data"); return ret; } queue_failed: { GST_ERROR_OBJECT (pool, "failed to queue buffer"); return ret; } start_failed: { GST_ERROR_OBJECT (pool, "failed to start streaming"); return GST_FLOW_ERROR; } } void gst_v4l2_buffer_pool_set_other_pool (GstV4l2BufferPool * pool, GstBufferPool * other_pool) { g_return_if_fail (!gst_buffer_pool_is_active (GST_BUFFER_POOL (pool))); if (pool->other_pool) gst_object_unref (pool->other_pool); pool->other_pool = gst_object_ref (other_pool); } void gst_v4l2_buffer_pool_copy_at_threshold (GstV4l2BufferPool * pool, gboolean copy) { GST_OBJECT_LOCK (pool); pool->enable_copy_threshold = copy; GST_OBJECT_UNLOCK (pool); } gboolean gst_v4l2_buffer_pool_flush (GstBufferPool * bpool) { GstV4l2BufferPool *pool = GST_V4L2_BUFFER_POOL (bpool); gboolean ret = TRUE; gst_v4l2_buffer_pool_streamoff (pool); if (!V4L2_TYPE_IS_OUTPUT (pool->obj->type)) ret = gst_v4l2_buffer_pool_streamon (pool); return ret; } #ifdef USE_V4L2_TARGET_NV void gst_v4l2_buffer_pool_enable_dynamic_allocation (GstV4l2BufferPool * pool, gboolean enable_dynamic_allocation) { GST_DEBUG_OBJECT (pool, "dynamic allocation enable %d", enable_dynamic_allocation); GST_OBJECT_LOCK (pool); pool->enable_dynamic_allocation = enable_dynamic_allocation; if (pool->vallocator) gst_v4l2_allocator_enable_dynamic_allocation (pool->vallocator, enable_dynamic_allocation); GST_OBJECT_UNLOCK (pool); } gint get_motion_vectors(GstV4l2Object *obj, guint32 bufferIndex, v4l2_ctrl_videoenc_outputbuf_metadata_MV *enc_mv_metadata) { v4l2_ctrl_video_metadata metadata; struct v4l2_ext_control control; struct v4l2_ext_controls ctrls; gint ret; ctrls.count = 1; ctrls.controls = &control; ctrls.ctrl_class = V4L2_CTRL_CLASS_MPEG; metadata.buffer_index = bufferIndex; metadata.VideoEncMetadataMV = enc_mv_metadata; control.id = V4L2_CID_MPEG_VIDEOENC_METADATA_MV; control.string = (gchar *)&metadata; if (!GST_V4L2_IS_OPEN (obj)) g_print ("V4L2 device is not open\n"); ret = obj->ioctl (obj->video_fd, VIDIOC_G_EXT_CTRLS, &ctrls); if (ret < 0) g_print ("Error getting dfata\n"); return ret; } static void report_metadata (GstV4l2Object * obj, guint32 buffer_index, v4l2_ctrl_videodec_outputbuf_metadata * metadata) { if (obj->Enable_frame_type_reporting) { switch (metadata->CodecParams.H264DecParams.FrameType) { case 0: g_print ("FrameType = B\n"); break; case 1: g_print ("FrameType = P\n"); break; case 2: g_print ("FrameType = I\n"); if (metadata->CodecParams.H264DecParams.dpbInfo.currentFrame.bIdrFrame) { g_print (" (IDR)\n"); } break; } g_print ("nActiveRefFrames = %d\n", metadata->CodecParams.H264DecParams.dpbInfo.nActiveRefFrames); } if (obj->Enable_error_check) { g_print ("ErrorType= %d Decoded MBs= %d Concealed MBs= %d FrameDecodeTime %d\n", metadata->FrameDecStats.DecodeError, metadata->FrameDecStats.DecodedMBs, metadata->FrameDecStats.ConcealedMBs, metadata->FrameDecStats.FrameDecodeTime); } } static void v4l2_video_dec_get_enable_frame_type_reporting (GstV4l2Object * obj, guint32 buffer_index, v4l2_ctrl_videodec_outputbuf_metadata * dec_metadata) { v4l2_ctrl_video_metadata metadata; struct v4l2_ext_control control; struct v4l2_ext_controls ctrls; gint ret = -1; ctrls.count = 1; ctrls.controls = &control; ctrls.ctrl_class = V4L2_CTRL_CLASS_MPEG; metadata.buffer_index = buffer_index; metadata.VideoDecMetadata = dec_metadata; control.id = V4L2_CID_MPEG_VIDEODEC_METADATA; control.string = (gchar *) &metadata; ret = obj->ioctl (obj->video_fd, VIDIOC_G_EXT_CTRLS, &ctrls); if (ret < 0) g_print ("Error while getting report metadata\n"); } static void v4l2_video_dec_clear_poll_interrupt (GstV4l2Object * obj) { struct v4l2_ext_control control; struct v4l2_ext_controls ctrls; gint ret = -1; ctrls.count = 1; ctrls.controls = &control; ctrls.ctrl_class = V4L2_CTRL_CLASS_MPEG; control.id = V4L2_CID_MPEG_SET_POLL_INTERRUPT; control.value = 0; ret = obj->ioctl (obj->video_fd, VIDIOC_S_EXT_CTRLS, &ctrls); if (ret < 0) g_print ("Error while clearing poll interrupt\n"); } #endif