Linux下实现视频读取,Linux实现视频读取


  V4L(video4linux是一些视频系统,视频软件、音频软件的基础,经常时候在需要采集图像的场合,如视频监控,webcam,可视电话,经常
使用在embedded linux中是linux嵌入式开发中经常使用的系统接口。它是linux内核提供给用户空间的编程接口,各种的视频和音频设备开
发相应的驱动程序后,就可以通过v4l提供的系统API来控制视频和音频设备,也就是说v4l分为两层,底层为音视频设备在内核中的驱动,上
层为系统提供的API,而对于我们来说需要的就是使用这些系统API。
V4L2是V4L的升级版本,为linux下视频设备程序提供了一套接口规范。包括一套数据结构和底层V4L2驱动接口。V4L2采用流水线的方式,
操作更简单直观,基本遵循打开视频设备、设置格式、处理数据、关闭设备,更多的具体操作通过ioctl函数来实现。

1、打开设备
  int open(const char *device_name, int flags);
  int fd = open("/dev/video0", O_RDONLY); //O_NONBLOCK --非阻塞(不推荐使用)
2、关闭设备 int close(int fd)
  int ret = close(fd);
3、v4l2_capability 查看属性
  int ioctl(int fd, int request, struct v4l2_capability *argp);
  struct v4l2_capability
  {
    u8 driver[16]; // 驱动名字
    u8 card[32]; // 设备名字
    u8 bus_info[32]; // 设备在系统中的位置
    u32 version; // 驱动版本号
    u32 capabilities; // 设备支持的操作
    u32 reserved[4]; // 保留字段
  };

4、设置视频格式与制式
相关函数:
int ioctl(int fd, int request, struct v4l2_fmtdesc *argp);
int ioctl(int fd, int request, struct v4l2_format *argp);

相关结构体:
v4l2_cropcap 结构体用来设置摄像头的捕捉能力,在捕捉上视频时应先先设置
v4l2_cropcap 的 type 域,再通过 VIDIO_CROPCAP 操作命令获取设备捕捉能力的参数,保存于 v4l2_cropcap 结构体中,包括 bounds
(最大捕捉方框的左上角坐标和宽高),defrect(默认捕捉方框的左上角坐标和宽高)等。
v4l2_format 结构体用来设置摄像头的视频制式、帧格式等,在设置这个参数时应先填 好 v4l2_format 的各个域,如 type(传输流类型),
fmt.pix.width(宽),fmt.pix.heigth(高),fmt.pix.field(采样区域,如隔行采样),fmt.pix.pixelformat(采样类型,如 YUV4:2:2),
然后通过 VIDIO_S_FMT 操作命令设置视频捕捉格式。
struct v4l2_fmtdesc
{
u32 index; // 要查询的格式序号,应用程序设置
enum v4l2_buf_type type; // 帧类型,应用程序设置
u32 flags; // 是否为压缩格式
u8 description[32]; // 格式名称
u32 pixelformat; // 格式
u32 reserved[4]; // 保留
};
所有的视频格式可以能下面的方法查看
#define v4l2_fourcc(a, b, c, d) ((a) | ((b) << 8) | ((c) << 16) | ((d) << 24))
All format : VIDIOC_ENUM_FMT, v4l2_fmtdesc

struct v4l2_format
{
enum v4l2_buf_type type; // 帧类型,应用程序设置
union fmt
{
struct v4l2_pix_format pix; // 视频设备使用
struct v4l2_window win;
struct v4l2_vbi_format vbi;
struct v4l2_sliced_vbi_format sliced;
u8 raw_data[200];
};
};

struct v4l2_pix_format
{
u32 width; // 帧宽,单位像素
u32 height; // 帧高,单位像素
u32 pixelformat; // 帧格式
enum v4l2_field field;
u32 bytesperline;
u32 sizeimage;
enum v4l2_colorspace colorspace;
u32 priv;
};

5、查看视频的帧率
int ioctl(int fd, int request, struct v4l2_streamparm parm* argp);
struct v4l2_streamparm parm;
parm.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
其中VIDIOC_G_PARM是用取帧率的,VIDIOC_S_PARM是用来设定帧率

6、定制ioctl函数
在内核目录下找到kernel/include/linux/videodev2.h头文件,你可查看所有的io控制的命令
/*
* Experimental, third param 0--video, 1--tracking
*/
#define VIDIOC_POCCESS_NOTIFY _IOW('V', 99, int) //add by Henry.Wen 20131126

实例:

#ifndef CAMERA_V4L2CAMERA_H
#define CAMERA_V4L2CAMERA_H

#include <stdio.h>
#include <stdlib.h>
#include <string.h>
#include <assert.h>
#include <getopt.h>             /* getopt_long() */
#include <fcntl.h>              /* low-level i/o */
#include <unistd.h>
#include <errno.h>
#include <malloc.h>
#include <sys/stat.h>
#include <sys/types.h>
#include <sys/time.h>
#include <sys/mman.h>
#include <sys/ioctl.h>
#include <asm/types.h>          /* for videodev2.h */
#include <linux/videodev2.h>

#define CONFIG_CAMERA_UVC_INVAL_FRAMECNT    5

namespace v4l2
{
	/**
	 * return error code
	 */
	enum _RET_ERROR_CODE
	{
		RET_ERROR_FAIL = -1,
		RET_ERROR_OK = 0,
		RET_ERROR_CAPTURE_NULL = -9999,
		RET_ERROR_CAPTURE_NAME,
		RET_ERROR_CAPTURE_CAPABILITY,
		RET_ERROR_CAPTURE_FORMAT,
		RET_ERROR_CAPTURE_BUFFER,
		RET_ERROR_CAPTURE_OUTMEMORY,
		RET_ERROR_CAPTURE_MMAP,
		RET_ERROR_CAPTURE_FORM,
		RET_ERROR_CAPTURE_MUMAP,
		RET_ERROR_CAPTURE_VIDIOC,
	};

	/**
	 *  Name: video_format enum
	 *  Function: Describe formats V4L2 will support
	 */
	typedef enum _pFormat
	{
		UNKNOWN,
		YUYV,
		MJPEG,
		YV12,
		YU12,
		NV12,
		NV21,
		H264,
	}pFormat;

	/**
	 * frame width and height infomation
	 */
	typedef struct _V4l2Info
	{
		unsigned int width;
		unsigned int height;
		unsigned int stepWidth;
		unsigned int length;
		void*        buffer;
	}V4l2Info;

	/**
	 * caputre properties
	 */
	typedef struct _V4l2Capture
	{
		pFormat      format;
		char         name[31];//dev_name
		int          fd;
		unsigned int rate;
		unsigned int quality;
		unsigned int brightness;
		V4l2Info     v4l2Info;
	}V4l2Capture;

	/**
	 *
	 */
	class V4l2Camera
	{
	public:
		V4l2Camera();
		virtual ~V4l2Camera();

	public:
		/**
		 * get the number of cameras
		 *
		 * @return the number of camera
		 */
		static int getNumberOfCameras();
		
		/**
		 * initialize v4l2 device
		 * @param capture v4l2 capture handl
		 * @param width frame width
		 * @param height fram height
		 *
		 * @return 0/other successful or failure
		 */
		int InitDevice(V4l2Capture *capture, pFormat format, const char* name, unsigned int rate, unsigned int width, unsigned int height);

		/**
		 * initialize v4l2 device
		 * @param capture v4l2 capture handl
		 * @param width frame width
		 * @param height fram height
		 *
		 * @return 0/other successful or failure
		 */
		int UninitDevice(V4l2Capture *capture);

		/**
		 * Set v4l2 device brightness
		 * @param capture v4l2 capture handl
		 * @param value brightness value
		 *
		 * @return 0/other successful or failure
		 */
		int SetBrightness(V4l2Capture *capture, unsigned int value);

		/**
		 * start v4l2 device
		 * @param fd v4l2 capture handl
		 *
		 * @return 0/other successful or failure
		 */
		int StartDevice(int fd);

		/**
		 * stop v4l2 device
		 * @fd capture v4l2 capture handl
		 *
		 * @return 0/other successful or failure
		 */
		int StopDevice(int fd);

		/**
		 * Get frame data
		 * @param capture v4l2 capture handl
		 *
		 * @return 0/other successful or failure
		 */
		int GetFrame(V4l2Capture *capture);

	private:
		int InitMmap(int fd);

		int xioctl(int fd, int request, void *arg);

		unsigned int GetCameraFormat(pFormat format);

		int AdjustV4l2Info(unsigned int& width, unsigned int& height);

		int MSleep(int fd, unsigned int msec);
		
		int MatchCameraAuto(int cameraId);

	private:
		typedef struct _Buffers
		{
			void *start;
			size_t length;
		}V4l2Buffers;

		int                    m_stime;
		V4l2Buffers*           m_buffers;
		unsigned int           m_nBuffers;
		V4l2Capture*           m_capture;
		
		static int mNumberOfCameras;
		static int mCameraIndex[10];
		int mUsbCameraIvalidFrameCnt;
		bool m_InitDevice;
	};
}//end namespace
#endif //CAMERA_V4L2CAMERA_H

  V4l2Camera.cpp

#include "V4l2Camera.h"

namespace v4l2
{
	#define BUFFERS_COUNT  4
	#define ARRAY_LEN(a) (sizeof(a) / sizeof(a[0]))
	#define MEMST_VALUE(x) memset(&(x), 0, sizeof (x))

	const V4l2Info g_v4l2Info[] = {{160, 120}, {320, 240}, {640, 480}, {1024, 768}, {1200, 900}, {1440, 1080}, {1600, 900}, {1600, 1200}};

	V4l2Camera::V4l2Camera()
	{
		// TODO Auto-generated constructor stub
		m_capture = NULL;
		mUsbCameraIvalidFrameCnt = 0;
		m_InitDevice = false;
	}

	V4l2Camera::~V4l2Camera()
	{
		// TODO Auto-generated destructor stub
		if(m_capture)
			UninitDevice(m_capture);
	}

	int V4l2Camera::xioctl(int fd, int request, void *arg)
	{
		int r;
		int nCount= 0;

		do
		{
			r = ioctl (fd, request, arg);
		}while (RET_ERROR_FAIL == r && (EINTR == errno) && (++nCount < 100));

		return r;
	}

	int V4l2Camera::mNumberOfCameras = 0;
	int V4l2Camera::mCameraIndex[] = {0};

	unsigned int V4l2Camera::GetCameraFormat(pFormat format)
	{
		unsigned int ret = UNKNOWN;
		switch(format)
		{
		case YUYV:
			ret = V4L2_PIX_FMT_YUYV;
			break;
		case MJPEG:
			ret = V4L2_PIX_FMT_MJPEG;
			break;
		case YV12:
			ret = V4L2_PIX_FMT_YVU420;
			break;
		case YU12:
			ret = V4L2_PIX_FMT_YUV420;
			break;
		case NV12:
			ret = V4L2_PIX_FMT_NV12;
			break;
		case NV21:
			ret = V4L2_PIX_FMT_NV21;
			break;
		case H264:
			ret = V4L2_PIX_FMT_MPEG;
			break;
		default:
			break;
		}
		return ret;
	}

	int V4l2Camera::InitMmap(int fd)
	{
		int ret = RET_ERROR_OK;
		struct v4l2_requestbuffers req;
		req.count = BUFFERS_COUNT;
		req.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
		req.memory = V4L2_MEMORY_MMAP;

		if (RET_ERROR_FAIL == xioctl (fd, VIDIOC_REQBUFS, &req) || req.count < 2)
		{
			return RET_ERROR_CAPTURE_BUFFER;
		}

		m_buffers = (V4l2Buffers*)calloc(req.count, sizeof(V4l2Buffers));

		if (!m_buffers)
		{
			return RET_ERROR_CAPTURE_OUTMEMORY;
		}

		for (m_nBuffers = 0; m_nBuffers < req.count; ++m_nBuffers)
		{
			struct v4l2_buffer buf;

			MEMST_VALUE (buf);

			buf.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
			buf.memory = V4L2_MEMORY_MMAP;
			buf.index = m_nBuffers;

			if (RET_ERROR_FAIL == xioctl (fd, VIDIOC_QUERYBUF, &buf))
			{
				ret = RET_ERROR_CAPTURE_BUFFER;
				break;
			}

			m_buffers[m_nBuffers].length = buf.length;
			m_buffers[m_nBuffers].start = mmap(NULL, buf.length, PROT_READ | PROT_WRITE, MAP_SHARED, fd, buf.m.offset);

			if (MAP_FAILED == m_buffers[m_nBuffers].start)
			{
				ret = RET_ERROR_CAPTURE_MMAP;
				break;
			}
		}
		return ret;
	}

	int V4l2Camera::AdjustV4l2Info(unsigned int& width, unsigned int& height)
	{
		int ret = RET_ERROR_FAIL;
		int index = 0;
		for(int nCount = ARRAY_LEN(g_v4l2Info) - 1, i = nCount; i >= 0; --i)
		{
			if(width <= (g_v4l2Info[i].width + 50))
			{
				index = i;
				ret = RET_ERROR_OK;
			}
			else if(0 != nCount)
			{
				width  = g_v4l2Info[index].width;
				height = g_v4l2Info[index].height;
				break;
			}
		}
		return ret;
	}

	int V4l2Camera::getNumberOfCameras()
	{
		char cam_path[20];
		int fd = -1, i=0;
		struct v4l2_capability capability;

		mNumberOfCameras = 0;
		memset(mCameraIndex,0x00,sizeof(mCameraIndex));

		for (i = 0; i < 10; ++i)
		{
			memset(cam_path,0x00,20);
			sprintf(cam_path, "/dev/video%d",i);
			fd = open(cam_path, O_RDONLY);
			if (fd < 0)
				continue;

			memset(&capability, 0, sizeof(struct v4l2_capability));
			if (ioctl(fd, VIDIOC_QUERYCAP, &capability) < 0)
			{
				//LOGE("Video device(%s): query capability not supported.\n", cam_path);
				goto loop_continue;
			}

			if ((capability.capabilities
					& (V4L2_CAP_VIDEO_CAPTURE | V4L2_CAP_STREAMING))
					!= (V4L2_CAP_VIDEO_CAPTURE | V4L2_CAP_STREAMING))
			{
			}
			else
			{
				mCameraIndex[mNumberOfCameras] = i;
				mNumberOfCameras++;
			}

	loop_continue:
			if (fd > 0)
			{
				close(fd);
				fd = -1;
			}
			continue;
		}

		return mNumberOfCameras;
	}
	
	int V4l2Camera::MatchCameraAuto(int cameraId)
	{
		if (mNumberOfCameras <= 0 || mNumberOfCameras > 10 || cameraId < 0)
		{
			return -1;
		}

		// search for camera ID normally
		for (int i = 0; i < mNumberOfCameras; ++i)
		{
			if (cameraId == mCameraIndex[i])
			{
				return cameraId;
			}
		}

		if (mNumberOfCameras == 1)
		{
			return mCameraIndex[0];
		}

		return cameraId > mCameraIndex[mNumberOfCameras -1] ? mCameraIndex[mNumberOfCameras -1] : mCameraIndex[0];
	}
	
	int V4l2Camera::InitDevice(V4l2Capture *capture, pFormat format, const char* name, unsigned int rate, unsigned int width, unsigned int height)
	{
		// TODO Auto-generated function stub
		struct stat st;
		int fd = 0, nRealCameraNameLen = 0, nMinCameraNameLen = 0;
		int ret = RET_ERROR_OK;
		char szcameraID[4], szCameraName[20];
		int cameraId = 0;
		struct v4l2_capability cap;
		struct v4l2_cropcap cropcap;
		struct v4l2_crop crop;
		
		if(NULL == capture || NULL == name || 0 == rate || 0 == width || 0 == height) {
			ret = RET_ERROR_CAPTURE_NULL;
			goto InitDeviceFAILED;
		}
	
		if ((nRealCameraNameLen = strlen(name)) < (nMinCameraNameLen =strlen("/dev/video0"))) {
			ret = RET_ERROR_CAPTURE_NULL;
			goto InitDeviceFAILED;
		}
		// Get camera ID

		memset(szcameraID,0x00,4);
		for (int i=0;i<3;i++) {
			if (nRealCameraNameLen >= (nMinCameraNameLen + i))
				szcameraID[i] = name[nMinCameraNameLen - 1 + i];
		}
		
		cameraId = atoi(szcameraID);
		mNumberOfCameras = 0;
		memset(mCameraIndex, 0x00, sizeof(mCameraIndex));
		if (0 == getNumberOfCameras()) {
			//LOGE("There is NO camera!");
			ret = RET_ERROR_CAPTURE_NAME;
			goto InitDeviceFAILED;
		}
	
		if (-1 == (cameraId = MatchCameraAuto(cameraId))) {
			//LOGE("There is NO camera!");
			ret = RET_ERROR_CAPTURE_NAME;
			goto InitDeviceFAILED;
		}
	
		memset(szCameraName, 0x00, 20);
		sprintf(szCameraName, "/dev/video%d", cameraId);
		//LOGI("camera name is %s.", name);
		//
		if ((RET_ERROR_FAIL == stat (szCameraName, &st)) || (!S_ISCHR (st.st_mode))
				|| (RET_ERROR_FAIL == (fd = open(szCameraName, O_RDWR | O_NONBLOCK, 0))))
		{
			ret = RET_ERROR_CAPTURE_NAME;
			goto InitDeviceFAILED;
		}

		if (RET_ERROR_FAIL == xioctl(fd, VIDIOC_QUERYCAP, &cap) || !(cap.capabilities & V4L2_CAP_VIDEO_CAPTURE)
				|| !(cap.capabilities & V4L2_CAP_STREAMING))
		{
			ret = RET_ERROR_CAPTURE_CAPABILITY;
			goto InitDeviceFAILED;
		}

		MEMST_VALUE(cropcap);
		cropcap.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;

		if(RET_ERROR_OK == xioctl (fd, VIDIOC_CROPCAP, &cropcap))
		{
			crop.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
			crop.c = cropcap.defrect; /* reset to default */

			xioctl(fd, VIDIOC_S_CROP, &crop);
		}

		struct v4l2_format fmt;
		MEMST_VALUE (fmt);

		ret = GetCameraFormat(format);
		if(UNKNOWN == ret) {
			return RET_ERROR_CAPTURE_FORMAT;
			goto InitDeviceFAILED;
		}

		if(RET_ERROR_OK != AdjustV4l2Info(width, height))
		{
			ret = RET_ERROR_CAPTURE_FORM;
			goto InitDeviceFAILED;
		}

		//SetBrightness(capture, capture->brightness);
		fmt.type                = V4L2_BUF_TYPE_VIDEO_CAPTURE;
		fmt.fmt.pix.width       = width;
		fmt.fmt.pix.height      = height;
		fmt.fmt.pix.pixelformat = ret;
		fmt.fmt.pix.field       = V4L2_FIELD_INTERLACED;
		printf("%s(%d) pixel format is %d\n", __FUNCTION__,__LINE__,ret);

		if (RET_ERROR_FAIL == xioctl (fd, VIDIOC_S_FMT, &fmt)) {
			ret = RET_ERROR_CAPTURE_FORMAT;
			goto InitDeviceFAILED;
		}

		if(RET_ERROR_OK != (ret = InitMmap(fd)) ) {// || RET_ERROR_OK != (ret = SetBrightness(capture, capture->brightness)))
			goto InitDeviceFAILED;
		}
	
		// set video frame rate
		struct v4l2_streamparm parm;
		parm.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
		if (RET_ERROR_OK != ioctl(fd, VIDIOC_G_PARM, &parm)) {
			//LOGI("VIDIOC_G_PARM fail....");
		}
	
		parm.parm.capture.timeperframe.numerator = 1;
		parm.parm.capture.timeperframe.denominator = rate;
		if (RET_ERROR_OK != ioctl(fd, VIDIOC_S_PARM, &parm)) {
			//LOGI("VIDIOC_S_PARM  Fail....");
		}
	
		//check setting of frame rate
		memset(&parm, 0x00, sizeof(v4l2_streamparm));
		parm.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
	
		ret = ioctl(fd, VIDIOC_G_PARM, &parm);
		if (ret != RET_ERROR_OK) {
			//LOGI("VIDIOC_G_PARM fail....");
		}
		
		capture->fd = fd;
		capture->v4l2Info.width = fmt.fmt.pix.width;
		capture->v4l2Info.height = fmt.fmt.pix.height;
		capture->format = format;
		capture->rate = rate;
		capture->v4l2Info.length = fmt.fmt.pix.sizeimage;
		capture->v4l2Info.stepWidth = fmt.fmt.pix.bytesperline;
		capture->v4l2Info.buffer = malloc(fmt.fmt.pix.sizeimage);
		strncpy(capture->name, name, sizeof(capture->name));

		m_stime = 1000 / rate;
		m_capture = capture;
		m_InitDevice = true;

		return ret;
		
InitDeviceFAILED:
		if (fd >= 0)
		{
			close(fd);
			fd = -1;
		}

	return ret;
	}

	int V4l2Camera::UninitDevice(V4l2Capture *capture)
	{
		// TODO Auto-generated function stub
		printf("%s(%d)...[BEGIN]\n", __FUNCTION__, __LINE__);
		int ret = RET_ERROR_OK;
		if (!m_InitDevice) {
			return ret;
		} else {
			m_InitDevice = false;
		}
		if (m_buffers) {
			for (unsigned int i = 0; i < m_nBuffers; ++i) {
				printf("%s(%d) munmap() i = %d\n", __FUNCTION__, __LINE__, i);
				if (RET_ERROR_FAIL == munmap(m_buffers[i].start, m_buffers[i].length)) {
					ret = RET_ERROR_CAPTURE_MUMAP;
					break;
				}
			}

			if (RET_ERROR_OK == ret) {
				printf("%s(%d) free(m_buffers)\n", __FUNCTION__, __LINE__);
				free(m_buffers);
			}
			m_buffers = NULL;
			if (capture) {
				if (capture->v4l2Info.buffer) {
					printf("%s(%d) free(capture->v4l2Info.buffer)\n", __FUNCTION__,
							__LINE__);
					free(capture->v4l2Info.buffer);
					capture->v4l2Info.buffer = NULL;
				}

				if (capture->fd >= 0) {
					printf("%s(%d) close(capture->fd)\n", __FUNCTION__, __LINE__);
					ret = close(capture->fd);
				}
			}

		}
		printf("%s(%d)...[END]\n", __FUNCTION__, __LINE__);
		return ret;
	}

	int V4l2Camera::SetBrightness(V4l2Capture *capture, unsigned int value)
	{
		// TODO Auto-generated function stub
		if(!capture || value > 10000)
			return RET_ERROR_FAIL;

		struct v4l2_control control;
		control.id = V4L2_CID_BRIGHTNESS;
		control.value = 255;

		if(RET_ERROR_FAIL == xioctl(capture->fd, VIDIOC_S_CTRL, &control))
		{
			return RET_ERROR_FAIL;
		}
		capture->brightness = control.value;

		return RET_ERROR_OK;
	}

	int V4l2Camera::StartDevice(int fd)
	{
		// TODO Auto-generated function stub
		if(fd < 0)
			return RET_ERROR_FAIL;

		int ret = RET_ERROR_OK;
		for (unsigned int i = 0; i < m_nBuffers; ++i)
		{
			struct v4l2_buffer buf;
			MEMST_VALUE(buf);

			buf.type        = V4L2_BUF_TYPE_VIDEO_CAPTURE;
			buf.memory      = V4L2_MEMORY_MMAP;
			buf.index       = i;

			if (RET_ERROR_FAIL == xioctl(fd, VIDIOC_QBUF, &buf))
			{
				ret = RET_ERROR_CAPTURE_VIDIOC;
				break;
			}
		}

		enum v4l2_buf_type type = V4L2_BUF_TYPE_VIDEO_CAPTURE;

		if (RET_ERROR_FAIL == xioctl(fd, VIDIOC_STREAMON, &type))
			ret = RET_ERROR_CAPTURE_VIDIOC;

		return ret;
	}

	int V4l2Camera::StopDevice(int fd)
	{
		// TODO Auto-generated function stub
		enum v4l2_buf_type type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
		if (RET_ERROR_FAIL == xioctl(fd, VIDIOC_STREAMOFF, &type))
		{
			return RET_ERROR_FAIL;
		}
		return 0;
	}

	int V4l2Camera::MSleep(int fd, unsigned int msec)
	{
		fd_set fds;
		FD_ZERO (&fds);
		FD_SET (fd, &fds);

		struct timeval tv;
		tv.tv_sec = msec;
		tv.tv_usec = 0;

		return select (fd + 1, &fds, NULL, NULL, &tv);
	}

	int V4l2Camera::GetFrame(V4l2Capture *capture)
	{
		// TODO Auto-generated function stub
		if(!capture)
			return RET_ERROR_FAIL;

		int fd = capture->fd;

		if(RET_ERROR_FAIL == MSleep(capture->fd, m_stime))
			return RET_ERROR_FAIL;

		struct v4l2_buffer buf;

		MEMST_VALUE(buf);
		buf.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
		buf.memory = V4L2_MEMORY_MMAP;
		buf.reserved = 0;

		// Skip the first CONFIG_CAMERA_UVC_INVAL_FRAMECNT video frames
		// because they are possibly invalid
		if (mUsbCameraIvalidFrameCnt< CONFIG_CAMERA_UVC_INVAL_FRAMECNT) {
			mUsbCameraIvalidFrameCnt++;
			if(xioctl(fd, VIDIOC_DQBUF, &buf) >= 0) {
				xioctl(fd, VIDIOC_QBUF, &buf);
			}
			return RET_ERROR_OK;
		}
		
		if(RET_ERROR_FAIL == xioctl(fd, VIDIOC_DQBUF, &buf) || buf.index >=  m_nBuffers)
		{
			return RET_ERROR_FAIL;
		}

		memcpy(capture->v4l2Info.buffer, m_buffers[buf.index].start, buf.bytesused);

		if(RET_ERROR_FAIL == xioctl(fd, VIDIOC_QBUF, &buf))
			return RET_ERROR_FAIL;

		return RET_ERROR_OK;
	}
}//end namespace

  


在linux下,怎利用QT实现从AVI文件里提取单帧图像数据

qt中的qmovie类进行视频播放,似乎只能读取gif和mng,不支持avi

所以要另外找库

avi是有版权的嘿,你只能找开源的解码器

推荐xvid

qt我没做过视频的,所以也没法提供实例代码

希望楼下来个高手解答下。。。。
 

linux下怎编程实现获取显卡的状态?

这个要获取网卡的状态的话,唯一的方法就是调用网卡驱动提供的服务也就是ioctl函数,ioctl函数可以接受网卡驱动提供的功能标识然后反馈网卡的状态,可以参见unix网络编程的第15或者16章。也可以参考ifconfig的源代码。
 

相关内容

    暂无相关文章