首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
基于多视定位算法的多摄像机标定   总被引:1,自引:0,他引:1       下载免费PDF全文
针对多摄像机系统标定,提出一种基于多视定位算法的标定方法,标定过程只需一块可自由移动的平面模板即可,利用约束优化思想,将各摄像机坐标系转换到参考摄像机坐标系下,从而得到摄像机之间相对位置关系。标定操作过程简单,易于实现。实验结果表明,该方法是一种有效的多摄像机标定方法。  相似文献   

2.
多摄像机系统具有摄像机数目多、空间位置分布复杂特点,导致多摄像机标定效率低。基本矩阵计算和非线性优化是摄像机标定算法的关键步骤。针对标定物空间位置相互独立性,改进随机抽样一致性(RANSAC)的基本矩阵计算和简化非线性优化的增量方程,提出多摄像机系统的并行标定算法。该算法挖掘多摄像机标定过程的内在并行化,从而提高了标定的时间效率。相比于传统的多摄像机标定算法,并行算法的时间复杂度从O(n3)降为O(n)。实验结果表明:使用多摄像机系统并行标定算法在不损失精度的同时能够减少标定时间,实现多摄像机系统的快速标定。  相似文献   

3.
In computer vision, camera calibration is a necessary process when the retrieval of information such as angles and distances is required. This paper addresses the multi-camera calibration problem with a single dimension calibration pattern under general motions. Currently, the known algorithms for solving this problem are based on the estimation of vanishing points. However, this estimate is very susceptible to noise, making the methods unsuitable for practical applications. Instead, this paper presents a new calibration algorithm, where the cameras are divided into binocular sets. The fundamental matrix of each binocular set is then estimated, allowing to perform a projective calibration of each camera. Then, the calibration is updated for the Euclidean space, ending the process. The calibration is possible without imposing any restrictions on the movement of the pattern and without any prior information about the cameras or motion. Experiments on synthetic and real images validate the new method and show that its accuracy makes it suitable also for practical applications.  相似文献   

4.
提供了一个无标记点的身体与面部运动同步捕获的方法.利用经过时间同步和空间标定的长焦彩色相机和Kinect相机来进行同步捕获.利用在环境中加入闪光来进行时间同步,使用张氏标定法进行空间标定,从而组成一组时间同步且空间对齐的混合相机(hybrid camera).然后利用Kinect fusion扫描用户的人体模型并嵌入骨骼.最后利用时间和空间都对齐好的两个相机来进行同步采集.首先从深度图像中得到人脸的平移参考值,然后在平移参考值的帮助下根据彩色图像的2D特征点重建人脸.随后,把彩色图像中得到的头部姿态传递给身体捕获结果.结果对比实验和用户调研实验均表明所提出的运动捕获的结果要好于单个的运动捕获结果.  相似文献   

5.
Camera calibration is to identify a model that infers 3-D space measurements from 2-D image observations. In this paper, the nonlinear mapping model of the camera is approximated by a series of linear input-output models defined on a set of local regions called receptive fields. Camera calibration is thus a learning procedure to evolve the size and shape of every receptive field as well as parameters of the associated linear model. Since the learning procedure can also provide an approximation extent measurement for the linear model on each of the receptive fields, calibration model is consequently obtained from a fusion framework integrated with all linear models weighted by their corresponding approximation measurements. Since each camera model is composed of several receptive fields, it is feasible to unitedly calibrate multiple cameras simultaneously. The 3-D measurements of a multi-camera vision system are produced from a weighted regression fusion on all receptive fields of cameras. Thanks to the fusion strategy, the resultant calibration model of a multi-camera system is expected to have higher accuracy than either of them. Moreover, the calibration model is very efficient to be updated whenever one or more cameras in the multi-camera vision system need to be activated or deactivated to adapt to variable sensing requirements at different stages of task fulfillment. Simulation and experiment results illustrate effectiveness and properties of the proposed method. Comparisons with neural network-based calibration method and Tsai's method are also provided to exhibit advantages of the method.  相似文献   

6.
This paper presents a novel approach for matching 2-D points between a video projector and a digital camera. Our method is motivated by camera–projector applications for which the projected image needs to be warped to prevent geometric distortion. Since the warping process often needs geometric information on the 3-D scene obtained from a triangulation, we propose a technique for matching points in the projector to points in the camera based on arbitrary video sequences. The novelty of our method lies in the fact that it does not require the use of pre-designed structured light patterns as is usually the case. The backbone of our application lies in a function that matches activity patterns instead of colors. This makes our method robust to pose, severe photometric and geometric distortions. It also does not require calibration of the color response curve of the camera–projector system. We present quantitative and qualitative results with synthetic and real-life examples, and compare the proposed method with the scale invariant feature transform (SIFT) method and with a state-of-the-art structured light technique. We show that our method performs almost as well as structured light methods and significantly outperforms SIFT when the contrast of the video captured by the camera is degraded.  相似文献   

7.
Self-identifying patterns for plane-based camera calibration   总被引:2,自引:0,他引:2  
Determining camera calibration parameters is a time-consuming task despite the availability of calibration algorithms and software. A set of correspondences between points on the calibration target and the camera image(s) must be found, usually a manual or manually guided process. Most calibration tools assume that the correspondences are already found. We present a system which allows a camera to be calibrated merely by passing it in front of a panel of self-identifying patterns. This calibration scheme uses an array of fiducial markers which are detected with a high degree of confidence, each detected marker provides one or four correspondence points. Experiments were performed calibrating several cameras in a short period of time with no manual intervention. This marker-based calibration system was compared to one using the OpenCV chessboard grid finder which also finds correspondences automatically. We show how our new marker-based system more robustly finds the calibration pattern and how it provides more accurate intrinsic camera parameters.  相似文献   

8.
视频监控摄像头在现代安防中起到了不可替代的作用,为适应更多环境,摄像头技术也一直在不断发展.本文提出了一种可实现多摄像头画面实时拼接的技术设计方案:基于加速稳健特征算法,实现对视频单帧画面间的特征点进行提取及拼接;使用FFmpeg(Fast Forward Mpeg)媒体处理库进行视频摄像头媒体流文件的分解及最终融合视...  相似文献   

9.
In this paper, we propose a new algorithm for dynamic calibration of multiple cameras. Based on the mapping between a horizontal plane in the 3-D space and the 2-D image plane on a panned and tilted camera, we utilize the displacement of feature points and the epipolar-plane constraint among multiple cameras to infer the changes of pan and tilt angles for each camera. This algorithm does not require a complicated correspondence of feature points. It can be applied to surveillance systems with wide-range coverage. It also allows the presence of moving objects in the captured scenes while performing dynamic calibration. The sensitivity analysis of our algorithm with respect to measurement errors and fluctuations in previous estimations is also discussed. The efficiency and feasibility of this approach has been demonstrated in some experiments over real scenery.  相似文献   

10.
Central catadioptric cameras are widely used in virtual reality and robot navigation,and the camera calibration is a prerequisite for these applications.In this paper,we propose an easy calibration method for central catadioptric cameras with a 2D calibration pattern.Firstly,the bounding ellipse of the catadioptric image and field of view (FOV) are used to obtain the initial estimation of the intrinsic parameters.Then,the explicit relationship between the central catadioptric and the pinhole model is used to initialize the extrinsic parameters.Finally,the intrinsic and extrinsic parameters are refined by nonlinear optimization.The proposed method does not need any fitting of partial visible conic,and the projected images of 2D calibration pattern can easily cover the whole image,so our method is easy and robust.Experiments with simulated data as well as real images show the satisfactory performance of our proposed calibration method.  相似文献   

11.
一种反射折射摄像机的简易标定方法   总被引:3,自引:0,他引:3  
Central catadioptric cameras are widely used in virtual reality and robot navigation, and the camera calibration is a prerequisite for these applications. In this paper, we propose an easy calibration method for central catadioptric cameras with a 2D calibration pattern. Firstly, the bounding ellipse of the catadioptric image and field of view (FOV) are used to obtain the initial estimation of the intrinsic parameters. Then, the explicit relationship between the central catadioptric and the pinhole model is used to initialize the extrinsic parameters. Finally, the intrinsic and extrinsic parameters are refined by nonlinear optimization. The proposed method does not need any fitting of partial visible conic, and the projected images of 2D calibration pattern can easily cover the whole image, so our method is easy and robust. Experiments with simulated data as well as real images show the satisfactory performance of our proposed calibration method.  相似文献   

12.
Hybrid central catadioptric and perspective cameras are desired in practice, because the hybrid camera system can capture large field of view as well as high-resolution images. However, the calibration of the system is challenging due to heavy distortions in catadioptric cameras. In addition, previous calibration methods are only suitable for the camera system consisting of perspective cameras and catadioptric cameras with only parabolic mirrors, in which priors about the intrinsic parameters of perspective cameras are required. In this work, we provide a new approach to handle the problems. We show that if the hybrid camera system consists of at least two central catadioptric and one perspective cameras, both the intrinsic and extrinsic parameters of the system can be calibrated linearly without priors about intrinsic parameters of the perspective cameras, and the supported central catadioptric cameras of our method can be more generic. In this work, an approximated polynomial model is derived and used for rectification of catadioptric image. Firstly, with the epipolar geometry between the perspective and rectified catadioptric images, the distortion parameters of the polynomial model can be estimated linearly. Then a new method is proposed to estimate the intrinsic parameters of a central catadioptric camera with the parameters in the polynomial model, and hence the catadioptric cameras can be calibrated. Finally, a linear self-calibration method for the hybrid system is given with the calibrated catadioptric cameras. The main advantage of our method is that it cannot only calibrate both the intrinsic and extrinsic parameters of the hybrid camera system, but also simplify a traditional nonlinear self-calibration of perspective cameras to a linear process. Experiments show that our proposed method is robust and reliable.  相似文献   

13.
摄像机标定是三维重建时的必要步骤。传统的标定方法对设备要求高、操作繁琐,而自标定方法虽然简便,但精度不高,会严重影响三维重建的效果。因此,越来越需要一种操作简便并且精度高的自标定方法。采用SIFT特征点匹配算法,根据多视序列图像中对应点间的相互关系,利用光束法平差,提出了一种基于局部-全局混合优化的迭代优化方法。针对图像匹配量大的问题,提出了一种邻域内图像互匹配方法来降低时间代价。实验表明,本文提出的多摄像机自标定方法是一种有效的高精度方法,采用的邻域内图像互匹配技术能很好地降低图像匹配的时间消耗。根据多视图像的对应点间相互关系,充分利用局部-全局优化的思想,通过混合优化的方法得到相机参数,对比现有自标定算法,本文给出的方法有较高的精度和鲁棒性。  相似文献   

14.
基于参考像面法的CCD摄像机标定新技术   总被引:4,自引:3,他引:1  
摄像机标定是计算机视觉检测中必不可少的步骤.在现有的摄像机参数标定算法中,纯线性算法标定速度快,但标定精度不是很高;相反如采用非线性搜索算法,标定精度高,但存在标定速度慢以及会出现不收敛现象.通过对现有线性标定算法的研究,提出一种既考虑摄像机的径向和切向畸变,又能实现全线性化求解的全新标定算法--参考像面法.经验证该方法标定速度快、精度较高、算法健壮,适用于大型视觉系统多摄像机的快速标定.  相似文献   

15.
激光雷达的点云和相机的图像经常被融合应用在多个领域。准确的外参标定是融合两者信息的前提。点云特征提取是外参标定的关键步骤。但是点云的低分辨率和低质量会影响标定结果的精度。针对这些问题,提出一种基于边缘关联点云的激光雷达与相机外参标定方法。首先,利用双回波提取标定板边缘关联点云;然后,通过优化方法从边缘关联点云中提取出与实际标定板尺寸大小兼容的标定板角点;最后,将点云中角点和图像中角点匹配。用多点透视方法求解激光雷达与相机之间的外参。实验结果表明,该方法的重投影误差为1.602px,低于同类对比方法,验证了该方法的有效性与准确性。  相似文献   

16.
针对线阵相机特殊使用场景中所需要的高精度图像,对线阵相机进行高精度标定。提出一种基于光束法平差的双线阵相机标定方法。通过背景差分法获取线阵相机的特征点的像素坐标。再利用已有的直接线性变换方法和非线性优化方法求出相机的内参,外参,畸变参数后,将得到的初始参数与世界坐标作为待优化集合,利用LM法和光束法平差对该集合进行更进一步的优化,使得双线阵系统的重投影误差降到最低。实验表明,该方法与传统的线阵相机标定方法相比重投影误差降低了75.01%。  相似文献   

17.
18.
由于传统线阵相机的标定过程复杂,且对标定物精度要求较高,难以保证缺陷的定位精度,本文提出一种线阵相机的圆环旋转标定方法以提高缺陷的定位精度。该方法设计一种新型的圆环形标定板,在静态标定基础上通过旋转线阵相机采集相机视线与圆的交点的坐标,得到旋转角度以及多组标定点,建立线阵相机的成像模型和径向畸变模型,通过非线性优化整体误差函数求解相机的内参和畸变参数,同时分析相机不同旋转角度对标定精度的影响。实验结果表明,当θ≤20°时,该方法的标定精度在0.35 pixel以内,满足实际检测的定位要求,并且在PCB缺陷检测中得到较好的验证。  相似文献   

19.
We present a method for active self-calibration of multi-camera systems consisting of pan-tilt zoom cameras. The main focus of this work is on extrinsic self-calibration using active camera control. Our novel probabilistic approach avoids multi-image point correspondences as far as possible. This allows an implicit treatment of ambiguities. The relative poses are optimized by actively rotating and zooming each camera pair in a way that significantly simplifies the problem of extracting correct point correspondences. In a final step we calibrate the entire system using a minimal number of relative poses. The selection of relative poses is based on their uncertainty. We exploit active camera control to estimate consistent translation scales for triplets of cameras. This allows us to estimate missing relative poses in the camera triplets. In addition to this active extrinsic self-calibration we present an extended method for the rotational intrinsic self-calibration of a camera that exploits the rotation knowledge provided by the camera’s pan-tilt unit to robustly estimate the intrinsic camera parameters for different zoom steps as well as the rotation between pan-tilt unit and camera. Quantitative experiments on real data demonstrate the robustness and high accuracy of our approach. We achieve a median reprojection error of $0.95$ pixel.  相似文献   

20.
Using vanishing points for camera calibration   总被引:42,自引:1,他引:42  
In this article a new method for the calibration of a vision system which consists of two (or more) cameras is presented. The proposed method, which uses simple properties of vanishing points, is divided into two steps. In the first step, the intrinsic parameters of each camera, that is, the focal length and the location of the intersection between the optical axis and the image plane, are recovered from a single image of a cube. In the second step, the extrinsic parameters of a pair of cameras, that is, the rotation matrix and the translation vector which describe the rigid motion between the coordinate systems fixed in the two cameras are estimated from an image stereo pair of a suitable planar pattern. Firstly, by matching the corresponding vanishing points in the two images the rotation matrix can be computed, then the translation vector is estimated by means of a simple triangulation. The robustness of the method against noise is discussed, and the conditions for optimal estimation of the rotation matrix are derived. Extensive experimentation shows that the precision that can be achieved with the proposed method is sufficient to efficiently perform machine vision tasks that require camera calibration, like depth from stereo and motion from image sequence.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号