[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

CN108093188B - A method for large-field video panorama stitching based on hybrid projection transformation model - Google Patents

A method for large-field video panorama stitching based on hybrid projection transformation model Download PDF

Info

Publication number
CN108093188B
CN108093188B CN201711419418.6A CN201711419418A CN108093188B CN 108093188 B CN108093188 B CN 108093188B CN 201711419418 A CN201711419418 A CN 201711419418A CN 108093188 B CN108093188 B CN 108093188B
Authority
CN
China
Prior art keywords
image
area
registration
pixel
regions
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201711419418.6A
Other languages
Chinese (zh)
Other versions
CN108093188A (en
Inventor
袁丁
胡晓辉
张弘
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beihang University
Original Assignee
Beihang University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beihang University filed Critical Beihang University
Priority to CN201711419418.6A priority Critical patent/CN108093188B/en
Publication of CN108093188A publication Critical patent/CN108093188A/en
Application granted granted Critical
Publication of CN108093188B publication Critical patent/CN108093188B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • H04N5/262Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
    • H04N5/265Mixing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2411Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/698Control of cameras or camera modules for achieving an enlarged field of view, e.g. panoramic image capture
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • H04N5/262Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Data Mining & Analysis (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Processing (AREA)
  • Image Analysis (AREA)

Abstract

The present invention provides a kind of joining method of big visual field video panorama based on hybrid projection transformation model, including five big steps, step 1: down-sampled, reduction operand is carried out to video;Step 2: the detection of Sift characteristic point is extracted;Step 3: to feature points clustering, region segmentation is carried out to image using SVM method, and carry out the matching of characteristic point;Step 4: hybrid projection matrix is calculated;Step 5: interative computation, realize the registration of subsequent image sequence with merge, the final splicing for realizing panorama sketch.Time-consuming problem in big, more projection array methods that present invention effectively avoids errors in traditional single projection array method has wide application background.

Description

一种基于混合投影变换模型的大视场视频全景图拼接的方法A method for large-field video panorama stitching based on hybrid projection transformation model

技术领域technical field

本发明涉及一种基于混合投影变换模型的大视场视频全景图拼接的方法,该方法可以实现视频图像序列的准确快速拼接,从而生成全景图,可以应用于虚拟现实等技术中,属于计算机视觉领域。The invention relates to a method for splicing a large field of view video panorama based on a hybrid projection transformation model. The method can realize accurate and fast splicing of video image sequences, thereby generating a panorama, which can be applied to technologies such as virtual reality and belongs to computer vision. field.

背景技术Background technique

大视场视频全景图拼接技术是当前视频处理领域中的一个热点研究问题,它可以将一段视频的图像序列进行拼接形成一张包含完整场景的广视角图像,解决了因摄像设备限制而存在的拍摄角度小、无法全景观测的问题。该技术已经广泛应用于遥感、海底探测和虚拟现实等领域之中。在遥感领域中,高分辨率图像的获取需要拼接建立大面积全景图像;在海底探测方面,同样需要对获取的小镜头视频信息进行拼接形成全景图以进行观测和处理;而虚拟现实更是需要通过全景图的建立来为用户提供更加直观立体的感受。The large-field video panorama stitching technology is a hot research problem in the current video processing field. It can stitch the image sequence of a video to form a wide-angle image containing a complete scene, which solves the problems existing due to the limitation of camera equipment. The problem is that the shooting angle is small and the panoramic observation cannot be performed. The technology has been widely used in remote sensing, seabed detection and virtual reality and other fields. In the field of remote sensing, the acquisition of high-resolution images requires splicing to build a large-area panoramic image; in seabed detection, it is also necessary to stitch the acquired small-lens video information to form a panoramic image for observation and processing; and virtual reality requires more Provide users with a more intuitive and three-dimensional experience through the establishment of panoramic images.

近年来,虽然全景图的拼接技术已经被广泛研究,并且已有很多成熟的方法,但是,仍然存在很多具有挑战性的问题需要去解决。最大的问题就是在目前大多数的全景图拼接中,都有一个假设前提:要么图像与摄像机距离很远,可以认为图像中的场景都处于同一平面中;要么图像是由摄像机沿投影中心旋转拍摄得到的。这样,在图像配准时,可以使用一个单应矩阵作为图像配准的投影变换模型。In recent years, although panorama stitching technology has been widely studied, and there are many mature methods, there are still many challenging problems to be solved. The biggest problem is that in most of the current panorama stitching, there is an assumption: either the image and the camera are far away, and the scenes in the image can be considered to be in the same plane; or the image is shot by the camera rotating along the projection center. owned. In this way, during image registration, a homography matrix can be used as a projection transformation model for image registration.

但是,这样的方法并不具有鲁棒性,因为在实际拍摄视频时具有很强的随机性和自由性,基本无法满足上述假设条件;因此,如果依然使用一个单应阵进行拼接,得到的全景图是存在误差的。目前,大多算法都是在后期处理时使用图像融合技术来弥补配准过程的误差,但这不能从根本上解决问题。However, this method is not robust, because it has strong randomness and freedom in actual video shooting, and basically cannot meet the above assumptions; therefore, if a homography is still used for stitching, the obtained panorama The picture is wrong. At present, most algorithms use image fusion technology in post-processing to compensate for the errors in the registration process, but this cannot fundamentally solve the problem.

基于此,使用多单应阵来进行视频全景图的拼接逐渐开始被研究。其中,比较典型的有2011年在CVPR上发表的Dual-Homography Warping(DHW)方法,它通过综合两个平面的单应矩阵来实现整幅图像的配准拼接工作。但是,在权重计算时,由于需要进行欧式距离的计算,并且随着特征点的增多,计算次数也会很大程度地增多,因此会耗费很大的时间。Based on this, the use of multi-homography arrays for video panorama stitching has gradually begun to be studied. Among them, a typical one is the Dual-Homography Warping (DHW) method published on CVPR in 2011, which realizes the registration and stitching of the entire image by synthesizing the homography matrix of two planes. However, in the weight calculation, since the Euclidean distance needs to be calculated, and with the increase of feature points, the number of calculations will also increase to a large extent, so it will take a lot of time.

发明内容SUMMARY OF THE INVENTION

本发明的技术解决问题:克服现有技术的不足,提供一种基于混合投影变换模型的大视场视频全景图拼接的方法,耗费时间更短,同时准确度也更高,且对拍摄视频方式无任何约束,能够具有广泛的应用。The technical solution of the present invention is to overcome the deficiencies of the prior art and provide a method for splicing a large field of view video panorama based on a hybrid projection transformation model, which consumes less time and has higher accuracy. Without any constraints, it can have a wide range of applications.

本发明技术解决方案:本发明旨在实现大视场视频全景图的拼接。其输入量是由摄像机获取的一段视频,输出为一幅全景图,本发明假设所处理的视频图像大致可分为两个平面(Background plane和foreground plane)。下面介绍本发明中实现大视场视频全景图拼接的技术解决方案:Technical solution of the present invention: The present invention aims to realize the splicing of video panorama images with a large field of view. The input is a video acquired by the camera, and the output is a panoramic image. The present invention assumes that the processed video image can be roughly divided into two planes (Background plane and foreground plane). The following introduces the technical solution for realizing large-field video panorama stitching in the present invention:

(1)对视频图像序列降采样,保证每两幅相邻图像间最少有1/3的重叠区域;(1) Downsampling the video image sequence to ensure that there is at least 1/3 of the overlapping area between every two adjacent images;

(2)对采样获得的前两幅图像的重叠区域进行Sift特征点的检测,获得的所述特征点集,并对特征点进行特征向量的提取;(2) The detection of Sift feature points is performed on the overlapping area of the first two images obtained by sampling, the obtained feature point set, and feature vector extraction is performed on the feature points;

(3)对特征点集进行聚类,使用支持向量机SVM的方法将图像进行划分为三个区域,记为区域Sb,Su和Sf,并进行特征点的匹配;(3) Clustering the feature point set, using the support vector machine SVM method to divide the image into three regions, denoted as regions S b , Su and S f , and matching the feature points;

(4)分别对Sb和Sf中的特征点使用Random Sample Consensus(RANSAC)方法来剔除误匹配点,并计算出单应矩阵,分别记为Hb、Hf(4) The Random Sample Consensus (RANSAC) method is used for the feature points in S b and S f respectively to eliminate the mismatched points, and the homography matrix is calculated, which are denoted as H b and H f respectively;

(5)对采样获得的前幅两图像进行配准,配准过程中,对三个区域Sb,Su和Sf中的像素采用混合投影变换模型进行第二幅图像到第一幅图像的配准;(5) The first two images obtained by sampling are registered. During the registration process, the pixels in the three regions S b , S u and S f are processed by the hybrid projection transformation model from the second image to the first image. registration;

(6)通过单应矩阵的迭代来实现第三幅到第二幅图像的配准,重复步骤(6),进而实现后续图像序列的配准,最后进行图像的融合,完成整个全景图的拼接任务。(6) Realize the registration of the third image to the second image through the iteration of the homography matrix, repeat step (6), and then realize the registration of subsequent image sequences, and finally perform image fusion to complete the stitching of the entire panorama Task.

本发明与现有技术相比的有益效果在于:The beneficial effects of the present invention compared with the prior art are:

(1)在使用点聚类方法后,本发明采用SVM方法来实现图像不同区域的划分,为后续的图像配准过程做准备;(1) After using the point clustering method, the present invention adopts the SVM method to realize the division of different areas of the image, so as to prepare for the subsequent image registration process;

(2)在图像配准过程中,使用了创新的混合投影变换模型,可以实现相对于单投影矩阵更加好的拼接效果;(2) In the process of image registration, an innovative hybrid projection transformation model is used, which can achieve a better stitching effect than a single projection matrix;

(3)在图像配准过程中,使用了创新的混合投影变换模型,不仅比传统的DHW等混合投影变换模型更贴近真实情况,而且还可以大大减少配准时间,使整个拼接过程更加高效。(3) In the process of image registration, an innovative hybrid projection transformation model is used, which is not only closer to the real situation than the traditional DHW and other hybrid projection transformation models, but also can greatly reduce the registration time and make the entire stitching process more efficient.

附图说明Description of drawings

图1为SVM方法分割图像为三个区域示意图;Fig. 1 is the schematic diagram that the SVM method divides the image into three regions;

图2本发明方法流程框图;Fig. 2 method flow chart of the present invention;

图3本发明实验结果图。Figure 3 is a graph of the experimental results of the present invention.

具体实施方式Detailed ways

下面结合附图及实施例对本发明进行详细说明。The present invention will be described in detail below with reference to the accompanying drawings and embodiments.

本发明通过传统的单反相机拍摄获取视频序列,在Matlab R2016b下实现视频全景图的拼接。本发明的流程图见图2所示,为了更好地理解本发明的技术方案,以下对本发明的具体实施方式作进一步描述:The present invention obtains video sequences by shooting with a traditional single-lens reflex camera, and realizes the stitching of video panoramas under Matlab R2016b. The flow chart of the present invention is shown in Figure 2. In order to better understand the technical scheme of the present invention, the specific embodiments of the present invention are further described below:

步骤一:读入视频序列,使用Matlab对获取的视频序列进行降采样,以减少图像配准的运算量。采样时,只需保证相邻采样图像中有约1/3的重叠区域即可。Step 1: Read in the video sequence, and use Matlab to downsample the acquired video sequence to reduce the computational complexity of image registration. When sampling, it is only necessary to ensure that there is about 1/3 of the overlapping area in the adjacent sampled images.

步骤二:本发明采用Sift算子对相邻两幅图像I1和I2进行特征点的检测以及Sift特征向量的计算。Step 2: The present invention uses the Sift operator to detect the feature points of the two adjacent images I 1 and I 2 and calculate the Sift feature vector.

步骤三:将二维图像平面当作二维特征空间,图像像素的u、v值分别代表不同的特征值,对获取的特征点集使用K-means方法进行聚类,其中K=2;接着,本发明采用支持向量机(SVM)的方法来实现对图像平面的划分,如图1所示,其中每个小圆圈代表一个Sift特征点。与支持向量相切的两条直线方程可分别记为:Step 3: The two-dimensional image plane is regarded as a two-dimensional feature space, and the u and v values of the image pixels represent different eigenvalues respectively, and the K-means method is used to cluster the obtained feature point set, where K=2; then , the present invention adopts the support vector machine (SVM) method to realize the division of the image plane, as shown in FIG. 1 , wherein each small circle represents a Sift feature point. The equations of the two straight lines tangent to the support vector can be written as:

wTx+b1=0 (1)w T x+b 1 =0 (1)

wTx+b2=0 (2)w T x+b 2 =0 (2)

其中,x=(u,v)Twhere x=(u,v) T .

接着,将满足wTx+b1<0的像素点区域记为Sb;将满足wTx+b2>0的像素点区域记为Sf;将满足wTx+b1>0且wTx+b2<0的像素点区域记为SnNext, the pixel point region satisfying w T x+b 1 <0 is denoted as S b ; the pixel point region satisfying w T x+b 2 >0 is denoted as S f ; the pixel point region satisfying w T x+b 1 >0 is denoted as S f ; And the pixel area where w T x+b 2 <0 is denoted as Sn .

对区域Sb和Sf中的Sift特征点进行匹配,匹配过程中,为了提高特征点的匹配准确度,定义I1中区域Sb(Sf)中的任一待匹配点(记为p1)与I2中区域Sb(Sf)中的特征点集中最近的特征点距离为d1,与I2中区域Sb(Sf)中的特征点集中次近的特征点的距离为d2,则距离比记为:Match the Sift feature points in the regions S b and S f . During the matching process, in order to improve the matching accuracy of the feature points, define any point to be matched in the region S b (S f ) in I 1 (denoted as p 1 ) The distance from the closest feature point in the feature point set in the area S b (S f ) in I 2 is d 1 , and the distance from the next closest feature point in the feature point set in the area S b (S f ) in I 2 is d 2 , then the distance ratio is recorded as:

如果Ratio>ε(ε是经验阈值,本发明经过大量反复试验确定为:5≤ε),则认为点p1找到了匹配点。If Ratio>ε (ε is an empirical threshold, the present invention determines after a lot of trial and error: 5≦ε), it is considered that point p 1 has found a matching point.

遍历I1中所有的Sift特征点,便得到了初始匹配点对集,分别记为Pb和PfBy traversing all the Sift feature points in I 1 , an initial set of matching point pairs is obtained, denoted as P b and P f respectively.

步骤四:用RANSAC方法分别对区域Sb和Sf进行单应矩阵H的计算。对于两幅二维图像,对应点之间满足单应变换(或投影变换),如(4)所示,其中(x,y,1)T和(x',y',1)T是空间点X在两幅图像上的投影齐次坐标值。对于区域Sb,每次随机从匹配点对集Pb中选择4组匹配点对计算Hb,并统计Pb中满足(4)的点个数比例,记为Inlier_Ratio。最后选择使得Inlier_Ratio达到最大的单应矩阵记为最终的HbStep 4: Use the RANSAC method to calculate the homography matrix H for the regions S b and S f respectively. For two two-dimensional images, the homography transformation (or projective transformation) is satisfied between the corresponding points, as shown in (4), where (x, y, 1) T and (x', y', 1) T are the space The projected homogeneous coordinates of point X on the two images. For the region S b , 4 sets of matching point pairs are randomly selected from the matching point pair set P b each time to calculate H b , and the ratio of the number of points satisfying (4) in P b is counted, denoted as Inlier_Ratio. Finally, the homography matrix that maximizes Inlier_Ratio is selected as the final H b .

其中,in,

同理,可以计算出Sf区域对应的单应矩阵记为Hf Similarly, the homography matrix corresponding to the S f region can be calculated as H f

步骤五:在得到了两个单应矩阵后,便可对图像进行配准。对于三个不同区域,本发明采用三种配准模型。对于区域Sb中的像素点,可以认为它们均属于Sb所在的那个平面,则这些像素点在配准时使用单应矩阵Hb,如(6)所示,其中(x,y,1)T为I2中某像素点的齐次坐标,(x',y',1)T为该点通过投影变换到I1中对应的点坐标。Step 5: After two homography matrices are obtained, the images can be registered. For the three different regions, the present invention employs three registration models. For the pixels in the area S b , it can be considered that they all belong to the plane where S b is located, then these pixels are registered using the homography matrix H b , as shown in (6), where (x, y, 1) T is the homogeneous coordinate of a pixel in I 2 , and (x', y', 1) T is the point that is transformed to the corresponding point coordinate in I 1 by projection.

对于区域Sf中的像素点,同理可使用(7)所示变换将点进行投影变换。For the pixel points in the area S f , similarly, the transformation shown in (7) can be used to perform projective transformation on the points.

对于Su中的像素点,本发明将它们称为“Undefined pixels”,即无法准确判断出这些点属于哪一个平面。针对此问题,本发明提出以下方法解决。For the pixels in Su, the present invention calls them " Undefined pixels", that is, it is impossible to accurately determine which plane these points belong to. In view of this problem, the present invention proposes the following methods to solve it.

(1)对于所有Su中的点,定义单应矩阵:(1) For all points in Su, define the homography matrix:

Hij=wbHb+wfHf (8)H ij = w b H b +w f H f (8)

其中Hij表示(i,j)处的像素对应的单应矩阵,Hb、Hf分别表示图像两个区域所对应的单应矩阵,wb和wf分别为这两个平面对应单应矩阵的权重系数。where H ij represents the homography matrix corresponding to the pixel at (i, j), H b and H f represent the homography matrix corresponding to the two regions of the image respectively, w b and w f are the homography corresponding to the two planes, respectively Weight coefficients for the matrix.

(2)计算每一个Su中像素点到区域Sb、Sf的特征点集中最近的特征点的距离,分别记为db、df,本发明提出在距离计算过程中使用曼哈顿距离来代替传统的欧式距离,这样可以很大程度的节约时间,权重计算公式如下:(2) Calculate the distance from the pixel point in each Su to the nearest feature point in the feature point set of the regions S b and S f , which are denoted as db and d f respectively . The present invention proposes to use the Manhattan distance in the process of distance calculation. Instead of the traditional Euclidean distance, this can save time to a great extent. The weight calculation formula is as follows:

wb=df/(db+df) (9)w b =d f /(d b +d f ) (9)

wf=db/(db+df) (10)w f =d b /(d b +d f ) (10)

综上所述,可以得到每个像素点的混合投影变换模型:In summary, the mixed projection transformation model of each pixel can be obtained:

其中,background表示背景区域,undefined area表示不明确区域,foreground表示前景区域。Among them, background represents the background area, undefined area represents the ambiguous area, and foreground represents the foreground area.

经过上述步骤,便实现了两幅图像的配准。After the above steps, the registration of the two images is achieved.

步骤六:经过上述五个步骤,实现了两幅图像序列的配准工作,接着,对第三幅I3以及后续的图像序列来实现配准拼接。本发明采用单应阵迭代的方法来实现此目的。为了方便清楚说明,本步骤将Hij的下标省去,记作H,Step 6: After the above five steps, the registration work of the two image sequences is realized, and then, the registration and splicing are realized for the third image I 3 and the subsequent image sequences. The present invention adopts the homography matrix iteration method to achieve this purpose. For the convenience of clear description, the subscript of H ij is omitted in this step, and it is denoted as H,

(1)对于I2和I3重叠的区域,可以使用式(12)直接进行迭代投影变换,(1) For the region where I 2 and I 3 overlap, Equation (12) can be used to directly perform iterative projection transformation,

H3→1=H2→1(H3→2(I3)) (12)H 3→1 =H 2→1 (H 3→2 (I 3 )) (12)

(2)对于I3中的I2和I3非重叠区域,可以使用式(13)进行加权迭代投影变换(2) For the non-overlapping regions of I 2 and I 3 in I 3 , the weighted iterative projection transformation can be performed using Eq. (13)

其中,本发明将K定义为I2和I3的重叠区域中的Sift特征点集,定义权重λq为λq=1/d_Manh,其中d_Manh表示H3→2(p)与q之间的曼哈顿距离。再对λq进行归一化处理,使得∑λq=1,Wherein, the present invention defines K as the Sift feature point set in the overlapping area of I 2 and I 3 , and defines the weight λ q as λ q =1/d_Manh, where d_Manh represents the relationship between H 3→2 (p) and q Manhattan distance. Then normalize λ q so that ∑λ q =1,

综上可得:In summary:

经过上述步骤可以计算出第三幅图像相对于第一幅图象的投影矩阵,然后进行投影变换,这样便实现了将第三幅图像的配准工作。对于后续图像序列,可循环迭代使用该方式。After the above steps, the projection matrix of the third image relative to the first image can be calculated, and then the projection transformation is performed, thus realizing the registration of the third image. This method can be used iteratively for subsequent image sequences.

最后,对配准的图像进行融合,本发明采用传统的渐入渐出融合方法(即线形插值融合算法),使用线形加权函数来实现在靠近重叠区域边界的地方做平滑过渡处理,加权函数如下式所示:Finally, the registered images are fused. The present invention adopts the traditional fade-in and fade-out fusion method (ie, linear interpolation fusion algorithm), and uses a linear weighting function to achieve smooth transition processing near the boundary of the overlapping area. The weighting function is as follows The formula shows:

其中I1和I2表示相邻两幅图像, where I 1 and I 2 represent two adjacent images,

本发明的有效性和准确性已经通过实验进行了验证,取得了很好的拼接结果。实验结果如图3所示,(a)表示视频序列中的四幅图像,(b)表示由这四幅图像计算得到的全景图。本发明的最大优势在于使用SVM方法对图像进行区域划分,并以此为基础,分区域进行图像的投影变换(单应变换),有效地避免了传统的单投影阵方法中误差大、多投影阵方法中费时的问题,在虚拟现实等技术领域中有着很好的前景。The validity and accuracy of the present invention have been verified through experiments, and good splicing results have been obtained. The experimental results are shown in Figure 3, where (a) represents the four images in the video sequence, and (b) represents the panorama calculated from these four images. The biggest advantage of the present invention is that the SVM method is used to divide the image into regions, and based on this, the projection transformation (homography transformation) of the image is performed in different regions, which effectively avoids the large error and multiple projections in the traditional single projection array method. The time-consuming problem in the matrix method has a good prospect in the technical fields such as virtual reality.

Claims (3)

1.一种基于混合投影变换模型的大视场视频全景图拼接的方法,其特征在于:包括以下步骤:1. a method for stitching a large field of view video panorama based on a hybrid projection transformation model, is characterized in that: comprise the following steps: (1)对视频图像序列降采样,保证每两幅相邻图像间最少有1/3的重叠区域;(1) Downsampling the video image sequence to ensure that there is at least 1/3 of the overlapping area between every two adjacent images; (2)对采样获得的前两幅图像的重叠区域进行Sift特征点的检测,获得的所述特征点集,并对特征点进行特征向量的提取;(2) The detection of Sift feature points is performed on the overlapping area of the first two images obtained by sampling, the obtained feature point set, and feature vector extraction is performed on the feature points; (3)对特征点集进行聚类,使用支持向量机SVM的方法将图像进行划分为三个区域,记为区域Sb,Su和Sf,并进行I1和I2中对应区域Sb和Sf特征点的匹配;(3) Cluster the feature point set, use the support vector machine SVM method to divide the image into three regions, denoted as regions S b , S u and S f , and perform the corresponding regions S in I 1 and I 2 The matching of b and S f feature points; (4)分别对Sb和Sf中的特征点使用Random Sample Consensus(RANSAC)方法来剔除误匹配点,并计算出两幅图像之间二维坐标的关系,即单应矩阵,分别记为Hb、Hf(4) The Random Sample Consensus (RANSAC) method is used for the feature points in S b and S f respectively to remove the mismatched points, and the relationship between the two-dimensional coordinates between the two images is calculated, that is, the homography matrix, which is recorded as H b , H f ; (5)对采样获得的前幅两图像进行配准,配准过程中,对三个区域Sb,Su和Sf中的像素采用混合投影变换模型进行第二幅图像到第一幅图像的配准;(5) The first two images obtained by sampling are registered. During the registration process, the pixels in the three regions S b , S u and S f are processed by the hybrid projection transformation model from the second image to the first image. registration; (6)通过单应矩阵的迭代来实现第三幅到第一幅图像的配准,重复步骤(6),进而实现后续图像序列的配准,最后进行图像的融合,完成整个全景图的拼接任务;(6) Realize the registration of the third image to the first image through the iteration of the homography matrix, repeat step (6), and then realize the registration of subsequent image sequences, and finally perform image fusion to complete the stitching of the entire panorama Task; 所述步骤(5)中,混合投影变换模型实现如下:In the step (5), the hybrid projection transformation model is implemented as follows: (51)对于区域Sb中的像素点,在配准时使用单应矩阵Hb,其中I1和I2分别代表第一副和第二副图像,(x,y,1)T为I2中某像素点的齐次坐标,(x',y',1)T为该点通过投影变换到I1中对应的点坐标,(51) For the pixels in the area S b , use the homography matrix H b during registration, where I 1 and I 2 represent the first and second images, respectively, (x, y, 1) T is I 2 The homogeneous coordinates of a pixel in I 1 , (x', y', 1) T is the corresponding point coordinate of the point transformed to I 1 by projection, (52)对于区域Sf中的像素点,使用(4)所示的变换将点进行投影变换;(52) For the pixel points in the area S f , use the transformation shown in (4) to perform projection transformation on the points; (53)对于Su中的像素点,定义单应矩阵:(53) For the pixels in Su, define the homography matrix: Hij=wbHb+wfHf (5)H ij = w b H b +w f H f (5) 其中Hij表示(i,j)处的像素对应的单应矩阵,Hb、Hf分别表示图像两个区域Sb、Sf所对应的单应矩阵,wb和wf分别为两个图像区域Sb、Sf对应单应矩阵的权重系数;Wherein H ij represents the homography matrix corresponding to the pixel at (i, j), H b and H f represent the homography matrix corresponding to the two image regions S b and S f respectively, w b and w f are two The weight coefficients of the image regions S b and S f corresponding to the homography matrix; 计算Su中每一个像素点到区域Sb、Sf的特征点集中最近的特征点的曼哈顿距离,分别记为db、df,则权重计算公式如下:Calculate the Manhattan distance from each pixel in S u to the nearest feature point in the feature point set of regions S b and S f , denoted as db and d f respectively , and the weight calculation formula is as follows: wb=df/(db+df) (6)w b =d f /(d b +d f ) (6) wf=db/(db+df) (7)w f =d b /(d b +d f ) (7) 混合投影模型表示为:The hybrid projection model is expressed as: 其中,foreground为前景区域,即Sf,undefined area为不明确区域,即Su,background为背景区域,即SbWherein, foreground is a foreground area, namely S f , undefined area is an ambiguous area, namely Su, and background is a background area, namely S b . 2.根据权利要求1所述的一种基于混合投影变换模型的大视场视频全景图拼接的方法,其特征在于:步骤(3)中,对图像进行区域划分时,2. a kind of method based on the large field of view video panorama stitching of hybrid projection transformation model according to claim 1, is characterized in that: in step (3), when image is carried out area division, (1)在使用K-means方法进行特征点聚类后,使用SVM方法对图像进行划分,与支持向量分别相切的两条直线方程为:(1) After using the K-means method for feature point clustering, the SVM method is used to divide the image, and the two straight line equations that are tangent to the support vector are: wTx+b1=0 (1)w T x+b 1 =0 (1) wTx+b2=0 (2)w T x+b 2 =0 (2) 其中,x=(u,v)T,w、b1和b2表示直线方程的系数,u代表像素点的横坐标,v代表像素点的纵坐标;Among them, x=(u,v) T , w, b 1 and b 2 represent the coefficients of the straight line equation, u represents the abscissa of the pixel, and v represents the ordinate of the pixel; (2)将满足wTx+b1<0的像素点区域记为Sb(2) Denote the pixel area that satisfies w T x+b 1 <0 as S b ; 将满足wTx+b2>0的像素点区域记为SfDenote the pixel area that satisfies w T x+b 2 >0 as S f ; 将满足wTx+b1>0且wTx+b2<0的像素点区域记为Su实现图像区域的划分。The pixel area that satisfies w T x+b 1 >0 and w T x+b 2 <0 is denoted as Su to realize the division of the image area. 3.根据权利要求1所述的一种基于混合投影变换模型的大视场视频全景图拼接的方法,其特征在于:所述步骤(6)中,第三幅及后续图像序列的配准如下:3. a kind of method based on the large field of view video panorama stitching of hybrid projection transformation model according to claim 1, is characterized in that: in described step (6), the registration of the 3rd and subsequent image sequence is as follows : 在I2和I3的非重叠区域中,使用加权单应矩阵迭代的方法,来实现图像的配准,如下:In the non-overlapping regions of I 2 and I 3 , the method of weighted homography matrix iteration is used to achieve image registration, as follows: 其中,将K定义为I2和I3的重叠区域中的Sift特征点集,定义权重λq为λq=1/d_Manh,其中d_Manh表示H3→2(p)与Sift特征点q之间的曼哈顿距离,再对λq进行归一化处理,使得∑λq=1。Among them, K is defined as the set of Sift feature points in the overlapping area of I 2 and I 3 , and the weight λ q is defined as λ q =1/d_Manh, where d_Manh represents the distance between H 3→2 (p) and the Sift feature point q The Manhattan distance of , and then normalize λ q to make Σλ q =1.
CN201711419418.6A 2017-12-25 2017-12-25 A method for large-field video panorama stitching based on hybrid projection transformation model Active CN108093188B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201711419418.6A CN108093188B (en) 2017-12-25 2017-12-25 A method for large-field video panorama stitching based on hybrid projection transformation model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201711419418.6A CN108093188B (en) 2017-12-25 2017-12-25 A method for large-field video panorama stitching based on hybrid projection transformation model

Publications (2)

Publication Number Publication Date
CN108093188A CN108093188A (en) 2018-05-29
CN108093188B true CN108093188B (en) 2019-01-25

Family

ID=62179005

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201711419418.6A Active CN108093188B (en) 2017-12-25 2017-12-25 A method for large-field video panorama stitching based on hybrid projection transformation model

Country Status (1)

Country Link
CN (1) CN108093188B (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109741289B (en) * 2019-01-25 2021-12-21 京东方科技集团股份有限公司 Image fusion method and VR equipment
CN110544202B (en) * 2019-05-13 2022-06-07 燕山大学 A method and system for disparity image stitching based on template matching and feature clustering
CN110223250B (en) * 2019-06-02 2021-11-30 西安电子科技大学 SAR geometric correction method based on homography transformation
CN113593426B (en) * 2021-09-28 2021-12-03 卡莱特云科技股份有限公司 Module division method for LED spliced screen image and LED screen correction method

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105678687A (en) * 2015-12-29 2016-06-15 天津大学 Stereo image stitching method based on content of images
CN107423008A (en) * 2017-03-10 2017-12-01 北京市中视典数字科技有限公司 A kind of multi-cam picture fusion method and scene display device in real time

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106886979B (en) * 2017-03-30 2020-10-20 深圳市未来媒体技术研究院 Image splicing device and image splicing method
CN107154022B (en) * 2017-05-10 2019-08-27 北京理工大学 A dynamic panorama stitching method suitable for trailers

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105678687A (en) * 2015-12-29 2016-06-15 天津大学 Stereo image stitching method based on content of images
CN107423008A (en) * 2017-03-10 2017-12-01 北京市中视典数字科技有限公司 A kind of multi-cam picture fusion method and scene display device in real time

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Constructing Image Panoramas using Dual-Homography Warping;JunHong Gao et al.;《CVPR 2011》;20110822;全文

Also Published As

Publication number Publication date
CN108093188A (en) 2018-05-29

Similar Documents

Publication Publication Date Title
Liu et al. Content-preserving warps for 3D video stabilization
Lin et al. Dynamic spatial propagation network for depth completion
Yao et al. Recurrent mvsnet for high-resolution multi-view stereo depth inference
CN104574311B (en) Image processing method and device
Adel et al. Image stitching based on feature extraction techniques: a survey
CN105245841B (en) A kind of panoramic video monitoring system based on CUDA
KR101175097B1 (en) Panorama image generating method
CN108093188B (en) A method for large-field video panorama stitching based on hybrid projection transformation model
CN106331480B (en) Video image stabilization method based on image splicing
CN112288628B (en) Aerial image splicing acceleration method and system based on optical flow tracking and frame extraction mapping
CN107274337B (en) Image splicing method based on improved optical flow
CN110580720B (en) A Camera Pose Estimation Method Based on Panorama
CN110753181A (en) A Video Stabilization Method Based on Feature Tracking and Grid Path Motion
CN109215118B (en) Incremental motion structure recovery optimization method based on image sequence
CN106101535A (en) A kind of based on local and the video stabilizing method of mass motion disparity compensation
KR20210133472A (en) Method of merging images and data processing device performing the same
CN111553841A (en) A Real-time Video Stitching Algorithm Based on Optimal Seam Update
CN109819158B (en) Video stabilization method based on light field imaging
GB2567245A (en) Methods and apparatuses for depth rectification processing
CN105282400B (en) A kind of efficient video antihunt means based on geometry interpolation
Chan et al. An object-based approach to image/video-based synthesis and processing for 3-D and multiview televisions
CN104036496B (en) Self-calibration method for radial distortion of fish-eye lens camera
CN108280805A (en) A kind of image split-joint method based on manifold optimization
Chiu et al. 360mvsnet: Deep multi-view stereo network with 360deg images for indoor scene reconstruction
Manda et al. Image stitching using ransac and bayesian refinement

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant