Next Article in Journal
GCM-SIV1.5: Optimal Tradeoff between GCM-SIV1 and GCM-SIV2
Previous Article in Journal
Partial Autocorrelation Diagnostics for Count Time Series
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Feature Correspondences Increase and Hybrid Terms Optimization Warp for Image Stitching

1
Key Laboratory of Symbolic Computation and Knowledge Engineering of Ministry of Education, School of Artificial Intelligence, Jilin University, Changchun 130012, China
2
College of Computer Science and Technology, Jilin University, Changchun 130012, China
3
School of Mathematical and Computer Sciences, Heriot-Watt University, Edinburgh EH14 4AS, UK
*
Author to whom correspondence should be addressed.
Entropy 2023, 25(1), 106; https://doi.org/10.3390/e25010106
Submission received: 12 November 2022 / Revised: 19 December 2022 / Accepted: 19 December 2022 / Published: 4 January 2023
(This article belongs to the Topic Recent Trends in Image Processing and Pattern Recognition)

Abstract

:
Feature detection and correct matching are the basis of the image stitching process. Whether the matching is correct and the number of matches directly affect the quality of the final stitching results. At present, almost all image stitching methods use SIFT+RANSAC pattern to extract and match feature points. However, it is difficult to obtain sufficient correct matching points in low-textured or repetitively-textured regions, resulting in insufficient matching points in the overlapping region, and this further leads to the warping model being estimated erroneously. In this paper, we propose a novel and flexible approach by increasing feature correspondences and optimizing hybrid terms. It can obtain sufficient correct feature correspondences in the overlapping region with low-textured or repetitively-textured areas to eliminate misalignment. When a weak texture and large parallax coexist in the overlapping region, the alignment and distortion often restrict each other and are difficult to balance. Accurate alignment is often accompanied by projection distortion and perspective distortion. Regarding this, we propose hybrid terms optimization warp, which combines global similarity transformations on the basis of initial global homography and estimates the optimal warping by adjusting various term parameters. By doing this, we can mitigate projection distortion and perspective distortion, while effectively balancing alignment and distortion. The experimental results demonstrate that the proposed method outperforms the state-of-the-art in accurate alignment on images with low-textured areas in the overlapping region, and the stitching results have less perspective and projection distortion.

1. Introduction

Image stitching is the process of merging a series of images with overlapping regions into a high-resolution image with a wider field of view [1]. It has been widely used in applications such as panorama [2], autonomous driving [3], video stabilization [4], virtual reality [5], and many others. Up till now, there are still great challenges for image stitching to align accurately, eliminate artifacts, and mitigate distortion, especially on low-textured and large-parallax images.
Traditional image stitching methods often use feature points to obtain a global homography matrix, and they map images with overlapping regions to the same plane. However, these approaches only work if the images are in the same plane or only rotated around the camera center. Under the condition of both parallax and rotation, global homography often produces misalignment and/or artifacts in the overlapping region and distortion in the non-overlapping region. Recently, spatially varying warping [6,7,8] has been widely used in image stitching. It divides an image into several grid cells, and then aligns each grid cell with local homography. It can effectively eliminate misalignment, but it may also produce distortions in the non-overlapping region. Since homography-based methods inevitably generate distortions, researchers have taken a different approach by using a seam-driven strategy [9,10,11], which finds the optimal seam according to a customized energy function. Although such an approach can effectively mitigate distortions, it usually cannot achieve the desired alignments. In addition, such a method relies too much on the energy function, and it is more frequently used as a post-processing blending method to eliminate artifacts. In recent years, deep learning-based image stitching frameworks [12,13,14] have been widely studied by researchers. However, they may still not be able to achieve the desired accurate alignment and distortion preservation performance compared to traditional methods.
Feature detection and correct matching are the basis of the image stitching process. SIFT [15] is a feature point detection and matching method used by almost all image stitching methods. After preliminary matching using SIFT, RANSA C [16] is used to remove outliers and obtain inliers. Although SIFT can extract a large number of feature points, the correct matching points after RANSAC iteration are greatly reduced compared with the detected feature points, which may lead to insufficient feature correspondence in the low-textured or repetitively-textured regions. Figure 1 shows the feature correspondences of SIFT+RANSAC under three different image types. It can be seen that the traditional SIFT+RANSAC pattern has sufficient inliers only under the condition of dense texture in the overlapping region, while obviously insufficient inliers in low-textured areas (see Figure 1a,c).
Grid-based image stitching methods are heavily dependent on the quantity and quality of feature points, which are crucial to the quality of image stitching results. The as-projective-as-possible (APAP) warp model [8] is a typical mesh-based local warping model. It uses SIFT+RANSAC as pre-processing. Figure 2a shows the inliers of APAP and the stitching result. The input images are with low-textured or repetitively-textured regions, which lead to insufficient inliers after RANSAC. It further causes misalignment of the stitching result. RANSAC can increase the number of inliers by increasing the threshold. Figure 2b shows the inliers and the stitching result after increasing the RANSAC threshold appropriately. Obviously, increasing the number of feature correspondences produces a more accurate alignment. However, limited by the local alignment of the APAP method, some unacceptable jagged misalignments appear unnaturally. Furthermore, limited by the characteristics of the RANSAC algorithm itself, when the RANSAC threshold is too high, it may still generate false matches if the overlapping region has many similar textures.
Considering that there may still be the issue of matching errors after RANSAC, elastic local alignment (ELA) [19] introduces Bayesian feature refinement to further remove possible outliers. Although outliers hardly exist after using ELA, ELA also removes some of the original normal values, and this may lead to the issue of insufficient feature points. Leveraging line-point consistence (LPC) [20] establishes a relationship between point and line dual features, and it applies a characteristic number (CN) [21] to add salient points to the feature point set to enrich the number of feature points. However, the feature points added by CN may not be correct matches, and they need to go through RANSAC, which will also make a low-textured region with insufficient feature points (see Figure 2c). Although some methods [18,22,23] that introduce line features and use LSD [24] to extract line segment features can cope well with a low-textured environment, line segments are mainly matched by endpoints, which are essentially point features, and they are subject to line segment length thresholds. Currently, line segment matching is still a challenging problem.
Grid-based Motion Statistics (GMS) [25] assume that adjacent pixels share similar motions, which is consistent with the principle of image stitching. It uses ORB [26] to extract feature points. Since ORB can set the initial feature point threshold, combined with GMS, it can increase the number of correct matching points in low-textured regions, thereby ensuring that the warped model has sufficient anchor points for local alignment. Compared with the SIFT+RANSAC pattern, the ORB+GMS method can not only make feature correspondences sufficient, but also ensure the correctness of the feature matching, which helps improve the quality of the stitching. The results of the proposed method are shown in Figure 2d.
In this paper, we propose a novel image stitching method based on feature correspondences increase and hybrid terms optimization warp. We first use the ORB+GMS algorithm to extract and match feature points to obtain sufficient feature correspondences, and at the same time we use line features to assist alignment and structure preservation; then, an inital homography combining global homography transformation and global similarity transformation is estimated as the initial homography to be optimized. On this basis, we estimate various optimization terms for the final image warping. Finally, we obtain the final stitching result by linear blending or seam-cutting blending. The overall process architecture of our method is shown in Figure 3.
The contributions of our approach are as follows:
  • A novel method to increase feature correspondences is proposed, which can be added to any image stitching model. It solves the problem of insufficient feature correspondence in low-textured or repetitively-textured regions, and it can effectively eliminate misalignment and artifacts.
  • A novel hybrid transformation combining global homography transformation and global similarity transformation is proposed to serve as the initial homography for structure preservation, which can flexibly fine-tune the structure of image stitching results compared to conventional global homography.
  • Various optimization terms are used to locally adjust the above hybrid warping model, which can effectively mitigate projection and perspective distortion. Our flexible and robust method can effectively balance alignment and distortion, especially on images with low-textured areas in the overlapping region. For images with large parallax or significant foreground-background relationship, seam-cutting blending instead of linear blending is used to eliminate inevitable misalignment or artifacts.

2. Related Work

2.1. Feature Extraction and Matching

Feature points refer to the points where the gray values of the image change drastically or the points with large curvature on the edges of the image, such as corners and spots. In the early days, common corner detection algorithms were Harris corner [27] and FAST [28]. Due to the feature limitations of corner points, Lowe et al. proposed SIFT [15], which is invariant to scale, rotation, and brightness. It is robust and fast, and it has been the most popular feature detector until now. SURF [29] improves the speed compared with SIFT, which is more suitable for practical applications. BRIEF [30] uses the Hamming Distance to achieve matching using the XOR operation between bits. ORB [26] combines the advantages of FAST and BRIEF. Compared with SIFT/SURF, it can detect denser matching points in areas with flat textures, and it has the best comprehensive performance among the current traditional feature detectors.
After feature detection, the Brute Force (BF) [31] algorithm is often used for preliminary matching, but there are often some false matches. In recent years, RANSAC [16] has been the most popular method for filtering out outliers. However, it is affected by the threshold of inliers; especially for locally aligned warping models, inliers are often mistakenly eliminated by RANSAC, resulting in an insufficient number of points in flat textured areas, and the rest cannot be guaranteed to be inliers, resulting in wrong stitching results. Some researchers have taken corresponding countermeasures to the problem of RANSAC. The triangular facet transformation (TFT) [32] uses bundle adjustment (BA) [33] to optimize for false matches that may be due to noise. LPC [21] uses the CN [20] point-line relationship to expand the matching point set, but the two-step RANSAC may lead to the loss of feature correspondences in the weakly-textured region, thus affecting the stitching results. Bian et al. proposed GMS [25], which assumes that adjacent pixels have similar motion and distinguishes between correct and incorrect matches by computing statistical scores within a small area around the matching point. It can be combined with ORB to increase the number of features by increasing the matching point threshold to help the matching quality, which makes it possible to solve challenging image stitching problems. Therefore, compared to RANSAC, GMS is more robust to address feature matching errors and insufficient features in low-texture regions.

2.2. Spatially Varying Warping and Seam Cutting

Traditional image stitching uses a global homography [2] to warp the image in the ideal case where the scene depth has little change in overlapping regions. In recent years, researchers have used multi-homography spatially varying warping to deal with parallax. The dual-homography warping (DHW) [34] divides the scene into a background plane and a foreground plane, and it aligns the background and foreground with two homography matrices, respectively. Smoothly varying affine (SVA) [7] improves local transformation and alignment with multiple affine transformations. APAP [8] divides the image into uniform grid cells, and it performs local homography alignment for each grid, which can better cope with parallax scenes. However, grid-based spatially varying warping often produces severe distortion in non-overlapping regions. The adaptive as-natural-as-possible (AANAP) warp [17] combines local homography transformation with global similarity transformation to effectively reduce distortion in non-overlapping regions. The shape-preserving half projective (SPHP) warp [35] uses subregional warping to smoothly transition overlapping and non-overlapping regions to reduce distortion.
Due to the difference in parallax and image capture time, artifacts inevitably appear in the overlapping areas of the stitching results. Gao et al. proposed a seam-driven [9] approach to deal with parallax and eliminate artifacts. Parallax-tolerant Image Stitching [10] is inspired by content-preserving warps (CPW) [36] and optimizes the use of seams to improve the stitching performance of large parallax scenes. Seam-guided local alignment (SEAGULL) [11] adds curved and straight structural protection constraints. Perception-based seam cutting for image stitching [37] introduces human perception and saliency detection to obtain the optimal seam, making the stitching result look more natural. Xue et al. proposed a stable seam measurement [38], which is used to estimate a stable seam cutting for mitigating parallax via a hybrid actor–critic architecture.

2.3. Structure Preservation and Distortion Mitigation

Lines are common features in natural images. Line structures bend unnaturally after image warping. The dual-feature warp (DFW) [18] introduces the line segment feature, and the constraint term constrains both points and lines to achieve better alignment. The global similarity prior (GSP) [39] constrains globally similar angles with line alignment, while local and global constraints are combined to reduce structural distortion. The quasi-homography warp (QHW) [40] uses intersecting lines to preserve the nature of homography warping. The single-perspective warp (SPW) [22] utilizes the QHW characteristic; it introduces a variety of protection terms and uses the least square method to obtain the optimal solution of the objective function to obtain the optimal homography. LPC [21] introduces the concept of global lines on the basis of SPW, and it merges short line segments into long ones to reduce the bending of line structures in natural images. Similarity transformation has been proved in [34] to reduce distortion in the overlapping region, and the reason being that similarity transformation are more rectangular and look more natural than homography. Analogously, the work in [41] introduces line and regular boundary preservation constraints to make the boundary of the stitching result as close as possible to a rectangle, while reducing unwanted distortions.
The remainder of this paper is organized as follows. Section 3 presents the proposed image stitching algorithm. The detailed experimental procedure and results are presented in Section 4. Finally, the conclusions are summarized in Section 5.

3. The Proposed Method

The homography transformation is the most common method to achieve image stitching. For the single image warp, global homography warping can handle the cases in the ideal conditions. However, under low texture and large parallax conditions, it is still a big challenge to align accurately even with local homography. It is very dependent on the quality and quantity of features correspondences in the pre-processing stage. The traditional SIFT+RANSAC pattern has been unable to satisfy the above requirements. Meanwhile, the homography transformation inevitably introduces distortion in the non-overlapping region. How to balance alignment and distortion has gradually become the key to single image warping. In this section, we first describe our feature correspondences increase module, and then we propose our hybrid terms optimization warp to address the balance of alignment and distortion.

3.1. Feature Correspondences Increase

Inspired by GMS, we merge the feature correspondences increase module into the image stitching process to solve the problem of insufficient feature correspondences in low-textured areas in the overlapping region. Compared with the one-step RANSAC, we try to get more correct matches in low-textured or repetitively-textured regions to satisfy the local warping model’s demand for sufficient feature points in overlapping regions to achieve more accurate alignment. Here, we use ORB with high robustness and dense feature points as the detector, BF for preliminary matching, and GMS refinement to distinguish between correct and incorrect matches.
GMS considers that there are several features that match the matching relationship in the correctly matched neighborhoods, while there are almost no features in the incorrectly matched neighborhoods. According to this feature, GMS counts the number of features that match the matching relationship to distinguish between correct and incorrect matches. Let { N , M } be the number of feature points of the input image pair { I a , I b } after ORB+BF, respectively. X = { x 1 , x 2 , . . . , x n } is all the feature matching neighborhoods from image I a to I b . X can be classified as true or false by measuring the local support of each matching pair through GMS. X i X is the subset of all neighborhoods. S i is the neighborhood support, which is expressed as follows:
S i = | X i | 1 ,
where −1 removes the original match from the sum.
GMS assumes that if the motion is smooth in a region, the correct matches have the same spatial locations across multiple neighborhood pairs, while false matches have relative spatial locations across multiple neighborhood pairs different. Equation (1) can be rewritten as follows:
S i = k = 1 K | X a k b k | 1 ,
where K is the number of small neighborhoods predicted to move along with feature matching. { a k , b k } is the predicted region pair, X a k b k belonging to X is the feature matching subset.
Let T a b and F a b be the same and different positions of the region { a , b } , respectively. f a b is the nearest neighbor feature in region b to one of the features in region a. p t = p ( f a b | T a b ) and p f = p ( f a b | F a b ) . Assuming that each region pair { a , b } has { n , m } feature points, then:
p t = t + ( 1 t ) β m / M p f = β ( 1 t ) ( m / M ) ,
where β is the adjustment factor. Thus, S i be the number of matches in the neighborhood of x i , the distribution of it follows the binomial distribution in Equation (2):
S i B ( K n , p t ) , if   x i is   ture B ( K n , p f ) , if   x i is   false .
where K is the number of disjoint regions which match i predicts move together.
To calculate S i more efficiently, the image is divided into 20 × 20 grids as in [25]. The score S i j for the cell pair { i , j } is calculated as follows:
S i j = k = 1 K = 9 | X i k j k | ,
where | X i k j k | is the number of feature matches in the nine-square grid. The score threshold τ is used to distinguish whether the feature matching is correct. If  S i j > τ , the matching x i at grid { i , j } is true; otherwise, x i is false.
In this way, we obtain a denser set of feature correspondences compared with the traditional SIFT+RANSAC pattern. It will be used as input for the subsequent warping.

3.2. Hybrid Terms Optimization Warp

3.2.1. Mathematical Preparation

Let X 1 = [ x 1 , y 1 , 1 ] T and X 2 = [ x 2 , y 2 , 1 ] T denote the feature points obtained by the feature correspondences increase in the input images I a and I b , respectively. The alignment of the overlapping region of the two images is a linear transformation from I a to I b in homogeneous coordinates, which is defined below:
X 2 = H X 1 ,
where H is defined as a 3 × 3 matrix. If it is a global homography transformation, then we have the following:
H g h = h 1 h 2 h 3 h 4 h 5 h 6 h 7 h 8 h 9 .
Similarly, if it is a global similarity transformation, we have the following:
H g s = s cos θ s sin θ t x s sin θ s cos θ t y 0 0 1 ,
where s and θ are the scale and angle of the transformation, respectively. t x and t y are the translation distances along the x and y directions, respectively.
In inhomogeneous coordinates:
x 2 = h 1 x 1 + h 2 y 1 + h 3 h 7 x 1 + h 8 y 1 + h 9 , y 2 = h 4 x 1 + h 5 y 1 + h 6 h 7 x 1 + h 8 y 1 + h 9 .
Next, we would like to solve the homography H g h and H g s . Even though these inhomogeneous equations involve the coordinates non-linearly, the coefficients of H appear linearly. Given N pairs of feature correspondences, we can form the following linear system of equations:
A h = 0 ,
where A R 2 N × 9 . It is obtained by transforming Equation (9). Equation (10) can be solved using homogeneous linear least square methods like Singular Value Decomposition (SVD). From SVD, the right singular vector that corresponds to the smallest singular value is the solution h , and then we reshape h into the matrix H g h . For similarity transformation, we filter out the similarity matrix corresponding to the rotation angle with the smallest angle according to the feature points as the global similarity H g s .
Global homography has difficulty producing an accurate alignment when the scenes are not coplanar. In contrast, APAP uses spatially varying warping H * , which expands the homography H into multiple local homographies applied to each grid cell, and it uses moving DLT for image alignment. This allows for better alignment in the overlapping region. Similar to the method in [42], assuming that the feature point p in the image I a can be represented by a linear combination of a vector composed of four vertices of the mesh V = [ v 1 , v 2 , v 3 , v 4 ] T . It can be expressed as: p = W V , and the bilinear interpolation weight of the four mesh vertices W = [ w 1 , w 2 , w 3 , w 4 ] T . Then, these mesh vertices are formed into the transformed mesh vertices V ^ = [ v 1 ^ , v 2 ^ , v 3 ^ , v 4 ^ ] T after mesh deformation. Therefore, the deformed vertex coordinate p ^ is calculated as the bilinear interpolation as follows:
p ^ = w 1 v 1 + w 2 v 2 + w 3 v 3 + w 4 v 4 .
As a result, we convert point correspondence constraints to mesh vertex correspondence constraints.
After the above preparation, the energy function can be defined as:
E = E a + E d + E s ,
where E a addresses the alignment issue by feature correspondences increase model to increase the number of correct matching (see Section 3.2.2), E d addresses the distortion issue by cross lines with a novel hybrid initial warp prior (see Section 3.2.3), and  E s addresses the salient content-preserving issue by protecting both local and global lines from being bent in the non-overlapping region (see Section 3.2.4). The mesh warping result is obtained by minimizing E (see Section 3.2.5). As all constraint terms are quadratic, they can be solved and minimized by any sparse linear solver.

3.2.2. Alignment Term

The point alignment item E p is used to align the two images { I a , I b } , which is defined as follows:
E p = i W i V i ^ p i 2 = | | p ^ p | | 2 ,
where p indicates the matching points in the image I b . Line alignment is also considered as an alignment supplement. Given a set of line correspondences { l a j , l b j } , where l a j I a is represented by the line segment with the endpoints P a k , and  P a k = W p k V p k ^ . l b j I b is represented by the line equation a j x + b j y + c j = 0 [18], then the line alignment term is represented as below:
E l = j , k l b j T P a k λ j 2 ,
where λ j = 1 a j 2 + b j 2 . With the scalar λ j , the weight between two endpoints is balanced in terms of geometric meaning.
In summary, the alignment term is represented as:
E a = λ p E p + λ l E l .

3.2.3. Distortion Term

Both global homography and local homography can cause distortion in the non-overlapping region of the image. This distortion is negligible for content close to the common main plane, but more severe for content further away from the main plane. In order to control the distortion of the target image I a , a series of horizontal and vertical lines are constructed, called cross lines. They are regarded as the intrinsic structure of the image and are used to reflect the overall warping degree of the target image I a .
In [40], s ( x , y , k ) is the slope of the line in I b corresponding to the line passing ( x , y ) with slope k in I a . l u is the warped horizontal line and l v is the vertical line that is closest to the boundary between overlapping and non-overlapping areas. If a homography warp is given, there is a unique set of parallel lines corresponding to it. The slopes of the parallel lines before and after warping are as follows:
k 1 = h 7 h 8 , s ( x , y , k 1 ) = h 4 h 8 h 5 h 7 h 1 h 8 h 2 h 7 .
Therefore, l v can be set to be the split line with slope k 1 that is closest to the boundary of the overlapping region and non-overlapping region, and  l u can be set to be orthogonal to l v before and after warping, as shown below:
k 1 · k 2 = 1 , s ( x , y , k 1 ) · s ( x , y , k 2 ) = 1 ,
Although cross lines can effectively alleviate the distortion of the global homography and balance the projection and perspective distortion in the non-overlapping region, it can produce a severely stretched single-perspective stitching result when the scene depth or parallax is large (see Figure 4b). Furthermore, it relies on the initial global homography. If the global homography transformation is misaligned in the overlapping region and severely distorted in the non-overlapping region, the cross lines can only alleviate less distortions and cannot flexibly deal with the distortion and misalignment problems without a fundamental solution that the distortion in the non-overlapping region caused by homography. In [17], the global similarity transformation can effectively mitigate the distortion in the non-overlapping region by iteratively finding the minimum rotation angle of the transformation. Therefore, we mix the global homography transformation with the global similarity transformation, and we adjust the rotation angle of the cross lines by adding weights, so as to reduce the distortion flexibly (as shown in Figure 4d).
In this paper, we change the initial global homography to a hybrid global homography transformation and a global similarity transformation as shown in Figure 5. This is defined as follows:
H i n i t = H g h + H h y ,
where H i n i t , H g h , H h y are the initial transformation, the global homography transformation, and the hybrid adjustment terms, respectively. They are all 3 × 3 matrices. The purpose is to accurately align in the overlapping region, reduce distortion in the non-overlapping region, and flexibly cope with large parallax and large scene depth. The hybrid adjustment term is defined as follows:
H h y = W h y ( H g s H g h ) ,
where H g s is the global similarity transformation, and  W h y is the weighting factor in the range of [ 0 , 1 ) for adjusting H g s and H g h . If  W h y is 0, H i n i t is the global homography transformation H g h , at this time the slope of the cross line in Equation (15) is the largest. Since the overall warping change is the largest, the distortion is the most serious. However, the alignment is better; if W h y is 1, H i n i t is the global similarity transformation H g s , but the slope of the cross lines does not exist. The closer the value of W h y is to 0, the higher the slope; the higher the degree of warp, the better the alignment, but the more serious the distortion. On the contrary, the closer the value of W h y to 1, the smaller the slope, the smaller the transformation angle, the less distortion, but the more serious the misalignment. Therefore, it is necessary to choose the appropriate value of W h y according to different types of images to be stitched. The stitching results at different W h y are shown in Figure 4.
By the method discussed above, we divide the distortion term E d into a perspective distortion term E p s that preserves the perspective given by the new hybrid homography warp in Equation (18) and a projection distortion term E p j that mitigates the projective distortion.
Given the set of cross line correspondences { l a i u , l b i u } and { l a j v , l b j v } that are parallel to l u and l v , the points of cross lines are recorded as { p k u i } and { p k v i } , then
E p s = i = 1 k = 1 | W p k + 1 u i V ^ p k + 1 u i W p k u i V ^ p k u i , n i u | 2 + j = 1 k = 1 | W p k + 1 v j V ^ p k + 1 v j W p k v j V ^ p k v j , n j v | 2 + j = 1 k = 1 W p k v j V ^ p k v j + W p k + 2 v j V ^ p k + 2 v j 2 W p k + 1 v j V ^ p k + 1 v j 2 ,
where n i u and n j v are the normal vectors of l b i u and l b j v . In Equation (19), the first two terms preserve the slopes of l b i u and l b j v , and the last term preserve the ratios of lengths on l a j v .
The projection distortion term for the non-overlapping region of image I a is defined as follows:
E p j = i = 1 k = 1 W p k u i V ^ p k u i + W p k + 2 u i V ^ p k + 2 u i 2 W p k + 1 u i V ^ p k + 1 u i 2 .
In fact, Equation (20) linearizes the scale on l a i u in the non-overlapping region of I a .
In summary, the distortion term is represented as:
E d = λ p s E p s + λ p j E p j .

3.2.4. Salient Term

The salient term is used to preserve the line structure in the non-overlapping region of I a . Given the set of salient lines { l a k s } , each l a k s I a is denoted by the set of endpoints p j k , then:
E l s = k = 1 j = 1 W p j + 1 k V ^ p j + 1 k W p j k V ^ p j k , n k 2 ,
where n k is the normal vector of { l b k s } that is calculated from the new homography H i n i t in Equation (17).
Due to the limitations of line segment detection, the detected line segment is generally not very long, and it may not be the actual line structure affected by the viewing direction or other factors. When some local lines are actually collinear, if they are not globally optimized, there will be significant global line bending after warping. Inspired by [20], we merge some apparently collinear local line segments into global lines, while adding the global line term E g s to the salient term.
Given the set of global lines { l a g s } , each l a g s I a is denoted by the set of endpoints p k g , then we have the following:
E g s = g = 1 k = 1 W p k + 1 g V ^ p k + 1 g W p k g V ^ p k g , n g 2 ,
where n g is the normal vector of { l b g s } . Figure 6 also shows the effect of adding the global line term on the stitching results.
Similarly, local lines term and global line term can be represented as follows:
E s = λ l s E l s + λ g s E g s .

3.2.5. Total Energy Function

The above terms can be written as a total energy function:
E = λ p E p + λ l E l + λ p s E p s + λ p j E p j + λ l s E l s + λ g s E g s .
The above function is quadratic, and this turns into an optimization problem which can be solved and minimized by any sparse linear solver.
To sum up, we define alignment terms, distortion terms, and salient terms based on a novel hybrid initial warp prior to address alignment and distortion issues. Among them, the point alignment term in the alignment term is based on our proposed feature correspondences increase model. The stitching process of the two images is shown in Algorithm 1.
Algorithm 1 Stitching two images.
Input: a target image I a and a reference image I b .
Output: a stitched image.
  1:
Match point and line features between I a and I b to obtain { p a , p b } and { l a , l b } via feature correspondences increase.
  2:
Calculate a global homography H g h and a global similarity matrix H g s via dual-feature.
  3:
Calculate a hybrid warp H i n i t by Equation (18) and Equation (19).
  4:
Calculate { l a i u , l b i u } and { l a j v , l b j v } from H i n i t by Equation (16) and Equation (17).
  5:
Detecting salient line segments in the non-overlapping region to obtain { l a k s } .
  6:
Merge collinear local line segments into global lines { l a g s } .
  7:
Uniformly sample { l a i u } , { l a i v } , { l a k s } , { l a g s } .
  8:
Solve V ^ via minimizing the total energy term in Equation (26).
  9:
Warp I a via the bilinear interpolation on the basis of V ^ .
  10:
Stitching the warped I a with I b via linear blending or seam-cutting blending.

4. Experiments

4.1. Experimental Setup

The proposed method has been tested on selected datasets, including some public datasets with 10 pairs of images in [8,10,17,20,34,39], and our own collected datasets with 5 pairs of images. Some input image pairs as examples are shown in Figure 7 for quantitative evaluation and qualitative comparison. All the images are natural RGB ones. In addition to regular natural images, we focus on images with low-textured or repetitively-textured regions in the overlapping region, which will better demonstrate the effectiveness of the proposed method.
We compare our approach with some state-of-the-art methods, including homography, APAP, AANAP, SPHP, SPW and LPC. The parameters of the existing methods are set as suggested by the original papers. In order to facilitate the comparison of alignment effects in the overlapping region and structure preservation effects in the non-overlapping region, while avoiding seam-cutting blending to mask the original misalignment, all the methods use linear blending. In addition, we also show the visual enhancement brought by seam-cutting blending in [37].
In the experiments, we first extract and match ORB feature points, then we remove mismatches with GMS, while using LSD [24] to detect line segments and match them by [21]. The number of ORB feature points is 30,000. Since the image stitching task does not involve rotation and scaling, the parameters of rotation and scaling in GMS are both 0. For the hybrid terms optimization parameters, the deformation mesh size is set to 40 × 40, and the inputs λ p , λ l , λ p s , λ p j , λ l s , and λ g s are set to 1, 5, 50, 100, 50, and 100, respectively. The initial homography fine-tuning weight W h y is set to 0.4. It is worth noting that W h y and λ p s are inversely proportional. If W h y is too large and the transformation is approximately similar to the global one, the perspective relationship will change significantly. Therefore, the perspective term should be reduced to ensure alignment of the overlapping regions. In this way, the non-overlapping region is close to the rectangle with the least distortion, and the distortion term and the salient term can be appropriately reduced, and vice versa. Since the salient term and the distortion term are competing in [22], the ration of the inputs λ p j and λ g s are 1:1.
The proposed model is mainly implemented in MATLAB (some portions of the code are written in C++), and all the experiments are carried out on a PC with AMD Ryzen 7 4800 H 2.90 GHz CPU and 16 GB RAM.

4.2. Quantitative Evaluation

We quantitatively evaluate the stitching quality of different single-perspective methods using SSIM [43] in the range of [0, 1]. The larger the value of SSIM, the more similar the structure after warping. The overlapping regions of each pair of images in the dataset are extracted to compute the corresponding SSIM after alignment.
Besides, we quantitatively evaluate the alignment accuracy of the proposed method by root mean squared error (RMSE) on feature correspondences:
R M S E ( f ) = 1 N i = 1 N f ( p i ) p i 2 ,
where N is the number of feature correspondences and f represents a planar warp. Table 1 shows the number of matches, SSIM and RMSE using different methods on different datasets. Among them, the first four rows are dense texture images with small parallax, the fifth to eighth rows are images with large parallax and no obvious low-textured areas, and the last seven rows are images with obvious low-textured areas in the overlapping region. The traditional SIFT+RANSAC model is used as a baseline model for comparison. Compared to it, our feature correspondences increase model significantly increases the number of matches generally. The more feature correspondences there are, the greater the image information entropy is, and the better the image detail performance is, so as to obtain a better stitching result. In particular, it allows sufficient matches for grid-based stitching algorithms, especially for low-textured images. Additionally, the values of SSIM and RMSE also show that our method outperforms other methods, especially that our method achieves the highest scores on images with significantly weakly textured or repetitively textured regions. On other types of images, although some scores are low, the scores of our method are comparable with other methods.

4.3. Qualitative Comparison

First of all, we replace the traditional SIFT/SURF + RANSAC pattern in each method with the proposed feature correspondences increase model. Figure 8 compares the stitching results of some methods before and after applying feature correspondences increase on the dataset intersection [17]. As in the original three methods shown in Figure 8a, the inliers are mostly concentrated in the house area as indicated by the blue circle, so the alignment around the house area is better. Since the feature points are only concentrated in the dense-textured region, the feature correspondences of low-textured region is insufficient, which leads to serious misalignment, as shown in the red rectangle. APAP is a grid-based local alignment method, and some pixels will destroy the continuity and consistency of the original texture. However, SPW, LPC and SPHP do not destroy the original texture of the image due to the restriction of alignment terms. Figure 8b shows the results after applying our proposed method. Due to the increase of feature correspondences, the alignment of the three methods is more accurate in low-textured regions. The local alignment property of APAP itself causes it to suffer structural damage, but with less distortion in the non-overlapping region. It also illustrates that grid-based local alignment relies on a sufficient and efficient distribution of the number of feature points. In the case of increased feature correspondences, SPW and LPC ignore the distortion of non-overlapping regions due to over-focusing on alignment. It results in a less smooth transition between overlapping and non-overlapping regions. LPC is better than SPW, but it still has artifacts. SPHP has more accurate alignment compared to using SIFT+RANSAC. However, due to its excellent distortion preserving properties, alignment is not as accurate as SPW and LPC. The above experiments indicate that our proposed feature correspondences increase model can be added to any feature point-based method to replace the original SIFT/SURF+RANSAC pattern to improve alignment accuracy. However, accurate alignment is often accompanied by inevitable distortions in the non-overlapping region. The next issue we want to focus on is how to balance alignment and distortion.
Figure 9 shows the stitching results of different methods on the dataset runway. The first column shows the stitching results. The runway areas in the red rectangles represent low-textured regions, and the magnified versions are shown in the second column. Similarly, the buildings in the blue rectangles represent the densely-textured regions, and their magnified versions are shown in the third column. Global homography does not yield satisfactory results due to the presence of parallax. There is obvious misalignment, either in the building or the runway. The results of APAP based on local alignment are slightly better than those of global alignment, but the feature points are concentrated in the densely-textured region, while they are insufficient in the low-textured regions. This leads to more accurate alignment in the building and serious misalignment in the runway. Further, global homography and APAP have the same projection and perspective distortion in the non-overlapping regions. The third and fourth rows demonstrate two methods of warping two images. While they effectively mitigate the distortion in the non-overlapping regions, insufficient inliers cause misalignment in low-textured regions, as shown in the red rectangles. The fifth row shows the results of TFA. Although it uses a two-step RANSAC outlier removal method to ensure the matching points as correct as possible, the characteristic of SIFT+RANSAC pattern leads to a serious lack of correct matches in the low-textured regions, which in turn leads to misalignment as shown in the red rectangles. Meanwhile, the triangulation method destroys the original structure, causing the line structure to be bent in the overlapping regions. In the non-overlapping region, its distortion preservation is better than global homography and APAP. In addition, SPW also shows misalignment in the overlapping regions. Although LPC aligned the densely-textured regions, it still caused serious misalignment of the runway areas due to insufficient feature points. Due to its more accurate alignment in the densely-textured regions in the blue rectangles, the distortion in the non-overlapping regions is more severe than SPW. To better demonstrate the effectiveness of our method, we provide three comparisons of results. The eighth row presents the results using only our feature correspondences increase (FCI) model. The ninth row shows the result of using only our hybrid terms optimization (HTO) warp but SIFT+RANSAC as the feature matching stage. The last row is the results of using both FCI and HTO. Using only FCI or HTO may not achieve the results we expect. Alignment using only FCI is significantly more accurate, but there is still slight distortion. Distortion is significantly alleviated by using only HTO but SIFT+RANSAC in the non-overlapping regions, and the target image is more rectangular, but the alignment of overlapping regions is less accurate. When both FCI and HTO are used, our method is aligned in the densely-textured regions. In addition, our alignment is very accurate in the low-textured regions compared to other methods. Furthermore, our method also outperforms SPW and LPC in terms of distortion in the non-overlapping regions. Our method achieves a balance between alignment and distortion to a certain extent.

4.4. Seam-Cutting Blending

The proposed method has achieved more accurate alignment and it effectively balances alignment and distortion by linear blending. Nonetheless, in some challenging large parallax scenes, our method may produce misalignment if only linear blending is used. Therefore, we introduce seam-cutting blending, which can effectively eliminate the artifacts and misalignment brought by linear blending.
The approach proposed in [10] is a typical seam-driven stitching method. Similar to the other methods we compared, it uses SIFT+RANSAC as feature matching, and it has similar optimization constraints. However, there is no line structure preservation term in its energy function. Figure 10 presents the comparison of [10] and our method on the dataset 025. The red circles highlight the differences between the two methods in the overlapping regions. The results of [10] present significant bending of the line structures at the transparent glass building. In contrast, our method outperforms this approach in terms of line structures preservation since our salient term can effectively protect the line structures from being bent. Besides, our method is also slightly better for distortion in the non-overlapping regions.
In addition to applying seam-cutting blending to handle large parallax scenes to a certain extent, our method can also deal with continuous structure damage and heavy occlusion issues due to a more accurate alignment. In fact, seam-cutting blending also depends on the pre-alignment. If the pre-alignment is severely misaligned, seam-cutting blending may destroy the original texture structure, resulting in inconsistent and discontinuous stitching results due to the inaccurate pre-alignment. Figure 11 shows the difference between LPC and our proposed method using seam-cutting blending on a large parallax example in our own dataset. It can be seen that the linear blending results of our proposed method are better than those of LPC with a different optimal seam. As a result, the seam-cutting blending of LPC destroys the original line structures, resulting in an obvious misalignment. This is shown in the blue circle in Figure 11e. In Figure 11f, our method achieves accurate alignment.
Likewise, if the object has a serious occlusion relationship between the foreground and the background, the degree of post-warp alignment will seriously affect the final results. Figure 12 shows an example with a front-to-back occlusion relationship via [19]. It can be seen that the optimal seam of our method effectively avoids the occlusion of the target foreground and background, and achieves accurate alignment (see Figure 12d,f).

5. Conclusions

In this paper, we propose a novel feature correspondences increase and hybrid terms optimization warp for image stitching. It can deal with the misalignment issue caused by insufficient feature correspondences of low-textured areas in the overlapping region and the distortion issue in non-overlapping regions, with better balancing alignment and distortion. The proposed method with seam-cutting blending can solve the issue of large parallax and occlusion. Both quantitative evaluation and qualitative comparison experiments show that the proposed method is more accurate in alignment and less distortion in the non-overlapping region compared with other methods, especially on images with low-textured areas in the overlapping region.

Author Contributions

Software, writing—original draft preparation, writing—review and editing, Y.C.; formal analysis, Y.W.; project administration, Y.W.; investigation, W.H.; validation, W.P.; supervision, Y.W. and W.P. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (No. 62072212), the Development Project of Jilin Province of China (No. 20220508125RC).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Szeliski, R. Image Alignment and Stitching: A Tutorial. Found. Trends Comput. Graph. Vis. 2007, 2, 1–104. [Google Scholar] [CrossRef]
  2. Brown, M.A.; Lowe, D.G. Automatic Panoramic Image Stitching using Invariant Features. Int. J. Comput. Vis. 2006, 74, 59–73. [Google Scholar] [CrossRef] [Green Version]
  3. Caesar, H.; Bankiti, V.; Lang, A.H.; Vora, S.; Liong, V.E.; Xu, Q.; Krishnan, A.; Pan, Y.; Baldan, G.; Beijbom, O. nuScenes: A Multimodal Dataset for Autonomous Driving. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 11618–11628. [Google Scholar]
  4. Luan, Y.; Han, C.; Wang, B. An Unsupervised Video Stabilization Algorithm Based on Key Point Detection. Entropy 2022, 24, 1326. [Google Scholar] [CrossRef]
  5. Madhusudana, P.C.; Soundararajan, R. Subjective and Objective Quality Assessment of Stitched Images for Virtual Reality. IEEE Trans. Image Process. 2019, 28, 5620–5635. [Google Scholar] [CrossRef] [PubMed]
  6. Chang, C.H.; Chen, C.J.; Chuang, Y.Y. Spatially-Varying Image Warps for Scene Alignment. In Proceedings of the 2014 22nd International Conference on Pattern Recognition, Stockholm, Sweden, 24–28 August 2014; pp. 64–69. [Google Scholar]
  7. Lin, W.Y.; Liu, S.; Matsushita, Y.; Ng, T.T.; Cheong, L.F. Smoothly varying affine stitching. In Proceedings of the CVPR 2011, Colorado Springs, CO, USA, 20–25 June 2011; pp. 345–352. [Google Scholar]
  8. Zaragoza, J.; Chin, T.J.; Tran, Q.H.; Brown, M.S.; Suter, D. As-Projective-As-Possible Image Stitching with Moving DLT. In Proceedings of the 2013 IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 2339–2346. [Google Scholar]
  9. Gao, J.; Li, Y.; Chin, T.J.; Brown, M.S. Seam-Driven Image Stitching. In Proceedings of the Eurographics 2013—Short Papers; Otaduy, M.A., Sorkine, O., Eds.; The Eurographics Association: Vienna, Austria, 2013. [Google Scholar] [CrossRef]
  10. Zhang, F.; Liu, F. Parallax-Tolerant Image Stitching. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 3262–3269. [Google Scholar]
  11. Lin, K.; Jiang, N.; Cheong, L.F.; Do, M.N.; Lu, J. SEAGULL: Seam-Guided Local Alignment for Parallax-Tolerant Image Stitching. In Proceedings of the ECCV, Amsterdam, The Netherlands, 11–14 October 2016. [Google Scholar]
  12. Nie, L.; Lin, C.; Liao, K.; Liu, S.; Zhao, Y. Unsupervised Deep Image Stitching: Reconstructing Stitched Features to Images. IEEE Trans. Image Process. 2021, 30, 6184–6197. [Google Scholar] [CrossRef] [PubMed]
  13. Zhang, L.; Huang, H. Image Stitching with Manifold Optimization. IEEE Trans. Multimed. 2022. [Google Scholar] [CrossRef]
  14. Nie, L.; Lin, C.; Liao, K.; Liu, S.; Zhao, Y. Depth-Aware Multi-Grid Deep Homography Estimation With Contextual Correlation. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 4460–4472. [Google Scholar] [CrossRef]
  15. LoweDavid, G. Distinctive Image Features from Scale-Invariant Keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar]
  16. Fischler, M.A.; Bolles, R.C. Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
  17. Lin, C.C.; Pankanti, S.; Ramamurthy, K.N.; Aravkin, A.Y. Adaptive as-natural-as-possible image stitching. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1155–1163. [Google Scholar]
  18. Li, S.; Yuan, L.; Sun, J.; Quan, L. Dual-Feature Warping-Based Motion Model Estimation. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 4283–4291. [Google Scholar]
  19. Li, J.; Wang, Z.; Lai, S.; Zhai, Y.; Zhang, M. Parallax-Tolerant Image Stitching Based on Robust Elastic Warping. IEEE Trans. Multimed. 2018, 20, 1672–1687. [Google Scholar] [CrossRef]
  20. Jia, Q.; Li, Z.; Fan, X.; Zhao, H.; Teng, S.; Ye, X.; Latecki, L.J. Leveraging Line-point Consistence to Preserve Structures for Wide Parallax Image Stitching. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 12181–12190. [Google Scholar]
  21. Jia, Q.; Gao, X.; Fan, X.; Luo, Z.; Li, H.; Chen, Z. Novel Coplanar Line-Points Invariants for Robust Line Matching Across Views. In Proceedings of the ECCV, Amsterdam, The Netherlands, 11–14 October 2016. [Google Scholar]
  22. Liao, T.; Li, N. Single-Perspective Warps in Natural Image Stitching. IEEE Trans. Image Process. 2020, 29, 724–735. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Joo, K.; Kim, N.; Oh, T.H.; Kweon, I.S. Line meets as-projective-as-possible image stitching with moving DLT. In Proceedings of the 2015 IEEE International Conference on Image Processing (ICIP), Bordeaux, France, 16–19 October 2015; pp. 1175–1179. [Google Scholar]
  24. von Gioi, R.G.; Jakubowicz, J.; Morel, J.M.; Randall, G. LSD: A Line Segment Detector. Image Process. Line 2012, 2, 35–55. [Google Scholar] [CrossRef] [Green Version]
  25. Bian, J.; Lin, W.Y.; Matsushita, Y.; Yeung, S.K.; Nguyen, T.D.; Cheng, M.M. GMS: Grid-Based Motion Statistics for Fast, Ultra-Robust Feature Correspondence. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 June 2017; pp. 2828–2837. [Google Scholar]
  26. Rublee, E.; Rabaud, V.; Konolige, K.; Bradski, G.R. ORB: An efficient alternative to SIFT or SURF. In Proceedings of the 2011 International Conference on Computer Vision, Tokyo, Japan, 25–27 May 2011; pp. 2564–2571. [Google Scholar]
  27. Harris, C.G.; Stephens, M.J. A Combined Corner and Edge Detector. In Proceedings of the Alvey Vision Conference, Manchester, UK, 31 August–2 September 1988. [Google Scholar]
  28. Rosten, E.; Drummond, T. Machine Learning for High-Speed Corner Detection. In Proceedings of the ECCV, Graz, Austria, 7–13 May 2006. [Google Scholar]
  29. Bay, H.; Tuytelaars, T.; Gool, L.V. SURF: Speeded Up Robust Features. In Proceedings of the ECCV, Graz, Austria, 7–13 May 2006. [Google Scholar]
  30. Calonder, M.; Lepetit, V.; Strecha, C.; Fua, P.V. BRIEF: Binary Robust Independent Elementary Features. In Proceedings of the ECCV, Crete, Greece, 5–11 September 2010. [Google Scholar]
  31. Lin, W.Y.; Cheng, M.M.; Lu, J.; Yang, H.; Do, M.N.; Torr, P.H.S. Bilateral Functions for Global Motion Modeling. In Proceedings of the ECCV, Zurich, Switzerland, 6–12 September 2014. [Google Scholar]
  32. Li, J.; Deng, B.; Tang, R.; Wang, Z.; Yan, Y. Local-Adaptive Image Alignment Based on Triangular Facet Approximation. IEEE Trans. Image Process. 2020, 29, 2356–2369. [Google Scholar] [CrossRef] [PubMed]
  33. Triggs, B.; McLauchlan, P.F.; Hartley, R.I.; Fitzgibbon, A.W. Bundle Adjustment—A Modern Synthesis. In Proceedings of the Workshop on Vision Algorithms, Corfu, Greece, 21–22 September 1999. [Google Scholar]
  34. Gao, J.; Kim, S.J.; Brown, M.S. Constructing image panoramas using dual-homography warping. In Proceedings of the CVPR, Colorado Springs, CO, USA, 20–25 June 2011; pp. 49–56. [Google Scholar]
  35. Chang, C.H.; Sato, Y.; Chuang, Y.Y. Shape-Preserving Half-Projective Warps for Image Stitching. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 3254–3261. [Google Scholar]
  36. Liu, F.; Gleicher, M.; Jin, H.; Agarwala, A. Content-preserving warps for 3D video stabilization. ACM Trans. Graph. 2009, 28, 44. [Google Scholar] [CrossRef] [Green Version]
  37. Li, N.; Liao, T.; Wang, C. Perception-based seam cutting for image stitching. Signal Image Video Process. 2018, 12, 967–974. [Google Scholar] [CrossRef]
  38. Xue, W.; Xie, W.; Zhang, Y.; Chen, S. Stable Linear Structures and Seam Measurements for Parallax Image Stitching. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 253–261. [Google Scholar] [CrossRef]
  39. Chen, Y.S.; Chuang, Y.Y. Natural Image Stitching with the Global Similarity Prior. In Proceedings of the ECCV, Amsterdam, The Netherlands, 11–14 October 2016. [Google Scholar]
  40. Li, N.; Xu, Y.; Wang, C. Quasi-Homography Warps in Image Stitching. IEEE Trans. Multimed. 2018, 20, 1365–1375. [Google Scholar] [CrossRef] [Green Version]
  41. Zhang, Y.; Lai, Y.K.; Zhang, F.L. Content-Preserving Image Stitching With Piecewise Rectangular Boundary Constraints. IEEE Trans. Vis. Comput. Graph. 2021, 27, 3198–3212. [Google Scholar] [CrossRef] [PubMed]
  42. Liu, S.; Yuan, L.; Tan, P.; Sun, J. Bundled camera paths for video stabilization. ACM Trans. Graph. (TOG) 2013, 32, 1–10. [Google Scholar] [CrossRef]
  43. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Comparing the number and distribution of inliers on different texture images. (a) Dense texture. (b) Semi-dense texture. (c) Low texture. The images are from [8,17,18], respectively.
Figure 1. Comparing the number and distribution of inliers on different texture images. (a) Dense texture. (b) Semi-dense texture. (c) Low texture. The images are from [8,17,18], respectively.
Entropy 25 00106 g001
Figure 2. Comparison of inliers and stitching results of different methods. For each subfigure, the feature correspondences are on the top, and the stitching results are on the bottom. (a) APAP with RANSAC threshold 0.1. (b) APAP with RANSAC threshold 0.5. (c) LPC. (d) Ours.
Figure 2. Comparison of inliers and stitching results of different methods. For each subfigure, the feature correspondences are on the top, and the stitching results are on the bottom. (a) APAP with RANSAC threshold 0.1. (b) APAP with RANSAC threshold 0.5. (c) LPC. (d) Ours.
Entropy 25 00106 g002
Figure 3. The overall process architecture of our method. The feature correspondences increase module mainly increases the number of feature points to ensure accurate alignment. The hybrid terms optimization warp is based on the above stage to obtain the optimal mesh deformation, which further balances the alignment and distortion.
Figure 3. The overall process architecture of our method. The feature correspondences increase module mainly increases the number of feature points to ensure accurate alignment. The hybrid terms optimization warp is based on the above stage to obtain the optimal mesh deformation, which further balances the alignment and distortion.
Entropy 25 00106 g003
Figure 4. The cross lines distortion mitigation results. (a) Warped mesh result of homography. (b) Warped mesh result of SPW. (c) Warped mesh result of LPC. (d) Warped mesh result of ours. The blue line is closest to the border of the overlapping region and the non-overlapping region, and the red line is perpendicular to it. The smaller the slope of the cross lines, the less distortion.
Figure 4. The cross lines distortion mitigation results. (a) Warped mesh result of homography. (b) Warped mesh result of SPW. (c) Warped mesh result of LPC. (d) Warped mesh result of ours. The blue line is closest to the border of the overlapping region and the non-overlapping region, and the red line is perpendicular to it. The smaller the slope of the cross lines, the less distortion.
Entropy 25 00106 g004
Figure 5. Hybrid global transformation. For the target image, the red dashed frame represents the global homography transformation and the blue dashed frame represents the global similarity transformation. The solid black frame in the middle represents the hybrid global transformation.
Figure 5. Hybrid global transformation. For the target image, the red dashed frame represents the global homography transformation and the blue dashed frame represents the global similarity transformation. The solid black frame in the middle represents the hybrid global transformation.
Entropy 25 00106 g005
Figure 6. Global lines and hybrid transformation with different W h y . (a) No global lines and W h y = 0 . (b) With global lines and W h y = 0 . (c) With global lines and W h y = 0.4 . (d) With global lines and W h y = 0.9 . The blue circles show the distortion mitigation effect with different W h y , and the red rectangles show the global lines for structure preservation.
Figure 6. Global lines and hybrid transformation with different W h y . (a) No global lines and W h y = 0 . (b) With global lines and W h y = 0 . (c) With global lines and W h y = 0.4 . (d) With global lines and W h y = 0.9 . The blue circles show the distortion mitigation effect with different W h y , and the red rectangles show the global lines for structure preservation.
Entropy 25 00106 g006
Figure 7. Datasets for experimental demonstration. (ae) are our own images. (f) is from [10].
Figure 7. Datasets for experimental demonstration. (ae) are our own images. (f) is from [10].
Entropy 25 00106 g007
Figure 8. Comparison of the stitching results for increasing the number of features. (a) The original methods of APAP, SPW, LPC and SPHP, respectively. (b) After applying our method. The red boxes represent the low texture region, and the blue circles represent the dense texture region.
Figure 8. Comparison of the stitching results for increasing the number of features. (a) The original methods of APAP, SPW, LPC and SPHP, respectively. (b) After applying our method. The red boxes represent the low texture region, and the blue circles represent the dense texture region.
Entropy 25 00106 g008aEntropy 25 00106 g008b
Figure 9. Comparison of stitching quality on runway. Row 1: Results using global homography. Row 2: Results using APAP. Row 3: Results of AANAP. Row 4: Results using SPHP. Row 5: Results of TFA. Row 6: Results of SPW. Row 7: Results of LPC. Row 8: Results only using our feature correspondences increase (FCI) model. Row 9: Results only using our hybrid terms optimization (HTO) warp but SIFT+RANSAC. Row 10: Results of our approach by using both FCI and HTO.
Figure 9. Comparison of stitching quality on runway. Row 1: Results using global homography. Row 2: Results using APAP. Row 3: Results of AANAP. Row 4: Results using SPHP. Row 5: Results of TFA. Row 6: Results of SPW. Row 7: Results of LPC. Row 8: Results only using our feature correspondences increase (FCI) model. Row 9: Results only using our hybrid terms optimization (HTO) warp but SIFT+RANSAC. Row 10: Results of our approach by using both FCI and HTO.
Entropy 25 00106 g009
Figure 10. Comparison of [10] and our method on dataset 025. (a) Stitching result of [10]. (b) Stitching result of our method. The red circles highlight the differences between the two methods in the overlapping regions. Our method protects the line structure better from being bent.
Figure 10. Comparison of [10] and our method on dataset 025. (a) Stitching result of [10]. (b) Stitching result of our method. The red circles highlight the differences between the two methods in the overlapping regions. Our method protects the line structure better from being bent.
Entropy 25 00106 g010
Figure 11. Comparison seam-cutting blending effects of LPC and ours. (a) Stitching result of LPC using linear blending. (b) Stitching result of ours using linear blending. (c) Optimal seam of LPC. (d) Optimal seam of ours. (e) Stitching result of LPC. (f) Stitching result of ours. The difference is shown in the blue circle. It is obvious to see that the seam blending of LPC destroys the original texture structure owing to misalignment seriously. In contrast, our method produces better seam-cutting blending due to more accurate alignment.
Figure 11. Comparison seam-cutting blending effects of LPC and ours. (a) Stitching result of LPC using linear blending. (b) Stitching result of ours using linear blending. (c) Optimal seam of LPC. (d) Optimal seam of ours. (e) Stitching result of LPC. (f) Stitching result of ours. The difference is shown in the blue circle. It is obvious to see that the seam blending of LPC destroys the original texture structure owing to misalignment seriously. In contrast, our method produces better seam-cutting blending due to more accurate alignment.
Entropy 25 00106 g011
Figure 12. (af) Seam-cutting blending results of images with severe occlusions. The red line indicates the optimal seam of LPC and ours. The seam of the LPC destroys the relationship between the foreground and background of the object in the blue circle. However, our method is capable of aligning accurately enough to avoid this problem.
Figure 12. (af) Seam-cutting blending results of images with severe occlusions. The red line indicates the optimal seam of LPC and ours. The seam of the LPC destroys the relationship between the foreground and background of the object in the blue circle. However, our method is capable of aligning accurately enough to avoid this problem.
Entropy 25 00106 g012
Table 1. Comparison on matches, SSIM and RMSE.
Table 1. Comparison on matches, SSIM and RMSE.
DatabasesSIFT+RANSACGlobal HomographyAPAPSPWLPCOURS
MatchesRMSESSIMRMSESSIMRMSESSIMRMSESSIMMatchesSSIMRMSE
fence  [17]5970.64891.76740.70111.77730.69521.45920.68071.336859110.71841.5014
Potberry  [20]3600.39712.34190.59042.34190.46482.05590.51441.892837720.61421.7384
railtracks  [8]6510.44742.57630.63542.57670.63612.2170.58711.515561760.67381.6670
DHW-temple  [34]3220.55172.66330.67992.64370.60572.26740.49992.319954740.69351.6570
MemorialHall  [39]640.53702.51690.55902.51690.50182.27100.52171.747116030.54171.8711
017  [10]3300.58713.05060.61513.30190.62542.35010.61392.615350640.62732.4313
cup  [20]1590.47472.64360.55092.11490.47782.48070.47782.606930410.57302.0057
office  [20]1810.54153.54230.65823.54230.61503.01660.61242.883929910.69041.9996
intersection  [17]4260.36123.46260.48093.50730.41522.55910.42852.886843140.55201.9684
tower  [19]6520.57343.29670.76013.29890.77022.20890.82591.639164090.85011.6145
runway2080.51803.08600.58923.08650.55102.74760.55842.015438460.66321.4540
car park2930.40442.65440.45732.65440.43132.46220.23283.750135650.49782.3074
football field2370.48813.11100.58283.11100.48722.52110.42511.941436410.58791.6641
sidewalk2450.72612.39330.76812.39330.72701.97040.47402.921131300.79531.9711
jump runway1170.72991.84740.74571.84700.75351.62500.66281.967623470.74991.7163
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cong, Y.; Wang, Y.; Hou, W.; Pang, W. Feature Correspondences Increase and Hybrid Terms Optimization Warp for Image Stitching. Entropy 2023, 25, 106. https://doi.org/10.3390/e25010106

AMA Style

Cong Y, Wang Y, Hou W, Pang W. Feature Correspondences Increase and Hybrid Terms Optimization Warp for Image Stitching. Entropy. 2023; 25(1):106. https://doi.org/10.3390/e25010106

Chicago/Turabian Style

Cong, Yizhi, Yan Wang, Wenju Hou, and Wei Pang. 2023. "Feature Correspondences Increase and Hybrid Terms Optimization Warp for Image Stitching" Entropy 25, no. 1: 106. https://doi.org/10.3390/e25010106

APA Style

Cong, Y., Wang, Y., Hou, W., & Pang, W. (2023). Feature Correspondences Increase and Hybrid Terms Optimization Warp for Image Stitching. Entropy, 25(1), 106. https://doi.org/10.3390/e25010106

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop