http://blog.csdn.net/xukaiwen_2016/article/details/53149794

看效果很好,但是我的机器opencv3.2 没有编译通过,遗憾,也可以作为一个思路!


Sift和Surf算法实现两幅图像拼接的过程是一样的,主要分为4大部分:
1. 特征点提取和描述
2. 特征点配对,找到两幅图像中匹配点的位置
3. 通过配对点,生成变换矩阵,并对图像1应用变换矩阵生成对图像2的映射图像
4. 图像2拼接到映射图像上,完成拼接

具体请转到http://m.blog.csdn.NET/article/details?id=52629856

代码如下:

[cpp] view plain copy
  1. #include "highgui/highgui.hpp"      
  2. #include "opencv2/nonfree/nonfree.hpp"      
  3. #include "opencv2/legacy/legacy.hpp"     
  4.   
  5. using namespace cv;  
  6.   
  7. //计算原始图像点位在经过矩阵变换后在目标图像上对应位置    
  8. Point2f getTransformPoint(const Point2f originalPoint, const Mat &transformMaxtri);  
  9.   
  10. int main(int argc, char *argv[])  
  11. {  
  12.     Mat image01,image02;  
  13.     if (argc < 2)  
  14.     {  
  15.         image01 = imread("left.jpg");  
  16.         image02 = imread("right.jpg");  
  17.     }  
  18.     else  
  19.     {  
  20.         image01 = imread(argv[1]);  
  21.         image02 = imread(argv[2]);  
  22.     }  
  23.     if (image01.empty() || image02.empty())  
  24.     {  
  25.         return 0;//图像没有全部读取成功  
  26.     }  
  27.     imshow("拼接图像1", image01);  
  28.     imshow("拼接图像2", image02);  
  29.     double time = getTickCount();  
  30.     //灰度图转换    
  31.     Mat image1, image2;  
  32.     cvtColor(image01, image1, CV_RGB2GRAY);  
  33.     cvtColor(image02, image2, CV_RGB2GRAY);  
  34.   
  35.     //提取特征点      
  36.     SiftFeatureDetector siftDetector(800);  // 海塞矩阵阈值    
  37.     vector<KeyPoint> keyPoint1, keyPoint2;  
  38.     siftDetector.detect(image1, keyPoint1);  
  39.     siftDetector.detect(image2, keyPoint2);  
  40.   
  41.     //特征点描述,为下边的特征点匹配做准备      
  42.     SiftDescriptorExtractor siftDescriptor;  
  43.     Mat imageDesc1, imageDesc2;  
  44.     siftDescriptor.compute(image1, keyPoint1, imageDesc1);  
  45.     siftDescriptor.compute(image2, keyPoint2, imageDesc2);  
  46.   
  47.     //获得匹配特征点,并提取最优配对       
  48.     FlannBasedMatcher matcher;  
  49.     vector<DMatch> matchePoints;  
  50.     matcher.match(imageDesc1, imageDesc2, matchePoints, Mat());  
  51.     if (matchePoints.size() < 10)  
  52.     {  
  53.         return 0;  
  54.     }  
  55.     sort(matchePoints.begin(), matchePoints.end()); //特征点排序,opencv按照匹配点准确度排序      
  56.     //获取排在前N个的最优匹配特征点    
  57.     vector<Point2f> imagePoints1, imagePoints2;  
  58.     for (int i = 0; i<10; i++)  
  59.     {  
  60.         imagePoints1.push_back(keyPoint1[matchePoints[i].queryIdx].pt);  
  61.         imagePoints2.push_back(keyPoint2[matchePoints[i].trainIdx].pt);  
  62.     }  
  63.   
  64.     //获取图像1到图像2的投影映射矩阵,尺寸为3*3    
  65.     Mat homo = findHomography(imagePoints1, imagePoints2, CV_RANSAC);  
  66.     Mat adjustMat = (Mat_<double>(3, 3) << 1.0, 0, image01.cols, 0, 1.0, 0, 0, 0, 1.0);//向后偏移image01.cols矩阵  
  67.     Mat adjustHomo = adjustMat*homo;//矩阵相乘,先偏移  
  68.   
  69.     //获取最强配对点(就是第一个配对点)在原始图像和矩阵变换后图像上的对应位置,用于图像拼接点的定位    
  70.     Point2f originalLinkPoint, targetLinkPoint, basedImagePoint;  
  71.     originalLinkPoint = keyPoint1[matchePoints[0].queryIdx].pt;  
  72.     targetLinkPoint = getTransformPoint(originalLinkPoint, adjustHomo);  
  73.     basedImagePoint = keyPoint2[matchePoints[0].trainIdx].pt;  
  74.   
  75.     //图像配准    
  76.     Mat imageTransform;  
  77.     //将图片1进行映射到图像2,本来映射后x值为负值,但是把映射矩阵向后偏移image01.cols矩阵  
  78.     //我们很难判断出拼接后图像的大小尺寸,为了尽可能保留原来的像素,我们尽可能的大一些,对于拼接后的图片可以进一步剪切无效或者不规则的边缘  
  79.     warpPerspective(image01, imageTransform, adjustMat*homo, Size(image02.cols + image01.cols+10, image02.rows));  
  80.   
  81.     //在最强匹配点的位置处衔接,最强匹配点左侧是图1,右侧是图2,这样直接替换图像衔接不好,光线有突变    
  82.     //Mat ROIMat = image02(Rect(Point(basedImagePoint.x, 0), Point(image02.cols, image02.rows)));  
  83.     //ROIMat.copyTo(Mat(imageTransform1, Rect(targetLinkPoint.x, 0, image02.cols - basedImagePoint.x + 1, image02.rows)));  
  84.   
  85.     //在最强匹配点左侧的重叠区域进行累加,是衔接稳定过渡,消除突变    
  86.     Mat image1Overlap, image2Overlap; //图1和图2的重叠部分       
  87.     image1Overlap = imageTransform(Rect(Point(targetLinkPoint.x - basedImagePoint.x, 0), Point(targetLinkPoint.x, image02.rows)));  
  88.     image2Overlap = image02(Rect(0, 0, image1Overlap.cols, image1Overlap.rows));  
  89.     Mat image1ROICopy = image1Overlap.clone();  //复制一份图1的重叠部分    
  90.     for (int i = 0; i<image1Overlap.rows; i++)  
  91.     {  
  92.         for (int j = 0; j<image1Overlap.cols; j++)  
  93.         {  
  94.             double weight;  
  95.             weight = (double)j / image1Overlap.cols;  //随距离改变而改变的叠加系数    
  96.             image1Overlap.at<Vec3b>(i, j)[0] = (1 - weight)*image1ROICopy.at<Vec3b>(i, j)[0] + weight*image2Overlap.at<Vec3b>(i, j)[0];  
  97.             image1Overlap.at<Vec3b>(i, j)[1] = (1 - weight)*image1ROICopy.at<Vec3b>(i, j)[1] + weight*image2Overlap.at<Vec3b>(i, j)[1];  
  98.             image1Overlap.at<Vec3b>(i, j)[2] = (1 - weight)*image1ROICopy.at<Vec3b>(i, j)[2] + weight*image2Overlap.at<Vec3b>(i, j)[2];  
  99.         }  
  100.     }  
  101.     Mat ROIMat = image02(Rect(Point(image1Overlap.cols, 0), Point(image02.cols, image02.rows)));  //图2中不重合的部分    
  102.     ROIMat.copyTo(Mat(imageTransform, Rect(targetLinkPoint.x, 0, ROIMat.cols, image02.rows))); //不重合的部分直接衔接上去    
  103.   
  104.     time = getTickCount() - time;  
  105.     time /= getTickFrequency();  
  106.     printf("match time=%f\n",time);  
  107.     namedWindow("拼接结果", 0);  
  108.     imshow("拼接结果", imageTransform);  
  109.     imwrite("matchResult.jpg",imageTransform);  
  110.     waitKey();  
  111.     return 0;  
  112. }  
  113.   
  114. //计算原始图像点位在经过矩阵变换后在目标图像上对应位置    
  115. Point2f getTransformPoint(const Point2f originalPoint, const Mat &transformMaxtri)  
  116. {  
  117.     Mat originelP, targetP;  
  118.     originelP = (Mat_<double>(3, 1) << originalPoint.x, originalPoint.y, 1.0);  
  119.     targetP = transformMaxtri*originelP;  
  120.     float x = targetP.at<double>(0, 0) / targetP.at<double>(2, 0);  
  121.     float y = targetP.at<double>(1, 0) / targetP.at<double>(2, 0);  
  122.     return Point2f(x, y);  
  123. }  
测试结果:

Opencv实现图像无缝拼接,Sift查找特征点,Flann进行匹配
                                               left左边图片

Opencv实现图像无缝拼接,Sift查找特征点,Flann进行匹配

                                               right右边图片

Opencv实现图像无缝拼接,Sift查找特征点,Flann进行匹配

                                             result拼接结果

测试结果能发现,合并后的图片两边会有黑色区域,如果相机位置不是同高的化,上下两边也会有黑色区域,需要对拼接后的图片进行二次剪切。算法思路是:分别扫描四个边缘,分别找到最大黑色区域长度,然后删掉就行了。


下面是我在opencv3.2上的代码,运行会出现异常。感觉是我的环境有问题。


#include "stdafx.h"
#include "opencv2/highgui/highgui.hpp"  
#include "opencv2/calib3d/calib3d.hpp"  
#include "opencv2/imgproc/imgproc.hpp"  
#include "opencv2/features2d/features2d.hpp"  
#include "opencv2/xfeatures2d/nonfree.hpp"     
using namespace cv;
using namespace std;




//计算原始图像点位在经过矩阵变换后在目标图像上对应位置    
Point2f getTransformPoint(const Point2f originalPoint, const Mat &transformMaxtri);


int main(int argc, char *argv[])
{
Mat image01, image02;
if (argc < 2)
{
image01 = imread("d:\\left.jpg");
image02 = imread("d:\\right.jpg");
}
else
{
image01 = imread(argv[1]);
image02 = imread(argv[2]);
}
if (image01.empty() || image02.empty())
{
printf("the loader the picture failed");
waitKey();
return 0;//图像没有全部读取成功  
}
imshow("拼接图像1", image01);
imshow("拼接图像2", image02);
double time = getTickCount();
//灰度图转换    
Mat image1, image2;
cvtColor(image01, image1, CV_RGB2GRAY);
cvtColor(image02, image2, CV_RGB2GRAY);


//提取特征点      
//SiftFeatureDetector siftDetector(800);  // 海塞矩阵阈值    
Ptr<FeatureDetector> siftDetector = xfeatures2d::SiftFeatureDetector::create();
vector<KeyPoint> keyPoint1, keyPoint2;
siftDetector->detect(image1, keyPoint1);
siftDetector->detect(image2, keyPoint2);


//特征点描述,为下边的特征点匹配做准备      
xfeatures2d::SiftDescriptorExtractor siftDescriptor;
Mat imageDesc1, imageDesc2;
siftDescriptor.compute(image1, keyPoint1, imageDesc1);
siftDescriptor.compute(image2, keyPoint2, imageDesc2);


//获得匹配特征点,并提取最优配对       
FlannBasedMatcher matcher;
vector<DMatch> matchePoints;
matcher.match(imageDesc1, imageDesc2, matchePoints, Mat());
if (matchePoints.size() < 10)
{
printf("the match point is below 10");
waitKey();
return 0;
}
sort(matchePoints.begin(), matchePoints.end()); //特征点排序,opencv按照匹配点准确度排序      
//获取排在前N个的最优匹配特征点    
vector<Point2f> imagePoints1, imagePoints2;
for (int i = 0; i<10; i++)
{
imagePoints1.push_back(keyPoint1[matchePoints[i].queryIdx].pt);
imagePoints2.push_back(keyPoint2[matchePoints[i].trainIdx].pt);
}


//获取图像1到图像2的投影映射矩阵,尺寸为3*3    
Mat homo = findHomography(imagePoints1, imagePoints2, CV_RANSAC);
Mat adjustMat = (Mat_<double>(3, 3) << 1.0, 0, image01.cols, 0, 1.0, 0, 0, 0, 1.0);//向后偏移image01.cols矩阵  
Mat adjustHomo = adjustMat*homo;//矩阵相乘,先偏移  


//获取最强配对点(就是第一个配对点)在原始图像和矩阵变换后图像上的对应位置,用于图像拼接点的定位    
Point2f originalLinkPoint, targetLinkPoint, basedImagePoint;
originalLinkPoint = keyPoint1[matchePoints[0].queryIdx].pt;
targetLinkPoint = getTransformPoint(originalLinkPoint, adjustHomo);
basedImagePoint = keyPoint2[matchePoints[0].trainIdx].pt;


//图像配准    
Mat imageTransform;
//将图片1进行映射到图像2,本来映射后x值为负值,但是把映射矩阵向后偏移image01.cols矩阵  
//我们很难判断出拼接后图像的大小尺寸,为了尽可能保留原来的像素,我们尽可能的大一些,对于拼接后的图片可以进一步剪切无效或者不规则的边缘  
warpPerspective(image01, imageTransform, adjustMat*homo, Size(image02.cols + image01.cols + 10, image02.rows));


//在最强匹配点的位置处衔接,最强匹配点左侧是图1,右侧是图2,这样直接替换图像衔接不好,光线有突变    
//Mat ROIMat = image02(Rect(Point(basedImagePoint.x, 0), Point(image02.cols, image02.rows)));  
//ROIMat.copyTo(Mat(imageTransform1, Rect(targetLinkPoint.x, 0, image02.cols - basedImagePoint.x + 1, image02.rows)));  


//在最强匹配点左侧的重叠区域进行累加,是衔接稳定过渡,消除突变    
Mat image1Overlap, image2Overlap; //图1和图2的重叠部分       
image1Overlap = imageTransform(Rect(Point(targetLinkPoint.x - basedImagePoint.x, 0), Point(targetLinkPoint.x, image02.rows)));
image2Overlap = image02(Rect(0, 0, image1Overlap.cols, image1Overlap.rows));
Mat image1ROICopy = image1Overlap.clone();  //复制一份图1的重叠部分    
for (int i = 0; i<image1Overlap.rows; i++)
{
for (int j = 0; j<image1Overlap.cols; j++)
{
double weight;
weight = (double)j / image1Overlap.cols;  //随距离改变而改变的叠加系数    
image1Overlap.at<Vec3b>(i, j)[0] = (1 - weight)*image1ROICopy.at<Vec3b>(i, j)[0] + weight*image2Overlap.at<Vec3b>(i, j)[0];
image1Overlap.at<Vec3b>(i, j)[1] = (1 - weight)*image1ROICopy.at<Vec3b>(i, j)[1] + weight*image2Overlap.at<Vec3b>(i, j)[1];
image1Overlap.at<Vec3b>(i, j)[2] = (1 - weight)*image1ROICopy.at<Vec3b>(i, j)[2] + weight*image2Overlap.at<Vec3b>(i, j)[2];
}
}
Mat ROIMat = image02(Rect(Point(image1Overlap.cols, 0), Point(image02.cols, image02.rows)));  //图2中不重合的部分    
ROIMat.copyTo(Mat(imageTransform, Rect(targetLinkPoint.x, 0, ROIMat.cols, image02.rows))); //不重合的部分直接衔接上去    


time = getTickCount() - time;
time /= getTickFrequency();
printf("match time=%f\n", time);
namedWindow("拼接结果", 0);
imshow("拼接结果", imageTransform);
imwrite("matchResult.jpg", imageTransform);
waitKey();
return 0;
}


//计算原始图像点位在经过矩阵变换后在目标图像上对应位置    
Point2f getTransformPoint(const Point2f originalPoint, const Mat &transformMaxtri)
{
Mat originelP, targetP;
originelP = (Mat_<double>(3, 1) << originalPoint.x, originalPoint.y, 1.0);
targetP = transformMaxtri*originelP;
float x = targetP.at<double>(0, 0) / targetP.at<double>(2, 0);
float y = targetP.at<double>(1, 0) / targetP.at<double>(2, 0);
return Point2f(x, y);
}


相关文章:

  • 2021-11-27
  • 2021-09-07
  • 2022-12-23
  • 2022-12-23
  • 2022-12-23
  • 2021-12-03
  • 2021-11-10
  • 2021-11-27
猜你喜欢
  • 2022-01-01
  • 2022-02-20
  • 2021-10-09
  • 2021-04-23
  • 2022-12-23
  • 2021-12-04
相关资源
相似解决方案