zoukankan      html  css  js  c++  java
  • java opencv使用相关

    Using OpenCV Java with Eclipse

    http://docs.opencv.org/2.4/doc/tutorials/introduction/java_eclipse/java_eclipse.html

    Since version 2.4.4 OpenCV supports Java. In this tutorial I will explain how to setup development environment for using OpenCV Java with Eclipse in Windows, so you can enjoy the benefits of garbage collected, very refactorable (rename variable, extract method and whatnot) modern language that enables you to write code with less effort and make less mistakes. Here we go.

    Configuring Eclipse

    First, obtain a fresh release of OpenCV from download page and extract it under a simple location like C:OpenCV-2.4.6. I am using version 2.4.6, but the steps are more or less the same for other versions.

    Now, we will define OpenCV as a user library in Eclipse, so we can reuse the configuration for any project. Launch Eclipse and select Window –> Preferencesfrom the menu.

    Eclipse preferences

    Navigate under Java –> Build Path –> User Libraries and click New....

    Creating a new library

    Enter a name, e.g. OpenCV-2.4.6, for your new library.

    Naming the new library

    Now select your new user library and click Add External JARs....

    Adding external jar

    Browse through C:OpenCV-2.4.6uildjava and select opencv-246.jar. After adding the jar, extend the opencv-246.jar and select Native library location and pressEdit....

    Selecting native library location 1

    Select External Folder... and browse to select the folder C:OpenCV-2.4.6uildjavax64. If you have a 32-bit system you need to select the x86 folder instead ofx64.

    Selecting native library location 2

    Your user library configuration should look like this:

    Selecting native library location 2

    Testing the configuration on a new Java project

    Now start creating a new Java project.

    Creating new Java project

    On the Java Settings step, under Libraries tab, select Add Library... and select OpenCV-2.4.6, then click Finish.

    Adding user defined library 1Adding user defined library 2

    Libraries should look like this:

    Adding user defined library

    Now you have created and configured a new Java project it is time to test it. Create a new java file. Here is a starter code for your convenience:

    import org.opencv.core.Core;
    import org.opencv.core.CvType;
    import org.opencv.core.Mat;
    
    public class Hello
    {
       public static void main( String[] args )
       {
          System.loadLibrary( Core.NATIVE_LIBRARY_NAME );
          Mat mat = Mat.eye( 3, 3, CvType.CV_8UC1 );
          System.out.println( "mat = " + mat.dump() );
       }
    }
    

    When you run the code you should see 3x3 identity matrix as output.

    Adding user defined library

    That is it, whenever you start a new project just add the OpenCV user library that you have defined to your project and you are good to go. Enjoy your powerful, less painful development environment :)

    http://www.cnblogs.com/lidabo/p/3501285.html

    Opencv3.1.0+opencv_contrib配置及使用SIFT测试

    因为需要用到一些比较新的跟踪算法,这两天装了opencv3.1并配置了opencv_contrib,并使用了SIFT算法测试是否配置成功。 
    1.opencv3.1安装与配置 
    这里不多言,不熟悉的可以参考浅墨的博客:http://blog.csdn.net/poem_qianmo/article/details/19809337 
    2.opencv_contrib安装与配置 
    从opencv3以来,一些比较新的功能都挪到了“opencv_contrib”库里。配置这个库需要重新编译OpenCV,关于此部分可以参考教程:http://blog.csdn.net/linshuhe1/article/details/51221015 
    关于此教程需要补充两点:A,使用cmake编译的过程中经常会失败,因为国内网络问题ippicv_windows_20151201.zip 文件下载失败导致,可以直接从这里下载:http://download.csdn.net/detail/qjj2857/9495013 B.教程最后配置包含目录、库目录时没有提及添加环境变量,这里也是同样需要的。还有一切配置完成后别忘了重启电脑哟。 
    3.写个程序测试一下配置是否成功吧 
    opencv3.1中SIFT匹配是在opencv_contrib库中的,这里我们就用它来做一个简单的测试。 
    参考: 
    1. cv::xfeatures2d::SIFT Class Reference:http://docs.opencv.org/3.1.0/d5/d3c/classcv_1_1xfeatures2d_1_1SIFT.html#gsc.tab=0 
    2. OpenCV3.1 xfeatures2d::SIFT 使用:http://blog.csdn.net/lijiang1991/article/details/50855279 
    程序:

    #include <iostream>
    #include <opencv2/opencv.hpp>  //头文件
    #include <opencv2/xfeatures2d.hpp>
    using namespace cv;  //包含cv命名空间
    using namespace std;
    
    int main()
    {
        //Create SIFT class pointer
        Ptr<Feature2D> f2d = xfeatures2d::SIFT::create();
        //读入图片
        Mat img_1 = imread("1.jpg");
        Mat img_2 = imread("2.jpg");
        //Detect the keypoints
        vector<KeyPoint> keypoints_1, keypoints_2;
        f2d->detect(img_1, keypoints_1);
        f2d->detect(img_2, keypoints_2);
        //Calculate descriptors (feature vectors)
        Mat descriptors_1, descriptors_2;
        f2d->compute(img_1, keypoints_1, descriptors_1);
        f2d->compute(img_2, keypoints_2, descriptors_2);    
        //Matching descriptor vector using BFMatcher
        BFMatcher matcher;
        vector<DMatch> matches;
        matcher.match(descriptors_1, descriptors_2, matches);
        //绘制匹配出的关键点
        Mat img_matches;
        drawMatches(img_1, keypoints_1, img_2, keypoints_2, matches, img_matches);
        imshow("【match图】", img_matches);
        //等待任意按键按下
        waitKey(0);
    }
    • 1
    • 2
    • 3
    • 4
    • 5
    • 6
    • 7
    • 8
    • 9
    • 10
    • 11
    • 12
    • 13
    • 14
    • 15
    • 16
    • 17
    • 18
    • 19
    • 20
    • 21
    • 22
    • 23
    • 24
    • 25
    • 26
    • 27
    • 28
    • 29
    • 30
    • 31
    • 32
    • 1
    • 2
    • 3
    • 4
    • 5
    • 6
    • 7
    • 8
    • 9
    • 10
    • 11
    • 12
    • 13
    • 14
    • 15
    • 16
    • 17
    • 18
    • 19
    • 20
    • 21
    • 22
    • 23
    • 24
    • 25
    • 26
    • 27
    • 28
    • 29
    • 30
    • 31
    • 32

    原始图片: 
    这里写图片描述 
    这里写图片描述 
    匹配结果: 
    这里写图片描述

    ———————————-2016/8/12——————————— 
    1.关于Ubuntu下opencv3.1及opencv_contrib的安装与配置可参考: 
    官网:Installation in Linux 
    http://www.cnblogs.com/asmer-stone/p/5089764.html 
    上博文中有两点需要注意: 
    A.按上文参考所述,第3步build文件夹需建在~/opencv/opencv文件夹中;且cmake时按照作者示例OPENCV_EXTRA_MODULES_PATH=~/opencv/opencv_contrib/modules ,注意”<>”需要去掉;末尾的.. 表示opencv源码在上一级目录中。当然如果你了解cmake的使用方法cmake [optional] <opencv source directory>, 可以任意设置文件夹目录。 
    B.与在windows下相同,cmake时会因为“ippicv_linux_20151201.tgz 无法下载”而导致失败。我们可以从http://download.csdn.net/download/lx928525166/9479919下载,并放入相应文件夹中。 
    2. 好,现在假设你已经安装配置好了。由于在windows下我们习惯了用一个IDE来编程,这里在Ubuntu下我选择使用eclipse来作为编程环境,下边简单说一下怎么在eclipse中配置opencv。 
    首先参考官网: http://docs.opencv.org/3.1.0/d7/d16/tutorial_linux_eclipse.html 你就应该能配置的差不多了,或者其他类似的吧网上一大堆。 
    但是中间可能会出现一些小问题,我个人配置的时候出现了两个小问题: 
    A. 错误 undefined reference to symbol ‘_ZN2cv6imreadERKNS_6StringEi’ ,参考:http://answers.opencv.org/question/46755/first-example-code-error/ 
    B. 错误 error while loading shared libraries: libopencv_core.so.3.0: cannot open shared object file: No such file or directory ,参考:http://stackoverflow.com/questions/27907343/error-while-loading-shared-libraries-libopencv-core-so-3-0 
    3. 所有配置均完成后,在上述windows下的代码可以在这里直接运行。见下图: 
    这里写图片描述
    4.关于在ubuntu下运行其他samples程序。这里以cpp为例,直接找到opencv/samples/cpp/example_cmake,这里有一个示例已经提供了Makefile文件,make一下即可生成可执行文件。其他cpp示例文件类似。

    OpenCV3如何使用SIFT和SURF Where did SIFT and SURF go in OpenCV 3?

    If you’ve had a chance to play around with OpenCV 3 (and do a lot of work with keypoint 

    If you’ve had a chance to play around with OpenCV 3 (and do a lot of work with keypoint detectors and feature descriptors) you may have noticed that the SIFT and SURF implementations are no longer included in the OpenCV 3 library by default.

    Unfortunately, you probably learned this lesson the hard way by opening up a terminal, importing OpenCV, and then trying to instantiate your favorite keypoint detector, perhaps using code like the following:

    Oh no! There is no longer a cv2.FeatureDetector_create  method!

    The same is true for our cv2.DescriptorExtractor_create  function as well:

    Furthermore, cv2.SIFT_create  and cv2.SURF_create  will fail as well:

    I’ll be honest — this had me scratching my head at first. How am I supposed to access SIFT, SURF, and my other favorite keypoint detectors and local invariant descriptors ifcv2.FeatureDetector_create  and cv2.DescriptorExtractor_create  have been removed?

    The cv2.FeatureDetector_create  and cv2.DescriptorExtractor_create  were (and still are) methods I used all the time. And personally, I really liked the OpenCV 2.4.X implementation. All you needed to do was pass in a string and the factory method would build the instantiation for you. You could then tune the parameters using the getter and setter methods of the keypoint detector or feature descriptor.

    Furthermore, these methods have been part of OpenCV 2.4.X for many years. Why in the world were they removed from the default install? And where were they moved to?

    In the remainder of this blog post, I’ll detail why certain keypoint detectors and local invariant descriptors were removed from OpenCV 3.0 by default. And I’ll also show you where you can find SIFT, SURF, and other detectors and descriptors in the new version of OpenCV.

    Why were SIFT and SURF removed from the default install of OpenCV 3.0?

    SIFT and SURF are examples of algorithms that OpenCV calls “non-free” modules. These algorithms are patented by their respective creators, and while they are free to use in academic and research settings, you should technically be obtaining a license/permission from the creators if you are using them in a commercial (i.e. for-profit) application.

    With OpenCV 3 came a big push to move many of these “non-free” modules out of the default OpenCV install and into the opencv_contrib package. The opencv_contrib  packages contains implementations of algorithms that are either patented or in experimental development.

    The algorithms and associated implementations in  opencv_contrib  are not installed by default and you need to explicitly enable them when compiling and installing OpenCV to obtain access to them.

    Personally, I’m not too crazy about this move.

    Yes, I understand including patented algorithms inside an open source library may raise a few eyebrows. But algorithms such as SIFT and SURF are pervasive across much of computer vision. And more importantly, the OpenCV implementations of SIFT and SURF are used by academics and researchers daily to evaluate new image classification, Content-Based Image Retrieval, etc. algorithms. By not including these algorithms by default, more harm than good is done (at least in my opinion).

    How do I get access to SIFT and SURF in OpenCV 3?

    To get access to the original SIFT and SURF implementations found in OpenCV 2.4.X, you’ll need to pull down both the opencv and opencv_contrib repositories from GitHub and then compile and install OpenCV 3 from source.

    Luckily, compiling OpenCV from source is easier than it used to be. I have gathered install instructions for Python and OpenCV for many popular operating systems over on the OpenCV 3 Tutorials, Resources, and Guides page — just scroll down the Install OpenCV 3 and Pythonsection and find the appropriate Python version (either Python 2.7+ or Python 3+) for your operating system.

    How do I use SIFT and SURF with OpenCV 3?

    So now that you have installed OpenCV 3 with the opencv_contrib  package, you should have access to the original SIFT and SURF implementations from OpenCV 2.4.X, only this time they’ll be in the xfeatures2d  sub-module through the cv2.SIFT_create  andcv2.SURF_create  functions.

    To confirm this, open up a shell, import OpenCV, and execute the following commands (assuming you have an image named test_image.jpg  in your current directory, of course):

    If all goes well, you should be able to instantiate the SIFT and SURF keypoint detectors and local invariant descriptors without error.

    It’s also important to note that by using opencv_contrib  you will not be interfering with any of the other keypoint detectors and local invariant descriptors included in OpenCV 3. You’ll still be able to access KAZE, AKAZE, BRISK, etc. without an issue:

    Summary

    In this blog post we learned that OpenCV has removed the cv2.FeatureDetector_create  andcv2.DescriptorExtractor_create  functions from the library. Furthermore, the SIFT and SURF implementations have also been removed from the default OpenCV 3 install.

    The reason for SIFT and SURF removal is due to what OpenCV calls “non-free” algorithms. Both SIFT and SURF are patented algorithms, meaning that you should technically be getting permission to use them in commercial algorithms (they are free to use for academic and research purposes though).

    Because of this, OpenCV has made the decision to move patented algorithms (along with experimental implementations) to the opencv_contrib package. This means that to obtain access to SIFT and SURF, you’ll need to compile and install OpenCV 3 from source withopencv_contrib  support enabled. Luckily, this isn’t too challenging with the help of myOpenCV 3 install guides.

    Once you have installed OpenCV 3 with opencv_contrib  support you’ll be able to find your favorite SIFT and SURF implementations in the xfeatures2d  package through thecv2.xfeatures2d.SIFT_create()  and cv2.xfeatures2d.SURF_create()  functions.

    from: If you’ve had a chance to play around with OpenCV 3 (and do a lot of work with keypoint 




    http://www.pyimagesearch.com/2015/07/16/where-did-sift-and-surf-go-in-opencv-3/

    If you’ve had a chance to play around with OpenCV 3 (and do a lot of work with keypoint 

    http://blog.csdn.net/garfielder007/article/details/51260087

    opencv java api提取图片sift特征 - anexplore

    import org.opencv.core.Core;
    import org.opencv.core.Mat;
    import org.opencv.core.MatOfKeyPoint;
    import org.opencv.highgui.Highgui;
    import org.opencv.features2d.*;
    public class ExtractSIFT
    {
     public static void main( String[] args )
     {
        System.loadLibrary( Core.NATIVE_LIBRARY_NAME );
        Mat test_mat = Highgui.imread("pfau.jpg");
        Mat desc = new Mat();
        FeatureDetector fd = FeatureDetector.create(FeatureDetector.SIFT);
        MatOfKeyPoint mkp =new MatOfKeyPoint();
        fd.detect(test_mat, mkp);
        DescriptorExtractor de = DescriptorExtractor.create(DescriptorExtractor.SIFT);
        de.compute(test_mat,mkp,desc );//提取sift特征
        System.out.println(desc.cols());
        System.out.println(desc.rows());
     
     }
    }



    学习OpenCV——KeyPoint Matching 优化方式

    今天读Mastering OpenCV with Practical Computer Vision Projects 中的第三章里面讲到了几种特征点匹配的优化方式,在此记录。

    在图像特征点检测完成后(特征点检测参考:学习OpenCV——BOW特征提取函数(特征点篇)),就会进入Matching  procedure。

     

     

    1. OpenCV提供了两种Matching方式

    • Brute-force matcher (cv::BFMatcher) 

    • Flann-based matcher (cv::FlannBasedMatcher)

    Brute-force matcher就是用暴力方法找到点集一中每个descriptor在点集二中距离最近的descriptor;

    Flann-based matcher 使用快速近似最近邻搜索算法寻找(用快速的第三方库近似最近邻搜索算法)

    一般把点集一称为 train set (训练集)对应模板图像,点集二称为 query set(查询集)对应查找模板图的目标图像。

    为了提高检测速度,你可以调用matching函数前,先训练一个matcher。训练过程可以首先使用cv::FlannBasedMatcher来优化,为descriptor建立索引树,这种操作将在匹配大量数据时发挥巨大作用(比如在上百幅图像的数据集中查找匹配图像)。而Brute-force matcher在这个过程并不进行操作,它只是将train descriptors保存在内存中。

     

     

    2. 在matching过程可以使用cv::DescriptorMatcher的如下功能来进行匹配:

     

    • 简单查找最优匹配:void match(const Mat& queryDescriptors, vector<DMatch>& matches,const vector<Mat>& masks=vector<Mat>() );
    • 为每个descriptor查找K-nearest-matches:void knnMatch(const Mat& queryDescriptors, vector<vector<DMatch> >& matches, int k,const vector<Mat>&masks=vector<Mat>(),bool compactResult=false );
    • 查找那些descriptors间距离小于特定距离的匹配:void radiusMatch(const Mat& queryDescriptors, vector<vector<DMatch> >& matches, maxDistance, const vector<Mat>& masks=vector<Mat>(), bool compactResult=false );
     

     

    3. matching结果包含许多错误匹配,错误的匹配分为两种:

     

    • False-positive matches: 将非对应特征点检测为匹配(我们可以对他做文章,尽量消除它)
    • False-negative matches: 未将匹配的特征点检测出来(无法处理,因为matching算法拒绝)
    为了消除False-positive matches采用如下两种方式:
    • Cross-match filter:
    在OpenCV中 cv::BFMatcher class已经支持交叉验证,建立 cv::BFMatcher将第二参数声明为true
    cv::Ptr<cv::DescriptorMatcher> matcher(new cv::BFMatcher(cv::NORM_HAMMING,true));
    经过Cross-match filter的结果:
    • Ratio test
    使用KNN-matching算法,令K=2。则每个match得到两个最接近的descriptor,然后计算最接近距离和次接近距离之间的比值,当比值大于既定值时,才作为最终match。

     

    [cpp] view plain copy
     
     print?在CODE上查看代码片派生到我的代码片
    1. void PatternDetector::getMatches(const cv::Mat& queryDescriptors,  std::vector<cv::DMatch>& matches)  
    2. {  
    3.     matches.clear();  
    4.     if (enableRatioTest)  
    5.     {  
    6.         // To avoid NaNs when best match has   
    7.         // zero distance we will use inverse ratio.   
    8.         const float minRatio = 1.f / 1.5f;  
    9.         // KNN match will return 2 nearest   
    10.         // matches for each query descriptor  
    11.         m_matcher->knnMatch(queryDescriptors, m_knnMatches, 2);  
    12.         for (size_t i=0; i<m_knnMatches.size(); i++)  
    13.         {  
    14.             const cv::DMatch& bestMatch = m_knnMatches[i][0];  
    15.             const cv::DMatch& betterMatch = m_knnMatches[i][1];  
    16.             float distanceRatio = bestMatch.distance /   
    17.                 betterMatch.distance;  
    18.             // Pass only matches where distance ratio between   
    19.             // nearest matches is greater than 1.5   
    20.             // (distinct criteria)  
    21.             if (distanceRatio < minRatio)  
    22.             {  
    23.                 matches.push_back(bestMatch);  
    24.             }  
    25.         }  
    26.     }  
    27.     else  
    28.     {  
    29.         // Perform regular match  
    30.         m_matcher->match(queryDescriptors, matches);  
    31.     }  
    32. }   


     
    4. Homography estimation

     

    为了进一步提升匹配精度,可以采用随机样本一致性(RANSAC)方法。
    因为我们是使用一幅图像(一个平面物体),我们可以将它定义为刚性的,可以在pattern image和query image的特征点之间找到单应性变换(homography transformation)。使用cv::findHomography找到这个单应性变换,使用RANSAC找到最佳单应性矩阵。(由于这个函数使用的特征点同时包含正确和错误匹配点,因此计算的单应性矩阵依赖于二次投影的准确性)
    [cpp] view plain copy
     
     print?在CODE上查看代码片派生到我的代码片
    1. bool PatternDetector::refineMatchesWithHomography  
    2. (  
    3. const std::vector<cv::KeyPoint>& queryKeypoints,  
    4. const std::vector<cv::KeyPoint>& trainKeypoints,   
    5. float reprojectionThreshold,  
    6. std::vector<cv::DMatch>& matches,  
    7. cv::Mat& homography  
    8. )  
    9. {  
    10. const int minNumberMatchesAllowed = 8;  
    11. if (matches.size() < minNumberMatchesAllowed)  
    12. return false;  
    13. // Prepare data for cv::findHomography  
    14. std::vector<cv::Point2f> srcPoints(matches.size());  
    15. std::vector<cv::Point2f> dstPoints(matches.size());  
    16. for (size_t i = 0; i < matches.size(); i++)  
    17. {  
    18. srcPoints[i] = trainKeypoints[matches[i].trainIdx].pt;  
    19. dstPoints[i] = queryKeypoints[matches[i].queryIdx].pt;  
    20. }  
    21. // Find homography matrix and get inliers mask  
    22. std::vector<unsigned char> inliersMask(srcPoints.size());  
    23. homography = cv::findHomography(srcPoints,   
    24. dstPoints,   
    25. CV_FM_RANSAC,   
    26. reprojectionThreshold,   
    27. inliersMask);  
    28. std::vector<cv::DMatch> inliers;  
    29. for (size_t i=0; i<inliersMask.size(); i++)  
    30. {  
    31. if (inliersMask[i])  
    32. inliers.push_back(matches[i]);  
    33. }  
    34. matches.swap(inliers);  
    35. return matches.size() > minNumberMatchesAllowed;  
    36. }   
     
    经过单应性变换的过滤结果

    今天读Mastering OpenCV with Practical Computer Vision Projects 中的第三章里面讲到了几种特征点匹配的优化方式,在此记录。

    在图像特征点检测完成后(特征点检测参考:学习OpenCV——BOW特征提取函数(特征点篇)),就会进入Matching  procedure。

     

     

    1. OpenCV提供了两种Matching方式

    • Brute-force matcher (cv::BFMatcher) 

    • Flann-based matcher (cv::FlannBasedMatcher)

    Brute-force matcher就是用暴力方法找到点集一中每个descriptor在点集二中距离最近的descriptor;

    Flann-based matcher 使用快速近似最近邻搜索算法寻找(用快速的第三方库近似最近邻搜索算法)

    一般把点集一称为 train set (训练集)对应模板图像,点集二称为 query set(查询集)对应查找模板图的目标图像。

    为了提高检测速度,你可以调用matching函数前,先训练一个matcher。训练过程可以首先使用cv::FlannBasedMatcher来优化,为descriptor建立索引树,这种操作将在匹配大量数据时发挥巨大作用(比如在上百幅图像的数据集中查找匹配图像)。而Brute-force matcher在这个过程并不进行操作,它只是将train descriptors保存在内存中。

     

     

    2. 在matching过程可以使用cv::DescriptorMatcher的如下功能来进行匹配:

     

    • 简单查找最优匹配:void match(const Mat& queryDescriptors, vector<DMatch>& matches,const vector<Mat>& masks=vector<Mat>() );
    • 为每个descriptor查找K-nearest-matches:void knnMatch(const Mat& queryDescriptors, vector<vector<DMatch> >& matches, int k,const vector<Mat>&masks=vector<Mat>(),bool compactResult=false );
    • 查找那些descriptors间距离小于特定距离的匹配:void radiusMatch(const Mat& queryDescriptors, vector<vector<DMatch> >& matches, maxDistance, const vector<Mat>& masks=vector<Mat>(), bool compactResult=false );
     

     

    3. matching结果包含许多错误匹配,错误的匹配分为两种:

     

    • False-positive matches: 将非对应特征点检测为匹配(我们可以对他做文章,尽量消除它)
    • False-negative matches: 未将匹配的特征点检测出来(无法处理,因为matching算法拒绝)
    为了消除False-positive matches采用如下两种方式:
    • Cross-match filter:
    在OpenCV中 cv::BFMatcher class已经支持交叉验证,建立 cv::BFMatcher将第二参数声明为true
    cv::Ptr<cv::DescriptorMatcher> matcher(new cv::BFMatcher(cv::NORM_HAMMING,true));
    经过Cross-match filter的结果:
    • Ratio test
    使用KNN-matching算法,令K=2。则每个match得到两个最接近的descriptor,然后计算最接近距离和次接近距离之间的比值,当比值大于既定值时,才作为最终match。

     

    [cpp] view plain copy
     
     print?在CODE上查看代码片派生到我的代码片
    1. void PatternDetector::getMatches(const cv::Mat& queryDescriptors,  std::vector<cv::DMatch>& matches)  
    2. {  
    3.     matches.clear();  
    4.     if (enableRatioTest)  
    5.     {  
    6.         // To avoid NaNs when best match has   
    7.         // zero distance we will use inverse ratio.   
    8.         const float minRatio = 1.f / 1.5f;  
    9.         // KNN match will return 2 nearest   
    10.         // matches for each query descriptor  
    11.         m_matcher->knnMatch(queryDescriptors, m_knnMatches, 2);  
    12.         for (size_t i=0; i<m_knnMatches.size(); i++)  
    13.         {  
    14.             const cv::DMatch& bestMatch = m_knnMatches[i][0];  
    15.             const cv::DMatch& betterMatch = m_knnMatches[i][1];  
    16.             float distanceRatio = bestMatch.distance /   
    17.                 betterMatch.distance;  
    18.             // Pass only matches where distance ratio between   
    19.             // nearest matches is greater than 1.5   
    20.             // (distinct criteria)  
    21.             if (distanceRatio < minRatio)  
    22.             {  
    23.                 matches.push_back(bestMatch);  
    24.             }  
    25.         }  
    26.     }  
    27.     else  
    28.     {  
    29.         // Perform regular match  
    30.         m_matcher->match(queryDescriptors, matches);  
    31.     }  
    32. }   


     
    4. Homography estimation

     

    为了进一步提升匹配精度,可以采用随机样本一致性(RANSAC)方法。
    因为我们是使用一幅图像(一个平面物体),我们可以将它定义为刚性的,可以在pattern image和query image的特征点之间找到单应性变换(homography transformation)。使用cv::findHomography找到这个单应性变换,使用RANSAC找到最佳单应性矩阵。(由于这个函数使用的特征点同时包含正确和错误匹配点,因此计算的单应性矩阵依赖于二次投影的准确性)
    [cpp] view plain copy
     
     print?在CODE上查看代码片派生到我的代码片
    1. bool PatternDetector::refineMatchesWithHomography  
    2. (  
    3. const std::vector<cv::KeyPoint>& queryKeypoints,  
    4. const std::vector<cv::KeyPoint>& trainKeypoints,   
    5. float reprojectionThreshold,  
    6. std::vector<cv::DMatch>& matches,  
    7. cv::Mat& homography  
    8. )  
    9. {  
    10. const int minNumberMatchesAllowed = 8;  
    11. if (matches.size() < minNumberMatchesAllowed)  
    12. return false;  
    13. // Prepare data for cv::findHomography  
    14. std::vector<cv::Point2f> srcPoints(matches.size());  
    15. std::vector<cv::Point2f> dstPoints(matches.size());  
    16. for (size_t i = 0; i < matches.size(); i++)  
    17. {  
    18. srcPoints[i] = trainKeypoints[matches[i].trainIdx].pt;  
    19. dstPoints[i] = queryKeypoints[matches[i].queryIdx].pt;  
    20. }  
    21. // Find homography matrix and get inliers mask  
    22. std::vector<unsigned char> inliersMask(srcPoints.size());  
    23. homography = cv::findHomography(srcPoints,   
    24. dstPoints,   
    25. CV_FM_RANSAC,   
    26. reprojectionThreshold,   
    27. inliersMask);  
    28. std::vector<cv::DMatch> inliers;  
    29. for (size_t i=0; i<inliersMask.size(); i++)  
    30. {  
    31. if (inliersMask[i])  
    32. inliers.push_back(matches[i]);  
    33. }  
    34. matches.swap(inliers);  
    35. return matches.size() > minNumberMatchesAllowed;  
    36. }   
     
    经过单应性变换的过滤结果
  • 相关阅读:
    推荐]历史上最强的绕口令
    超级经理人的关系学:打造黄金人脉
    个人创业的难点和解决之道
    你的人脉关系中不可缺少的十种人[推荐]
    哲理短文一则:揭示最好的成功法则
    [经验交流]太精彩,太有启发性了(经典经典) 转
    权力领域是人才浪费的致命区域
    2006创业完全手册
    爱你我的宝贝(转)
    最远的你是我最近的爱
  • 原文地址:https://www.cnblogs.com/bnuvincent/p/6548735.html
Copyright © 2011-2022 走看看