在Android上使用FAST检测器进行实时目标识别



这项工作的目的是从视频场景中的现有图像中提取实时关键点processFrame由于匹配而无法执行。我可以用圆圈的形式在图像上实时显示对应的关键点

class Sample1View extends SampleViewBase {
    public static final int     VIEW_MODE_RGBA   = 0;
    public static final int  VIEW_MODE_BLUE  = 1; 
    public static final int  VIEW_MODE_YELLOW = 2;
    public static final int  VIEW_MODE_DE = 3;
    private Mat mYuv;
    private Mat mRgba;
    private Mat mGraySubmat;
    private Mat mResult;
    private Mat mIntermediateMat;
 private Bitmap mBitmap;
 private int mViewMode;
 private Mat mColor;
 private Mat mHsv;
 TimingLogger timings;
   private Mat img1;
    private Mat descriptors;
    private MatOfKeyPoint keypoints;
    private FeatureDetector detector;
    private DescriptorExtractor descriptor;
    private DescriptorMatcher matcher;
 private static final String TAG ="Sample::View";
    public Sample1View(Context context) {
        super(context);
        mViewMode = VIEW_MODE_RGBA;
 try {
 img1=Utils.loadResource(getContext(), R.drawable.wings);
 } catch (IOException e) {
 // TODO Auto-generated catch block
 Log.w("Activity::LoadResource","Unable to load resource R.drawable.wings");
 e.printStackTrace();
 }
        descriptors = new Mat();
        keypoints = new MatOfKeyPoint();
        detector = FeatureDetector.create(FeatureDetector.FAST);
        detector.detect(img1, keypoints);
        descriptor = DescriptorExtractor.create(DescriptorExtractor.ORB);
        descriptor.compute(img1, keypoints, descriptors);
        matcher = DescriptorMatcher.create(DescriptorMatcher.BRUTEFORCE_HAMMING);
    }
 @Override
 protected void onPreviewStarted(int previewWidth, int previewHeight) {
     Log.i(TAG, "preview Started");
     synchronized (this) {
         mYuv = new Mat(getFrameHeight() + getFrameHeight() / 2, getFrameWidth(), CvType.CV_8UC1);
         mGraySubmat = mYuv.submat(0, getFrameHeight(), 0, getFrameWidth());
         mRgba = new Mat();
         mIntermediateMat = new Mat();
         mBitmap = Bitmap.createBitmap(previewWidth, previewHeight, Bitmap.Config.ARGB_8888); 
         mHsv = new Mat();
         mColor = new Mat();
         mResult = new Mat();
         }
 }
 @Override
 protected void onPreviewStopped() {
     Log.i(TAG, "preview Stopped");
     if(mBitmap != null) {
   mBitmap.recycle();
  }
  synchronized (this) {
            // Explicitly deallocate Mats
            if (mYuv != null)
                mYuv.release();
            if (mRgba != null)
                mRgba.release();
            if (mGraySubmat != null)
                mGraySubmat.release();
            if (mIntermediateMat != null)
                mIntermediateMat.release();
            mYuv = null;
            mRgba = null;
            mGraySubmat = null;
            mIntermediateMat = null;
            if (mResult != null)
             mResult.release();
            if (mHsv != null)
             mHsv.release();
            if (mColor != null)
             mColor.release();
            mColor = null;
         mResult = null;
         mHsv = null;
        }
    }

//cvt_YUVtoRGBtoHSV:

    @Override
    protected Bitmap processFrame(byte[] data) {
        mYuv.put(0, 0, data);
        final int viewMode = mViewMode;
                 ColorDetection.cvt_YUVtoRGBtoHSV(mYuv,mGraySubmat);
      MatOfKeyPoint mKeyPoints = new MatOfKeyPoint();
    MatOfDMatch  matches = new MatOfDMatch();
      detector.detect(mGraySubmat, mKeyPoints);
      descriptor.compute(mGraySubmat, mKeyPoints, mIntermediateMat);
      matcher.match(mIntermediateMat, descriptors, matches);
      mIntermediateMat2.create(resultSize, CvType.CV_8UC1);
      Features2d.drawMatches(mGraySubmat, mKeyPoints, mGraySubmat, mKeyPoints, matches, mIntermediateMat2);
      Imgproc.resize(mIntermediateMat2, mIntermediateMat2, mRgba.size());
      Imgproc.cvtColor(mIntermediateMat2, mRgba, Imgproc.COLOR_RGBA2BGRA, 4);
        break;
        }
    Bitmap bmp = mBitmap;
    try {
      Utils.matToBitmap(mRgba, bmp);
    } catch(Exception e) {
        Log.e("org.opencv.samples.*", "Utils.matToBitmap() throws an exception: " + e.getMessage());
        bmp.recycle();
        bmp = null;
    }
    return bmp;
    }
    public void setViewMode(int viewMode) {
     mViewMode = viewMode;
    }
}

和日志

CvException [org.opencv.core。CvException:/home/reports/ci/slave/50-SDK/opencv/modules/features2d/src/draw.cpp:207: error: (-215) i>= 0 &&i1 & lt;static_cast(keypoints1.size())在函数void cv::drawMatches(const cv::Mat&, const std::vector&, const std::vector&, cv::Mat&, const Scalar&, const std::vector&, int)

这篇文章有点老了,但我还是给了一个答案。

你的drawMatches函数的前四个参数不正确。前两个和后两个是一样的。
你可以改变drawMatches的前四个参数(mGraySubmat, mKeyPoints, mGraySubmat, mKeyPoints,…);

& # 160; & # 160; & # 160; & # 160; (mGraySubmat、mKeyPoints img1,要点、…);

& # 160; & # 160; & # 160; & # 160; (img1,要点,mGraySubmat mKeyPoints,…),
使用适合你的那个。


对于其他读者,如果他们有同样的问题,他们可以尝试用后两个参数改变前两个参数。
例:
   (image1, keypointsImage1, image2, keypointsImage2,…);

   (image2, keypointsImage2, image1, keypointsImage1,…);

相关内容

  • 没有找到相关文章