首页 > 其他分享 >AnswerOpenCV一周佳作欣赏(0615-0622)

AnswerOpenCV一周佳作欣赏(0615-0622)

时间:2022-12-26 23:07:44浏览次数:60  
标签:Mat img image cv 0622 alpha 0615 AnswerOpenCV CV

i'm using OpenCV for Android.
I would like to know,how to make image correction(auto adjustments of brightness/contrast) for image(bitmap) in android via OpenCV or that can be done by native ColorMatrixFilter from Android!??

I tried to google,but didn't found good tutorials/examples.
So how can i achieve my goal? Any ideas?
Thanks!


算法问题,需要寻找到自动调整亮度和对比度的算法。

回答:

Brightness and contrast is linear operator with parameter ​​alpha​​​ and ​​beta​



​= alpha * I(x,y) +​



In OpenCV you can do this with ​​convertTo​​.

The question here is how to calculate alpha and ​beta​ automatically ?

Looking at histogram, alpha operates as color range amplifier, beta operates as range shift.

Automatic brightness and contrast optimization calculates alpha and beta so that the output range is 0..255.

​​= max(I) 
- min(I)



wanted output range
=
255;



alpha
= output range
/ input range
=
255
/ ( max(I)
- min(I) )


min(O)
= alpha
* min(I)
+ beta

beta
=
-min(I)
*​​


Histogram Wings Cut (clipping)

To maximize the result of it's useful to cut out some color with few pixels.

This is done cutting left right and wings of histogram where color frequency is less than a value (typically 1%). Calculating cumulative distribution from the histogram you can easly find where to cut.

may be sample chart helps to understand:


AnswerOpenCV一周佳作欣赏(0615-0622)_Image


By the way ​​BrightnessAndContrastAuto​​​ could be named ​normalizeHist

Histogram equalization and CLAE works only on gray images and they change grays level balancing. look at the images below:



AnswerOpenCV一周佳作欣赏(0615-0622)_Image_02



也就是实现了彩色的颜色增强算法

​​void BrightnessAndContrastAuto(      const cv      :
:Mat
&src, cv
:
:Mat
&dst,
float clipHistPercent)
{
CV_Assert(clipHistPercent >
=
0);

CV_Assert((src.type() == CV_8UC1)
|| (src.type()
== CV_8UC3)
|| (src.type()
== CV_8UC4));

int histSize =
256;

float alpha, beta;
double minGray =
0, maxGray
=
0;

//to calculate grayscale histogram

cv :
:Mat gray;

if (src.type() == CV_8UC1) gray
= src;

else if (src.type() == CV_8UC3) cvtColor(src, gray, CV_BGR2GRAY);

else if (src.type() == CV_8UC4) cvtColor(src, gray, CV_BGRA2GRAY);

if (clipHistPercent ==
0)

{
// keep full available range

cv :
:minMaxLoc(gray,
&minGray,
&maxGray);

}
else

{
cv :
:Mat hist;
//the grayscale histogram

float range[] = {
0,
256 };

const float
* histRange
= { range };

bool uniform =
true;
bool accumulate =
false;
calcHist( &gray,
1,
0, cv
:
:Mat (), hist,
1,
&histSize,
&histRange, uniform, accumulate);

// calculate cumulative distribution from the histogram

std :
:vector
<
float
> accumulator(histSize);

accumulator[ 0]
= hist.at
<
float
>(
0);

for ( int i =
1; i
< histSize; i
++)

{
accumulator[i] = accumulator[i
-
1]
+ hist.at
<
float
>(i);

}
// locate points that cuts at required value

float max = accumulator.back();

clipHistPercent *= (max
/
100.
0);
//make percent as absolute

clipHistPercent /=
2.
0;
// left and right wings

// locate left cut

minGray =
0;

while (accumulator[minGray] < clipHistPercent)

minGray ++;

// locate right cut

maxGray = histSize
-
1;

while (accumulator[maxGray] >
= (max
- clipHistPercent))

maxGray --;

}
// current range

float inputRange = maxGray
- minGray;

alpha = (histSize
-
1)
/ inputRange;
// alpha expands current range to histsize range

beta =
-minGray
* alpha;
// beta shifts current range so that minGray will go to 0

// Apply brightness and contrast normalization

// convertTo operates with saurate_cast

src.convertTo(dst, -
1, alpha, beta);

// restore alpha channel from source

if (dst.type() == CV_8UC4)

{
int from_to[] = {
3,
3};

cv :
:mixChannels(
&src,
4,
&dst,
1, from_to,
1);

}
return;
}


AnswerOpenCV一周佳作欣赏(0615-0622)_Image_03



AnswerOpenCV一周佳作欣赏(0615-0622)_#include_04


 



二、 Template matching behavior - Color


I am evaluating template matching algorithm to differentiate similar and dissimilar objects. What I found is confusing, I had an impression of template matching is a method which compares raw pixel intensity values. Hence when the pixel value varies I expected Template Matching to give a less match percentage.

I have a template and search image having same shape and size differing only in color(Images attached). When I did template matching surprisingly I am getting match percentage greater than 90%.


​​                  img 
= cv2.imread(
'./images/searchtest.png', cv2.IMREAD_COLOR)template
= cv2.imread(
'./images/template.png', cv2.IMREAD_COLOR)


res = cv2.matchTemplate(img, template, cv2.TM_CCORR_NORMED)

min_val, max_val, min_loc, max_loc = cv2.minMaxLoc(res) print(max_val)
​​



​ Template Image :

AnswerOpenCV一周佳作欣赏(0615-0622)_#include_05


Search Image :

AnswerOpenCV一周佳作欣赏(0615-0622)_Image_06


Can someone give me an insight why it is happening so? I have even tried this in HSV color space, Full BGR image, Full HSV image, Individual channels of B,G,R and Individual channels of H,S,V. In all the cases I am getting a good percentage. Any help could be really appreciated.

AnswerOpenCV一周佳作欣赏(0615-0622)_#include_07



  这个问题的核心就是彩色模板匹配。很有趣的问题。回答直接给出了source code, ​​https://github.com/LaurentBerger/ColorMatchTemplate​​ 关于这个问题,我认为,彩色模板匹配的意义不是很大,毕竟用于定位的时候,黑白效果更好。 三、 I am using Brute Force Matcher with L2 norm. Referring this link ​​https://docs.opencv.org/2.4/doc/tutor...​​After the process, I get below image as output

AnswerOpenCV一周佳作欣赏(0615-0622)_2d_08



What is the likely position of the object suggested by the feature matching?

I don't understand how to choose the likely position using this image :(

这是一个只知其一不知其二的问题,他能够找到旋转的地方,但是对于下一步如何做没有思路。其实,解决此类问题,最好的方法就是参考教程做一遍。

当然,管理员的回答非常明确:

to retrieve the position of your matched object, you need ​​some further steps​​ :

  • filter the matches for outliers
  • extract the 2d point locations from the keypoints
  • apply ​​findHomography()​​ on the matched 2d points to get a transformation matrix between your query and the scene image
  • apply ​​perspectiveTransform​​ on the boundingbox of the query object, to see, where it is located in the scene image.

我也给出具体回答:


AnswerOpenCV一周佳作欣赏(0615-0622)_2d_09


​​        //used surf      ​​

​​
//


#include "stdafx.h"


#include <iostream>


#include "opencv2/core/core.hpp"


#include "opencv2/imgproc/imgproc.hpp"


#include "opencv2/features2d/features2d.hpp"


#include "opencv2/highgui/highgui.hpp"


#include "opencv2/nonfree/features2d.hpp"


#include "opencv2/calib3d/calib3d.hpp"


using namespace std;


using namespace cv;


int main( int argc, char** argv )


{





Mat img_1 ;


Mat img_2 ;


Mat img_raw_1 = imread("c1.bmp");


Mat img_raw_2 = imread("c3.bmp");


cvtColor(img_raw_1,img_1,CV_BGR2GRAY);


cvtColor(img_raw_2,img_2,CV_BGR2GRAY);


//-- Step 1: 使用SURF识别出特征点


int minHessian = 400;


SurfFeatureDetector detector( minHessian );


std::vector<KeyPoint> keypoints_1, keypoints_2;


detector.detect( img_1, keypoints_1 );


detector.detect( img_2, keypoints_2 );


//-- Step 2: 描述SURF特征


SurfDescriptorExtractor extractor;


Mat descriptors_1, descriptors_2;


extractor.compute( img_1, keypoints_1, descriptors_1 );


extractor.compute( img_2, keypoints_2, descriptors_2 );


//-- Step 3: 匹配


FlannBasedMatcher matcher;//BFMatcher为强制匹配


std::vector< DMatch > matches;


matcher.match( descriptors_1, descriptors_2, matches );


//取最大最小距离


double max_dist = 0; double min_dist = 100;


for( int i = 0; i < descriptors_1.rows; i++ )


{


double dist = matches[i].distance;


if( dist < min_dist ) min_dist = dist;


if( dist > max_dist ) max_dist = dist;


}


std::vector< DMatch > good_matches;


for( int i = 0; i < descriptors_1.rows; i++ )


{


if( matches[i].distance <= 3*min_dist )//这里的阈值选择了3倍的min_dist


{


good_matches.push_back( matches[i]);


}


}


//-- Localize the object from img_1 in img_2


std::vector<Point2f> obj;


std::vector<Point2f> scene;


for( int i = 0; i < (int)good_matches.size(); i++ )


{


//这里采用“帧向拼接图像中添加的方法”,因此左边的是scene,右边的是obj


scene.push_back( keypoints_1[ good_matches[i].queryIdx ].pt );


obj.push_back( keypoints_2[ good_matches[i].trainIdx ].pt );


}


//直接调用ransac,计算单应矩阵


Mat H = findHomography( obj, scene, CV_RANSAC );


//图像对准


Mat result;


warpPerspective(img_raw_2,result,H,Size(2*img_2.cols,img_2.rows));


Mat half(result,cv::Rect(0,0,img_2.cols,img_2.rows));


img_raw_1.copyTo(half);


imshow("result",result);


waitKey(0);


return 0;


}
​​

四、 Run OpenCV to MFC

  I have reproduced ​​this​​ sample, in a MFC app.


The cv::Mat is a CDOcument variable member:


​// Attributes
public :
std : :vector <CBlob > m_blobs;
cv : :Mat m_Mat; ​


and I draw rectangles over image, with an method (called at some time intervals, according FPS rate):


​ DrawBlobInfoOnImage(m_blobs, m_Mat); ​


Here is the code of this method:


​​void CMyDoc       :       :DrawBlobInfoOnImage(std       :
:vector
<CBlob
>
& blobs, cv
:
:Mat
& Mat)

{
for ( unsigned int i = 0;i < blobs.size();
++i)

{
if (blobs[i].m_bStillBeingTracked)
{
cv : :rectangle(Mat, blobs[i].m_rectCurrentBounding, SCALAR_RED, 2);

double dFontScale = blobs[i].m_dCurrentDiagonalSize / 60.
0;

int nFontThickness = ( int)roundf(dFontScale * 1. 0);

cv : :putText(Mat, (LPCTSTR)IntToString(i), blobs[i].m_vecPointCenterPositions.back(), CV_FONT_HERSHEY_SIMPLEX, dFontScale, SCALAR_GREEN, nFontThickness);
}
}
} ​​


but the result of this method is something like that:

AnswerOpenCV一周佳作欣赏(0615-0622)_#include_10

My question is: how can I draw only the last blobs result over my image ?

I have tried to clean up m_Mat, and to enable to draw only blobs.size() - 1 blob over image, none of this worked ...


非常有意思的问题,主要就是说他能够用mfc调用oepncv了,但是不知道如何将视频中的框子正确显示(也就是不要有拖尾)

这个也是对问题思考不是很深造成的问题。其实,解决的方法无外乎两个

一是直接修改视频流,也就是在原来的“读取-显示”循环里面添加一些东西(也就是框子),如果是这种方法,你使用或者不使用mfc基本没有什么影响;

比如<​​PariticalFilter在MFC上的运行,源代码公开​​>

AnswerOpenCV一周佳作欣赏(0615-0622)_2d_11

二是采用mfc的机制。mfc不是都是对话框吗?那就创建一个窗体,专门用来显示这个矩形就好啦。 

比如<​​GreenOpenPaint的实现(五)矩形框​​>​

AnswerOpenCV一周佳作欣赏(0615-0622)_#include_12



 

五、How to find the thickness of the red color sealent in the image ????


Hi,

I want to find the thickness of the red colored sealent in the image.

First I'm extracting the sealent portion by using findcontours by having minimum and maximum conotur area.And then check the Area,length and thickness of the sealent,i can find the area as well as length but i m not able to find the thickness of the sealent portion.

Please help me guys........below is the example image.

AnswerOpenCV一周佳作欣赏(0615-0622)_Image_13


提示一下:

do a ​​distance transform​​​ and a ​​skeleton​​ on the binary image.

这是一个算法问题,具体解决,下周给出实现。大家可以先研究一下。

 


标签:Mat,img,image,cv,0622,alpha,0615,AnswerOpenCV,CV
From: https://blog.51cto.com/jsxyhelu2017/5971115

相关文章

  • answerOpenCV轮廓类问题解析
    contour在opencv中是一个基础的数据结构,灵活运用的话,作用很大。以contour为关键字,在answerOpenCV中能够发现很多有趣的东西。 1、无法解决的问题​​......
  • AnswerOpenCV学习_Opencv multiple circle detection in a image
    ​​https://answers.opencv.org/question/234920/opencv-multiple-circle-detection-in-a-image/​​​​​​原图如上,目标是从这副图片中寻找“细胞”......
  • AnswerOpenCV(0416)一周佳作欣赏
    2017年4月16日一周AnswerOpenCV佳作赏析1、HelloHowtosmoothedgeoftextinbinaryimage,basedonthreshold.Somethinglikeanti-aliasingbyopenCv?example......
  • 20220615 Servlet 相关的 Spring Boot 自动配置类
    自动配置类Servlet相关的匹配的自动配置类:org.springframework.boot.autoconfigure.web.servlet.DispatcherServletAutoConfigurationorg.springframework.boot.autoc......
  • Tomcat-CVE-2017-20615
    环境搭建:切换到vulhub对应目录docker-composebuilddocker-composeup-d漏洞产生原因Tomcat配置问题,配置了可写(readonly=false),导致我们可以往服务器写文件。  ......
  • 软件工程 张雯雯 223201062202
    2.1实验过程2.1.1实验代码给出博客园截图和网址,并附上github的代码地址。GitHub代码地址: 2.1.2实验过程(1) 本人角色自己在本次结对项目中所担任角色,以及结对伙......
  • 软件工程 张雯雯223201062202
    请阅读北航陈彦吉同学的这篇博客中的各参考资料,并回答如下问题:(1)回顾你过去将近3年的学习经历当初你报考的时候,是真正喜欢软件工程这个专业吗?你现在后悔选择了这个专......