1. 数学模型

2.整体流程

获取标定与图像数据==>stereoRectify==>initUndistortRectifyMap==>remap==>bg/sgbm恢复出视差图==>l利用数学模型求解深度图==》深度图相关的应用/点云为基础的应用。(2D==>3D的转换)

3. 接口解析

3.1

#include <opencv2/highgui/highgui.hpp>
#include <opencv2/core/core.hpp>
#include <opencv2/opencv.hpp>using namespace cv;
using namespace std;int blockSize = 0, uniquenessRatio = 0, numDisparities = 0;
Ptr<StereoBM> bm = StereoBM::create(16, 9);
Mat xyz;              //三维坐标Mat Q;//深度视差映射矩阵
Rect left_valid_roi, right_valid_roi;//图像校正之后,会对图像进行裁剪,这里的validRoi就是指裁剪之后的区域
Mat left_rectified_gray, right_rectified_gray;//左右摄像头校准后图像
Mat img;
Mat left_img, right_img;//左右摄像头原始图像
Mat left_rectified_img, right_rectified_img;//左右摄像头校准后图像Vec3f  point3;
float d;
Point origin;         //鼠标按下的起始点
Rect selection;      //定义矩形选框
bool selectObject = false;    //是否选择对象
int mindisparity = 0;
int ndisparities = 64;
int SADWindowSize = 5;void stereo_BM_match(int, void*)
{bm->setBlockSize(2 * blockSize + 5);     //SAD窗口大小,5~21之间为宜bm->setROI1(left_valid_roi);bm->setROI2(right_valid_roi);bm->setPreFilterCap(31);bm->setMinDisparity(0);  //最小视差,默认值为0, 可以是负值,int型bm->setNumDisparities(numDisparities * 16 + 16);//视差窗口,即最大视差值与最小视差值之差,窗口大小必须是16的整数倍,int型bm->setTextureThreshold(10);bm->setUniquenessRatio(uniquenessRatio);//uniquenessRatio主要可以防止误匹配bm->setSpeckleWindowSize(100);bm->setSpeckleRange(32);bm->setDisp12MaxDiff(-1);Mat disp, disp8;bm->compute(left_rectified_gray, right_rectified_gray, disp);//输入图像必须为灰度图disp.convertTo(disp8, CV_8UC1, 255 / ((numDisparities * 16 + 16) * 16.));//计算出的视差是CV_16S格式reprojectImageTo3D(disp, xyz, Q, true); //在实际求距离时,ReprojectTo3D出来的X / W, Y / W, Z / W都要乘以16(也就是W除以16),才能得到正确的三维坐标信息。xyz = xyz * 16;imshow("disparity_BM", disp8);
}void stereo_SGBM_match(int, void*)
{Mat disp;//SGBMcv::Ptr<cv::StereoSGBM> sgbm = cv::StereoSGBM::create(mindisparity, ndisparities, SADWindowSize);int P1 = 8 * left_rectified_img.channels() * SADWindowSize * SADWindowSize;int P2 = 10 * left_rectified_img.channels() * SADWindowSize * SADWindowSize;sgbm->setP1(P1);sgbm->setP2(P2);sgbm->setPreFilterCap(15);sgbm->setUniquenessRatio(10);sgbm->setSpeckleRange(2);sgbm->setSpeckleWindowSize(100);sgbm->setDisp12MaxDiff(1);//sgbm->setMode(cv::StereoSGBM::MODE_HH);sgbm->compute(left_rectified_img, right_rectified_img, disp);disp.convertTo(disp, CV_32F, 1.0 / 16);                //除以16得到真实视差值Mat disp8U = Mat(disp.rows, disp.cols, CV_8UC1);       //显示normalize(disp, disp8U, 0, 255, NORM_MINMAX, CV_8UC1);reprojectImageTo3D(disp8U, xyz, Q, true); //在实际求距离时,ReprojectTo3D出来的X / W, Y / W, Z / W都要乘以16(也就是W除以16),才能得到正确的三维坐标信息。xyz = xyz * 16;namedWindow("disparity_SGBM", WINDOW_AUTOSIZE);imshow("disparity_SGBM", disp8U);}/*void stereo_GC_match(int, void*)
{CvStereoGCState* state = cvCreateStereoGCState(16, 2);left_disp_ = cvCreateMat(left->height, left->width, CV_32F);right_disp_ = cvCreateMat(right->height, right->width, CV_32F);cvFindStereoCorrespondenceGC(left, right, left_disp_, right_disp_, state, 0);cvReleaseStereoGCState(&state);CvMat* disparity_left_visual = cvCreateMat(size.height, size.width, CV_8U); cvConvertScale(disparity_left, disparity_left_visual, -16);
}
*//*****描述:鼠标操作回调*****/
static void onMouse(int event, int x, int y, int, void*)
{if (selectObject){selection.x = MIN(x, origin.x);selection.y = MIN(y, origin.y);selection.width = std::abs(x - origin.x);selection.height = std::abs(y - origin.y);}switch (event){case EVENT_LBUTTONDOWN:   //鼠标左按钮按下的事件origin = Point(x, y);selection = Rect(x, y, 0, 0);selectObject = true;//cout << origin << "in world coordinate is: " << xyz.at<Vec3f>(origin) << endl;point3 = xyz.at<Vec3f>(origin);point3[0];//cout << "point3[0]:" << point3[0] << "point3[1]:" << point3[1] << "point3[2]:" << point3[2]<<endl;cout << "世界坐标:" << endl;cout << "x: " << point3[0] << "  y: " << point3[1] << "  z: " << point3[2] << endl;d = point3[0] * point3[0] + point3[1] * point3[1] + point3[2] * point3[2];d = sqrt(d);   //mm// cout << "距离是:" << d << "mm" << endl;d = d / 10.0;   //cmcout << "距离是:" << d << "cm" << endl;// d = d/1000.0;   //m// cout << "距离是:" << d << "m" << endl;break;case EVENT_LBUTTONUP:    //鼠标左按钮释放的事件selectObject = false;if (selection.width > 0 && selection.height > 0)break;}
}int main()
{VideoCapture cap(1);int FRAME_WIDTH = cap.get(CAP_PROP_FRAME_WIDTH);int FRAME_HEIGHT = cap.get(CAP_PROP_FRAME_HEIGHT);Mat aligned_rectified_img(FRAME_HEIGHT, FRAME_WIDTH, CV_8UC3);//校准+对齐后的图像cout << "Resolution:<" << cap.get(CAP_PROP_FRAME_WIDTH) << "," << cap.get(CAP_PROP_FRAME_HEIGHT) << ">\n";namedWindow("camera", WINDOW_AUTOSIZE);//    namedWindow("left_img",CV_WINDOW_NORMAL);//    namedWindow("right_img",CV_WINDOW_NORMAL);//    namedWindow("left_rectified_img",CV_WINDOW_NORMAL);//    namedWindow("right_rectified_img",CV_WINDOW_NORMAL);//    namedWindow("rectified_img",CV_WINDOW_NORMAL);namedWindow("aligned_rectified_img", WINDOW_AUTOSIZE);//resizeWindow("camera", 1200, 600);// resizeWindow("aligned_rectified_img", 1200, 600);Mat left_cameraMatrix = Mat::eye(3, 3, CV_64F);//左相机内参矩阵left_cameraMatrix.at<double>(0, 0) = 2.762165790037839e+02;//Fxleft_cameraMatrix.at<double>(0, 1) = 0;left_cameraMatrix.at<double>(0, 2) = 1.765880468329375e+02;//Cxleft_cameraMatrix.at<double>(1, 1) = 2.762317738515432e+02;//Fyleft_cameraMatrix.at<double>(1, 2) = 1.272320444598781e+02;//CyMat left_distCoeffs = Mat::zeros(5, 1, CV_64F);//畸变系数left_distCoeffs.at<double>(0, 0) = 0.065542106666972;//k1left_distCoeffs.at<double>(1, 0) = -0.099179821896270;//k2left_distCoeffs.at<double>(2, 0) = 0;//p1left_distCoeffs.at<double>(3, 0) = 0;//p2left_distCoeffs.at<double>(4, 0) = 0;Mat right_cameraMatrix = Mat::eye(3, 3, CV_64F);//右相机内参矩阵right_cameraMatrix.at<double>(0, 0) = 2.734695143541476e+02;//Fxright_cameraMatrix.at<double>(0, 1) = 0;right_cameraMatrix.at<double>(0, 2) = 1.724017536155269e+02;//Cxright_cameraMatrix.at<double>(1, 1) = 2.733548075965133e+02;//Fyright_cameraMatrix.at<double>(1, 2) = 1.255695004672240e+02;//CyMat right_distCoeffs = Mat::zeros(5, 1, CV_64F);//畸变系数right_distCoeffs.at<double>(0, 0) = 0.067619149443979;//k1right_distCoeffs.at<double>(1, 0) = -0.104286472771764;//k2right_distCoeffs.at<double>(2, 0) = 0;//p1right_distCoeffs.at<double>(3, 0) = 0;//p2right_distCoeffs.at<double>(4, 0) = 0;Mat rotation_matrix = Mat::zeros(3, 3, CV_64F);//旋转矩阵rotation_matrix.at<double>(0, 0) = 0.999997933684708;rotation_matrix.at<double>(0, 1) = 5.541735042905797e-04;rotation_matrix.at<double>(0, 2) = -0.001955893157243;rotation_matrix.at<double>(1, 0) = -5.560064711997943e-04;rotation_matrix.at<double>(1, 1) = 0.999999406695233;rotation_matrix.at<double>(1, 2) = -9.367315446314680e-04;rotation_matrix.at<double>(2, 0) = 0.001955372884999;rotation_matrix.at<double>(2, 1) = 9.378170983011573e-04;rotation_matrix.at<double>(2, 2) = 0.999997648505221;Mat rotation_vector;//旋转矩阵Rodrigues(rotation_matrix, rotation_vector);//旋转矩阵转化为旋转向量,罗德里格斯变换Mat translation_vector = Mat::zeros(3, 1, CV_64F);//平移向量translation_vector.at<double>(0, 0) = -74.303646210221200;translation_vector.at<double>(1, 0) = -0.208289299602746;translation_vector.at<double>(2, 0) = -1.203122420388863;Mat R1, R2;//左右相机的3x3矫正变换(旋转矩阵)Mat P1, P2;//左右相机新的坐标系统(矫正过的)输出 3x4 的投影矩阵// Mat Q;//深度视差映射矩阵// Rect left_valid_roi, right_valid_roi;//图像校正之后,会对图像进行裁剪,这里的validRoi就是指裁剪之后的区域Mat rmap[2][2];//映射表 必须为:CV_16SC2/CV_32FC1/CV_32FC2Size imageSize;imageSize = Size(FRAME_WIDTH >> 1, FRAME_HEIGHT);/*立体校正的时候需要两幅图像共面并且行对准 以使得立体匹配更加的可靠使得两幅图像共面的方法就是把两个摄像头的图像投影到一个公共成像面上,这样每幅图像从本图像平面投影到公共图像平面都需要一个旋转矩阵RstereoRectify 这个函数计算的就是从图像平面投影都公共成像平面的旋转矩阵R1,R2。 R1,R2即为左右相机平面行对准的校正旋转矩阵。左相机经过R1旋转,右相机经过R2旋转之后,两幅图像就已经共面并且行对准了。其中P1,P2为两个相机的投影矩阵,其作用是将3D点的坐标转换到图像的2D点的坐标:P*[X Y Z 1]' =[x y w]Q矩阵为重投影矩阵,即矩阵Q可以把2维平面(图像平面)上的点投影到3维空间的点:Q*[x y d 1] = [X Y Z W]。其中d为左右两幅图像的视差*///双目校准stereoRectify(left_cameraMatrix, left_distCoeffs,//左摄像头内参和畸变系数right_cameraMatrix, right_distCoeffs,//右摄像头内参和畸变系数imageSize, rotation_vector, translation_vector,//图像大小,右摄像头相对于左摄像头旋转矩阵,平移向量R1, R2, P1, P2, Q,//输出的参数CALIB_ZERO_DISPARITY, -1, imageSize, &left_valid_roi, &right_valid_roi);//Precompute maps for cv::remap().//用来计算畸变映射,输出的X / Y坐标重映射参数,remap把求得的映射应用到图像上。initUndistortRectifyMap(left_cameraMatrix, left_distCoeffs, R1, P1, imageSize, CV_16SC2, rmap[0][0], rmap[0][1]);initUndistortRectifyMap(right_cameraMatrix, right_distCoeffs, R2, P2, imageSize, CV_16SC2, rmap[1][0], rmap[1][1]);while (1){cap >> img;left_img = img(Rect(0, 0, FRAME_WIDTH >> 1, FRAME_HEIGHT));right_img = img(Rect(FRAME_WIDTH >> 1, 0, FRAME_WIDTH >> 1, FRAME_HEIGHT));//经过remap之后,左右相机的图像已经共面并且行对准了remap(left_img, left_rectified_img, rmap[0][0], rmap[0][1], INTER_LINEAR);remap(right_img, right_rectified_img, rmap[1][0], rmap[1][1], INTER_LINEAR);//立体匹配生成深度图需要用到灰度图cvtColor(left_rectified_img, left_rectified_gray, COLOR_BGR2GRAY);cvtColor(right_rectified_img, right_rectified_gray, COLOR_BGR2GRAY);//复制左相机校准图像到总图像上for (int i = 0; i < left_rectified_img.rows; i++){for (int j = 0; j < left_rectified_img.cols; j++){aligned_rectified_img.at<Vec3b>(i, j)[0] = left_rectified_img.at<Vec3b>(i, j)[0];aligned_rectified_img.at<Vec3b>(i, j)[1] = left_rectified_img.at<Vec3b>(i, j)[1];aligned_rectified_img.at<Vec3b>(i, j)[2] = left_rectified_img.at<Vec3b>(i, j)[2];}}//复制右相机校准图像到总图像上for (int i = 0; i < right_rectified_img.rows; i++){for (int j = 0; j < right_rectified_img.cols; j++){aligned_rectified_img.at<Vec3b>(i, j + (FRAME_WIDTH >> 1))[0] = right_rectified_img.at<Vec3b>(i, j)[0];aligned_rectified_img.at<Vec3b>(i, j + (FRAME_WIDTH >> 1))[1] = right_rectified_img.at<Vec3b>(i, j)[1];aligned_rectified_img.at<Vec3b>(i, j + (FRAME_WIDTH >> 1))[2] = right_rectified_img.at<Vec3b>(i, j)[2];}}//红色框出校正后的左右图像rectangle(img, left_valid_roi, Scalar(0, 0, 255), 2, 8);rectangle(img, Rect(right_valid_roi.x + (FRAME_WIDTH >> 1), right_valid_roi.y, right_valid_roi.width, right_valid_roi.height), Scalar(0, 0, 255), 2, 8);//做绿色线增强对比for (int i = 0; i < aligned_rectified_img.rows; i += 32){line(aligned_rectified_img, Point(0, i), Point(aligned_rectified_img.cols, i), Scalar(0, 255, 0), 1, LINE_8);}imshow("camera", img);imshow("left_img",left_img);imshow("right_img",right_img);imshow("left_rectified_img",left_rectified_img);imshow("right_rectified_img",right_rectified_img);imshow("aligned_rectified_img", aligned_rectified_img);stereo_BM_match(0, 0);stereo_SGBM_match(0, 0);//创建SAD窗口 TrackbarcreateTrackbar("BlockSize:\n", "disparity_BM", &blockSize, 8, stereo_BM_match);// 创建视差唯一性百分比窗口 TrackbarcreateTrackbar("UniquenessRatio:\n", "disparity_BM", &uniquenessRatio, 50, stereo_BM_match);// 创建视差窗口 TrackbarcreateTrackbar("NumDisparities:\n", "disparity_BM", &numDisparities, 16, stereo_BM_match);createTrackbar("SADWindowSize:\n", "disparity_SGBM", &SADWindowSize, 30, stereo_SGBM_match);setMouseCallback("disparity_SGBM", onMouse, 0);setMouseCallback("disparity_BM", onMouse, 0);int key = waitKey(30);if (key == 27)//按下Esc退出return 0;}
}
#include <iostream>
#include <fstream>
#include <map>
#include <vector>
#include <cmath>#include <opencv2/highgui.hpp>
#include <opencv2/calib3d.hpp>
#include <opencv2/imgproc.hpp>using namespace std;//****这里设置三种颜色来让程序运行的更醒目****
#define WHITE "\033[37m"
#define BOLDRED "\033[1m\033[31m"
#define BOLDGREEN "\033[1m\033[32m"//****相机内参****
struct CAMERA_INTRINSIC_PARAMETERS
{double cx,cy,fx,fy,baseline,scale;
};struct FILE_FORM
{cv::Mat left,right;//存放左右视图数据string dispname,depthname,colorname;//存放三种输出数据的文件名
};//****读取配置文件****
class ParameterReader
{public:ParameterReader(string filename = "./parameters.txt")//配置文件目录{ifstream fin(filename.c_str());if(!fin){cerr<<BOLDRED"can't find parameters file!"<<endl;return;}while(!fin.eof()){string str;getline(fin,str);if(str[0] == '#')//遇到’#‘视为注释{continue;}int pos = str.find("=");//遇到’=‘将其左右值分别赋给key和alueif (pos == -1){continue;}string key = str.substr(0,pos);string value = str.substr(pos+1,str.length());data[key] = value;if (!fin.good()){break;}}}string getData(string key)//获取配置文件参数值{map<string,string>::iterator iter = data.find(key);if (iter == data.end()){cerr<<BOLDRED"can't find:"<<key<<" parameters!"<<endl;return string("NOT_FOUND");}return iter->second;}
public:map<string,string> data;
};FILE_FORM readForm(int index,ParameterReader pd);//存入当前序列左右视图数据和三种输出结果文件名
void stereoSGBM(cv::Mat lpng,cv::Mat rpng,string filename,cv::Mat &disp);//SGBM方法获取视差图
void stereoBM(cv::Mat lpng,cv::Mat rpng,string filename,cv::Mat &disp);//BM方法获取视差图
void disp2Depth(cv::Mat disp,cv::Mat &depth,CAMERA_INTRINSIC_PARAMETERS camera);//由视察图计算深度图int main(int argc, char const *argv[])
{ParameterReader pd;//读取配置文件CAMERA_INTRINSIC_PARAMETERS camera;//相机内参结构赋值camera.fx = atof(pd.getData("camera.fx").c_str());camera.fy = atof(pd.getData("camera.fy").c_str());camera.cx = atof(pd.getData("camera.cx").c_str());camera.cy = atof(pd.getData("camera.cy").c_str());camera.baseline = atof(pd.getData("camera.baseline").c_str());camera.scale = atof(pd.getData("camera.scale").c_str());int startIndex = atoi(pd.getData("start_index").c_str());//起始序列int endIndex = atoi(pd.getData("end_index").c_str());//截止序列bool is_color = pd.getData("is_color") == string("yes");//判断是否要输出彩色深度图cout<<BOLDRED"......START......"<<endl;for (int currIndex = startIndex;currIndex < endIndex;currIndex++)//从起始序列循环至截止序列{cout<<BOLDGREEN"Reading file: "<<currIndex<<endl;FILE_FORM form = readForm(currIndex,pd);//获取当前序列的左右视图以及输出结果文件名cv::Mat disp,depth,color;//****判断使用何种算法计算视差图****if (pd.getData("algorithm") == string("SGBM")){stereoSGBM(form.left,form.right,form.dispname,disp);}else if (pd.getData("algorithm") == string("BM")){stereoBM(form.left,form.right,form.dispname,disp);}else{cout<<BOLDRED"Algorithm is wrong!"<<endl;return 0;}disp2Depth(disp,depth,camera);//输出深度图cv::imwrite(form.depthname,depth);cout<<WHITE"Depth saved!"<<endl;//****判断是否输出彩色深度图****if (is_color){cv::applyColorMap(depth,color,cv::COLORMAP_JET);//转彩色图cv::imwrite(form.colorname,color);cout<<WHITE"Color saved!"<<endl;}}cout<<BOLDRED"......END......"<<endl;return 0;
}FILE_FORM readForm(int index,ParameterReader pd)
{FILE_FORM f;string lpngDir = pd.getData("left_dir");//获取左视图输入目录名string rpngDir = pd.getData("right_dir");//获取右视图输入目录名string dispDir = pd.getData("disp_dir");//获取视差图输出目录名string depthDir = pd.getData("depth_dir");//获取深度图输出目录名string colorDir = pd.getData("color_dir");//获取彩色深度图输出目录名string rgbExt = pd.getData("rgb_extension");//获取图片数据格式后缀名//输出当前文件序号(使用的TUM数据集,其双目视图命名从000000至004540,详情参看博文末尾ps)string numzero;if ( index >= 0 && index <= 9 ){numzero = "00000";}else if ( index >= 10 && index <= 99 ){numzero = "0000";}else if ( index >= 100 && index <= 999 ){numzero = "000";}else if ( index >= 1000 && index <= 9999 ){numzero = "00";}else if ( index >= 10000 && index <= 99999 ){numzero = "0";}//获取左视图文件名stringstream ss;ss<<lpngDir<<numzero<<index<<rgbExt;string filename;ss>>filename;f.left = cv::imread(filename,0);//这里要获取单通道数据//获取右视图文件名ss.clear();filename.clear();ss<<rpngDir<<numzero<<index<<rgbExt;ss>>filename;f.right = cv::imread(filename,0);//这里要获取单通道数据//获取深度图输出文件名ss.clear();filename.clear();ss<<depthDir<<index<<rgbExt;ss>>filename;f.depthname = filename;//获取视差图输出文件名ss.clear();filename.clear();ss<<dispDir<<index<<rgbExt;ss>>filename;f.dispname = filename;//获取彩色深度图输出文件名ss.clear();filename.clear();ss<<colorDir<<index<<rgbExt;ss>>filename;f.colorname = filename;return f;
}void stereoSGBM(cv::Mat lpng,cv::Mat rpng,string filename,cv::Mat &disp)
{disp.create(lpng.rows,lpng.cols,CV_16S);cv::Mat disp1 = cv::Mat(lpng.rows,lpng.cols,CV_8UC1);cv::Size imgSize = lpng.size();cv::Ptr<cv::StereoSGBM> sgbm = cv::StereoSGBM::create();int nmDisparities = ((imgSize.width / 8) + 15) & -16;//视差搜索范围int pngChannels = lpng.channels();//获取左视图通道数int winSize = 5;sgbm->setPreFilterCap(31);//预处理滤波器截断值sgbm->setBlockSize(winSize);//SAD窗口大小sgbm->setP1(8*pngChannels*winSize*winSize);//控制视差平滑度第一参数sgbm->setP2(32*pngChannels*winSize*winSize);//控制视差平滑度第二参数sgbm->setMinDisparity(0);//最小视差sgbm->setNumDisparities(nmDisparities);//视差搜索范围sgbm->setUniquenessRatio(5);//视差唯一性百分比sgbm->setSpeckleWindowSize(100);//检查视差连通区域变化度的窗口大小sgbm->setSpeckleRange(32);//视差变化阈值sgbm->setDisp12MaxDiff(1);//左右视差图最大容许差异sgbm->setMode(cv::StereoSGBM::MODE_SGBM);//采用全尺寸双通道动态编程算法sgbm->compute(lpng,rpng,disp);disp.convertTo(disp1,CV_8U,255 / (nmDisparities*16.));//转8位cv::imwrite(filename,disp1);cout<<WHITE"Disp saved!"<<endl;
}void stereoBM(cv::Mat lpng,cv::Mat rpng,string filename,cv::Mat &disp)
{disp.create(lpng.rows,lpng.cols,CV_16S);cv::Mat disp1 = cv::Mat(lpng.rows,lpng.cols,CV_8UC1);cv::Size imgSize = lpng.size();cv::Rect roi1,roi2;cv::Ptr<cv::StereoBM> bm = cv::StereoBM::create(16,9);int nmDisparities = ((imgSize.width / 8) + 15) & -16;//视差搜索范围bm->setPreFilterType(CV_STEREO_BM_NORMALIZED_RESPONSE);//预处理滤波器类型bm->setPreFilterSize(9);//预处理滤波器窗口大小bm->setPreFilterCap(31);//预处理滤波器截断值bm->setBlockSize(9);//SAD窗口大小bm->setMinDisparity(0);//最小视差bm->setNumDisparities(nmDisparities);//视差搜索范围bm->setTextureThreshold(10);//低纹理区域的判断阈值bm->setUniquenessRatio(5);//视差唯一性百分比bm->setSpeckleWindowSize(100);//检查视差连通区域变化度窗口大小bm->setSpeckleRange(32);//视差变化阈值bm->setROI1(roi1);bm->setROI2(roi2);bm->setDisp12MaxDiff(1);//左右视差图最大容许差异bm->compute(lpng,rpng,disp);disp.convertTo(disp1,CV_8U,255 / (nmDisparities*16.));cv::imwrite(filename,disp1);cout<<WHITE"Disp saved!"<<endl;
}void disp2Depth(cv::Mat disp,cv::Mat &depth,CAMERA_INTRINSIC_PARAMETERS camera)
{depth.create(disp.rows,disp.cols,CV_8UC1);cv::Mat depth1 = cv::Mat(disp.rows,disp.cols,CV_16S);for (int i = 0;i < disp.rows;i++){for (int j = 0;j < disp.cols;j++){if (!disp.ptr<ushort>(i)[j])//防止除0中断continue;depth1.ptr<ushort>(i)[j] = camera.scale * camera.fx * camera.baseline / disp.ptr<ushort>(i)[j];}}depth1.convertTo(depth,CV_8U,1./256);//转8位
}

3.2 stereoRectify

void stereoRectify( InputArray cameraMatrix1, InputArray distCoeffs1,InputArray cameraMatrix2, InputArray distCoeffs2,Size imageSize, InputArray R, InputArray T,OutputArray R1, OutputArray R2,OutputArray P1, OutputArray P2,OutputArray Q, int flags = CALIB_ZERO_DISPARITY,double alpha = -1, Size newImageSize = Size(),CV_OUT Rect* validPixROI1 = 0, CV_OUT Rect* validPixROI2 = 0 );
参数:R,tvec: 是由相机1变换到相机2的变换矩阵,按照视觉这边的习惯,一般世界坐标系下的点变换到相机坐标系下的点
R1,R2:对两个相机进行旋转矫正的旋转变换矩阵,为世界系到两个相机的变换
P1,P2:两个矫正畸变后的相机投影矩阵,相对的坐标系还是原来的相机1所在坐标系
flags: 一般CALIB_ZERO_DISPARITY
alpha: -1或者不设置则使用自动剪切,0的时候没有黑边,1的时候保留所有原图像素,会有黑边
newImageSize: 默认与原图相同,所以会有剪切,该参数需要跟initUndistortRectifyMap相同,设置大一些会在大畸变的时候保留更多细节.
使用默认参数,即自动裁剪,保留原始图像大小,结果是图像自动缩放,无黑边。
使用alpha=1.0时,原图所有像素都被保留,即不裁剪,保留黑边,但是图像会缩放,结果中P矩阵中K明显进行了缩放。
alpha: 0.0 表示去除黑边后的最大四边形,1.0表示保留原图所有像素,即保留黑边

undistort()函数:

主要针对单张图像进行去畸变操作,使用默认参数的时候主要控制参数是利用newCameraMatrix来完成,而newCameraMatrix一般由getOptimalNewCameraMatrix()函数得到,getOptimalNewCameraMatrix()函数可以控制的参数有:

alpha: 0.0 表示去除黑边后的最大四边形,1.0表示保留原图所有像素,即保留黑边
newImageSize:不设置的话,默认与原图相等,设置了的话会在alpha作用之后对图像进行缩放
总的来说alpha和newImageSize是两个互不干扰的参数,alpha只管是否对图像进行裁剪,而newImageSize只负责把图像进行缩放,这二者都会对相机参数造成影响.

主要完成对双目图像进行调整,计算出用于立体校正的参数,即:给定两个相机的K、畸变参数、外参,计算出相应的R1,R2,P1,P2 矩阵. 作用如公式所示:

在initUndistortRectifyMap的时候,对归一化平面上的点坐标进行旋转的修正,效果如下图,P1,P2矩阵与给定的alpha和newImageSize参数有关,alpha只控制有无黑边(即有效像素)newImageSize控制图像缩放.


3.3 reprojectImageTo3D函数

该函数将视差图,通过投影矩阵Q,得到一副映射图,图像大小与视差图相同,且每个像素具有三个通道,分别存储了该像素位置在相机坐标系下的三维点坐标在x, y, z,三个轴上的值,即每个像素的在相机坐标系下的三维坐标。

void cv::reprojectImageTo3D( InputArray disparity,   //视差图像OutputArray _3dImage, //映射后存储三维坐标的图像InputArray Q,         //重投影矩阵 通过stereoRectify得到bool handleMissingValues = false, //计算得到的非正常值是否给值,如果为true则给值10000int ddepth = -1          //输出类型 -1 即默认为CV_32FC3 还可以是 CV_16S, CV_32S, CV_32F
)

4. 参考资料

  1. https://stackoverflow.com/questions/22418846/reprojectimageto3d-in-opencv
  2. https://docs.opencv.org/3.4/d9/d0c/group__calib3d.html#ga617b1685d4059c6040827800e72ad2b6
  3. https://www.coder.work/article/114127

【OpenCV】双目相机测距及其深度恢复原理及其算法流程相关推荐

  1. 双目相机计算稠密深度点云(一)

    双目相机计算稠密深度点云 1.双目立体匹配原理 1.1 图像矫正 1.2 视差计算 2.elas_ros 包运行 3.KITTI数据集运行 3.1 kitti数据集转换为rosbag 3.2 运行KI ...

  2. OpenCV双目相机拍照及图片分割

    最近在做双目相机测距及三维重建,我从tb买了一个双目相机,第一步需要通过棋盘图来对双目相机进行标定,由于双目相机拍出来的左右相机的图片是一张图,需要进行分割. #include<iostream ...

  3. 双目相机计算稠密深度点云(二)

    双目相机计算稠密深度点云 1.ZED相机驱动包 2.校准ZED相机 3.运行ELAS 参考资料 接着上一篇的文章实现利用真实的相机计算3D点云,实验中我们使用的是ZED的相机. 1.ZED相机驱动包 ...

  4. 双目相机测距代码演示

    双目测距的操作流程有四步:相机标定--双目校正--双目匹配--计算深度,具体内容参考 : https://blog.csdn.net/qq_38236355/article/details/88933 ...

  5. OpenCV双目相机的标定C++

    代码来和数据集来自于"小白公众号" 链接:https://pan.baidu.com/s/1bYrj57MEdhwWp45XXVbdpA 提取码:b3m6 双目相机标定的过程: 1 ...

  6. OpenCV | 双目相机标定之OpenCV获取左右相机图像+MATLAB单目标定+双目标定

    博主github:https://github.com/MichaelBeechan 博主CSDN:https://blog.csdn.net/u011344545 原本网上可以搜到很多关于双目相机标 ...

  7. OpenCV 双目相机录制的视频,左右图像分割

    双目相机输出将左右摄像头拼接在一起输出 思路: 拿opencv读取视频每一帧(33ms) 拿Rect分割一下就好了 将left right图像保存到不同文件(这里我将提前保存好的时间戳与采集的图像一一 ...

  8. 崔岩的笔记——粒子滤波原理及应用(3)粒子滤波原理及算法流程

    崔岩的笔记--粒子滤波原理及应用(1)概率论与数理统计基础_今天也是睡觉的一天的博客-CSDN博客 崔岩的笔记--粒子滤波原理及应用(2)蒙特卡洛法与贝叶斯滤波_今天也是睡觉的一天的博客-CSDN博客 ...

  9. opencv双目相机标定-示例代码分析

      在这里我使用的是Learning OpenCV3的示例,本节使用的项目代码可以在这里下载到. 一.运行示例   在下载完整个工程以后,按照工程使用说明,下载配置Opencv,运行VS2019项目即 ...

最新文章

  1. C++异常 返回错误码
  2. dbvis 数据库连接工具-更新数据库驱动方法示例演示,驱动与数据库版本不匹配问题:Unknown system variable ‘query_cache_size‘解决方法
  3. db2怎么限定查询条数_如何查询各国进口关税税率!
  4. 新年寄语 | 2018 以及 Oracle 18c 一个时代的开启
  5. 基于XML的AOP配置
  6. HttpClient4.5.2 使用cookie保持会话——连接池的实现结束篇(4)
  7. 驱动设计ARM(6410)-按键驱动0基础知识点
  8. Mac 使用rz sz 命令
  9. 自由软件运动与GNU项目
  10. android relativelayout 垂直居中,RelativeLayout子控件垂直居中
  11. MySQL重复数据排序_排序数据以在MySQL中重复记录
  12. 在线思维导图的制作教程分享,帮你快速掌握绘制要领
  13. 网站托管服务器配置,了解托管服务器的三个小技巧
  14. RTKlib软件学习(观测文件与星历文件读取)
  15. 安然公司特殊目的实体(SPEs)解读
  16. STC12C5A60S2 双串口
  17. javafx自定义分页控件的实现
  18. Android Studio运行app启动模拟器一直卡在“Wating for target device to come online”的解决办法
  19. 短视频课堂丨传统工厂企业短视频打造之“三好人设”
  20. win10锁屏聚集图片不更新,右上角没有出现喜欢的按钮,怎么办?

热门文章

  1. 提升工作效率的软件及网站(不断更新中)
  2. 中枢神经系统解剖结构图,神经系统的解剖结构
  3. YFX云分享软件库驾培不可或缺的教学工具。
  4. 我的世界服务器自定义怪物怎么用,我的世界完全自定义怪物教程攻略
  5. pmp-相关方权利/利益方格
  6. ubuntu 刷新频率 如何查看_Ubuntu分辨率和刷新频率设置
  7. 算法设计与分析第一章作业
  8. 音乐专业如何利用计算机思维,利用信息技术培养学生音乐创造思维能力
  9. 服装ERP软件首要的好处都有哪些?
  10. 自旋锁(spin lock)学习