摘要:在ROS kinetic下,利用realsense D435深度相机采集校准的RGBD图片,合成点云,在rviz中查看点云,最后保存成pcd文件。

一、 各种bug

代码编译成功后,打开rviz添加pointcloud2选项卡,当我订阅合成点云时,可视化失败,选项卡报错:

1)Data size (9394656 bytes) does not match width (640) times height (480) times point_step (32).  Dropping message.

解读:通过  rostopic echo /pointcloud_topic  读取相机节点发布的原始点云的相关数据,可以发现每一帧原始点云的点数量为307200。合成点云的点数量为   /  ,约26万。可以猜测由于某种原因,系统把每一帧合成点云的数据都丢弃了。

原因:我事先给定合成点云的大小为,height = 480,width = 640. 然而在合成点云的过程中,剔除了部分违法值(d=0),由此导致合成点云的点数量与指定的点数量不匹配,合成点云的数据因此被丢弃。

解决方法:采用如下方法给定点云大小, cloud->height = ; cloud->width = cloud->points.size();

2)transform xxxxx;

解读:通过  rostopic echo /pointcloud_topic  ,发现原始点云数据具有如下信息,

header:
seq: 50114
stamp:
secs: 1528439158
nsecs: 557543003
frame_id: "camera_color_optical_frame"

由此推断,合成点云缺失参考坐标系header.frame_id。点云中点的XYZ属性是相对于某个坐标系来描述的,因此,需要指定点云的参考坐标系。

解决方法:添加点云的header信息,

pub_pointcloud.header.frame_id = "camera_color_optical_frame";
pub_pointcloud.header.stamp = ros::Time::now();

3)将合成的点云存储成pcd文件时遇到如下报错:

[ INFO] [1528442016.931874649]: point cloud size = 0
terminate called after throwing an instance of 'pcl::IOException'
what(): : [pcl::PCDWriter::writeASCII] Input point cloud has no data!
Aborted (core dumped)

经过多方查找,摸索了一步trick,分享给大家。真实报错原因仍未查明,期待前辈的指点

高博的源代码如下所示,里面的cloud是pcl的数据类型,

pcl::io::savePCDFile( "./pointcloud.pcd", *cloud );  。

我的源代码如下面所示,先通过 pcl::toROSMsg() 将pcl的数据类型转换成ros的数据类型,再写入pcd中,即可跳过报错。

4)相机内参

由于在彩色图和深度图配准的过程中,选用的是彩色图的坐标系,因此在合成点云(像素坐标在变换到相机坐标)时应该选用彩色图的相机内参。

realsense官方提供了一个应用程序可以查看所有视频流的内参。

gordon@gordon-:/usr/local/bin$ ./rs-sensor-control 

如下所示

84 : Color #0 (Video Stream: Y16 640x480@ 60Hz)
85 : Color #0 (Video Stream: BGRA8 640x480@ 60Hz)
86 : Color #0 (Video Stream: RGBA8 640x480@ 60Hz)
87 : Color #0 (Video Stream: BGR8 640x480@ 60Hz)
88 : Color #0 (Video Stream: RGB8 640x480@ 60Hz)
89 : Color #0 (Video Stream: YUYV 640x480@ 60Hz)
90 : Color #0 (Video Stream: Y16 640x480@ 30Hz)
91 : Color #0 (Video Stream: BGRA8 640x480@ 30Hz)
92 : Color #0 (Video Stream: RGBA8 640x480@ 30Hz)
93 : Color #0 (Video Stream: BGR8 640x480@ 30Hz)
94 : Color #0 (Video Stream: RGB8 640x480@ 30Hz)
95 : Color #0 (Video Stream: YUYV 640x480@ 30Hz)
96 : Color #0 (Video Stream: Y16 640x480@ 15Hz)
97 : Color #0 (Video Stream: BGRA8 640x480@ 15Hz)
98 : Color #0 (Video Stream: RGBA8 640x480@ 15Hz)
99 : Color #0 (Video Stream: BGR8 640x480@ 15Hz)
100: Color #0 (Video Stream: RGB8 640x480@ 15Hz)
101: Color #0 (Video Stream: YUYV 640x480@ 15Hz)
102: Color #0 (Video Stream: Y16 640x480@ 6Hz)
103: Color #0 (Video Stream: BGRA8 640x480@ 6Hz)
104: Color #0 (Video Stream: RGBA8 640x480@ 6Hz)
105: Color #0 (Video Stream: BGR8 640x480@ 6Hz)
106: Color #0 (Video Stream: RGB8 640x480@ 6Hz)
107: Color #0 (Video Stream: YUYV 640x480@ 6Hz)

5)深度图从ROS的数据类型转换为CV的数据类型

参看链接

 

二、程序代码

#include <ros/ros.h>
#include <image_transport/image_transport.h>
#include <opencv2/highgui/highgui.hpp>
#include <cv_bridge/cv_bridge.h>
#include <sensor_msgs/image_encodings.h>
#include <sensor_msgs/PointCloud2.h> // PCL 库
#include <pcl/io/pcd_io.h>
#include <pcl/point_types.h>
#include <pcl/visualization/pcl_visualizer.h>
#include <pcl_conversions/pcl_conversions.h> // 定义点云类型
typedef pcl::PointCloud<pcl::PointXYZRGB> PointCloud; using namespace std;
//namespace enc = sensor_msgs::image_encodings; // 相机内参
const double camera_factor = ;
const double camera_cx = 321.798;
const double camera_cy = 239.607;
const double camera_fx = 615.899;
const double camera_fy = 616.468; // 全局变量:图像矩阵和点云
cv_bridge::CvImagePtr color_ptr, depth_ptr;
cv::Mat color_pic, depth_pic; void color_Callback(const sensor_msgs::ImageConstPtr& color_msg)
{
//cv_bridge::CvImagePtr color_ptr;
try
{
cv::imshow("color_view", cv_bridge::toCvShare(color_msg, sensor_msgs::image_encodings::BGR8)->image);
color_ptr = cv_bridge::toCvCopy(color_msg, sensor_msgs::image_encodings::BGR8); cv::waitKey(); // 不断刷新图像,频率时间为int delay,单位为ms
}
catch (cv_bridge::Exception& e )
{
ROS_ERROR("Could not convert from '%s' to 'bgr8'.", color_msg->encoding.c_str());
}
color_pic = color_ptr->image; // output some info about the rgb image in cv format
cout<<"output some info about the rgb image in cv format"<<endl;
cout<<"rows of the rgb image = "<<color_pic.rows<<endl;
cout<<"cols of the rgb image = "<<color_pic.cols<<endl;
cout<<"type of rgb_pic's element = "<<color_pic.type()<<endl;
} void depth_Callback(const sensor_msgs::ImageConstPtr& depth_msg)
{
//cv_bridge::CvImagePtr depth_ptr;
try
{
//cv::imshow("depth_view", cv_bridge::toCvShare(depth_msg, sensor_msgs::image_encodings::TYPE_16UC1)->image);
//depth_ptr = cv_bridge::toCvCopy(depth_msg, sensor_msgs::image_encodings::TYPE_16UC1);
cv::imshow("depth_view", cv_bridge::toCvShare(depth_msg, sensor_msgs::image_encodings::TYPE_32FC1)->image);
depth_ptr = cv_bridge::toCvCopy(depth_msg, sensor_msgs::image_encodings::TYPE_32FC1); cv::waitKey();
}
catch (cv_bridge::Exception& e)
{
ROS_ERROR("Could not convert from '%s' to 'mono16'.", depth_msg->encoding.c_str());
} depth_pic = depth_ptr->image; // output some info about the depth image in cv format
cout<<"output some info about the depth image in cv format"<<endl;
cout<<"rows of the depth image = "<<depth_pic.rows<<endl;
cout<<"cols of the depth image = "<<depth_pic.cols<<endl;
cout<<"type of depth_pic's element = "<<depth_pic.type()<<endl;
}
int main(int argc, char **argv)
{
ros::init(argc, argv, "image_listener");
ros::NodeHandle nh;
cv::namedWindow("color_view");
cv::namedWindow("depth_view");
cv::startWindowThread();
image_transport::ImageTransport it(nh);
image_transport::Subscriber sub = it.subscribe("/camera/color/image_raw", , color_Callback);
image_transport::Subscriber sub1 = it.subscribe("/camera/aligned_depth_to_color/image_raw", , depth_Callback);
ros::Publisher pointcloud_publisher = nh.advertise<sensor_msgs::PointCloud2>("generated_pc", );
// 点云变量
// 使用智能指针,创建一个空点云。这种指针用完会自动释放。
PointCloud::Ptr cloud ( new PointCloud );
sensor_msgs::PointCloud2 pub_pointcloud; double sample_rate = 1.0; // 1HZ,1秒发1次
ros::Rate naptime(sample_rate); // use to regulate loop rate cout<<"depth value of depth map : "<<endl; while (ros::ok()) {
// 遍历深度图
for (int m = ; m < depth_pic.rows; m++){
for (int n = ; n < depth_pic.cols; n++){
// 获取深度图中(m,n)处的值
float d = depth_pic.ptr<float>(m)[n];//ushort d = depth_pic.ptr<ushort>(m)[n];
// d 可能没有值,若如此,跳过此点
if (d == )
continue;
// d 存在值,则向点云增加一个点
pcl::PointXYZRGB p; // 计算这个点的空间坐标
p.z = double(d) / camera_factor;
p.x = (n - camera_cx) * p.z / camera_fx;
p.y = (m - camera_cy) * p.z / camera_fy; // 从rgb图像中获取它的颜色
// rgb是三通道的BGR格式图,所以按下面的顺序获取颜色
p.b = color_pic.ptr<uchar>(m)[n*];
p.g = color_pic.ptr<uchar>(m)[n*+];
p.r = color_pic.ptr<uchar>(m)[n*+]; // 把p加入到点云中
cloud->points.push_back( p );
}
} // 设置并保存点云
cloud->height = ;
cloud->width = cloud->points.size();
ROS_INFO("point cloud size = %d",cloud->width);
cloud->is_dense = false;// 转换点云的数据类型并存储成pcd文件
pcl::toROSMsg(*cloud,pub_pointcloud);
pub_pointcloud.header.frame_id = "camera_color_optical_frame";
pub_pointcloud.header.stamp = ros::Time::now();
pcl::io::savePCDFile("./pointcloud.pcd", pub_pointcloud);
cout<<"publish point_cloud height = "<<pub_pointcloud.height<<endl;
cout<<"publish point_cloud width = "<<pub_pointcloud.width<<endl; // 发布合成点云和原始点云
pointcloud_publisher.publish(pub_pointcloud);
ori_pointcloud_publisher.publish(cloud_msg); // 清除数据并退出
cloud->points.clear(); ros::spinOnce(); //allow data update from callback;
naptime.sleep(); // wait for remainder of specified period;
} cv::destroyWindow("color_view");
cv::destroyWindow("depth_view");
}
05-11 17:27