<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>S. Wang); zheng.rong@nlpr.ia.ac.cn (Z. Rong); pengju.zhang@nlpr.ia.ac.cn (P. Zhang);
yhwu@nlpr.ia.ac.cn (Y. Wu)</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Accurate Extrinsic Calibration of LiDAR and Camera with Refined Vertex Features</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Shuo Wang</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Zheng Rong</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Pengju Zhang</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Yihong Wu</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>National Laboratory of Pattern Recognition, Institute of Automation, Chinese Academy of Sciences</institution>
          ,
          <addr-line>Beijing</addr-line>
          ,
          <country country="CN">China</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>School of Artificial Intelligence, University of Chinese Academy of Sciences</institution>
          ,
          <addr-line>Beijing</addr-line>
          ,
          <country country="CN">China</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2022</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0003</lpage>
      <abstract>
        <p>LiDARs and cameras are widely used in many research fields, due to the complementarity of their data. Calibrating the extrinsic parameters of LiDAR frame and camera frame is essential for fusing the two kinds of data. Calibration methods based on a calibration board extract geometry features from point clouds and images, then build geometry constraints to estimate the extrinsic parameters. In this paper, we exhaustively analyze the measurement characteristics of LiDARs that would introduce notable negative efect on the calibration, including the noise of depth measurement and the divergence of laser beams. Therefore, we propose a refining method for vertex features from LiDAR point clouds using the prior information of the board, which can efectively mitigate the efect of systematic measurement errors and improve the accuracy of the calibration results. Meanwhile, our calibration method reduces the minimal number of calibration datasets to one, which promotes the eficiency of calibration processes. Besides, we propose an objective and independent evaluation method for target-based calibration methods. Extensive experiments and comparisons with state-of-the-art methods show that using refined vertex features can notably improve the accuracy and eficiency of extrinsic parameter calibration.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Extrinsic Calibration</kwd>
        <kwd>Vertex Feature</kwd>
        <kwd>Calibration Evaluation</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>LiDARs and cameras are indispensable and ubiquitous in positioning and navigation, which
are used in many fields of applications, such as autonomous drivings and industrial robotics.
LiDARs measure the structure of surroundings in the form of point clouds and cameras measure
the texture of surroundings in the form of images. The information fusion of LiDAR and camera
can strengthen the images with accurate depth information and provide color information for
point clouds. The prerequisite of information fusion is the calibration of extrinsic parameters
between a LiDAR frame and a camera frame.</p>
      <p>
        Calibration board-based methods are the most popular calibration methods nowadays [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2,
3, 4, 5, 6, 7, 8</xref>
        ]. In these methods, one or more boards are deployed in a static environment. By
extracting and matching geometry features of the calibration board from point cloud and image,
the constraints of extrinsic parameters between LiDAR and camera can be constructed, and the
extrinsic parameters can be estimated using these constraints.
      </p>
      <p>In most of existing calibration methods, geometry features are extracted directly from point
clouds and refined with fitting algorithms, such as plane fitting algorithms and line fitting
algorithms, to reduce sensor noises. But these fitting methods only focus on the geometry
characteristics of calibration boards and ignore the measurement characteristics of LiDARs. In
this paper, we analyze the measurement characteristics of LiDARs and the systematic errors
raised by these characteristics, including the noise of range measurements and the divergence
of LiDAR beams. Then we propose a refining method for vertex features to reduce the negative
impact aroused by the systematic errors. We calibrate the extrinsic parameters of LiDAR and
camera using the refined vertex features to improve the accuracy of calibration results. Due to
the high accuracy of refined vertex features, our proposed method can calibrate the extrinsic
parameters using only one calibration dataset.</p>
      <p>To evaluate calibration results, some papers compare their estimated extrinsic parameters with
the ground truth, for example, simulated values. Some papers calculate the re-projection errors
of geometry features for evaluation, such as point features or line features from calibration
boards. However, in real world, the ground truth is hard to get and diferent kinds of
reprojection errors are not holistic measurements for two 3D frames. To evaluate our calibration
method efectively and compare our method with other calibration methods fairly, we propose
an evaluation method using raw LiDAR measurements and 3D space distances of point-to-line
that are logically independent of any method used in the calibration process.</p>
      <p>The contributions of this paper are as follows:
1) A refining method for vertex features extracted from LiDAR point clouds yielding notable
accuracy improvement of calibration results;
2) A fair and independent evaluation method using 3D space distances;
3) Extensive experiments and comparisons between our proposed method and other
state-ofthe-art calibration methods.</p>
      <p>The rest of this paper is organized as follows. Section 2 reviews calibration methods. Section 3
analyzes measurement characteristics of LiDARs, proposes a refining method for vertex features,
and introduces a calibration method based on refined vertex features. An evaluation method
using 3D space distances is also presented in this part. Section 4 shows the experimental results.
Section 5 concludes this paper.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        Extrinsic calibration methods between LiDAR and camera can be classified as three kinds,
target-based calibration methods, target-less calibration methods and motion-based calibration
methods. In target-based calibration methods, there are two kinds of targets: calibration
boards [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2, 3, 4, 5, 6, 7, 8</xref>
        ] and other artificial markers that are diferent from typical calibration
boards, such as a board with diferent patterns like a circle [ 9, 10], or a target with diferent
shapes like a ball [11, 12]. The calibration boards are main calibration targets. With explicit
data association between LiDAR and camera, the calibration board-based methods usually use
the corresponding relationship between features extracted from point clouds and images to
estimate the extrinsic parameters.
      </p>
      <p>
        Calibration board-based methods can be classified as plane-based, edge-based and
pointbased methods, according to their used features. The plane-based constraint uses the plane
features of calibration boards. This kind constraint can be constructed by the points on the
board extracted from point clouds and plane parameters calculated from images in the camera
frame [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2, 3</xref>
        ]. Another way to construct the plane-based constraint is to compute the plane
parameters from point clouds [
        <xref ref-type="bibr" rid="ref1">1, 4, 8</xref>
        ]. The edge-based constraint uses the edge features of
calibration boards. For point clouds, a board edge can be modeled as a series of points on the
edge [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] or line parameters fitted by these points [
        <xref ref-type="bibr" rid="ref2">2, 4, 5</xref>
        ]. For cameras, a board edge can be
modeled as a back-projected plane[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] or a 3D line in the camera frame [
        <xref ref-type="bibr" rid="ref2">2, 4, 5</xref>
        ]. The point-based
constraint uses the point features of board vertices. The vertices in the LiDAR frame can
be computed by the points on the calibration board. From images, the projection of board
vertices can be extracted [6, 7], and the coordinates of board vertices in the camera frame can
be derived according to the relative pose between the calibration board and the camera [4, 8].
The calibration board-based methods usually perform better than other calibration methods
because the extraction and matching of features are explicit and clear.
      </p>
      <p>Target-less calibration methods depend on the information extracted from the
environment rather than artificial targets. Structure information [ 13, 14, 15], semantic information [16]
and mutual information [17, 18] are widely used in the target-less calibration methods. The
weakness of the target-less calibration methods is that there is no explicit corresponding
relationship between point clouds and images compared with target-based methods.</p>
      <p>Motion-based calibration methods [19, 20, 21, 22, 23, 24] leverage odometry information
to estimate the extrinsic parameters. Motion-based methods simplify the procedure of data
collections but the accuracy of extrinsic parameters greatly depends on the performance of the
odometry algorithms.</p>
      <p>In this paper, we focus on calibration board-based methods. Vertex features extracted from
point clouds are refined according to the prior information of the calibration board to reduce
the systematic errors of LiDAR measurements and improve the accuracy and eficiency of
calibration results. We compare our calibration method with other state-of-the-art calibration
methods using our proposed fair evaluation method based on space distances.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Calibration Method</title>
      <sec id="sec-3-1">
        <title>3.1. Overview of Our Calibration Method</title>
        <p>We denote the coordinates of a point in the LiDAR coordinate frame {} as   =
(, , )T, and the coordinates of the identical point in the camera coordinate frame
{} as   = ( ,  ,  )T. The transformation between   and   can be written as
can be written as
where  is the rotation matrix,  is the translation vector. The transformation matrix  
  =    +  ,
  =
The goal of the extrinsic calibration is to estimate  and  .</p>
        <p>In this paper, two ArUCO markers are used as calibration targets to implement the extrinsic
calibration between a LiDAR and a camera, which are pasted on two separate boards. A
calibration board coordinate frame {} is built to represent the calibration features as Figure 1.
The  plane of the calibration board frame is the board plane. The  axis is perpendicular
to the  plane and points out of the marker plane.</p>
        <p>The whole calibration process is described below. First, we analyze the measurement
characteristics and systematic errors of LiDARs. Second, in the LiDAR frame, according to extracted
points on the board from point clouds and the known size of the calibration board, the points on
vertices are refined. Third, the relative transformation between the calibration board frame and
the camera frame is calculated with the known size of the marker, and the vertices of calibration
board are estimated in the camera frame. Finally, with the information of vertices in these two
sensor frames, the corresponding relationship between frames is constructed and the extrinsic
parameters are calibrated. The details of each stage are introduced in the following paragraphs.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Measurement Characteristics and Systematic Errors of LiDARs</title>
        <p>There are two kinds of systematic errors of LiDARs.</p>
        <sec id="sec-3-2-1">
          <title>3.2.1. Range Error</title>
          <p>The LiDAR measures the surroundings in spherical coordinates with the range , elevation 
and azimuth  . Its Cartesian coordinates are calculated as
 =  *  * ,
 =  *  * ,
 =  * .
(3)
The elevation  and azimuth  are accurate, but there is error in the range . This paper uses a
Velodyne VLP-16 LiDAR sensor, whose range error can reach up to ± 3 typically. As we can
see from Figure 2, when we project LiDAR points on a board plane about 2 away from the
LiDAR, the thickness of the point cloud can be up to 4. The true thickness is zero.
(a) Front View
(b) Top View</p>
        </sec>
        <sec id="sec-3-2-2">
          <title>3.2.2. Divergence Error</title>
          <p>As shown in Figure 3a, the LiDAR beam becomes divergent and the LiDAR spot becomes larger
as distance becomes larger. For example, the horizontal size of spot is about 18.2 and the
vertical size is about 12.5 when the range is 2.</p>
          <p>Therefore, when we project a LiDAR point on the edge of the calibration board, part of the
LiDAR spots is on the board and other part is on the background. For these edge points, when
the reflectivity of the calibration board is stronger than that of the background, the returned
range measurement is the distance between the LiDAR and the board edge, even if the centerline
of the LiDAR beam is out of the board (the LiDAR spot is mostly on the background). The
resulting point measurement is called ghost point and makes the point cloud of the calibration
board larger than its true size.</p>
          <p>When the reflectivity of the calibration board is similar with that of the background, the
returned ranges of the edge points will be a weighted average of the range measurement of
the calibration board and the background. This kind of LiDAR point is also the ghost point
and makes a series of points behind the calibration board, whose shape is like a streamline.
As shown in Figure 3b, we can see that the points in the highlight ellipse are the points near
the board edge which is represented by the bright line. Due to the divergence error of LiDAR
(a) Divergence of LiDAR Beam</p>
          <p>(b) Ghost Points Caused by Divergence
beams, the LiDAR points on the edge are of the true edge of the calibration board and there are
a series of points behind of the board edge.</p>
          <p>Because of the range error and divergence error of LiDAR measurements, the points on the
board plane are not co-planar and the points on the edge are of the true edge, which makes the
calculation results of the points on the calibration board vertex not accurate. These problems
further afect the calibration results between LiDAR and camera.</p>
          <p>To solve these problems, we use the prior information of the calibration board to find the
true position of the board in LiDAR point clouds. As a board can be defined by its four vertices,
we propose a refining method to extract the accurate vertices from LiDAR point clouds and use
the vertex information to calibrate the extrinsic parameters between LiDAR and camera.</p>
        </sec>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Calibration Information in LiDAR Frame</title>
        <sec id="sec-3-3-1">
          <title>3.3.1. Extraction of Plane Points</title>
          <p>With the known size of a calibration board and the relative pose between calibration board
and LiDAR, the points located on the board can be segmented out from the LiDAR point cloud.
The parameters of board plane in the LiDAR frame,  , = (, , , )T, are fitted by
RANSAC method with the fitting threshold of 0.01 in this paper. The inlier points of the
iftted plane parameters are considered as the LiDAR points on the board plane. We denote these
points as  ,(1 ≤  ≤  ), where  is the number of inlier points.</p>
        </sec>
        <sec id="sec-3-3-2">
          <title>3.3.2. Extraction of Edge Points</title>
          <p>The points lying on the edge are extracted from the board plane points and divided into four
groups which corresponds to the four edges of the board. We denote the edge points in LiDAR
frame as  ,,( = 1, 2, 3, 4, 1 ≤  ≤  ), where  is the number of points on each edge.
With the points on the edge, the line parameters of calibration board edges in the LiDAR frame
are fitted by RANSAC method with the threshold of 0.01. The line parameters can be denoted
as , = ((,)T, ( ,)T)T, where 
, is the direction vector of the edge line
and  , is a point on the edge. The inlier points of the edge line fitting are considered as
the edge points.</p>
        </sec>
        <sec id="sec-3-3-3">
          <title>3.3.3. Extraction and Refinement of Vertex Points</title>
          <p>With the above resulting edge line parameters, the coordinates of board vertices in the LiDAR
frame  ,( = 1, 2, 3, 4) can be calculated by the parameters of edge lines. Due to the
LiDAR measurement errors discussed in Section 3.2, the two neighboring edge lines are not
co-planar and not intersecting in fact. We calculate the shortest segment line between the two
neighboring edge lines and the midpoint of the segment line is regarded as the vertex of the
calibration board. These coordinates can be used as initial values in the further refinement
using the prior information of the calibration board. The proposed method uses two kinds of
prior information: geometry-based information and pose-based information.</p>
          <p>Geometry-based constraint can be used to construct two kinds of constraints. The first one
is the constraint of the length of board edges, i.e. the distance of each pair of neighboring board
vertices should be equal to the length of board edges . We denote the neighboring vertices as
the th, th (1 ≤  ≤ 4, 1 ≤  ≤ 4,  ̸= ) vertices, and this constraint can be constructed
as
 = || , −
 ,||.</p>
          <p>The second one is the constraint of the perpendicular relationship between neighboring
edges.</p>
          <p>We denote the two perpendicular edge lines is determined by the th,th,th(1 ≤  ≤
4, 1 ≤  ≤ 4, 1 ≤  ≤ 4,  ̸=  ̸= ) vertex points, and the constraint relationship can be
written as
0 = ( , −
 ,)T( , −
 ,).</p>
          <p>Using these constraints, the four vertices can be fully guaranteed co-planar and correct size.
The shape and size of the board determined by these four vertices are the same as the actual
calibration board. Consequently, the measurement errors caused by the LiDAR are eliminated,
including the range error and divergence error. The Geometry-based information can constrain
the accurate geometry of the vertices, but during the optimization the pose of the vertices in
the LiDAR frame may drift.</p>
          <p>Pose-based constraint can be used to further constrain the pose of the board determined
by the four vertices in the LiDAR frame.</p>
          <p>We use the laser points on the edge to constrain the pose of the board. The edge point
 ,,( = 1, 2, 3, 4, 1 ≤  ≤  ) should be on the edge line determined by the vertex points,
which can be formulated as
0 =
||( , −
 ,,) × ,|| ,</p>
          <p>||,||
where</p>
          <p>,( = 1, 2, 3, 4) is the direction vector of the th edge line in the LiDAR frame and
can be calculated as</p>
          <p>, =  , −
 ,.</p>
          <p>(4)
(5)
(6)
(7)
 and  are the indices of vertex points that are on the end of the th board edge.</p>
          <p>The vertex points are refined with these constraints in the LiDAR frame. We solve this
optimization problem using Levenberg-Marquardt algorithm implemented by Ceres.</p>
        </sec>
      </sec>
      <sec id="sec-3-4">
        <title>3.4. Calibration Information in Camera Frame</title>
        <p>According to ArUCO Library [25, 26], with known camera intrinsic matrix  and distortion
coeficients , the relative pose   between calibration board frame and camera frame can be
determined as
  =
︂( 
0T
)︂ .
1</p>
        <p>Denoting the coordinates of the four board vertices in the calibration board frame as
 ,( = 1, 2, 3, 4), their corresponding coordinates in the camera frame can be written
as
 , =</p>
        <p>,,  = 1, 2, 3, 4.</p>
      </sec>
      <sec id="sec-3-5">
        <title>3.5. Extrinsic Parameter Estimation</title>
        <p>After the vertex information of the calibration board accurately extracted from point clouds
(Section 3.3) and images (Section 3.4), the extrinsic parameters between the LiDAR frame and
the camera frame can be estimated using the point-to-point correspondence.</p>
        <p>With the calculated coordinates of vertex points in LiDAR frame and camera frame, the
point-to-point constraint can be constructed as
(8)
(9)
(10)
 4
 = ∑︁ ∑︁ ||   ,, −  ,,||,</p>
        <p>=1 =1
where  is the number of the calibration datasets, (1 ≤  ≤ 4) is the index of board vertices.</p>
        <p>The optimization problem is solved using Levenberg-Marquardt algorithm implemented by
Ceres. We summarize our proposed calibration method in Algorithm 1.</p>
        <p>Algorithm 1 Calibration of extrinsic parameters between LiDAR and camera
Input:</p>
        <p>Camera intrinsic matrix  and distortion coeficients matrix ;</p>
        <p>Calibration data collected from  board poses;
Output:</p>
        <p>Extrinsic parameters   ;
1: Extract the LiDAR points on board and edge from point clouds;
2: Calculate and refine the board vertices  ,,,  = 1, 2, 3, 4;
3: Calculate the vertices in the camera frame  ,,,  = 1, 2, 3, 4;
4: Estimate the extrinsic parameters   using (10);
5: return   .</p>
      </sec>
      <sec id="sec-3-6">
        <title>3.6. Evaluation Method of Calibration Results</title>
        <p>Because the ground truth is not available in most cases, some papers used re-projection errors to
evaluate the accuracy of calibration results. By transforming the points lying the edge or vertex
from LiDAR frame to camera frame with the estimated extrinsic parameters and projecting
them on images, the re-projection error can be defined as the distance between the projected
point coordinates and its corresponding line or vertex. However, the re-projection error is not
an independent method for the evaluation because it is relative with the specific calibration
method. For example, the evaluation using re-projection error of vertex points is prone to better
results with point-based calibration methods. Intuitively, the re-projection error reflects the
alignment between the objects in 3D space and 2D image space, but it is not a suficient and
necessary condition for the alignment in two 3D frames.</p>
        <p>To evaluate various calibration methods fairly, we propose a 3D space distance based method
to evaluate the calibration results as (11), which is independent of the specific calibration
method.</p>
        <p>=1</p>
        <p>||,||
 =
1 ∑︁ ||(   , −  ,) × ,|| ,
(11)
where   is the calibration result,  , is the coordinates of a LiDAR point on the edge in
the LiDAR frame,  , is the coordinates of a point on the edge in the camera frame, ,
is the direction vector of the edge in the camera frame and  is the point number.</p>
        <p>, is directly extracted from the point cloud as introduced in Section 3.2. Although
these laser points on the edge are not the ideally accurate points on the edge as we discussed in
Section 3.2, using these raw sensor measurements to evaluate the calibration results is fair and
direct to reflect calibration quality.</p>
        <p>This 3D space distance can’t be zero even if we calculate it with ideal calibration results
because of the systematic errors of LiDAR measurements, but smaller space distance directly
represents better calibration results.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experiment</title>
      <sec id="sec-4-1">
        <title>4.1. Calibration Data Collection</title>
        <p>In this paper, we use two multi-sensors device sets to collect calibration data, which are denoted
as Device-1 and Device-2. Each device set is comprised of a Velodyne VLP-16 LiDAR and an
RGB camera with the resolution of 752 × 480. The LiDAR and camera are mounted on a base
using rigid connection. As introduced in Section 3.1, two boards with diferent ArUCO markers
are used as calibration targets to make full use of the camera field of view and shorten the
tedious and laborious calibration process.</p>
        <p>For each device, we collect the calibration datasets from 27 diferent poses of calibration
boards with respective to the sensor, extract the feature information of two calibration boards
as introduced in Section 3.3 and Section 3.4, and finally get 54 sets of board information for
experiments.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Experiment of Vertex Refinement</title>
        <p>The refinement results of vertex points are evaluated by computing the length of each board
edge and the angle between neighboring edges, before and after the refinement introduced in
Section 3.3.</p>
        <p>We denote the true length of board edges as , the calculated length using the vertices as
, the true angle of neighboring edges as  , and the calculated angle using the vertices as
 . Then the error can be formulated as
 = || −</p>
        <p>||,
 = ||  −  ||.
(12)
(13)</p>
        <p>We compute the board edge length and the board angle of all 54 board poses for each device
with and without the vertex refinement, and show the mean and standard deviation of the
errors in Table 1 and Table 2.</p>
        <p>After refining the vertex of calibration boards using the prior information, the error caused
by the measurement characteristics of LiDARs can reach to a remarkably low level. Without
the vertex refinement, the length error can be up to 4 and the angle error can be up to 2,
which will introduce error into the points extraction from board edges and the calculation of
vertex points, and deteriorate the extrinsic calibration between LiDAR and camera. Eliminating
the error of points extracted from point clouds can efectively improve the accuracy of extrinsic
parameters calibration, and we use the refined vertex points to implement the following extrinsic
calibration.</p>
        <p>"!" #
"!"#
#"%!#&amp;#
$! ! "#
"!" #
"!"#
#"%!#&amp;#
$! ! "#
$$"
&amp;
$$"
&amp;</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Experiment Results of Calibration</title>
        <p>
          We evaluate the performance of our proposed method and compare it with other state-of-the-art
methods [
          <xref ref-type="bibr" rid="ref1">1, 6</xref>
          ]. [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ] introduced two kinds of geometry constraints to calibrate the extrinsic
parameters. The first one is the point-to-plane correspondences which uses the LiDAR points
on the board plane and the plane parameters estimated in the camera frame to construct the
constraint. We denote this calibration method as Mishra’s plane method. The second one is the
point-to-line correspondences which uses the LiDAR points on the edge and the back-projected
plane parameters estimated in the camera frame to construct the constraint. We denote this
calibration method as Mishra’s edge method. [6] used the point-to-point correspondences to
calibrate the extrinsic parameters. The points used in [6] are the vertex points of the board,
which is denoted as Ankit’s vertex method, but this paper only calculates the vertices by the
edge line parameters without the refinement like our methods.
#"#!$
#"#$
$#&amp;"$’$
%"!" ! #$
#"#!$
#"#$
$#&amp;"$’$
%"!" ! #$
$!!##"#"
(a) Mean of space distances for Device-1
%" "$ $#$#
(b) Standard deviation of space distances for Device-1
$!!##"#"
(c) Mean of space distances for Device-2
%" "$ $#$#
(d) Standard deviation of space distances for Device-2
        </p>
        <p>
          We use the collected 54 sets of board information to implement the calibration for each
LiDAR-camera device set, as shown in Figure 4. The experiments are conducted in the form
of cross-validation. We calibrate the extrinsic parameters using these diferent methods with
two or more datasets, and evaluate the calibration results with the remaining datasets. In detail,
we randomly choose  ∈ [
          <xref ref-type="bibr" rid="ref2">2, 50</xref>
          ] datasets to calibrate the extrinsic parameters. For each , we
execute each calibration method 100 times and compute the mean and standard deviation of the
space distance (11) of each execution for evaluation.
        </p>
        <p>Figure 4a and Figure 4c show that the mean of the 3D space distance of our proposed method
is much smaller than that of other methods, which means our proposed calibration method
performs better on the accuracy of the calibration results. It is worth to mention that even using
only 2 datasets, our proposed calibration method has little loss on the accuracy performance.
Figure 4b and Figure 4d show that other state-of-the-art methods get more stable calibration
results when they use more calibration datasets, but our proposed method can maintain the
stability of calibration even using a few of datasets.</p>
        <p>For Mishra’s plane method, its space distance is much larger than that of other methods,
because the point-to-plane correspondence constructed by one dataset can only provide 3 DoF
(Degree of Freedom) constraints, as shown in Figure 5. The cost function used in point-to-plane
based calibration method is the distance between LiDAR points and the board in the camera
frame. The rotation of the laser points along the normal vector of the board plane or the
translation of the laser points on the board plane won’t violate the above mentioned constraint,
i.e. the cost function will not reflect this change. This is the reason why in Mishra’s plane
method at least three non-parallel board poses are necessary for extrinsic parameters calibration,
and more datasets lead to better results as shown in Figure 4a and Figure 4c. The performance of
Mishra’s plane method depends on the quantity and quality of the collected calibration datasets.
Notice that the range error discussed in Section 3.2 also reduces the accuracy of this calibration
method.</p>
        <p>For Mishra’s edge method, as the point-to-line based calibration method needs at least two
datasets collected at diferent poses to calibrate the extrinsic parameters, the 3D space distance
errors are smaller than that of Mishra’s plane method. At the same time, the divergence error
of LiDAR measurements discussed in Section 3.2 can also influence the calibration results of
this kind of methods.</p>
        <p>Ankit’s vertex method performs better than Mishra’s plane method and Mishra’s edge method,
because the dataset collected from only one pose is enough for the point-to-point based method
to complete the extrinsic calibration. When the number of datasets increases, the 3D space
distance error of Ankit’s vertex method decreases but still much larger than that of our method.
The reason is that the vertex points are calculated form the line parameters of the board edge,
which is estimated using the LiDAR points on the board edge, but the range error and divergence
error of LiDAR measurements discussed in Section 3.2 can notably decrease the accuracy of the
edge line estimation.</p>
        <p>Our proposed method uses the similar point-to-point correspondence to estimate extrinsic
parameters as Ankit’s vertex method, but the board vertices used for the estimation are well
refined using the prior information of the calibration board to mitigate the influence of the
systematic errors of LiDAR measurements. The 3D space distance error of our proposed method
is remarkably smaller than that of other method, and remains at a stable and low level even using
just a few calibration datasets. In fact, our proposed method can accurately calibrate the extrinsic
parameters with only one dataset collected at one pose without any loss of performances.
(a) Mishra’s plane method (b) Mishra’s edge method (c) Ankit’s vertex method (d) Our proposed method
(e) Mishra’s plane method (f) Mishra’s edge method (g) Ankit’s vertex method (h) Our proposed method</p>
        <p>To intuitively show the diference of the calibration results using diferent methods, we
re-project the raw laser scan and calculated vertices on the image using the estimated extrinsic
parameters, as shown in Figure 6. Diferent colors of the projection of LiDAR points indicates
the diferent point depth away from the origin of the camera frame. From the figure we can
clearly see that our proposed method fits the laser points better with the calibration board and
other foreground objects, such as the tripod, compared with other state-of-the-art methods.
Further we zoom in the re-projection results of the vertices in the highlight ellipse, as shown
in the second row of Figure 6. The red point on the board vertex is used as the true location
of the board vertex for reference. The blue point close to the board vertex is the projection of
refined board vertex extracted from point clouds. The distance between these two points can
clearly reflect the calibration performance. As shown in Figure 6h, the distance of these two
points using our proposed method is smaller than those of other methods, which indicates our
calibration methods is more accurate than other calibration methods.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>In this paper, we analyze the measurement characteristics and systematic errors of a LiDAR
sensor, propose a refining method for vertex points towards eliminating the errors in LiDAR
measurements, and further propose a calibration method for the extrinsic parameters using
the refined vertex points. Through quantitative and qualitative experiments, we demonstrate
the performance of our method and compare it with other three state-of-the-art calibration
methods. Experiment results show that the refinement of vertex points can efectively mitigate
the influence of the measurement error, and consequently our proposed calibration method
performs notably better than the state-of-the-art methods on the calibration accuracy and
stability.</p>
      <p>To fairly evaluate the calibration results, we also introduce a 3D space distance based
evaluation method, which uses raw LiDAR measurements and is fully independent of the diferent
calibration methods mentioned in this paper.</p>
      <p>The future research work includes the error analysis of calibration information from images,
and a refinement method for the geometry features extracted from images to improve the
accuracy of calibration results further.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>This work was supported by the National Natural Science Foundation of China under Grant
No.62002359 and 61836015.
[3] G. Koo, J. Kang, B. Jang, N. Doh, Analytic plane covariances construction for precise
planarity-based extrinsic calibration of camera and lidar, in: 2020 IEEE International
Conference on Robotics and Automation (ICRA), 2020, pp. 6042–6048. doi:10.1109/
ICRA40945.2020.9197149.
[4] R. Voges, B. Wagner, Set-membership extrinsic calibration of a 3d lidar and a camera, in:
2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2020,
pp. 9012–9019. doi:10.1109/IROS45743.2020.9341266.
[5] Y. Bok, D. Choi, Y. Jeong, I. S. Kweon, Capturing village-level heritages with a
handheld camera-laser fusion sensor, in: 2009 IEEE 12th International Conference on
Computer Vision Workshops, ICCV Workshops, 2009, pp. 947–954. doi:10.1109/ICCVW.2009.
5457600.
[6] A. Dhall, K. Chelani, V. Radhakrishnan, K. M. Krishna, Lidar-camera calibration using 3d-3d
point correspondences, CoRR abs/1705.09785 (2017). URL: http://arxiv.org/abs/1705.09785.
arXiv:1705.09785.
[7] Y. Xie, R. Shao, P. Guli, B. Li, L. Wang, Infrastructure based calibration of a multi-camera
and multi-lidar system using apriltags, in: 2018 IEEE Intelligent Vehicles Symposium (IV),
2018, pp. 605–610. doi:10.1109/IVS.2018.8500646.
[8] S. Verma, J. S. Berrio, S. Worrall, E. Nebot, Automatic extrinsic calibration between a
camera and a 3d lidar using 3d point and plane correspondences, in: 2019 IEEE Intelligent
Transportation Systems Conference (ITSC), 2019, pp. 3906–3912. doi:10.1109/ITSC.
2019.8917108.
[9] H. Alismail, L. D. Baker, B. Browning, Automatic calibration of a range sensor and camera
system, in: 2012 Second International Conference on 3D Imaging, Modeling, Processing,
Visualization &amp; Transmission, 2012, pp. 286–292. doi:10.1109/3DIMPVT.2012.52.
[10] S. A. Rodriguez F., V. Fremont, P. Bonnifait, Extrinsic calibration between a multi-layer
lidar and a camera, in: 2008 IEEE International Conference on Multisensor Fusion and
Integration for Intelligent Systems, 2008, pp. 214–219. doi:10.1109/MFI.2008.4648067.
[11] J. Kümmerle, T. Kühner, M. Lauer, Automatic calibration of multiple cameras and depth
sensors with a spherical target, in: 2018 IEEE/RSJ International Conference on Intelligent
Robots and Systems (IROS), 2018, pp. 1–8. doi:10.1109/IROS.2018.8593955.
[12] T. Tóth, Z. Pusztai, L. Hajder, Automatic lidar-camera calibration of extrinsic
parameters using a spherical target, in: 2020 IEEE International Conference on Robotics and
Automation (ICRA), 2020, pp. 8580–8586. doi:10.1109/ICRA40945.2020.9197316.
[13] J. Kang, N. L. Doh, Automatic targetless camera-lidar calibration by aligning edge with
gaussian mixture model, Journal of Field Robotics 37 (2020) 158–179. URL: https://onlinelibrary.
wiley.com/doi/abs/10.1002/rob.21893. doi:https://doi.org/10.1002/rob.21893.
arXiv:https://onlinelibrary.wiley.com/doi/pdf/10.1002/rob.21893.
[14] P. Moghadam, M. Bosse, R. Zlot, Line-based extrinsic calibration of range and image
sensors, in: 2013 IEEE International Conference on Robotics and Automation, 2013, pp.
3685–3691. doi:10.1109/ICRA.2013.6631095.
[15] J. Castorena, U. S. Kamilov, P. T. Boufounos, Autocalibration of lidar and optical cameras
via edge alignment, in: 2016 IEEE International Conference on Acoustics, Speech and
Signal Processing (ICASSP), 2016, pp. 2862–2866. doi:10.1109/ICASSP.2016.7472200.
[16] Y. Zhu, C. Li, Y. Zhang, Online camera-lidar calibration with sensor semantic information,
in: 2020 IEEE International Conference on Robotics and Automation (ICRA), 2020, pp.
4970–4976. doi:10.1109/ICRA40945.2020.9196627.
[17] G. Pandey, J. R. McBride, S. Savarese, R. M. Eustice, Automatic targetless extrinsic
calibration of a 3d lidar and camera by maximizing mutual information, in: Proceedings of the
Twenty-Sixth AAAI Conference on Artificial Intelligence, AAAI’12, AAAI Press, 2012, pp.
2053–2059.
[18] A. Mastin, J. Kepner, J. Fisher, Automatic registration of lidar and optical images of urban
scenes, in: 2009 IEEE Conference on Computer Vision and Pattern Recognition, 2009, pp.
2639–2646. doi:10.1109/CVPR.2009.5206539.
[19] C. Park, P. Moghadam, S. Kim, S. Sridharan, C. Fookes, Spatiotemporal camera-lidar
calibration: A targetless and structureless approach, IEEE Robotics and Automation
Letters 5 (2020) 1556–1563. doi:10.1109/LRA.2020.2969164.
[20] Z. Taylor, J. Nieto, Motion-based calibration of multimodal sensor extrinsics and timing
ofset estimation, IEEE Transactions on Robotics 32 (2016) 1215–1229. doi: 10.1109/TRO.
2016.2596771.
[21] T. Scott, A. A. Morye, P. Piniés, L. M. Paz, I. Posner, P. Newman, Choosing a time and place
for calibration of lidar-camera systems, in: 2016 IEEE International Conference on Robotics
and Automation (ICRA), 2016, pp. 4349–4356. doi:10.1109/ICRA.2016.7487634.
[22] T. Scott, A. A. Morye, P. Piniés, L. M. Paz, I. Posner, P. Newman, Exploiting known
unknowns: Scene induced cross-calibration of lidar-stereo systems, in: 2015 IEEE/RSJ
International Conference on Intelligent Robots and Systems (IROS), 2015, pp. 3647–3653.
doi:10.1109/IROS.2015.7353887.
[23] Z. Xiao, H. Li, D. Zhou, Y. Dai, B. Dai, Accurate extrinsic calibration between monocular
camera and sparse 3d lidar points without markers, in: 2017 IEEE Intelligent Vehicles
Symposium (IV), 2017, pp. 424–429. doi:10.1109/IVS.2017.7995755.
[24] G. Pascoe, W. Maddern, P. Newman, Direct visual localisation and calibration for road
vehicles in changing city environments, in: 2015 IEEE International Conference on
Computer Vision Workshop (ICCVW), 2015, pp. 98–105. doi:10.1109/ICCVW.2015.23.
[25] F. J. Romero-Ramirez, R. Muñoz-Salinas, R. Medina-Carnicer, Speeded up detection
of squared fiducial markers, Image and Vision Computing 76 (2018) 38–47. URL: https:
//www.sciencedirect.com/science/article/pii/S0262885618300799. doi:https://doi.org/
10.1016/j.imavis.2018.05.004.
[26] S. Garrido-Jurado, R. Muñoz-Salinas, F. Madrid-Cuevas, R. Medina-Carnicer,
Generation of fiducial marker dictionaries using mixed integer linear programming, Pattern
Recognition 51 (2016) 481–491. URL: https://www.sciencedirect.com/science/article/pii/
S0031320315003544. doi:https://doi.org/10.1016/j.patcog.2015.09.023.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>S.</given-names>
            <surname>Mishra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. R.</given-names>
            <surname>Osteen</surname>
          </string-name>
          , G. Pandey,
          <string-name>
            <given-names>S.</given-names>
            <surname>Saripalli</surname>
          </string-name>
          ,
          <article-title>Experimental evaluation of 3d-lidar camera extrinsic calibration</article-title>
          ,
          <source>in: 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>9020</fpage>
          -
          <lpage>9026</lpage>
          . doi:
          <volume>10</volume>
          .1109/IROS45743.
          <year>2020</year>
          .
          <volume>9340911</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>L.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kaess</surname>
          </string-name>
          ,
          <article-title>Automatic extrinsic calibration of a camera and a 3d lidar using line and plane correspondences</article-title>
          ,
          <source>in: 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)</source>
          ,
          <year>2018</year>
          , pp.
          <fpage>5562</fpage>
          -
          <lpage>5569</lpage>
          . doi:
          <volume>10</volume>
          .1109/IROS.
          <year>2018</year>
          .
          <volume>8593660</volume>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>