A Novel Representation of Parts for Accurate 3D Object Detection and Tracking in Monocular Images
2015 IEEE International Conference on Computer Vision 文章地址

Abstract:We present a method that estimates in real-time and under challenging conditions the 3D pose of a known object. Our method relies only on grayscale images since depth cameras fail on metallic objects; it can handle poorly textured objects, and cluttered, changing environments; the pose it predicts degrades gracefully in presence of large occlusions. As a result, by contrast with the state-of-the-art, our method is suitable for practical Augmented Reality applications even in industrial environments. To be robust to occlusions, we first learn to detect some parts of the target object. Our key idea is to then predict the 3D pose of each part in the form of the 2D projections of a few control points. The advantages of this representation is three-fold: We can predict the 3D pose of the object even when only one part is visible; when several parts are visible, we can combine them easily to compute a better pose of the object; the 3D pose we obtain is usually very accurate, even when only few parts are visible.

1、文章提出了一种在具有挑战性的条件下实时估计已知对象的3D姿态的方法,该方法仅使用灰度图像进行预测。
2、为了在遮挡情况下保持方法的鲁棒性,首先学习检测目标对象的某些部分。关键思想是以几个控制点的2D投影预测每个part的3D姿态。 优点在于:即使只有一个part可见,也可以预测物体的3D姿态;当多个parts可见时,也可以很容易地将它们组合起来,以计算物体更准确的姿态。

1 Introduce

3D object detection and tracking methods have undergone impressive improvements in recent years [5, 27, 13, 4, 30, 25, 39, 17, 1, 45, 38, 20, 35, 44]. However, each of the current approaches has its own weaknesses: Many of these approaches [5, 13, 1, 35] rely on a depth sensor, which
would fail on metallic objects or outdoor scenes; methods based on feature points [25, 17] expect textured objects; those based on edges [4, 39] are sensitive to cluttered background; most of these methods [13, 27, 30, 38, 11, 45, 20] are not robust to occlusion. We also want a method fast enough for interactive 3D applications.
近年来,3D对象检测和跟踪方法已经取得了令人印象深刻的进步[5,27,13,4,30,25,39,17,1,45,38,20,35,44]。然而,目前的每一种方法都有自己的缺点:这些方法中的许多[5,13,1,35]依赖于深度传感器,这会在金属物体或户外场景中失败;基于特征点的方法[25,17]需要有纹理的对象;基于边缘的方法[4,39]对杂乱背景敏感;这些方法[13,27,30,38,11,45,20]中的大多数对遮挡不鲁棒。我们还需要一种足够快的方法用于交互式3D应用程序。

图1:在具有挑战性的条件下,用常规相机检测3D的盒子。在该数据集上,用户从盒子中取出物品,并将它们放在桌面上,通常情况下会遮挡大部分盒子。尽管存在这些困难,但我们可以在每个图像中独立准确地估计盒子的3D姿态。

As Fig. 1 shows 1, we are interested in scenes with poorly textured objects, possibly visible only under heavy occlusions, drastic light changes, and changing background. A depth sensor is not an option in our setup, as the target objects often have specular surfaces. Feature point-based methods also fail because of the lack of texture. These are typical conditions of many Augmented Reality applications.
如图1所示,我们对纹理较模糊的物体感兴趣,这些物体可能仅在严重遮挡、剧烈光线变化和背景变化的情况下可见。深度传感器不在我们设置的选项中,因为目标对象通常具有镜面。由于缺少纹理,基于特征点的方法也会失败。这些是许多增强现实应用的典型条件。

At the core of our approach is the efficient detection of discriminative parts of the target object. Relying on parts for 3D object detection is not new [12, 27, 33, 20, 45]. The novelty in our approach is a powerful representation of the pose of each part.
我们方法的核心是有效检测目标对象具有识别性的part,[12、27、33、20、45]也应用了类似的方法。 我们方法的创新性在于对每个part姿势的有效表示。

Some previous methods used homographies [16, 12, 45] to represent a part pose, however this assumes that the object is piece-wise planar, and it is not easy to combine the homographies from several parts together to compute a better pose for the target object. Feature point-based methods
simply use the 2D locations of the feature points, which wastes very useful information.
以前的一些方法使用单应图[16,12,45]来表示part的姿态,然而这些方法假设对象是逐段平面的,并且不容易将多个部分的单应图组合在一起以计算目标对象的更好姿态。基于特征点的方法简单地使用特征点的2D位置,浪费了非常有用的信息。


图2.对物体part的三维姿态的表示。(a)每个part设置七个3D控制点,这些控制点排列成三个正交方向。(b)给定物体part的image patch,使用回归器预测这些控制点的2D重投影,以及预测的不确定性。

As shown in Fig 2, we therefore propose to represent the pose of each part by the 2D reprojections of a small set of 3D control points. The control points are only“virtual”, and do not have to correspond to specific image features. This representation is invariant to the part’s image location and only depends on its appearance. We show that a Convolutional Neural Network [19] (CNN) can predict the locations of these reprojections very accurately, and can also be used to predict the uncertainty of these location estimates.
如图2所示,我们建议通过一小组3D控制点的2D重投影来表示每个part的姿态。控制点只是“虚拟”的,不必与特定的图像特征相对应。该表示对于part的图像位置是不变的,并且仅取决于其外观。 我们表明,卷积神经网络[19](CNN)可以非常准确地预测这些重投影的位置,也可以用于预测这些位置估计的不确定性。

Given an input image, we run a detector to obtain a few hypotheses on the image locations of each part. We also use a CNN for this task, but another detection method could be used. We then predict the reprojections of the control points by applying a specific CNN to each hypothesis. This gives us a set of 3D-2D correspondences, some of which may be erroneous, but from which we can compute the 3D pose of the target object with a simple robust algorithm.
给定一个输入图像,我们运行一个检测器来获得关于每个part的图像位置的假设。 我们也使用CNN来完成这项任务,但也可以使用另一种检测方法。然后,我们通过对每个假设位置应用特定的CNN来预测控制点的2D重投影。 这为我们提供了一组3D-2D对应关系,其中一些可能是错误的,但从中我们可以使用简单的鲁棒算法计算目标对象的3D姿态。

This approach has several advantages:
• We do not need to assume the parts are planar, as was done in some previous work;
• we can predict the 3D pose of the object even when only one part is visible;
• when several parts are visible, we can combine them easily to compute a better pose of the object;
• the 3D pose we obtain is usually very accurate, even when only few parts (or a single one) are visible.
这种方法有几个优点:
•我们不需要假设每part都是平面的,就像之前的一些工作中所做的那样;
•即使只有一part可见,我们也可以预测物体的3D姿态;
•当多个parts可见时,我们可以很容易地将它们组合起来,以计算对象的更佳姿态;
•即使只有少数part(或单个part)可见,我们获得的3D姿态通常也非常准确。

2 Related Work

3 Part Pose Representation and Estimation

Given an input grayscale image2, we want to estimate the 3D pose p of a calibrated projective camera with respect to a known rigid object. We assume that we are given a 3D model of the object, for example in the form of a triangular mesh, and a set of manually labelled parts on the object. A very small number of parts is required by our framework; in all our tests we employed at most 4 parts for an object. We currently select the parts by hand (automatic selection is left to future work). Ideally, the parts should be easy to detect in images, and spread over the object.
给定输入灰度图像(所有实验都使用VGA图像),我们希望估计校准投影相机相对于已知刚性物体的3D姿态p我们假设我们得到了物体的3D模型,例如,三角形网格的形式,以及物体上一组手动标记的part。我们的框架需要非常少的part;在我们的所有测试中,一个物体最多使用4个part。我们目前手动选择物体被识别的part(自动选择留待将来的工作)。理想情况下,这个部分应易于在图像中检测,并扩展到整个物体。

In this section, we justify our choice for the representation of the pose parts, and we explain how we detect the parts and predict their poses. The next section describes our algorithm to compute the pose of the camera based on the predicted pose parts. The main notations are resumed in Table 1.
在本节中,我们证明了我们对用于姿态表示的part的选择,并解释了我们如何检测这些parts并预测它们的姿态。下一节描述了基于已预测的part姿势来计算相机姿势的算法。表1中列出了主要符号。

3.1. Representing the Part Poses

One can think of different ways to represent the 3D poses of parts of objects. For example, it is possible to use homographies [16, 12, 45]. However, this assumes that the part surface is planar, and makes it difficult to merge the contributions of the different parts.
人们可以想到不同的方法来表示物体的部分3D姿态。例如,可以使用单应图[16,12,45]。但是,这些文章假设part表面是平面的,因此很难合并不同的part。

Another possibility we considered is to predict from the appearance of the part, a 3D rotation matrix and the depth value of its center. Assuming an orthogonal projection, it is possible to retrieve the 3D translation as well, from the patch center image location and the predicted depth. However, this representation is not translation invariant in a full perspective model. Also it is not clear how to merge rotations for estimating the pose of the whole target object. Finally, it is difficult to predict the depth accurately from the image patch, as our results will demonstrate.
我们考虑的另一种可能的方法是从物体的外观、3D旋转矩阵及其中心的深度值来估计。假设正交投影,也可以从patch中心图像位置和预测的深度值检测3D平移。然而,这种表示在全透视模型中不是平移不变的。此外,还不清楚如何合并旋转以估计整个目标物体的姿态。最后,如我们的结果所示,很难从图像块准确预测深度。

Since ourfinal solution is based on 3D control points, as already mentioned, we could also directly predict the 3D locations of the 3D Control Points in the camera reference system: This makes combining the poses simpler, as this only involves computing the rigid motion between two sets of 3D points [40]. Unfortunately, this representation is not translation invariant. Moreover, as for the previous option, it requires to directly predict the depths of the points, which is far from accurate in our experiments.
由于我们的最终解决方案基于3D控制点,如前所述,我们还可以直接预测相机参考系统中3D控制点的3D位置:这使得组合姿势更简单,因为这只涉及计算两组3D点之间的刚性运动[40]。不幸的是,这种表示不是平移不变的。此外,对于前面的选项,它需要直接预测点的深度,这在我们的实验中远远不够准确。

This is why we propose to represent the part pose as the 2D reprojections of a set of 3D control points. This representation is fully translation invariant; it is straightforward to combine the poses of an arbitrary number of parts, by simply grouping all the 2D reprojections together and solving a PnP problem; we do not need to predict the depth of the 3D points, which is difficult to do accurately. These advantages entail a tremendous accuracy gain, as showed by our results in Section 5.2. In our experiments, we used 7 control points for each part, spanning 3 orthogonal directions, as shown in Fig. 2(a), however other configurations could probably be used.
这就是为什么我们计划将part姿态表示为一组3D控制点的2D重投影。这种表示是完全平移不变的;通过将所有2D重投影简单地组合在一起并求解PnP问题,可以简单地组合任意数量的part的姿态;我们不需要预测3D点的深度,这很难准确地做到。如我们在第5.2节中的结果所示,这些优势带来了巨大的精度增益。在我们的实验中,我们为每个部件使用了7个控制点,跨越3个正交方向,如图2(a)所示,但也可能使用其他配置。

3.2. Detecting the Parts


图3.检测part.(a) 盒子的输入图像。(b) 每个图像位置的CNNpart−detCNN^{part-det}CNNpart−det的输出。每种颜色对应不同的part。(c) 高斯平滑后的输出。(d) 检测到的部分,对应于(c)中的局部最大值

We use a set of registered training images of the target object under different poses and lighting (as the one shown in Fig. 3(a)) to learn to detect the parts and predict their control points. We will denote our training data as:
我们使用目标物体在不同姿态和照明下的一组训练图像(如图3(a)所示)来学习检测part并预测其控制点。我们将训练数据表示为
T={(Ii,{cij}j,{vijk}jk}i,T = \{( I_i,\{c_{ij}\}_j,\{v_{ijk}\}_{jk}\}_i,T={(Ii​,{cij​}j​,{vijk​}jk​}i​,
where IiI_iIi​denotes the i-th training image, cijc_{ij}cij​ the projectionof the center CjC_jCj​ of the jthj_{th}jth​ part on IiI_iIi​, and vijkv_{ijk}vijk​ the projection of the kthk_{th}kth​ control point of the jthj_{th}jth​ part in this image.
其中IiI_iIi​表示第iii个训练图像,cijc_{ij}cij​表示第jjj个部分的中心CjC_jCj​ 在IiI_iIi​上的投影,vijkv_{ijk}vijk​表示该图像中第jjj个part的第k个控制点的投影。


图4.用于part检测的CNNpart−detCNN^{part-det}CNNpart−det架构。最后一层输出与每个part或背景相对应的image patch的似然性。

During an offline stage, we train afirst CNN with a standard multi-class architecture shown in Fig. 4 to detect the parts. The input to this CNN is a 32 × 32 image patch q\bm{q}q, its output consists of the likelihoods P(J=j∣q)P(J = j | \bm{q})P(J=j∣q) of the patch to correspond to one of the NPN_PNP​ parts. We train the CNN with patches randomly extracted around the centers cijc_{ij}cij​ of the parts in images IiI_iIi​ and patches extracted from the background, and by optimizing the negative log-likelihood over the parameters ω\omegaω of the CNN:
在离线阶段,我们使用图4所示的标准多类架构训练第一个CNN,以检测part。该CNN的输入是一个32×32的image patch q\bm{q}q,其输出由对应于NPN_PNP​ parts之一的image patch的似然性P(J=j∣q)P(J = j | \bm{q})P(J=j∣q)组成。我们使用围绕图像IiI_iIi​中的part的中心cijc_{ij}cij​随机提取的image patch和从背景中提取的image patch来训练CNN,并通过优化CNN的参数ω\omegaω上的负对数似然性:
ω^=argmin∑j=0NP∑q∈Tj−logsoftmax(CNNωpart−det(q))[j]\hat \omega = arg\ min \sum_{j=0}^{N_P}\sum_{\bm{q}\in T_j} - log\ softmax(CNN^{part-det}_\omega (\bm{q})) [j]ω^=arg minj=0∑NP​​q∈Tj​∑​−log softmax(CNNωpart−det​(q))[j]
where TjT_jTj​ is a training set made of image patches centered on part jjj and T0T_0T0​ is a training set made of image patches from the background, CNNωpart−det(q)CNN^{part-det}_\omega (\bm{q})CNNωpart−det​(q) is the NP+1N_P + 1NP​+1-vector output by the CNN when applied to patch q\bm{q}q, and softmaxCNNωpart−det(q)[j]CNN^{part-det}_\omega (\bm{q})[j]CNNωpart−det​(q)[j] is the j-th coordinate of vector softmaxCNNωpart−det(q)CNN^{part-det}_\omega (\bm{q})CNNωpart−det​(q).
其中,TjT_jTj​是由以part jjj为中心的image patch组成的训练集, T0T_0T0​是由来自背景的image patch构成的训练集。CNNωpart−det(q)CNN^{part-det}_\omega (\bm{q})CNNωpart−det​(q)是当应用于image patchq\bm{q}q时由CNN输出的NP+1−vectorN_P + 1 - vectorNP​+1−vector,CNNωpart−det(q)[j]CNN^{part-det}_\omega (\bm{q})[j]CNNωpart−det​(q)[j]是矢量softmax的第jjj个坐标CNNωpart−det(q)CNN^{part-det}_\omega (\bm{q})CNNωpart−det​(q)。

At run time, we apply this CNN to each 32 × 32 patch in the input images captured by the camera. This can be done very efficiently as the convolutions performed by the CNN can be shared between the patches [9]. As shown in Fig. 3, we typically obtain clusters of large values for the likelihood of each part around the centers of the parts. We therefore apply a smoothing Gaussianfilter on the output of the CNN, and retain only the local maximums of these values as candidates for the locations of the parts.
在运行时,我们将此CNN应用于相机捕获的输入图像中的每个32×32 patch。这可以非常有效地完成,因为CNN执行的卷积可以在image patch之间共享。如图3所示,我们通常获得围绕part中心的每个part的似然性的大值聚类。因此,我们在CNN的输出上应用平滑高斯滤波器,并仅保留这些值的局部最大值作为零件位置的候选值。

The result of this step is, for each part jjj, a set Sj=(c^jl,sjl)lS_j ={(\hat c_{jl}, s_{jl})}_lSj​=(c^jl​,sjl​)l​ of 2D location candidate c^jl\hat c_{jl}c^jl​ for the part together with a score sjls_{jl}sjl​ that is the value of the local maxima returned by the CNN. We will exploit this score in our pose estimation algorithm described in Section 4. We typically get up to 4 detections for each part in a given input image.
该步骤的结果是,对于每个部分 jjj,该部分的2D位置候选的集合Sj=(c^jl,sjl)lS_j ={(\hat c_{jl}, s_{jl})}_lSj​=(c^jl​,sjl​)l​以及分数sjls_{jl}sjl​,该分数sjls_{jl}sjl​是CNN返回的局部最大值的值。我们将在第4节中描述的姿态估计算法中利用这个分数。我们通常对给定输入图像中的每个part进行最多4次检测。

3.3. Predicting the Reprojections of the Control Points and their Uncertainty

Once the parts are detected, we apply a second CNN to the patches centered on the candidates c^jl\hat c_{jl}c^jl​ to predict the projections of the control points for these candidates. Each part has its specific CNN. As shown in Fig. 5, these networks take as input a patch of size of 64 × 64. The output layer is made of 2NVN_VNV​ neurons, with NVN_VNV​ the number of control points of the part, which predicts the 2D locations of the control points. We train each of these CNNs during an offline stage by simply minimizing over the parameters ω\omegaω of the CNN the squared loss of the predictions:
一旦检测到物体的part,我们将第二个CNN应用于以候选点c^jl\hat c_{jl}c^jl​为中心的image patch,以预测这些候选控制点的投影。每个part都有其特定的CNN。如图5所示,这些网络将大小为64×64的patch作为输入。输出层由2NVN_VNV​个神经元组成,用来预测控制点的2D位置,NVN_VNV​是part中控制点的数量。我们在离线阶段通过简单地最小化CNN的参数ω\omegaω来训练这些CNN中的每一个预测的平方损失:
ω^=argmin⁡∑(q,w)∈Vj∣∣w−CNNωcp−pred−j(q)∣∣2\hat \omega = arg \min \sum_{\bm{(q,w)}\in V_j} ||\bm w - CNN^{cp-pred-j}_ \omega(\bm q)||^2ω^=argmin(q,w)∈Vj​∑​∣∣w−CNNωcp−pred−j​(q)∣∣2
where VjV_jVj​ is a training set of image patches q\bm qq centered on part jjj and the corresponding 2D locations of the control points concatenated in a (2NVN_VNV​)-vector w\bm ww, and CNNwcp−pred−j(q)CNN^{cp-pred-j}_w (\bm q)CNNwcp−pred−j​(q) is the prediction for these locations made by the CNN specific for part jjj, given patch q\bm qq as input.
其中VjV_jVj​是以part jjj为中心的图像块q\bm qq的训练集,以及以(2NVN_VNV​)的向量w\bm ww连接的控制点的对应2D位置,并且CNNwcp−pred−j(q)CNN^{cp-pred-j}_w (\bm q)CNNwcp−pred−j​(q) 是给定图像块q\bm qq作为输入,CNN对part jjj做出的位置预测。

At run-time, we obtain for each c^jl\hat c_{jl}c^jl​ candidate, predictions {v^jkl}\{\hat v_{jkl}\}{v^jkl​} for the control points projections. In addition, we estimate the 2D uncertainty for the predictions, by propagating the image noise through the CNN that predicts the control point projections . Let us consider the matrix:
在运行时,我们获得了每个候选的控制点c^jl\hat c_{jl}c^jl​的预测{v^jkl}\{\hat v_{jkl}\}{v^jkl​}。此外,我们通过通过预测控制点投影的CNN传播图像噪声来估计预测的2D不确定性[43]。让我们考虑以下矩阵:
SV=Jc^(σI)Jc^⊤(4)S_V=J_{\hat c}(\sigma I)J_{\hat c}^{\top} (4)SV​=Jc^​(σI)Jc^⊤​(4)

where σ\sigmaσ is the standard deviation of the image noise assumed to be Gaussian and affect each image pixel independently, III is the 642×64264^2×64^2642×642 Identity matrix, and Jc^J_{\hat c}Jc^​ is the Jacobian of the function computed by the CNN, evaluated at the patch centered on the candidate c^\hat cc^. Such a Jacobian matrix can be computed easily with a Deep Learning framework such as Theano, by composing the Jacobians of the successive layers of the network. We neglect the correlation between the different control points tofinally extract from the block diagonal of SVS_VSV​ the 2 × 2 uncertainty matrix noted SjklS_{jkl}Sjkl​ below for each control point. An example of predicted control points and their uncertainties is shown in Fig. 2(b). Note that we can easily compute the SjklS_{jkl}Sjkl​ matrices without having to compute the entire, and very large, product in Eq. (4)
其中σ是假设为高斯并独立影响每个图像像素的图像噪声的标准偏差,III为642×64264^2×64^2642×642恒等矩阵,Jc^J_{\hat c}Jc^​ 为CNN计算的函数的雅可比矩阵,在以候选点为中心的图像块处进行评估。通过组合网络的连续层的雅可比矩阵,可以使用深度学习框架(如Theano)轻松计算此类雅可比矩阵。我们忽略了不同控制点之间的相关性,最终从SVS_VSV​的块对角线中提取每个控制点的2×2。预测控制点及其不确定性的示例如图2(b)所示。注意,我们可以很容易地计算SjklS_{jkl}Sjkl​矩阵,而不必计算公式(4)中的整个非常大的乘积。

4. Estimating the Object Pose

Thanks to our representation for the part poses, estimating the object pose is straightforward, since each control point provides a 3D-2D correspondence. We describe here the method we use, other methods are probably possible.
由于我们对part姿态的表示,使得估计物体姿态很简单,因为每个控制点都提供了3D-2D对应关系。我们在这里描述我们使用的方法,其他方法也可能有效。

We assume that we are given a prior on the pose p\bm pp, in the form of a Mixture-of-Gaussians{(pˉm,Sm)}\{(\bar p_m, S_m)\}{(pˉ​m​,Sm​)}, as [23] was done e.g. in. This prior is very general, and allows us to define the normal action range of the camera. Moreover, the pose computed for the previous frames can be easily incorporated within this framework to exploit temporal consistency.
我们假设我们以Mixture-of-Gaussian{(pˉm,Sm)}\{(\bar p_m, S_m)\}{(pˉ​m​,Sm​)}的形式给出了姿态p\bm pp的先验,如[23]中所做的。这个先验是非常通用的,允许我们定义相机的正常动作范围。此外,为先前帧计算的姿态可以容易地并入这个框架中,以利用时间一致性。

In the following, we willfirst assume that this prior is defined as a single Gaussian distribution of mean and covariance (pˉ0,S0)(\bar p_0, S_0)(pˉ​0​,S0​). We will extend our approach to the Mixture-of-Gaussians in Section 4.3.
在下文中,我们将首先假设该先验被定义为均值和协方差(pˉ0,S0)(\bar p_0, S_0)(pˉ​0​,S0​)的单一高斯分布。我们将在第4.3节中将我们的方法扩展到Mixture-of-Gaussians。

4.1. Using a single Gaussian Pose Prior

Let usfirst assume there is no outlier returned by the part detection process or by the control point prediction, and that all the parts are visible. Then, the object pose p^\hat pp^​, or equivalently the camera pose, can be estimated as the minimizer of F(p)F(p)F(p), with F(p)F(p)F(p)=
让我们首先假设part检测过程或控制点预测没有返回异常值,并且所有部分都是可见的。然后,可以将物体姿态(或等效的相机姿态)估计为F(p)F(p)F(p)的最小值,其中F(p)F(p)F(p)=
1NP∑j,kdist2(Sjk,Γp(Vjk),v^jk)+(p−pˉ0)⊤S0−1(p−pˉ0)(5)\frac{1}{N_P}\sum_{j,k}dist^2(S_{jk},Γp(V_{jk}),\hat v_{jk})+(\bm{p}-\bm{\bar p_0})\top S^{-1}_0(\bm{p}-\bm{\bar p_0}) \ (5)NP​1​j,k∑​dist2(Sjk​,Γp(Vjk​),v^jk​)+(p−pˉ​0​)⊤S0−1​(p−pˉ​0​) (5)

where the sum is over all the control points of all the parts, and Γp(V)Γp(V)Γp(V) is the 2D projection of V under pose p.v^jkp.\ \hat v_{jk}p. v^jk​ is the projection of control point VjkV_{jk}Vjk​ and SjkS_{jk}Sjk​ its uncertainty estimated as explained in Section 3.3—since we assume there is no outlier, we dropped here the lll index corresponding to the multiple detections. dist(.) is the Mahalanobis distance:
其中,和在所有部分的所有控制点上,Γp(V)Γp(V)Γp(V)是VVV在姿态ppp下的2D投影。 v^jk\hat v_{jk}v^jk​是控制点VjkV_{jk}Vjk​的投影,SjkS_{jk}Sjk​为不确定性,如第3.3节所述 ——因为我们假设没有异常值,所以我们在这里删除了与多次检测相对应的lll。dist(.)是马氏距离。

F(p)F(p)F(p) is minimized using the Gauss-Newton algorithm initialized with p0p_0p0​.
F(p)F(p)F(p)使用初始化为p0p_0p0​的高斯-牛顿算法最小化。

4.2. Robust detection of parts

In practice, for the location of the jthj_{th}jth​ part, the detection procedure described in Section 3.2 can get a set of hypotheses SjS_jSj​, and at most one is correct.
实际上,对于第jjj个part的位置,第3.2节中描述的检测程序可以得到一组假设SjS_jSj​,并且最多有一个是正确的。

Checking all the possible combinations would be time consuming, so we rank the candidates according to their score sjls_{jl}sjl​, keep the best four candidates for each part and greedily examine the possible sets CCC of correspondences between a part and the candidate detections.
检查所有可能的组合将是耗时的,因此我们根据候选part的分数sjls_{jl}sjl​对其进行排名,为每个part保留最好的四个,并使用贪婪算法检查一个part和候选检测之间的可能的对应集合CCC。

Similarly to [23], we exploit the pose prior forfirst quickly evaluating if the correspondences in CCC can yield a good pose estimate.
类似于[23],我们利用先验姿态,首先快速评估 CCC 中对应的姿态是否能够产生良好的姿势估计。

We only consider a set CCC if

where T=40T = 40T=40, and whereS^0(Cj)=JS0J⊤\hat S_0(C_j) = J S_0J^\topS^0​(Cj​)=JS0​J⊤, with JJJ the jacobian of Γp0(Cj)Γ_{p_0} (C_j)Γp0​​(Cj​), is the covariance of the projection Γp0(Cj)Γ_{p_0} (C_j)Γp0​​(Cj​) of CjC_jCj​, and ρj^ρ_{\hat j}ρj^​​ is a random candidate of the set (since it is reasonable to suppose that at least one candidate of a part has been reliably detected).
其中T=40T = 40T=40,且其中S^0(Cj)=JS0J⊤\hat S_0(C_j) = J S_0J^\topS^0​(Cj​)=JS0​J⊤,是CjC_jCj​的投影Γp0(Cj)Γ_{p_0} (C_j)Γp0​​(Cj​)的协方差,其中JJJ是Γp0(Cj)Γ_{p_0} (C_j)Γp0​​(Cj​)的雅可比,ρj^ρ_{\hat j}ρj^​​是集合的随机候选(因为有理由假设至少一个part的候选控制点已经被可靠地检测到)。

If CCC passes this test, we compute the average distanceρˉ=1∣C∣∑jρj∼\bar ρ = \frac{1}{|C|}\sum_jρ^{\sim}_jρˉ​=∣C∣1​∑j​ρj∼​ of its points. We keep the NCN_CNC​ sets with the lowest average distance (in practice, we set NC=4N_C = 4NC​=4 for all our experiments); we run the Gauss-Newton optimization of Eq. (5) using each CCC to obtain a pose estimate, and evaluate it as explained in Section 4.3.
如果 CCC 通过该测试,我们计算平均距离ρˉ=1∣C∣∑jρj∼\bar ρ = \frac{1}{|C|}\sum_jρ^{\sim}_jρˉ​=∣C∣1​∑j​ρj∼​。我们保持具有最低平均距离的NCN_CNC​集(在实践中,我们为所有实验设置了NC=4N_C = 4NC​=4);我们使用每个 CCC 运行方程(5)的高斯-牛顿优化,以获得姿态估计,并按照第4.3节所述对其进行评估。

4.3. Using a Mixture-of-Gaussians for the Pose Prior

In practice, the prior for the pose is in the form of a Mixture-of-Gaussians {(pm,Σm)}m with M = 9 components (the prior employed for the BOX dataset is shown in Fig.6). We apply the method described above to each
component, and obtain MNCMN_CMNC​ possible pose estimates:p^(1),...,p^(MNC)\hat p^{(1)},...,\hat p^(M N_C)p^​(1),...,p^​(MNC​).
在实践中,姿态的先验是Mixture-of-Gaussians{(pm,∑m)}m\{(p_m,∑_m)\}_m{(pm​,∑m​)}m​和 M=9M=9M=9 个分量(BOX数据集使用的先验如图6所示)。我们将上述方法应用于每个分量,并获得MNCMN_CMNC​可能的姿态估计:p^(1),...,p^(MNC)\hat p^{(1)},...,\hat p^{(M N_C)}p^​(1),...,p^​(MNC​)。

To finally identify the best pose estimate, we evaluate each p^(n)\hat p^{(n)}p^​(n), employing a weighted sum of several cues: the angle between the quaternions for p^(n)\hat p^{(n)}p^​(n) and the corresponding pˉm\bar p_mpˉ​m​ prior; the average reprojection error of the set of control points C according to p^(n)\hat p^{(n)}p^​(n); the correlation between the object contours after projection by p^(n)\hat p^{(n)}p^​(n) and the edges detected in the image. For setting the weights, we train a simple linear regressor on the training video sequences to predict the Euclidean distance between p^(n)\hat p^{(n)}p^​(n)and the groundtruth. At testing time, we use the linear regressor to evaluate the quality of the computed pose, i.e we keep the pose that gives the smallest predicted distance.
为了最终确定最佳姿态估计,我们使用几个线索的加权和来评估每个 p^(n)\hat p^{(n)}p^​(n):p^(n)\hat p^{(n)}p^​(n)的四元数与相应的 pˉm\bar p_mpˉ​m​ 先验之间的角度;根据 p^(n)\hat p^{(n)}p^​(n) 得到的控制点C集合的平均重投影误差;投影后的物体轮廓与图像中检测到的边缘之间的相关性。为了设置权重,我们在训练视频序列上训练一个简单的线性回归器,以预测 p^(n)\hat p^{(n)}p^​(n) 和 groundtruth 之间的欧几里得距离。在测试时,我们使用线性回归器来评估计算出的姿势的质量,即我们保持给出最小预测距离的姿势。

If the optimization of Eq. (5) converges, we add to the initial prior the estimated pose and its covariance as part of the pose prior for the next frame. This helps enforcing temporal consistency. The pose covariance is obtained using the Extended Kalman Filter update formula [43] when optimizing Eq. (5).
如果等式(5)的优化收敛,我们将估计的姿态及其协方差作为下一帧的姿态先验的一部分添加到初始先验。这有助于加强时间一致性。当优化方程(5)时,使用扩展卡尔曼滤波器更新公式[43]获得姿态协方差。

5. Experimental Results

笔记:A Novel Representation of Parts for Accurate 3D Object Detection and Tracking in Monocular Images相关推荐

  1. 【论文学习】Towards Accurate Oriented Object Detection in Aerial Images with Adaptive Multi-level Feature

    AFF-Det: 通过自适应多级特征融合实现航空图像中的精确定向目标检测 论文链接:<Towards Accurate Oriented Object Detection in Aerial I ...

  2. 【论文学习】MKIoU Loss: Towards Accurate Oriented Object Detection in Aerial Images

    [论文学习]MKIoU Loss: Towards Accurate Oriented Object Detection in Aerial Images 在本文中,提出了一种近似 SkewIoU 的 ...

  3. Cascaded Partial Decoder for Fast and Accurate Salient Object Detection

    CVPR2019发布,在看参考文献时发现的一篇显著性目标检测文章 有些博客讲的还是蛮详细的,放在下面了 Cascaded Partial Decoder for Fast and Accurate S ...

  4. 论文报告笔记(四) Towards Precise End-to-end Weakly Supervised Object Detection Network

    文章目录 论文报告笔记(四) Towards Precise End-to-end Weakly Supervised Object Detection Network 弱监督学习 WSDDN OIC ...

  5. [CVPR2019]Cascaded Partial Decoder for Fast and Accurate Salient Object Detection

    论文地址:Cascaded Partial Decoder for Fast and Accurate Salient Object Detection 代码地址:https://github.com ...

  6. 【文献阅读】Cascaded Partial Decoder for Fast and Accurate Salient Object Detection

    文章目录 写在前面 整体框架 部分解码器 整体注意力 整体代码 写在前面 这是2019CVPR的一篇文章,关于显著目标检测的.提出了一种CPD框架 大概记录一下对这个文章的理解,做了一些图,如有不对, ...

  7. 论文笔记-F3Net:Fusion, Feedback and Focus for Salient Object Detection

    论文笔记之2020-AAAI-F3Net-F3Net:Fusion, Feedback and Focus for Salient Object Detection 论文地址:https://arxi ...

  8. Unifying Voxel-based Representation with Transformer for 3D Object Detection (UVTR)论文笔记

    原文链接:https://arxiv.org/abs/2206.00630 1.引言 统一表达对多模态的知识迁移和特征融合是必不可少的. 目前的统一表达可大致分为输入级和特征级的.输入级方法在网络开始 ...

  9. CFT:Multi-Camera Calibration Free BEV Representation for 3D Object Detection——论文笔记

    参考代码:暂无 1. 概述 介绍:在相机数据作为输入的BEV感知算法中很多是需要显式或是隐式使用相机内外参数的,但是相机的参数自标定之后并不是一直保持不变的,这就对依赖相机标定参数的算法带来了麻烦.如 ...

最新文章

  1. 巧用宏定义进行调试 (转)
  2. 为什么两个数异或结果在与其中一个数异或得到的是另一个数?(a^b)^a=b
  3. 递归和迭代路由_静态路由在以太网接口中的不同书写会导致路由器怎样的操作结果?...
  4. 项目管理基础:项目的生存周期模型
  5. python去掉列表中的单引号_Python:如何从列表项中删除单引号(Python: How to remove single quotes from list item)...
  6. BCGControlBar入门使用手册
  7. jquery +ajax 上传加预览
  8. 弹幕,是怎样练成的?
  9. office插件开发_理工科公式编辑工具——Aurora插件
  10. js图片上传(配合七牛云)
  11. zuc算法代码详解_ZUC算法了解
  12. 软件工程pert图_《软件工程》软件项目管理实验
  13. 在什么情况下需要用矿物质防火电
  14. Android手机SD卡分区
  15. python学习-day9内置函数(高阶)
  16. 如何让自己打计算机打字快起来,电脑快速打字的技巧
  17. MongoDB 学习笔记
  18. 自定义域名:为自己的CSDN博客添加自定义域名吧!
  19. WCF学习——掀开你的盖头来
  20. jointjs教程例子

热门文章

  1. 张丽俊最新演讲:要像竹子一样扎根,你终会一飞冲天
  2. 匿名内部类以及Lambda表达式(Java和C#)
  3. excel与access结合运用_access和excel结合应用
  4. visual studio登陆界面空白_电脑同时登陆两个微信,你会吗?
  5. css斜切角/内阴影
  6. 终于有人把股市集合竞价的秘密说清楚了
  7. 英语学习——学英语的心态(2)
  8. excel调整所有行高列宽
  9. 领域分类的问题_别人的问题领域
  10. 2015美国大学计算机科学专业排名,2015年USNews美国大学计算机科学专业研究生排名...