170 Matching Annotations
  1. Sep 2021
    1. 2D convolutions are not designedto directly leverage the tiled structure of this feature space.Instead, we propose to first learn to expand this structured

      2d卷积并不适用

    2. The result-ing tensor is at a reduced resolution, but in contrast to strid-ing or pooling

      经过此操作之后的tensor降低了分辨率,但与stride和pooling相对,这样的操作没有损失

    3. Lv between the magnitudeof the pose-translation component of the pose network pre-diction ˆtand the measured instantaneous velocity scalar vmultiplied by the time difference between target and sourceframes ∆Tt→s,

      新的对于速度的损失函数,两帧之间的t变换和速度v,时间△t之间的关系

    4. Mt is a binary mask that avoids computing the pho-tometric loss on the pixels that do not have a valid map-ping

      没有可用地图时候就不用计算光度什么的loss

    5. and rely on the ground-truth LiDARmeasu,rements to scale their depth estimates appropriatelyfor evaluation purpose

      用雷达的测量来验证结果

    6. where a depth and pose network are simultaneouslylearned from unlabeled monocular videos

      从无标签单目视频中同时学习深度和相机运动

    7. vided an alternative strategy involving training a monocu-lar depth network with stereo cameras, without requiringground-truth depth labels

      好思路

    8. estimation have mostly focused on engineering the lossfunction

      醉经的在自监督单目深度估计的工作大多集中在loss function的设计

    1. Higher resolution gives the Base-*methods an advantage in depth accuracy, but on the otherhand these methods are more prone to outliers.

      吹嘘自己

    2. This enhances the smooth-ness of estimated flow fields and the sharpness of motiondiscontinuities

      就是这个意思,约束周围的深度不能过分差异

    3. Therotation r = θv is an angle axis representation with angle θand axis v. The translation t is given in Cartesian coordi-nates.

      r是由angle和axis表示的,t是直角坐标表示的

    4. By feeding optical flow estimate into the secondencoder-decoder we let it make use of motion parallax

      怎么样使得光溜能够在第二个ed中使用,利用了视差

    5. takes as input the optical flow, its confidence, the im-age pair, and the second image warped with the estimatedflow field. Based on these inputs it estimates depth, sur-face normals, and camera motion.

      第2个encoder-decoder input:第一个encoder的光流,置信度,一对图像,用估计的flow wrap的第二张图 output:depth,表面法向量,相机运动

    6. The network estimates not only depthand motion, but additionally surface normals, optical flowbetween the images and confidence of the matching

      该网络可以同时估计: 1.深度 2.相机运动 3.平面法向量 4.光溜 5.匹配置信度

    7. DeMoN takes an image pair as input and predicts the depth map of the first image and the relativepose of the second camera.

      输入是image pair, 预测出第一幅图的depth,以及第2张图的相对位姿

    8. In this paper, we succeed for the first time in training aconvolutional network to jointly estimate the depth and thecamera motion from an unconstrained pair of images.

      主要工作

    1. XY Z only

      定义两个只包含xyz的点云,这两个点云的数量可能不一样,或者是点云之间没有对应?(或者是指单个point并没有严格的对应)

    2. given input point clouds from two consec-utive frames (point cloud 1 and point cloud 2), our networkestimates a translational flow vector for every point in thefirst frame to indicate its motion between the two frames

      给定两连续帧的点云,得出点云1中每个点到点云2的rt

    1. To increase the localization and mapping accuracy, weremove any of the following points

      对点云进行的筛选操作,很有借鉴意义 1.视场边界的地方,解释说有较大的曲率,>17° 2.算出来的I(P)值过大或者过小,可能觉得异常 3.算出来的角度Θ过大或者过小 4.在边界处去除那个较远的点

    2. Our con-tributions are

      主要贡献 1.LOAM算法改写,使他适合于livox这种模式的雷达 2.小trick优化LOAM算法 3.(重点)运动问题解决

    1. Fig. 9. Statistics from real sensor data: Dispersion in the localization of thereference points (a) and calibration deviation from the final result (linear, inm, and angular, in rad) (b)

      实验结论

    2. Note thatthis condition would not be fulfilled when calibrating a front-facing 360° LiDAR scanner and a rear-looking camera, forinstance

      啥意思

    3. The goal of the registration step is to find the optimaltransformation parametersˆθso that when the resulting trans-formationˆTis applied

      目的

    4. To increase the robustness of thealgorithm,

      增加鲁棒性,采取了几个措施,肯定要多帧,单帧按照上面的提取方法完全有可能提取不到信息。

    5. 2D circle segmentation isused to extract a model of the pattern holes present inP4

      3.P4上2d圆检测,设定很多个阈值来判断4个hole,繁琐

    6. P3contains only points representing the edges of the calibrationtarget; that is, the outer borders and the holes.

      只包含目标hole边界的点,首先P2是边界点云,第一步又拟合出了target周围的平面,两个求个交集,同时定义一个阈值提高一下精度

    7. The intended outcome is an estimate ofthe 3D location of the centers in sensor coordinates.

      开始对立体性质的4个hole中心点位置进行估计

    8. pass-through filters are applied in the threecartesian coordinates to remove points outside the area wherethe target is to be placed

      对点云精心预处理,剔除掉不是target周围的点?

    9. In all cases, the output ofthis stage is a set of four 3D points representing the center ofthe holes in the target, in local coordinates

      定位输出的结果就是4个hole中心点的三维坐标

    10. As noticeable in the two different embodiments shown inFig. 2,

      标定板的样子,4个aruco marker在四周,中间4个圆圈,由后文可知,这个摆放方式需要固定