Learning-Based Feature Detection and Matching
Learn to detect features Learn to correspond features between frames via their descriptors
Deep Visual Odometry
Monocular Depth Estimation End-to-End Visual Odometry (CNN-LSTM architecture capable of predicting pose from image sequences)
Learning-Based Calibration
Predict camera intrinsics from a single image, collects features prone to distortion and FOV (outputs intrinsics).
- trains on synthetic data with known intrinsics
- real data with already estimated intrinsics from SfM
- useful for autocalibration Extrinsic Calibration (predict extrinsics of camera in lidar frame)
Loop Closure Detection
Bag of words Place recognition
End-to-End Learning for SLAM
Take in an image sequence and predict depth and pose for each frame.
- Build a map from those depths and poses which can be differentiable Learn a map representation through pose net and a branched off NeRF model to do neural rendering
Outlier Rejection
Learned RANSAC
