An Invitation to 3D Vision is an introductory tutorial on 3D computer vision (a.k.a. geometric vision or visual geometry or multiple-view geometry). It aims to make beginners understand basic theories on 3D vision and implement its applications using OpenCV. In addition to tutorial slides, example codes are provided in the purpose of education. They include simple but interesting and practical applications. The example codes are written as short as possible (mostly less than 100 lines) to be clear and easy to understand.
- To clone this repository (codes and slides):
git clone https://github.com/mint-lab/3dv_tutorial.git
- To fork this repository to your Github: Click here
- To download codes and slides as a ZIP file: Click here
- 📝 How to run example codes in Python
- 📝 How to run example codes in C++
- The main title, An Invitation to 3D Vision, came from a legendary book by Yi Ma, Stefano Soatto, Jana Kosecka, and Shankar S. Sastry. We wish that our tutorial will be the first gentle invitation card for beginners to 3D vision and its applications.
- The subtitle, for everyone, was inspired from Prof. Kim's online lecture (in Korean). Our tutorial is also intended not only for students and researchers in academia, but also for hobbyists and developers in industries. We tried to describe important and typical problems and their solutions in OpenCV. We hope readers understand it easily without serious mathematical background.
- Section 1. Introduction
- Section 2. Single-view Geometry
- Section 3. Two-view Geometry
- Section 4. Solving Problems
- Section 5. Finding Correspondence
- Section 6. Multiple-view Geometry
- Special Topic) Bayesian Filtering
- Section 7. Visual SLAM and Odometry
- Section 1. Introduction [slides]
-
Section 2. Single-view Geometry [slides]
- Getting Started with 2D
- 3D rotation conversion [python]
- Pinhole Camera Model
- Geometric Distortion Models
- Geometric distortion visualization [python]
- Geometric distortion correction [python] [cpp] [result video]
- Camera Calibration
- Absolute Camera Pose Estimation (a.k.a. perspective-n-point; PnP)
- Pose estimation (chessboard) [python] [cpp] [result video]
- Pose estimation (book) [python] [cpp]
- Pose estimation (book) with camera calibration [python] [cpp]
- Pose estimation (book) with camera calibration without initial
$K$ [python] [cpp] [result video]
- Getting Started with 2D
-
Section 3. Two-view Geometry [slides]
- Planar Homography
- Epipolar Geometry
- Epipolar line visualization [python]
- Relative Camera Pose Estimation
- Fundamental matrix estimation [python]
- Monocular visual odometry (epipolar version) [python] [cpp] [result video]
- Triangulation
-
Section 4. Solving Problems [slides]
- Solving Linear Equations in 3D Vision
- Solving Nonlinear Equations in 3D Vision
-
Section 5. Finding Correspondence [slides]
- Feature Points and Descriptors
- Harris corner
- Feature point comparison
- Feature Matching and Tracking
- Feature matching comparison
- Feature tracking with KLT tracker
- Outlier Rejection
- Feature Points and Descriptors
-
Section 6. Multiple-view Geometry
- Bundle Adjustment
- Structure-from-Motion (SfM)
- Global SfM
- Incremental SfM
- Section 7. Visual SLAM and Odometry
The authors thank the following contributors and projects.
- Jae-Yeong Lee: He motivated many examples.
- Giseop Kim: He contributed the initial version of SfM codes with Toy-SfM and cvsba.
- The KITTI Vision Benchmark Suite: The KITTI odometry dataset #07 was used to demonstrate visual odometry and SLAM.
- Russell Hewett: His two hill images were used to demonstrate image stitching.
- Kang Li: His shaking CCTV video was used to demonstrate video stabilization.
- Richard Blais: His book cover and video in the OpenCV tutorial were used to demonstrate camera pose estimation and augmented reality.