Skip to content

Implementation of Visual SLAM concepts for course Deep Learning for Robot Perception.

Notifications You must be signed in to change notification settings

kumar-akshay324/visual-odom-dl-project

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

16 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Visual SLAM - Visual Odometry Estimation using Deep Learning on the KITTI Dataset

Course Project for Deep Learning for Robot Perception

THIS PROJECT IS STILL UNDER HEAVY DEVELOPMENT

Dependencies

  • Numpy
  • Keras
  • Tensorflow(Ideal backend to go with the data_format used in Keras implementation)
  • OpenCV
  • transform3d

Data Folder Structure

├── dataset
	├── poses
	│   ├── 00.txt
	│   ├── 01.txt
	│   ├── 02.txt
	│   ├── 03.txt
	│   ├── 04.txt
	│   ├── 05.txt
	│   ├── 06.txt
	│   ├── 07.txt
	│   ├── 08.txt
	│   ├── 09.txt
	│   └── 10.txt
	└── sequences
	    └── 00
	        ├── calib.txt
	        └── image_0
	        	├──00000.png
	        	└──00001.png
	        ├── image_1
	        └── times.txt

Note: You should not use the claibration files provided by default inside the camera image dataset, but one after merging the calibration data into it.

File Description

  • odometry.py - Extract the poses for corresponding stacked images as a stacked batch
  • vo_estimation.py - Process the stacked images and run the DL model on it
  • process_images.py - pyKITTI based implemention - Still under development

Usage

The complete implementation is to be run using the make filesystem with commands inside the Makefile

The parameters to be modified in the script are: * data_params = [path_to_poses, image_sequences, pose_verbosity] where image_sequences = [training sequences list] * cnn_model_params = [batch_size] * rnn_model_params = [time_steps]

  • Run make run to obtain a batch of training images. Further integration with training and testing model is in progress.

The parameters to be modified in the script are: * data_params = [path_to_dataset, image_sequences, image_ratio,, image_verbosity] where image_sequences = [training sequences list, test sequences list] * rnn_model_params = [time_step, LSTM_nodes] * cnn_model_params = [batch_size, ] * operation_flags = [debug_verbosity, [train_image_start_pointer, test_image_start_pointer]]

Authors

About

Implementation of Visual SLAM concepts for course Deep Learning for Robot Perception.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published