SoK: Vehicle Orientation Representations for Deep Rotation Estimation

Overview

SoK: Vehicle Orientation Representations for Deep Rotation Estimation

Raymond H. Tu, Siyuan Peng, Valdimir Leung, Richard Gao, Jerry Lan

This is the official implementation for the paper SoK: Vehicle Orientation Representations for Deep Rotation Estimation

Model Diagram

Table of Conents

Envrionment Setup

Install required packages via conda

# create conda environment based on yml file
conda env update --file environment.yml
# activate conda environment
conda activate KITTI-Orientation

Clone git repo:

git clone [email protected]:umd-fire-coml/KITTI-orientation-learning.git

Training

Check training.sh for example training script

Training Parameter setup:

Training parameters can be configured using cmd arguments

  • --predict: Specify prediction target. Options are rot-y, alpha
  • --converter: Specify prediction method. Options are alpha, rot-y, tricosine, multibin, voting-bin, single-bin
  • --kitti_dir: path to kitti dataset directory. Its subdirectory should have training/ and testing/ Default path is dataset/
  • --training_record: root directory of all training record, parent of weights and logs directory. Default path is training_record
  • --resume: Resume from previous training under training_record directory
  • --add_pos_enc: Add positional encoding to input
  • --add_depth_map: Add depth map information to input

For all the training parameter setup, please using

python3 model/training.py -h

Training Result

Exp ID Target Loss Functions Additional Inputs Accuracy (%)
E1 rot-y L2 Loss - 90.490
E2 rot-y Angle Loss - 89.052
E3 alpha L2 Loss - 90.132
E4 Single Bin L2 Loss - 94.815
E5 Single Bin L2 Loss Pos Enc 94.277
E6 Single Bin L2 Loss Dep Map 93.952
E7 Voting Bins (4-Bin) L2 Loss - 93.609
E8 Tricosine L2 Loss - 94.249
E9 Tricosine L2 Loss Pos Enc 94.351
E10 Tricosine L2 Loss Dep Map 94.384
E11 2 Conf Bins L2(Bins,Confs) - 83.304
E12 4 Conf Bins L2(Bins,Confs) - 88.071
Owner
FIRE Capital One Machine Learning of the University of Maryland
FIRE Capital One Machine Learning is a Course-based Undergrad Research Experience that provides undergrad students with research experience in Machine Learning.
FIRE Capital One Machine Learning of the University of Maryland
Unity Propagation in Bayesian Networks Handling Inconsistency via Unity Smoothing

This repository contains the scripts needed to generate the results from the paper Unity Propagation in Bayesian Networks Handling Inconsistency via U

0 Jan 19, 2022
Code for DeepCurrents: Learning Implicit Representations of Shapes with Boundaries

DeepCurrents | Webpage | Paper DeepCurrents: Learning Implicit Representations of Shapes with Boundaries David Palmer*, Dmitriy Smirnov*, Stephanie Wa

Dima Smirnov 36 Dec 08, 2022
This repository contains the code needed to train Mega-NeRF models and generate the sparse voxel octrees

Mega-NeRF This repository contains the code needed to train Mega-NeRF models and generate the sparse voxel octrees used by the Mega-NeRF-Dynamic viewe

cmusatyalab 260 Dec 28, 2022
A library for hidden semi-Markov models with explicit durations

hsmmlearn hsmmlearn is a library for unsupervised learning of hidden semi-Markov models with explicit durations. It is a port of the hsmm package for

Joris Vankerschaver 69 Dec 20, 2022
Code for EmBERT, a transformer model for embodied, language-guided visual task completion.

Code for EmBERT, a transformer model for embodied, language-guided visual task completion.

41 Jan 03, 2023
Code for "Unsupervised Layered Image Decomposition into Object Prototypes" paper

DTI-Sprites Pytorch implementation of "Unsupervised Layered Image Decomposition into Object Prototypes" paper Check out our paper and webpage for deta

40 Dec 22, 2022
tinykernel - A minimal Python kernel so you can run Python in your Python

tinykernel - A minimal Python kernel so you can run Python in your Python

fast.ai 37 Dec 02, 2022
Source code for our paper "Do Not Trust Prediction Scores for Membership Inference Attacks"

Do Not Trust Prediction Scores for Membership Inference Attacks Abstract: Membership inference attacks (MIAs) aim to determine whether a specific samp

<a href=[email protected]"> 3 Oct 25, 2022
an implementation of softmax splatting for differentiable forward warping using PyTorch

softmax-splatting This is a reference implementation of the softmax splatting operator, which has been proposed in Softmax Splatting for Video Frame I

Simon Niklaus 338 Dec 28, 2022
MatchGAN: A Self-supervised Semi-supervised Conditional Generative Adversarial Network

MatchGAN: A Self-supervised Semi-supervised Conditional Generative Adversarial Network This repository is the official implementation of MatchGAN: A S

Justin Sun 12 Dec 27, 2022
Official Pytorch implementation of "Learning Debiased Representation via Disentangled Feature Augmentation (Neurips 2021, Oral)"

Learning Debiased Representation via Disentangled Feature Augmentation (Neurips 2021, Oral): Official Project Webpage This repository provides the off

Kakao Enterprise Corp. 68 Dec 17, 2022
Lightweight stereo matching network based on MobileNetV1 and MobileNetV2

MobileStereoNet: Towards Lightweight Deep Networks for Stereo Matching

Cognitive Systems Research Group 139 Nov 30, 2022
Resco: A simple python package that report the effect of deep residual learning

resco Description resco is a simple python package that report the effect of dee

Pierre-Arthur Claudé 1 Jun 28, 2022
[CVPR22] Official codebase of Semantic Segmentation by Early Region Proxy.

RegionProxy Figure 2. Performance vs. GFLOPs on ADE20K val split. Semantic Segmentation by Early Region Proxy Yifan Zhang, Bo Pang, Cewu Lu CVPR 2022

Yifan 54 Nov 29, 2022
A small demonstration of using WebDataset with ImageNet and PyTorch Lightning

A small demonstration of using WebDataset with ImageNet and PyTorch Lightning This is a small repo illustrating how to use WebDataset on ImageNet. usi

50 Dec 16, 2022
Instance-wise Occlusion and Depth Orders in Natural Scenes (CVPR 2022)

Instance-wise Occlusion and Depth Orders in Natural Scenes Official source code. Appears at CVPR 2022 This repository provides a new dataset, named In

27 Dec 27, 2022
Predicting path with preference based on user demonstration using Maximum Entropy Deep Inverse Reinforcement Learning in a continuous environment

Preference-Planning-Deep-IRL Introduction Check my portfolio post Dependencies Gym stable-baselines3 PyTorch Usage Take Demonstration python3 record.

Tianyu Li 9 Oct 26, 2022
darija <-> english dictionary

darija-dictionary Having advanced IT solutions that are well adapted to the Moroccan context passes inevitably through understanding Moroccan dialect.

DODa 102 Jan 01, 2023
RE3: State Entropy Maximization with Random Encoders for Efficient Exploration

State Entropy Maximization with Random Encoders for Efficient Exploration (RE3) (ICML 2021) Code for State Entropy Maximization with Random Encoders f

Younggyo Seo 47 Nov 29, 2022
Using multidimensional LSTM neural networks to create a forecast for Bitcoin price

Multidimensional LSTM BitCoin Time Series Using multidimensional LSTM neural networks to create a forecast for Bitcoin price. For notes around this co

Jakob Aungiers 318 Dec 14, 2022