Skip to content

Tools to evaluate object detection results using the KITTI dataset.

Notifications You must be signed in to change notification settings

cguindel/eval_kitti

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

28 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

eval_kitti

Build Status License: CC BY-NC-SA

The eval_kitti software contains tools to evaluate object detection results using the KITTI dataset. The code is based on the KITTI object development kit. This version is useful to validate object detection approaches before submitting results to the official KITTI ranking.

Tools

  • evaluate_object is an improved version of the official KITTI evaluation that enables multi-class evaluation and splits of the training set for validation. It's updated according to the modifications introduced in 2017 by the KITTI authors.
  • parser is meant to provide mAP and mAOS stats from the precision-recall curves obtained with the evaluation script (updated on November 2019 to reflect changes in the official evaluation)
  • create_link is a helper that can be used to create a link to the results obtained with lsi-faster-rcnn.

Usage

Build evaluate_object with CMake:

mkdir build
cd build
cmake ..
make

The evaluate_object executable will be then created inside build. The following folders are also required to be placed there in order to perform the evaluation:

  • data/object/label_2, with the KITTI dataset labels.
  • lists, containing the .txt files with the train/validation splits. These files are expected to contain a list of the used image indices, one per row, following these examples.
  • results, in which a subfolder should be created for every test, including a second-level data folder with the resulting .txt files to be evaluated.

Typical evaluation pipeline

  1. Get results using the object detector that you want to evaluate. Results must follow the KITTI format; please refer to the KITTI object detection devkit for further information. You should use a different set of frames for training and validation. As a result, you will have a folder with thousands of txt files, one per validation frame, and typically with non-correlative filenames.
  2. Create a folder within eval_litti/build/results with the name of the experiment. Create a symbolic link to the folder containing your resulting txt files inside the newly created folder and rename it as data; e.g., if the experiment is named exp1, txt files to be evaluated should be accessible at eval_litti/build/results/exp1/data.
  3. Go to eval_kitti/build and run evaluate_object with the name of the experiment and the txt file containing the validation split as arguments. For instance, if you created the folder eval_litti/build/results/exp1 and are using the valsplit validation set from here, you should run ./evaluate_object exp1 valsplit.
  4. Wait until the script is completed. Modify lines 8 and 11 of eval_kitti/parser.py to specify the classes and parameters for which you want average results; by default, you will obtain the same stats as in the KITTI benchmark. Then, go back to eval_kitti and run ./parser.py passing the name of the experiment as an argument. In our example, python parser.py exp1. Average stats (AP for detection, AOS for orientation, etc.) will be printed.

Copyright

This work is a derivative of The KITTI Vision Benchmark Suite by A. Geiger, P. Lenz, C. Stiller and R. Urtasun, used under CC BY-NC-SA. Consequently, code in this repository is published under the same Creative Commons Attribution-NonCommercial-ShareAlike 3.0 License. This means that you must attribute the work in the manner specified by the authors, you may not use this work for commercial purposes and if you alter, transform, or build upon this work, you may distribute the resulting work only under the same license.

About

Tools to evaluate object detection results using the KITTI dataset.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published