Skip to content

A Python/C library for decoding DVS encodings from binary files to structured NumPy arrays of events.

License

Notifications You must be signed in to change notification settings

stevenabreu7/expelliarmus

 
 

Repository files navigation

expelliarmus

PyPI codecov contributors Documentation Status Discord

A Python/C library for decoding DVS binary data formats to NumPy structured arrays.

Supported formats

  • DAT (Prophesee).
  • EVT2 (Prophesee).
  • EVT3 (Prophesee).

Installation

You can install the library through pip:

pip install expelliarmus 

The package is tested on Windows, MacOS and Linux. Join us on Discord to propose features or to signal bugs!

Documentation

Check out readthedocs!

Quickstart

Shall we start practicing some spells? For that, we need a Wizard!

from expelliarmus import Wizard

Let's cast a spell called read() and read this RAW file to a structured NumPy array!

wizard = Wizard(encoding="evt3", fpath="./pedestrians.raw")
arr = wizard.read()
print(arr.shape) # Number of events encoded to the NumPy array.
(39297796,)

The array is a collection of (timestamp, x_address, y_address, polarity) tuples.

print(arr.dtype)
[('t', '<i8'), ('x', '<i2'), ('y', '<i2'), ('p', 'u1')]

A typical sample looks like this:

print(arr[0])
(5840504, 707, 297, 0)

If we would like to reduce the EVT3 file size, we can use the cut(fpath_in, fpath_out, new_duration) spell to limit the recording time duration to 12ms, for instance:

nevents = wizard.cut(fpath_out="./pedestrians_cut.raw", new_duration=12)
print(f"Number of events embedded in the cut file: {nevents}.") # The number of events embedded in the output file.
Number of events embedded in the cut file: 540.

This can be verified by reading the new file in an array.

cut_arr = wizard.read(fpath="./pedestrians_cut.raw")
print(f"Length of array extracted from the cut recording: {len(cut_arr)}.")
Length of array extracted from the cut recording: 540.

The files are consistent:

print(f"First original sample: {arr[0]} | First cut sample: {cut_arr[0]}.")
print(f"{nevents}th original sample: {arr[nevents-1]} | Last cut sample: {cut_arr[-1]}.")
print((arr[:nevents]==cut_arr[:]).all())
First original sample: (5840504, 707, 297, 0) | First cut sample: (5840504, 707, 297, 0).
540th original sample: (5853218, 1208, 253, 0) | Last cut sample: (5853218, 1208, 253, 0).
True

The time duration is, more or less, the desired one (the events are discrete, hence we have not a fine control over them).

print(f"New recording duration: {((cut_arr['t'][-1] - cut_arr['t'][0])/1000):.2f} ms") 
New recording duration: 12.71 ms

What if you wand is not strong enough for handling spells on very large recordings? Well, we can try to read the files one chunk at time...

wizard.set_chunk_size(chunk_size=512)
print(f"Length of the chunk: {len(next(wizard.read_chunk()))}.")
Length of the chunk: 512.

Let's read less events, so that we are able to visualize them

wizard.set_chunk_size(chunk_size=16)
print(next(wizard.read_chunk()))
[(5848837,  610, 296, 1) (5848843,  834, 302, 1) (5848846,  593, 254, 1)
 (5848846, 1003, 298, 1) (5848859,  610, 299, 1) (5848887,  709, 306, 0)
 (5848888,  756, 292, 0) (5848895,  704, 300, 0) (5848903,  744, 169, 1)
 (5848904, 1209, 252, 0) (5848905,  709, 307, 0) (5848911,  139, 315, 0)
 (5848918,  603, 301, 1) (5848918,  708, 299, 1) (5848924,  778, 295, 1)
 (5848967,  140, 315, 0)]

A small benchmark

Here it is a small benchmark using expelliarmus on the file formats supported. The data shows the file size, read time for the full file and read time for reading the file in chunks. The performance is compared against HDF5, HDF5 LZF, HDF5 GZIP and NumPy.

full_read

==================================================
Full file read
==================================================
DAT (413MB), execution time: 0.248s.
HDF5 (826MB,  100.00%), execution time: 0.361s,  45.78%.
HDF5 GZIP (163MB, -60.53%), execution time: 2.348s,  847.56%.
HDF5 LZF (316MB, -23.49%), execution time: 1.419s,  472.77%.
NumPy (826MB,  100.00%), execution time: 0.138s, -44.16%.
==================================================
EVT2 (157MB), execution time: 0.221s.
HDF5 (621MB,  295.54%), execution time: 0.252s,  14.00%.
HDF5 GZIP (156MB, -0.64%), execution time: 2.111s,  854.72%.
HDF5 LZF (276MB,  75.80%), execution time: 1.206s,  445.73%.
NumPy (621MB,  295.54%), execution time: 0.092s, -58.25%.
==================================================
EVT3 (350MB), execution time: 1.824s.
HDF5 (1701MB,  386.00%), execution time: 0.690s, -62.18%.
HDF5 GZIP (419MB,  19.71%), execution time: 5.533s,  203.37%.
HDF5 LZF (746MB,  113.14%), execution time: 3.009s,  64.99%.
NumPy (1701MB,  386.00%), execution time: 0.259s, -85.79%.

chunk_read

==================================================
Chunk reading.
==================================================
DAT (413MB), execution time: 0.625s.
HDF5 (826MB,  100.00%), execution time: 1.992s,  218.60%.
HDF5 LZF (316MB, -23.49%), execution time: 3.939s,  530.10%.
HDF5 GZIP (163MB, -60.53%), execution time: 5.675s,  807.75%.
==================================================
EVT2 (157MB), execution time: 0.282s.
HDF5 (621MB,  295.54%), execution time: 1.298s,  360.61%.
HDF5 LZF (276MB,  75.80%), execution time: 3.511s,  1146.43%.
HDF5 GZIP (156MB, -0.64%), execution time: 5.488s,  1848.33%.
==================================================
EVT3 (350MB), execution time: 1.795s.
HDF5 (1701MB,  386.00%), execution time: 3.944s,  119.79%.
HDF5 LZF (746MB,  113.14%), execution time: 10.360s,  477.28%.
HDF5 GZIP (419MB,  19.71%), execution time: 17.359s,  867.31%.

Contributing

Please check our documentation page for more details on contributing.

About

A Python/C library for decoding DVS encodings from binary files to structured NumPy arrays of events.

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages

  • Python 57.3%
  • C 42.7%