Lossless Compression of Neuromorphic Vision Sensor Data Based on Point Cloud Representation
Affiliation
Kingston University London; University of ChesterPublication Date
2022-11-14
Metadata
Show full item recordAbstract
Visual information varying over time is typically captured by cameras that acquire data via images (frames) equally spaced in time. Using a different approach, Neuromorphic Vision Sensors (NVSs) are emerging visual capturing devices that only acquire information when changes occur in the scene. This results in major advantages in terms of low power consumption, wide dynamic range, high temporal resolution, and lower data rates than conventional video. Although the acquisition strategy already results in much lower data rates than conventional video, such data can be further compressed. To this end, in this paper we propose a lossless compression strategy based on point cloud compression, inspired by the observation that, by appropriately reporting NVS data in a $(x,y,t)$ tridimensional space, we have a point cloud representation of NVS data. The proposed strategy outperforms the benchmark strategies resulting in a compression ratio up to 30% higher for the considered.Citation
Martini, M., Adhuran, J., & Khan, K. (2022). Lossless compression of neuromorphic vision sensor data based on point cloud representation. IEEE Access, 10, 121352-121364. https://doi.org/10.1109/ACCESS.2022.3222330.Publisher
IEEEJournal
IEEE AccessAdditional Links
https://ieeexplore.ieee.org/document/9950514Type
ArticleEISSN
2169-3536ae974a485f413a2113503eed53cd6c53
10.1109/ACCESS.2022.3222330
Scopus Count
Collections
The following license files are associated with this item:
- Creative Commons