Export Publication

The publication can be exported in the following formats: APA (American Psychological Association) reference format, IEEE (Institute of Electrical and Electronics Engineers) reference format, BibTeX and RIS.

Export Reference (APA)
Ramna Maqsood, Nunes, P., Conti, C. & Soares, L. D. (2025). WaveE2VID: Frequency-Aware Event-Based Video Reconstruction. In 2025 IEEE International Conference on Image Processing (ICIP). (pp. 570-575). Anchorage, AK, USA: IEEE.
Export Reference (IEEE)
R. Maqsood et al.,  "WaveE2VID: Frequency-Aware Event-Based Video Reconstruction", in 2025 IEEE Int. Conf. on Image Processing (ICIP), Anchorage, AK, USA, IEEE, 2025, pp. 570-575
Export BibTeX
@inproceedings{maqsood2025_1764931232414,
	author = "Ramna Maqsood and Nunes, P. and Conti, C. and Soares, L. D.",
	title = "WaveE2VID: Frequency-Aware Event-Based Video Reconstruction",
	booktitle = "2025 IEEE International Conference on Image Processing (ICIP)",
	year = "2025",
	editor = "",
	volume = "",
	number = "",
	series = "",
	doi = "10.1109/ICIP55913.2025.11084548",
	pages = "570-575",
	publisher = "IEEE",
	address = "Anchorage, AK, USA",
	organization = ""
}
Export RIS
TY  - CPAPER
TI  - WaveE2VID: Frequency-Aware Event-Based Video Reconstruction
T2  - 2025 IEEE International Conference on Image Processing (ICIP)
AU  - Ramna Maqsood
AU  - Nunes, P.
AU  - Conti, C.
AU  - Soares, L. D.
PY  - 2025
SP  - 570-575
DO  - 10.1109/ICIP55913.2025.11084548
CY  - Anchorage, AK, USA
AB  - Event cameras, which detect local brightness changes instead of capturing full-frame images, offer high temporal resolution and low latency. Although existing convolutional neural networks (CNNs) and transformer-based methods for event-based video reconstruction have achieved impressive results, they suffer from high computational costs due to their linear operations. These methods often require 10M-30M parameters and inference times of 30-110 ms per forward pass at a resolution of 640 × 480 on modern GPUs. Furthermore, to reduce computational costs, these methods apply CNN-based downsampling, which leads to the loss of fine details. To address these challenges, we propose an efficient hybrid model, WaveE2VID, which combines the frequency-domain analysis of the wavelet transform with the spatio-temporal context modeling of a deep convolutional recurrent network. Our model achieves 50% faster inference speed and lower GPU memory usage than CNN and transformer-based methods, maintaining reconstruction performance on par with state-of-the-art approaches
across benchmark datasets.
ER  -