Skip to content

Commit 9e6c707

Browse files
committed
added figure to README
1 parent 848865d commit 9e6c707

File tree

2 files changed

+16
-2
lines changed

2 files changed

+16
-2
lines changed

README.md

Lines changed: 16 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,20 @@
11
# Masked Event Modelling: Self-Supervised Pretraining for Event Cameras
2+
Link to paper: https://arxiv.org/abs/2212.10368v1
23

3-
Event cameras asynchronously capture brightness changes with low latency, high temporal resolution, and high dynamic range. Deploying deep learning methods for classification or other tasks to these sensors typically requires large labeled datasets. However, annotation of event data is a costly and laborious process. To reduce the dependency on labeled event data, we introduce Masked Event Modeling (MEM), a self-supervised pretraining framework for events. Our method pretrains a neural network on unlabeled events, which can originate from any event camera recording. Subsequently, the pretrained model is finetuned on a downstream task, leading to a consistently improved performance on the task while requiring fewer labels. Our method outperforms the state-of-the-art object classification across three datasets, N-ImageNet, N-Cars, and N-Caltech101, increasing the top-1 accuracy of previous work by significant margins. We further find that MEM is even superior to supervised RGB-based pretraining when tested on real-world event data. Models pretrained with MEM exhibit improved label efficiency, especially in low data regimes, and generalize well to the dense task of semantic image segmentation.
4+
Event cameras asynchronously capture brightness changes with low latency, high temporal resolution, and high dynamic range. However, annotation of event data is a costly and laborious process, which limits the use of deep learning methods for classification and other semantic tasks with the event modality. To reduce the dependency on labeled event data, we introduce Masked Event Modeling (MEM), a self-supervised framework for events. Our method pretrains a neural network on unlabeled events, which can originate from any event camera recording. Subsequently, the pretrained model is finetuned on a downstream task, leading to a consistent improvement of the task accuracy.
5+
For example, our method reaches state-of-the-art classification accuracy across three datasets, N-ImageNet, N-Cars, and N-Caltech101, increasing the top-1 accuracy of previous work by significant margins. When tested on real-world event data, MEM is even superior to supervised RGB-based pretraining. The models pretrained with MEM are also label-efficient and generalize well to the dense task of semantic image segmentation.
6+
7+
![MEM](mem-pipeline-figure.png)
8+
9+
Citation:
10+
```
11+
@inproceedings{masked-event-modeling,
12+
title={Masked Event Modeling: Self-Supervised Pretraining for Event Cameras},
13+
author={Simon Klenk, David Bonello, Lukas Koestler, Nikita Araslanov and Daniel Cremers},
14+
booktitle=WACV,
15+
year={2024}
16+
}
17+
```
418

519
## Setup
620

@@ -97,7 +111,7 @@ cd mem
97111
Follow the README in `mem/semantic_segmentation` to train a semantic segmentation model.
98112

99113

100-
## Atribution
114+
## Attribution
101115

102116
- This code is based on Microsoft's BEiT code (https://github.com/microsoft/unilm/tree/master/beit).
103117
- The eventVAE code is based on Phil Wang's code (https://github.com/lucidrains/DALLE-pytorch)

mem-pipeline-figure.png

202 KB
Loading

0 commit comments

Comments
 (0)