This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

EventCLIP: Adapting CLIP for Event-based Object Recognition

Ziyi Wu1,2, Xudong Liu1, Igor Gilitschenski1,2
1University of Toronto, 2Vector Institute
{ziyiwu,xudong,gilitschenski}@cs.toronto.edu
Abstract

Recent advances in zero-shot and few-shot classification heavily rely on the success of pre-trained vision-language models (VLMs) such as CLIP. Due to a shortage of large-scale datasets, training such models for event camera data remains infeasible. Thus, adapting existing VLMs across modalities to event vision is an important research challenge. In this work, we introduce EventCLIP, a novel approach that utilizes CLIP for zero-shot and few-shot event-based object recognition. We first generalize CLIP’s image encoder to event data by converting raw events to 2D grid-based representations. To further enhance performance, we propose a feature adapter to aggregate temporal information over event frames and refine text embeddings to better align with the visual inputs. We evaluate EventCLIP on N-Caltech, N-Cars, and N-ImageNet datasets, achieving state-of-the-art few-shot performance. When fine-tuned on the entire dataset, our method outperforms all existing event classifiers. Moreover, we explore practical applications of EventCLIP including robust event classification and label-free event recognition, where our approach surpasses previous baselines designed specifically for these tasks.

1 Introduction

Event-based cameras have recently gained significant interest in computer vision due to their high temporal resolution, low energy consumption, and high dynamic range properties [13]. Event-based vision has shown promising results in various applications, such as object recognition [16, 32, 58], detection [48, 19, 38], tracking [12, 15, 18], and optical flow estimation [85, 20, 3]. However, this novel imaging modality poses unique challenges, including the need for specialized models to handle the asynchronous nature of events, and the lack of large-scale datasets. As in classical recognition problems, newly captured event data can contain objects from categories that are not present in the training set of deployed models. In such cases, trained models will fail, and it may be infeasible to re-train the model every time a new object category is introduced, motivating the need for event-based zero-shot and few-shot recognition systems.

Refer to caption
(a) Open-set and robust classification
Refer to caption
(b) Learning from unlabeled data
Figure 1: Existing event-based classifiers are trained from scratch on closed-set datasets. They fail on events with unseen categories, camera motions, or lighting conditions. Instead, our method utilizes pre-trained CLIP for zero-shot or few-shot open-world event recognition. Moreover, EventCLIP can be applied to the task of (a) robust event classification by ensembling with trained models, and (b) label-free learning by training on generated pseudo labels.

In frame-based vision, pre-trained vision-language models (VLMs) such as CLIP [50] have shown remarkable success in zero-shot and few-shot learning tasks. Trained on large-scale datasets, these models try to map paired images and texts to an aligned feature space. Open-world zero-shot classification is made possible by leveraging the feature similarity between unseen objects and texts with novel category names [50, 29]. Recently, several works have designed data-efficient methods to adapt CLIP under the few-shot learning setting for better accuracy [14, 83, 82, 78, 69]. However, there is currently no large-scale event-text dataset available, making it impossible to train such event-language models from scratch. This motivates us to ask the question: can 2D pre-trained VLMs be transferred to event-based vision and realize zero-shot or few-shot object recognition?

In this work, we propose EventCLIP as the first attempt introducing CLIP to event-based visual understanding. To bridge the gap between asynchronous event data and CLIP’s frame-based input representation, we split an event stream into multiple time windows, and convert each of them into a 2D frame. Following Radford et al. [50], text prompts are constructed by placing class names into hand-crafted templates, and text features are extracted as the zero-shot classifier weight. Each event frame is classified by CLIP individually, and the final result is obtained by simple voting.

Although EventCLIP can achieve zero-shot recognition, its performance still lags behind existing classifiers trained on event domain data. We thus propose to learn lightweight adapters to refine the pre-trained CLIP features. Different from previous work which only adapts one image feature [14] or a fixed number of features in pre-defined orders [79], the number and order of our event features depend on the camera trajectory. We thus design a Transformer-based adapter to aggregate temporal information from multiple frames. We also fine-tune text features as the weight of the output fully-connected layer in a classifier. With these designs, EventCLIP achieves more data-efficient few-shot learning compared to existing event-based classifiers.

Finally, we explore more applications of EventCLIP. We discover that the 2D pre-trained knowledge in CLIP, acquired from Internet-scale data, synergizes with the domain knowledge in models trained purely on event data. Therefore, we directly ensemble EventCLIP and existing event classifiers, which consistently improves their accuracy by more than 10% on N-ImageNet robustness subsets featuring unseen camera motions and lightnings [32]. In addition, since EventCLIP can classify unseen objects, we use it to label raw events. Leveraging the spatio-temporal property of events, we are able to select high-quality pseudo labels and achieve state-of-the-art unsupervised classification results.

In summary, this work makes four main contributions: (i) the first zero-shot open-world event-based object recognition method using CLIP, (ii) a Transformer-based feature adapter tailored to event temporal information aggregation, (iii) state-of-the-art few-shot and fine-tuning results on three datasets, and (iv) significantly improved robustness and unsupervised classification accuracy on N-ImageNet subsets.

2 Related Work

We briefly review recent works in event-based recognition, bridging general and event-based vision, and CLIP-based transfer learning, which is further expanded in Appendix A.

Deep Learning for Event-based Classification. Depending on the utilization of the asynchronous nature of events, existing event-based classifiers can be mainly categorized into two classes, namely, synchronous and asynchronous methods. Synchronous models aggregate events to a grid-based representation, followed by standard modules such as Convolutional Neural Networks (CNNs) [58, 16, 17, 32, 27]. Significant efforts have been made to achieve efficient and expressive event-to-frame conversion, such as binarized event occurrence [6], event counts [41], and sorted event timestamps [1]. Recently, EST [16] has achieved state-of-the-art results with an end-to-end learnable event-to-frame conversion pipeline. To improve the robustness against data noise, DiST [32] proposes to suppress noisy events with their spatio-temporal relationships, which has proved effective under camera motion and lighting change in data capture. Compared to asynchronous methods [39, 56, 42], synchronous models achieve consistently better results across datasets. As our primary goal is to achieve high accuracy instead of efficiency, we adopt representative synchronous event-based classifiers as our baselines in the experiments.

Bridging General and Event-based Vision. Inspired by the great success of classical computer vision, several works have introduced techniques from frame-based vision to process event data. Some papers focus on reconstructing natural images from events, and then apply conventional deep models on the converted frames [53, 54, 60, 57, 68, 5]. Yet, they often introduce large computational overhead which is at odds with event cameras’ low-latency nature. Closer to ours are methods that transfer knowledge learned from images to event-based models [27, 65, 61, 76, 77, 43]. However, they either require paired recordings of images and events, or massive labels on image data. In this work, we utilize CLIP pre-trained on RGB image-text pairs for data-efficient event-based classification. Our method converts events into frames via simple counting, and directly applies CLIP for zero-shot classification. We can further boost its performance via few-shot feature adaptation, without the need for paired RGB images or large amounts of labels.

CLIP-based Few-Shot Transfer Learning. Transfer learning aims to leverage large pre-trained models to facilitate learning in data-scarce scenarios. In event-based object recognition, existing methods also utilize models pre-trained on RGB images from ImageNet [9] as their backbones to improve performance [41, 16, 32, 1]. Trained on millions of image-text pairs, CLIP [50] learns transferable representations that are useful for downstream tasks. To further enhance its accuracy, some methods [83, 82, 73] insert learnable text tokens to perform task-specific prompt tuning, which requires backpropagation over the heavy text encoder. CLIP-Adapter [14], Tip-Adapter [78], and WiSE-FT [69] instead learn lightweight CLIP feature adapters.

Besides 2D image classification, CLIP has also been extended to 2D detection [23, 84, 31], segmentation [51, 80], and video analysis [66, 30]. Our work is inspired by PointCLIP [79, 87], which projects point clouds to multi-view images for CLIP-based zero-shot and few-shot 3D shape recognition. Unlike PointCLIP, event data usually only capture the boundary information of objects, which poses a greater domain gap than point clouds as point clouds often depict complete object surfaces. In addition, we design a Transformer-based adapter for event temporal information fusion, while PointCLIP simply uses an MLP since their multi-view projections follow a pre-defined order.

3 Method

EventCLIP builds upon large-scale pre-trained CLIP [50] (Sec. 3.1) and converts event streams to 2D images for zero-shot event understanding (Sec. 3.2). When a few labeled examples are provided, we learn lightweight feature adapters to further improve the few-shot accuracy (Sec. 3.3). Finally, we extend EventCLIP to more tasks including robust event classification and label-free event recognition in Sec. 3.4. The overall pipeline of EventCLIP is illustrated in Fig. 2.

3.1 Background: CLIP-based Image Classification

The training objective of CLIP is to map images and texts to a joint embedding space. CLIP consists of two encoders for image and text inputs, respectively. During training, given a batch of image-text pairs, CLIP maximizes the cosine similarity between embeddings of positive pairs, while minimizing it for negative pairs using a contrastive loss. CLIP is trained on a collection of 400 million web-crawled image-text data. The large pre-training dataset enables the incorporation of diverse visual concepts, thereby enhancing the transferability of the learned features to downstream tasks.

As CLIP is trained to match image and text features, it naturally lends itself to zero-shot classification. Formally, let 𝒇𝒙\bm{f}_{\bm{x}} be the image feature extracted by CLIP’s image encoder for an image 𝒙\bm{x}. Meanwhile, we construct text inputs by placing category names into a pre-defined template such as “a photo of a [CLASS]”, and leverage CLIP’s text encoder to extract a set of KK features W={𝒘i}i=1KW=\{\bm{w}_{i}\}_{i=1}^{K}, where KK denotes the number of classes. The probability of predicting class ii given image 𝒙\bm{x} is then computed as:

p(y=i|𝒙)=exp(cos(𝒘i,𝒇𝒙)/τ)j=1Kexp(cos(𝒘j,𝒇𝒙)/τ),p(y=i|\bm{x})=\frac{\exp(\cos(\bm{w}_{i},\bm{f}_{\bm{x}})/\tau)}{\sum_{j=1}^{K}\exp(\cos(\bm{w}_{j},\bm{f}_{\bm{x}})/\tau)}, (1)

where cos(,)\cos(\cdot,\cdot) denotes the cosine similarity between two vectors, and τ\tau is a scaling factor learned by CLIP. The zero-shot inference does not require any in-domain training data to fine-tune the model, but can achieve competitive results with fully supervised baselines on 2D image datasets [50].

3.2 Zero-Shot Event Understanding

Prior work in event recognition has shown that raw events can be converted to meaningful 2D frames such as edge maps [13]. This motivates us to prompt VLMs pre-trained in the RGB domain to process event camera data. We adopt CLIP due to its public availability and importance in the literature. However, the presented approach also generalizes to other VLMs as we will show in the experiments.

Bridging the Modality Gap. Event cameras record brightness changes at each sensing pixel, and output a sequence of events ={ei=(xi,yi,ti,pi)}\mathcal{E}=\{e_{i}=(x_{i},y_{i},t_{i},p_{i})\}, each parameterized by its spatial position (xi,yi)(x_{i},y_{i}), triggered timestamp tit_{i}, and polarity pi{1,1}p_{i}\in\{-1,1\}. Since raw events are asynchronous and sparse, they are represented as a set, which differs from the grid-like representations CLIP requires. To bridge this modality gap, we convert raw events into 2D frames. Specifically, we split an event stream \mathcal{E} to MM time windows {i}i=1M\{\mathcal{E}_{i}\}_{i=1}^{M} by grouping every NN consecutive events. Grouping with event count ensures better robustness against camera speed compared to using fixed time intervals [32]. For each i\mathcal{E}_{i}, we construct a 2-channel histogram by counting the number of positive and negative events at each pixel. To obtain a 3-channel image, we first normalize the histogram to the range of [0,1][0,1], and then colorize it with a pre-defined RGB color map. Finally, we set the empty pixels to pure white for better visual quality following Klenk et al. [35].

Zero-Shot Classification. After converting raw events to MM 2D frames, we leverage CLIP’s image encoder to extract their features F={𝒇i}i=1MF=\{\bm{f}_{i}\}_{i=1}^{M}. Following Radford et al. [50], we then construct text prompts with class names and hand-crafted templates, and use CLIP to extract text embeddings W={𝒘i}i=1KW=\{\bm{w}_{i}\}_{i=1}^{K}. The zero-shot prediction for each time window can be computed with Eq. 1. Here, the text template should reflect the domain-specific knowledge about event data, such as the visual property of the converted frames.

To obtain the final classification output, we need to aggregate predictions from MM time windows. PointCLIP [79, 87] also faces this problem as they project a point cloud to multiple views. They simply assign hyper-parameters to weigh the importance of each view, which are fixed through the dataset. This is feasible since the 3D point clouds they considered are all aligned to a canonical pose, and thus the projected views for different data follow the same order (e.g., {front, right, back, left, top, bottom}). However, the temporal order of our time windows depends on the event camera’s trajectory, which varies a lot across data samples, making a pre-defined set of weights sub-optimal. Inspired by DeepSets [75], we select the order-invariant set operation mean-pooling to average the classification probabilities from all time windows as the final prediction output.

Refer to caption
Figure 2: EventCLIP overview. Given raw events \mathcal{E}, we split it to MM time windows and construct MM 2-channel event histograms. They are converted to 3-channel RGB frames with a color map, and fed to CLIP to obtain image features FF. Meanwhile, we construct text prompts with KK class names, and utilize CLIP to extract text features as the classifier weight WW. FF is updated with a feature adapter 𝒯\mathcal{T} to fuse temporal information, and its cosine similarity with WW serves as the classification output. 𝒯\mathcal{T} is an identity function under the zero-shot setting, and a Transformer encoder in few-shot learning. The final result is obtained by averaging predictions from MM frames.

3.3 Few-Shot Feature Adaptation

With the event-to-frame conversion pipeline, we successfully transform a 2D CLIP into a zero-shot event classifier for the “unseen” event camera data. However, zero-shot EventCLIP still underperforms domain-specifically trained classifiers. To close the accuracy gap, we consider the few-shot setting, where a few labeled examples are available per class. With limited data, it is impossible to fine-tune the entire model which will lead to severe overfitting. Instead, we only refine the features extracted by a frozen CLIP model.

Image Feature Adapter. Our goal is to incorporate the event domain knowledge into the extracted image features {𝒇i}i=1M\{\bm{f}_{i}\}_{i=1}^{M} to obtain a refined representation {𝒇i}i=1M\{\bm{f}^{*}_{i}\}_{i=1}^{M}. Prior works simply apply an MLP to update features since there is either only one feature vector per sample (2D image) [14, 69], or the visual features follow a fixed order (3D point cloud) [79, 87]. For example, PointCLIP concatenates the multi-view features into one vector and feeds it to the MLP-based adapter. Their performance will degrade significantly if we shuffle the order of projection views. In contrast, as discussed above, the final prediction of EventCLIP should be order-invariant. Therefore, the output of our image feature adapter {𝒇i}i=1M\{\bm{f}^{*}_{i}\}_{i=1}^{M} should be permutation-equivariant to the input features {𝒇i}i=1M\{\bm{f}_{i}\}_{i=1}^{M}. In addition, we need an architecture that can process an arbitrary number of inputs as the number of time windows MM varies across samples. Inspired by a recent CLIP-based video classifier [30], we apply a lightweight 2-layer Transformer encoder [64], 𝒯\mathcal{T}, to aggregate the temporal information of event streams. To avoid overfitting, we employ residual connections from CLIP features to the Transformer output features {𝒇~i}i=1M\{\tilde{\bm{f}}_{i}\}_{i=1}^{M}:

{𝒇~i}i=1M=𝒯({𝒇i}i=1M),𝒇i=α𝒇i+(1α)𝒇~i,\{\tilde{\bm{f}}_{i}\}_{i=1}^{M}=\mathcal{T}(\{\bm{f}_{i}\}_{i=1}^{M}),\ \ \bm{f}^{*}_{i}=\alpha\bm{f}_{i}+(1-\alpha)\tilde{\bm{f}}_{i}, (2)

where α\alpha is a hyper-parameter controlling the portion of the original CLIP knowledge. After applying the visual adapter, we use the updated image features {𝒇i}i=1M\{\bm{f}^{*}_{i}\}_{i=1}^{M} and the text features WW to perform classification as done in the zero-shot setting. Thanks to the order invariance property, our few-shot EventCLIP is also more robust against different camera motions during the data capture process.

Text Feature Adapter. Recently, several works have studied data-efficient fine-tuning of CLIP’s text branch [83, 82]. As pointed out by He et al. [25], all these methods aim at learning a better classifier weight W={𝒘i}i=1KW=\{\bm{w}_{i}\}_{i=1}^{K}. Therefore, we follow them to adopt the simple Classifier Tuning method [69], by fine-tuning WW with gradient descent. In our initial experiments, Classifier Tuning indeed achieves competitive performance with more complicated tuning methods [83, 73], while requiring much less computation.

3.4 Extensions of EventCLIP

Robust classification. Existing event-based classifiers are trained from scratch on event data. These datasets are often captured under limited environment variations [32]. Thus, the model performance degrades drastically when tested on unseen settings, such as changes in lighting or camera motion. On the contrary, CLIP is trained on Internet-scale data, thus exhibiting high robustness against data corruption. A natural idea is thus to ensemble the two models for joint prediction. Specifically, we simply average the predicted logits from a pre-trained event-based classifier and a zero-shot or few-shot EventCLIP as the final output. As we will show in the experiments, the domain-specific event knowledge and the 2D pre-trained knowledge are able to complement each other, leading to state-of-the-art model robustness.

Label-free learning. In many practical scenarios, we have access to not only a few labeled data but also massive unlabeled events. The extreme setting in this line is label-free learning, where we only have raw events without any labels. In both cases, we can leverage EventCLIP to create pseudo labels, and then fine-tune the model on them. To generate reliable pseudo labels, we run predictions on multiple augmented versions of the event, and only select data with a consistent predicted label. Formally, based on the fact that an event stream should remain the same class after horizontal flip and temporal reverse, given an event, we create four versions of it by applying the augmentations combinatorially. Then, we discard events with an inconsistent predicted class. To further improve the label quality, we employ a threshold τ\tau to select high-confident samples, and only take the top-kk predictions per class for balanced model training. See Appendix C for implementation details in this setting.

4 Experiments

In Sec. 4.2, we study the best design choices to transfer CLIP’s pre-trained knowledge to event camera data. Then, we show the performance gain from limited training data in few-shot learning (Sec. 4.3). When more data are available, EventCLIP can achieve state-of-the-art accuracy by fine-tuning the entire model (Sec. 4.4). In Sec. 4.5, we leverage our method to improve the robustness of existing event classifiers by ensemble. Finally, in Sec. 4.6, we demonstrate unsupervised learning from raw events with our method.

4.1 Experimental Setup

Datasets. We use three public datasets in our experiments: N-Caltech [47], N-Cars [58], and N-ImageNet [32]. N-Caltech contains 8,246 samples from 101 classes, recorded by a moving 180 ×\times 240 resolution ATIS system [49] in front of a monitor displaying still images from the original RGB Caltech101 dataset [11]. In contrast, N-Cars provides event streams recorded by the ATIS system in a real-world urban environment. It contains 12,336 samples of the class car and 11,693 samples of the class background. Similar to N-Caltech, N-ImageNet is the event camera version of ImageNet [9]. As the largest event camera dataset, it contains 1.78 million event streams and 1,000 classes. The data were captured with a moving 480 ×\times 640 resolution Samsung DVS Gen3 event camera [59]. N-ImageNet also provides variants of the test set captured with different camera motions and brightness, serving as a benchmark to evaluate the robustness of event classifiers. See Appendix B for detailed descriptions of each variant. For few-shot training, we randomly sample a subset of data from each category. We always report the results on the entire test set.

Baselines. We compare EventCLIP with current state-of-the-art event-based classifiers, namely, EST [16], Event Histogram [41], Sorted Time Surface [1], and DiST [32]. See Appendix C for their implementation details. Notice that, we use ResNet34 [24] pre-trained on the RGB ImageNet [9] as the backbone for all the baselines following their original paper. For DiST and EST, we also tested larger backbones such as ResNet101 and ViT-L [10], but did not observe clear improvement as will be shown later. We will introduce other baselines in each task below.

Our Implementation Details. To convert event streams into frames, we set the number of events per time window NN as 20,000, 10,000, and 70,000 on N-Caltech, N-Cars, and N-ImageNet, respectively. This accounts for each dataset’s event camera resolution. For colorizing the events, i.e., converting the 2-channel event histograms to 3-channel RGB images, we simply use a gray-scale color map by multiplying both positive and negative event counts with [127, 127, 127]. For the pre-trained CLIP, we adopt the variant with the ViT-L/14 [10] image encoder. We select “a point cloud image of a [CLASS]” as the text template.

Dataset N-Caltech N-Cars N-ImageNet
Acc. 69.67 82.28 20.78
Table 1: Zero-shot classification accuracy (%) of EventCLIP on N-Caltech, N-Cars, and N-ImageNet with our best settings.
Dataset N-Caltech N-ImageNet
NN (×103\times 10^{3}) 15 20 25 50 70 80
0-shot Acc. 69.87 69.67 69.33 20.04 20.78 20.61
10-shot Acc. 84.98 85.62 84.93 27.83 28.63 27.62
Table 2: Ablation on the event time window size NN. We report zero-shot and ten-shot accuracy (%) as measurements.
Dataset N-Caltech N-ImageNet
Method Gray R-B Learn Gray R-B Learn
0-shot Acc. 69.67 65.93 - 20.78 17.49 -
10-shot Acc. 85.62 82.87 85.69 28.63 25.23 28.55
Table 3: Ablation on the event histograms colorization methods. See text for implementation details of the ablated methods.

4.2 Zero-Shot Classification

Results. Tab. 1 presents the zero-shot classification accuracy of EventCLIP. Without any in-domain training, our method achieves an accuracy of 69.67% on N-Caltech which has 101 classes. This proves the effectiveness of our event-to-frame conversion pipeline in bridging the RGB and event camera domains. In addition, our model scores a higher 82.28% accuracy on N-Cars which is captured in the real world, demonstrating its generalizability. On the challenging N-ImageNet dataset, we achieve a lower accuracy of 20.78% due to the lack of event domain knowledge.

Refer to caption
Refer to caption
Refer to caption
Figure 3: Few-shot classification accuracy on N-Caltech, N-Cars, and N-ImageNet. All baselines build upon the ResNet34 backbone pre-trained on RGB ImageNet. For EST and DiST, we try additional variants with larger ResNet101 and ViT-L backbones. We also report EventCLIP with another VLM FLIP [40]. Our method consistently outperforms state-of-the-art event-based classifiers across all settings.

Ablation Study. We first study the event time window size NN in Tab. 2. Since event streams from N-Cars are generally sparse, we convert all events to one frame without ablation. On N-Caltech, a smaller NN achieves the best zero-shot accuracy, but we select 20,000 since it strikes a better balance between zero- and few-shot results. We need a higher N=70,000N=70,000 for optimal performance on N-ImageNet due to its higher camera resolution. See Appendix D.1 for more ablations on NN. Tab. 3 ablates different ways of colorizing the event histogram. We test the red-blue color map (dubbed R-B) commonly used to visualize events, which multiplies positive and negative event counts with [255, 0, 0] and [0, 0, 255], respectively. It leads to much worse results because the color statistics of their converted images are distinct from the natural images CLIP is trained on. We also design a learnable method by initializing the color map with two vectors and optimizing them jointly during training. This leads to similar accuracies, but requires 10×\times the computation since it needs to backpropagate through the heavy ViT image encoder. Overall, our gray-scale color map is efficient and reduces the domain gap of input to CLIP.

Model RN RN×\times4 RN×\times16 RN×\times64 ViT-B ViT-L
Size (MB) 244 403 631 1300 335 890
0-shot Acc. 44.34 51.41 60.83 61.92 61.11 69.67
10-shot Acc. 74.21 77.43 79.95 81.23 80.70 85.62
Table 4: Ablation on the image encoder of CLIP on N-Caltech. We test the ResNet50 (RN) family and Vision Transformers (ViT-B/16, ViT-L/14). Model sizes and accuracies are reported.
Prompt 0-shot Acc. 10-shot Acc.
A photo of a [CLASS] 66.57 82.18
An event camera photo of [CLASS] 64.73 78.34
An edge map of a [CLASS] 68.70 84.07
A sketch image of a [CLASS] 69.64 85.16
A point cloud image of a [CLASS] 69.67 85.62
[Learnable Tokens] + [CLASS] - 85.37
Table 5: Ablation on the text prompt templates on N-Caltech. “Learnable Tokens” trains 16 context vectors as in CoOp [83].

One important property of foundation models is the scalability of their performance with model size. We study this effect using different image encoders of CLIP in Tab. 4. Within the ResNet [24] and the ViT [10] family, EventCLIP achieves higher accuracy as the model size grows. Notably, ViT-L significantly outperforms RN50×\times64, despite having much fewer parameters. The reason might be the converted event frames mostly capture the object boundary, and are thus biased towards the shape information. Studies have shown that CNNs are usually texture-biased [21], while ViTs are better at processing shape information [45, 63]. This observation may serve as future guidelines in designing event-based vision architectures. Tab. 5 compares the template used to construct text prompts. The template “A photo of a [CLASS]” commonly used in 2D vision tasks achieves 66.57% zero-shot accuracy. Simply prefixing “photo” with “event camera” leads to worse results as CLIP is not trained on event data. Instead, we explicitly describe the visual property of event frames. Since events are mostly triggered by object boundaries, “edge map” and “sketch image” both lead to better results. Surprisingly, describing event frames with “point cloud image” achieves the highest accuracy, which aligns with previous works that treat raw events as spatio-temporal points [67]. We also tried prompt tuning with learnable textual tokens [83], which achieves similar performance. However, it trains 5×\times slower as it requires backpropagation through the heavy text encoder.

4.3 Few-Shot Classification

Settings. We experiment with 1, 3, 5, 10, 20 shots on N-Caltech and N-ImageNet. Since N-Cars only has two categories, we multiply the number of shots by 10. We test EventCLIP with four variants of adapters: (i) PointCLIP’s MLP visual adapter [79], (ii) our proposed Transformer visual adapter, (iii) the Classifier Tuning text adapter [69], and (iv) a joint adapter combining (ii) and (iii). We also test our generality by replacing CLIP with another VLM FLIP [40].

Results. We first compare EventCLIP using joint adapter with baselines in Fig. 3. All baselines with ResNet34 backbone achieve similar performance across datasets, which aligns with previous observation [32]. In contrast, EventCLIP achieves significantly higher few-shot accuracy. Our 20-shot accuracy on N-Caltech (85.62%) surpasses EST trained on the entire N-Caltech (81.7%) by around 4%. Notice that, all baselines are initialized with backbones pre-trained on RGB ImageNet, which is the source for creating N-ImageNet. Still, EventCLIP achieves consistently higher accuracy across all numbers of shots. Overall, the results prove that CLIP’s large-scale pre-training learns generalizable representations, enabling quick adaptation to the new event camera domain with limited training data.

For a fair comparison, we also evaluated baselines with larger backbones ResNet101 and ViT-L. Naively fine-tuning ViTs on limited data usually leads to severe overfitting [10], so we adopt the state-of-the-art data-efficient ViT training strategy [62]. As shown in the figures, DiST with ResNet101 leads to similar performance, while ViT-L results in much worse accuracy even with the advanced training strategy. This indicates the superiority of our EventCLIP framework as it scales well with larger models.

Finally, we replace CLIP with another VLM FLIP [40] (dubbed EventCLIP (FLIP)). This variant achieves slightly lower performance compared to using CLIP, as FLIP also achieves lower zero-shot accuracy than CLIP on RGB image datasets. Nonetheless, EventCLIP with FLIP still outperforms baselines, showing its generality with base VLMs.

Refer to caption
Figure 4: Ablation on feature adapters of EventCLIP on N-ImageNet. Trans. stands for the Transformer-based visual adapter.

Ablation Study. Fig. 4 compares the performance of EventCLIP with four feature adapters on N-ImageNet. The Transformer-based visual adapter consistently outperforms the MLP-based counterpart by a sizeable margin, showing the importance of its permutation-equivariant property. Interestingly, adapting the text features along yields clearly better results than adapting the image features along, which is distinct from prior works on images [14, 78]. This may be due to the larger domain gaps event frames pose compared to images. Finally, adapting both branches leads to the best performance without suffering from overfitting. See Appendix D.2 for more ablations on N-Caltech and N-Cars.

4.4 Fine-Tuning EventCLIP

Settings. Few-shot feature adaptation is able to improve the model accuracy with minor computation and time overhead. However, it still underperforms event-based classifiers trained on the entire dataset. To show that EventCLIP can also achieve state-of-the-art performance when more data are available, we propose to fine-tune the CLIP’s image encoder jointly with the text feature adapter.

Baseline. We compare with concurrent work E-CLIP [81], which also uses CLIP but requires paired RGB images with events during training. E-CLIP is the current state-of-the-art method using full datasets. For a fair comparison, we adopt CLIP with the same ViT-B/16 backbone as they do.

Results. Tab. 6 presents the fine-tuning results of EventCLIP. Following the protocol of E-CLIP, we also report the accuracy in the few-shot setting. On N-Caltech, we achieve better results in the low data regime. With more data, EventCLIP is competitive with E-CLIP. We hypothesize that this is because N-Caltech is extremely class-imbalanced where some categories only have 20 samples. As a result, the paired RGB images provide much useful information for E-CLIP. This is verified by the results on N-ImageNet with abundant samples per class. Without using additional data, EventCLIP still outperforms E-CLIP consistently over all settings, achieving new state-of-the-art performance.

Data per Class 1 2 5 10 20 All
E-CLIP 66.72 75.87 82.35 86.92 90.51 93.89
EventCLIP 75.82 78.86 83.57 87.42 90.41 93.57
(a) Fine-tuning results on N-Caltech
Data per Class 1 2 5 10 20 All
E-CLIP 22.22 26.85 28.70 30.56 35.11 51.85
EventCLIP 24.39 27.23 31.12 34.24 38.28 53.20
(b) Fine-tuning results on N-ImageNet
Table 6: Fine-tuning results. We compare EventCLIP with E-CLIP [81] which fine-tunes the CLIP image encoder. We report results under the few-shot setting or trained on all data in the dataset.
Data Variation None Trajectory Brightness Average
Variant ID. Orig. 1 2 3 4 5 6 7 8 9 All
Event Histogram 47.73 43.73 33.72 37.69 24.56 35.24 20.89 29.68 36.33 34.56 32.93
+ DiST 51.67 48.02 38.18 43.16 27.56 40.02 25.19 34.22 40.63 38.83 37.31
+ Ev-TTA (online) - 44.94 44.63 43.31 41.48 43.46 26.89 34.71 43.86 43.42 40.86
+ Ev-TTA (offline) - 48.64 48.01 47.24 44.49 47.06 30.08 38.34 47.37 46.58 44.20
+ EventCLIP (0-shot) 50.03 48.49 43.33 41.57 37.90 40.14 25.72 34.28 44.33 44.65 40.05
+ EventCLIP (20-shot) 51.68 51.06 46.58 43.63 42.59 42.94 27.64 37.18 47.65 46.93 42.91
+ EventCLIP (100-shot) 52.72 51.98 48.82 44.98 44.75 44.79 30.26 38.84 49.68 48.20 44.70
Sorted Time Surface 47.90 44.33 33.50 40.17 23.72 37.19 21.57 30.31 36.63 35.18 33.62
+ DiST 51.56 47.92 37.92 43.84 27.07 40.64 25.38 34.35 40.49 38.87 37.39
+ Ev-TTA (online) - 46.02 45.29 45.91 42.53 43.90 26.70 36.17 45.00 45.22 41.86
+ Ev-TTA (offline) - 49.58 47.67 48.36 45.59 46.72 30.07 39.30 48.24 47.76 44.81
+ EventCLIP (0-shot) 50.25 48.78 43.12 44.03 37.24 41.66 26.23 35.21 44.39 43.82 40.50
+ EventCLIP (20-shot) 52.23 51.27 46.65 46.07 42.38 44.22 28.22 37.95 48.04 47.22 43.56
+ EventCLIP (100-shot) 52.85 52.78 48.92 47.44 44.68 45.55 30.53 39.68 49.63 49.19 45.38
DiST 48.43 45.17 36.58 42.28 26.57 38.70 24.39 32.76 38.99 37.37 35.89
+ Ev-TTA (online) - 46.32 46.05 46.57 43.23 44.58 28.05 36.98 46.03 45.64 42.61
+ Ev-TTA (offline) - 48.53 47.75 48.38 45.35 47.26 31.02 39.07 48.19 47.66 44.80
+ EventCLIP (0-shot) 50.53 49.36 44.47 45.12 37.83 43.15 28.01 36.79 45.72 44.58 41.67
+ EventCLIP (20-shot) 52.28 51.42 47.53 46.77 42.34 45.05 29.62 39.05 49.00 47.52 44.26
+ EventCLIP (100-shot) 53.12 52.45 49.01 47.78 43.85 46.22 31.00 39.92 49.77 48.74 45.42
Table 7: Classification accuracy on N-ImageNet and its robustness variants. The numbers of baseline event-based classifiers and Ev-TTA are copied from Kim et al. [33]. Average is computed over 9 variants. Best results are bold and the second-best results are underlined.

4.5 Robust Event Classification

Settings. Our goal is to evaluate whether the large-scale 2D pre-trained knowledge in CLIP is complementary to existing event-based classifiers. Therefore, we perform model ensemble using 0-, 20-, and 100-shot EventCLIP (few-shot trained with the joint feature adapter) with baselines trained on the entire N-ImageNet by simply averaging the predicted class logits. We test the original and the ensembled model on the normal test set of N-ImageNet, as well as its 9 variants [32] which are captured under out-of-distribution camera motions and lighting conditions.

Baselines. We adopt pre-trained weights of baselines from the official codebase of N-ImageNet. EST is excluded since there is no pre-trained weight available. We report the accuracy of model ensemble with DiST, as DiST is currently the most robust event-based classifier. We also compare with Ev-TTA [33], which is designed specifically for event domain adaptation. Importantly, Ev-TTA requires access to data in the new domain to perform test-time adaptation to update model weights. It has an online version where novel testing data are used only once and an offline version where the out-of-distribution data are used multiple times.

Results. As shown in Tab. 7, model ensemble with 0-shot EventCLIP already increases the accuracy of baselines by more than 5%, which is higher than ensemble with a fully trained DiST. This indicates that pre-trained CLIP contains information that cannot be effectively learned from event camera datasets only. Such information complements pre-trained event-based classifiers, making them more robust against data corruption. Compared to ensemble with DiST trained on the entire N-ImageNet, ensemble with 20-shot EventCLIP improves the average accuracy on robustness variants by more than 5%. It is worth noting that DiST achieves 48.43% accuracy on the original test set, while our 20-shot EventCLIP scores a much lower 30.57%. Besides, ensemble DiST with a worse performing EventCLIP still greatly improves the model performance on both the original and the robustness variants of N-ImageNet.

Next, we compare our ensemble method with the sophisticated test-time adaptation method Ev-TTA, which requires access to additional out-of-distribution data. Surprisingly, EventCLIP trained on 20 samples per category (less than 2% of all training data) outperforms the online version of Ev-TTA on most of the subsets. With 8% of training data, our 100-shot EventCLIP outperforms the offline version of Ev-TTA, achieving new state-of-the-art robustness results. See Appendix D.3 for additional qualitative analysis.

4.6 Label-Free Event Recognition

Settings and Baseline. We follow Ev-LaFOR [5] to perform fully unsupervised learning on a 100-class subset of N-ImageNet. Please refer to their paper for details on the N-ImageNet (Mini) split. Ev-LaFOR reconstructs images from events and uses CLIP for classification. It has two variants, and we compare with the better-performing one that uses unlabeled events and RGB images. For a fair comparison, we adopt CLIP with the same ViT-B/32 backbone.

Results. EventCLIP with ViT-B/32 backbone achieves a zero-shot accuracy of 27.08% on the test set of N-ImageNet (Mini). Naive self-training degrades it to 26.43%. With our consistency-based labeling process, we can improve the performance to 35.26%, surpassing Ev-LaFOR’s accuracy of 31.28% by around 4%. See Appendix D.4 for additional results under the semi-supervised learning setting.

5 Conclusion

In this paper, we propose EventCLIP which adapts CLIP for open-set event recognition. With event-to-frame conversion, we successfully transfer CLIP’s 2D pre-trained knowledge to the event camera domain. To further enhance the performance, we develop lightweight adapters to refine the pre-trained CLIP embeddings. Moreover, EventCLIP can be employed to improve the robustness of existing classifiers via model ensemble, or learn from unlabeled data with self-training. Our work opens up new possibilities to apply recent advances in foundation models to event-based vision. We discuss the limitations of this work in Appendix E.

Acknowledgments

We acknowledge the support of the Natural Sciences and Engineering Research Council of Canada (NSERC). This research was enabled in part by support provided by Compute Ontario (www.computeontario.ca), the Digital Research Alliance of Canada (https://alliancecan.ca/), the Province of Ontario, and the Government of Canada through CIFAR, and companies sponsoring the Vector Institute (www.vectorinstitute.ai/partnerships/current-partners/). We would also like to acknowledge Vector Institute for computation support.

We would like to thank Yang Zheng for the help with data preparation, Xiaoshi Wu for discussions about CLIP models, and Mathias Gehrig, Yash Kant, Xuanchi Ren, Liquan Wang for valuable discussions and support.

References

  • Alzugaray and Chli [2018] Ignacio Alzugaray and Margarita Chli. Ace: An efficient asynchronous corner tracker for event cameras. In 3DV, 2018.
  • Amir et al. [2017] Arnon Amir, Brian Taba, David Berg, Timothy Melano, Jeffrey McKinstry, Carmelo Di Nolfo, Tapan Nayak, Alexander Andreopoulos, Guillaume Garreau, Marcela Mendoza, et al. A low power, fully event-based gesture recognition system. In CVPR, 2017.
  • Benosman et al. [2013] Ryad Benosman, Charles Clercq, Xavier Lagorce, Sio-Hoi Ieng, and Chiara Bartolozzi. Event-based visual flow. TNNLS, 2013.
  • Cannici et al. [2020] Marco Cannici, Marco Ciccone, Andrea Romanoni, and Matteo Matteucci. A differentiable recurrent surface for asynchronous event-based data. In ECCV, 2020.
  • Cho et al. [2023] Hoonhee Cho, Hyeonseong Kim, Yujeong Chae, and Kuk-Jin Yoon. Label-free event-based object recognition via joint learning with image reconstruction from events. In ICCV, 2023.
  • Cohen et al. [2018] Gregory Cohen, Saeed Afshar, Garrick Orchard, Jonathan Tapson, Ryad Benosman, and Andre van Schaik. Spatial and temporal downsampling in event-based visual classification. IEEE Transactions on Neural Networks and Learning Systems, 2018.
  • Cubuk et al. [2020] Ekin D Cubuk, Barret Zoph, Jonathon Shlens, and Quoc V Le. Randaugment: Practical automated data augmentation with a reduced search space. In CVPR Workshop, 2020.
  • De Tournemire et al. [2020] Pierre De Tournemire, Davide Nitti, Etienne Perot, Davide Migliore, and Amos Sironi. A large scale event-based detection dataset for automotive. arXiv preprint arXiv:2001.08499, 2020.
  • Deng et al. [2009] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. ImageNet: A large-scale hierarchical image database. In CVPR, 2009.
  • Dosovitskiy et al. [2021] Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, et al. An image is worth 16x16 words: Transformers for image recognition at scale. In ICLR, 2021.
  • Fei-Fei et al. [2004] Li Fei-Fei, Rob Fergus, and Pietro Perona. Learning generative visual models from few training examples: An incremental bayesian approach tested on 101 object categories. In CVPR Workshop, 2004.
  • Gallego et al. [2017] Guillermo Gallego, Jon EA Lund, Elias Mueggler, Henri Rebecq, Tobi Delbruck, and Davide Scaramuzza. Event-based, 6-dof camera tracking from photometric depth maps. TPAMI, 2017.
  • Gallego et al. [2020] Guillermo Gallego, Tobi Delbrück, Garrick Orchard, Chiara Bartolozzi, Brian Taba, Andrea Censi, Stefan Leutenegger, Andrew J Davison, Jörg Conradt, Kostas Daniilidis, et al. Event-based vision: A survey. TPAMI, 2020.
  • Gao et al. [2021] Peng Gao, Shijie Geng, Renrui Zhang, Teli Ma, Rongyao Fang, Yongfeng Zhang, Hongsheng Li, and Yu Qiao. Clip-adapter: Better vision-language models with feature adapters. arXiv preprint arXiv:2110.04544, 2021.
  • Gehrig et al. [2018] Daniel Gehrig, Henri Rebecq, Guillermo Gallego, and Davide Scaramuzza. Asynchronous, photometric feature tracking using events and frames. In ECCV, 2018.
  • Gehrig et al. [2019] Daniel Gehrig, Antonio Loquercio, Konstantinos G Derpanis, and Davide Scaramuzza. End-to-end learning of representations for asynchronous event-based data. In ICCV, 2019.
  • Gehrig et al. [2020a] Daniel Gehrig, Mathias Gehrig, Javier Hidalgo-Carrió, and Davide Scaramuzza. Video to events: Recycling video datasets for event cameras. In CVPR, 2020a.
  • Gehrig et al. [2020b] Daniel Gehrig, Henri Rebecq, Guillermo Gallego, and Davide Scaramuzza. Eklt: Asynchronous photometric feature tracking using events and frames. IJCV, 2020b.
  • Gehrig and Scaramuzza [2023] Mathias Gehrig and Davide Scaramuzza. Recurrent vision transformers for object detection with event cameras. In CVPR, 2023.
  • Gehrig et al. [2021] Mathias Gehrig, Mario Millhäusler, Daniel Gehrig, and Davide Scaramuzza. E-raft: Dense optical flow from event cameras. In 3DV, 2021.
  • Geirhos et al. [2019] Robert Geirhos, Patricia Rubisch, Claudio Michaelis, Matthias Bethge, Felix A Wichmann, and Wieland Brendel. Imagenet-trained cnns are biased towards texture; increasing shape bias improves accuracy and robustness. In ICLR, 2019.
  • Girdhar et al. [2023] Rohit Girdhar, Alaaeldin El-Nouby, Zhuang Liu, Mannat Singh, Kalyan Vasudev Alwala, Armand Joulin, and Ishan Misra. Imagebind: One embedding space to bind them all. In CVPR, 2023.
  • Gu et al. [2022] Xiuye Gu, Tsung-Yi Lin, Weicheng Kuo, and Yin Cui. Open-vocabulary object detection via vision and language knowledge distillation. In ICLR, 2022.
  • He et al. [2016] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In CVPR, 2016.
  • He et al. [2023] Ruifei He, Shuyang Sun, Xin Yu, Chuhui Xue, Wenqing Zhang, Philip Torr, Song Bai, and Xiaojuan Qi. Is synthetic data from generative models ready for image recognition? In ICLR, 2023.
  • Hu et al. [2021a] Edward J Hu, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, Weizhu Chen, et al. Lora: Low-rank adaptation of large language models. In ICLR, 2021a.
  • Hu et al. [2020] Yuhuang Hu, Tobi Delbruck, and Shih-Chii Liu. Learning to exploit multiple vision modalities by using grafted networks. In ECCV, 2020.
  • Hu et al. [2021b] Yuhuang Hu, Shih-Chii Liu, and Tobi Delbruck. v2e: From video frames to realistic dvs events. In CVPR, 2021b.
  • Jia et al. [2021] Chao Jia, Yinfei Yang, Ye Xia, Yi-Ting Chen, Zarana Parekh, Hieu Pham, Quoc Le, Yun-Hsuan Sung, Zhen Li, and Tom Duerig. Scaling up visual and vision-language representation learning with noisy text supervision. In ICML, 2021.
  • Ju et al. [2022] Chen Ju, Tengda Han, Kunhao Zheng, Ya Zhang, and Weidi Xie. Prompting visual-language models for efficient video understanding. In ECCV, 2022.
  • Kamath et al. [2021] Aishwarya Kamath, Mannat Singh, Yann LeCun, Gabriel Synnaeve, Ishan Misra, and Nicolas Carion. Mdetr-modulated detection for end-to-end multi-modal understanding. In ICCV, 2021.
  • Kim et al. [2021] Junho Kim, Jaehyeok Bae, Gangin Park, Dongsu Zhang, and Young Min Kim. N-imagenet: Towards robust, fine-grained object recognition with event cameras. In ICCV, 2021.
  • Kim et al. [2022] Junho Kim, Inwoo Hwang, and Young Min Kim. Ev-tta: Test-time adaptation for event-based object recognition. In CVPR, 2022.
  • Kingma and Ba [2014] Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
  • Klenk et al. [2022] Simon Klenk, David Bonello, Lukas Koestler, and Daniel Cremers. Masked event modeling: Self-supervised pretraining for event cameras. arXiv preprint arXiv:2212.10368, 2022.
  • Lagorce et al. [2016] Xavier Lagorce, Garrick Orchard, Francesco Galluppi, Bertram E Shi, and Ryad B Benosman. Hots: a hierarchy of event-based time-surfaces for pattern recognition. TPAMI, 2016.
  • Lee et al. [2016] Jun Haeng Lee, Tobi Delbruck, and Michael Pfeiffer. Training deep spiking neural networks using backpropagation. Frontiers in neuroscience, 2016.
  • Li et al. [2022] Jianing Li, Jia Li, Lin Zhu, Xijie Xiang, Tiejun Huang, and Yonghong Tian. Asynchronous spatio-temporal memory network for continuous event-based object detection. TIP, 2022.
  • Li et al. [2021] Yijin Li, Han Zhou, Bangbang Yang, Ye Zhang, Zhaopeng Cui, Hujun Bao, and Guofeng Zhang. Graph-based asynchronous event processing for rapid object recognition. In ICCV, 2021.
  • Li et al. [2023] Yanghao Li, Haoqi Fan, Ronghang Hu, Christoph Feichtenhofer, and Kaiming He. Scaling language-image pre-training via masking. In CVPR, 2023.
  • Maqueda et al. [2018] Ana I Maqueda, Antonio Loquercio, Guillermo Gallego, Narciso García, and Davide Scaramuzza. Event-based vision meets deep learning on steering prediction for self-driving cars. In CVPR, 2018.
  • Messikommer et al. [2020] Nico Messikommer, Daniel Gehrig, Antonio Loquercio, and Davide Scaramuzza. Event-based asynchronous sparse convolutional networks. In ECCV, 2020.
  • Messikommer et al. [2022] Nico Messikommer, Daniel Gehrig, Mathias Gehrig, and Davide Scaramuzza. Bridging the gap between events and frames through unsupervised domain adaptation. RA-L, 2022.
  • Mueggler et al. [2017] Elias Mueggler, Henri Rebecq, Guillermo Gallego, Tobi Delbruck, and Davide Scaramuzza. The event-camera dataset and simulator: Event-based data for pose estimation, visual odometry, and slam. IJRR, 2017.
  • Naseer et al. [2021] Muhammad Muzammal Naseer, Kanchana Ranasinghe, Salman H Khan, Munawar Hayat, Fahad Shahbaz Khan, and Ming-Hsuan Yang. Intriguing properties of vision transformers. NeurIPS, 2021.
  • Neil et al. [2016] Daniel Neil, Michael Pfeiffer, and Shih-Chii Liu. Phased lstm: Accelerating recurrent network training for long or event-based sequences. NeurIPS, 2016.
  • Orchard et al. [2015] Garrick Orchard, Ajinkya Jayawant, Gregory K Cohen, and Nitish Thakor. Converting static image datasets to spiking neuromorphic datasets using saccades. Frontiers in neuroscience, 2015.
  • Perot et al. [2020] Etienne Perot, Pierre De Tournemire, Davide Nitti, Jonathan Masci, and Amos Sironi. Learning to detect objects with a 1 megapixel event camera. NeurIPS, 2020.
  • Posch et al. [2010] Christoph Posch, Daniel Matolin, and Rainer Wohlgenannt. A qvga 143 db dynamic range frame-free pwm image sensor with lossless pixel-level video compression and time-domain cds. IEEE Journal of Solid-State Circuits, 2010.
  • Radford et al. [2021] Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. Learning transferable visual models from natural language supervision. In ICML, 2021.
  • Rao et al. [2022] Yongming Rao, Wenliang Zhao, Guangyi Chen, Yansong Tang, Zheng Zhu, Guan Huang, Jie Zhou, and Jiwen Lu. Denseclip: Language-guided dense prediction with context-aware prompting. In CVPR, 2022.
  • Rebecq et al. [2018] Henri Rebecq, Daniel Gehrig, and Davide Scaramuzza. Esim: an open event camera simulator. In CoRL, 2018.
  • Rebecq et al. [2019a] Henri Rebecq, René Ranftl, Vladlen Koltun, and Davide Scaramuzza. Events-to-video: Bringing modern computer vision to event cameras. In CVPR, 2019a.
  • Rebecq et al. [2019b] Henri Rebecq, René Ranftl, Vladlen Koltun, and Davide Scaramuzza. High speed and high dynamic range video with an event camera. TPAMI, 2019b.
  • Reizenstein et al. [2021] Jeremy Reizenstein, Roman Shapovalov, Philipp Henzler, Luca Sbordone, Patrick Labatut, and David Novotny. Common objects in 3d: Large-scale learning and evaluation of real-life 3d category reconstruction. In ICCV, 2021.
  • Schaefer et al. [2022] Simon Schaefer, Daniel Gehrig, and Davide Scaramuzza. Aegnn: Asynchronous event-based graph neural networks. In CVPR, 2022.
  • Scheerlinck et al. [2020] Cedric Scheerlinck, Henri Rebecq, Daniel Gehrig, Nick Barnes, Robert Mahony, and Davide Scaramuzza. Fast image reconstruction with an event camera. In WACV, 2020.
  • Sironi et al. [2018] Amos Sironi, Manuele Brambilla, Nicolas Bourdis, Xavier Lagorce, and Ryad Benosman. HATS: Histograms of averaged time surfaces for robust event-based object classification. In CVPR, 2018.
  • Son et al. [2017] Bongki Son, Yunjae Suh, Sungho Kim, Heejae Jung, Jun-Seok Kim, Changwoo Shin, Keunju Park, Kyoobin Lee, Jinman Park, Jooyeon Woo, et al. 4.1 a 640×\times 480 dynamic vision sensor with a 9μ\mum pixel and 300meps address-event representation. In 2017 IEEE International Solid-State Circuits Conference (ISSCC), 2017.
  • Stoffregen et al. [2020] Timo Stoffregen, Cedric Scheerlinck, Davide Scaramuzza, Tom Drummond, Nick Barnes, Lindsay Kleeman, and Robert Mahony. Reducing the sim-to-real gap for event cameras. In ECCV, 2020.
  • Sun et al. [2022] Zhaoning Sun, Nico Messikommer, Daniel Gehrig, and Davide Scaramuzza. Ess: Learning event-based semantic segmentation from still images. In ECCV, 2022.
  • Touvron et al. [2021] Hugo Touvron, Matthieu Cord, Matthijs Douze, Francisco Massa, Alexandre Sablayrolles, and Hervé Jégou. Training data-efficient image transformers & distillation through attention. In ICML, 2021.
  • Tuli et al. [2021] Shikhar Tuli, Ishita Dasgupta, Erin Grant, and Thomas L Griffiths. Are convolutional neural networks or transformers more like human vision? arXiv preprint arXiv:2105.07197, 2021.
  • Vaswani et al. [2017] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. NeurIPS, 2017.
  • Wang et al. [2021a] Lin Wang, Yujeong Chae, Sung-Hoon Yoon, Tae-Kyun Kim, and Kuk-Jin Yoon. Evdistill: Asynchronous events to end-task learning via bidirectional reconstruction-guided cross-modal knowledge distillation. In CVPR, 2021a.
  • Wang et al. [2021b] Mengmeng Wang, Jiazheng Xing, and Yong Liu. Actionclip: A new paradigm for video action recognition. arXiv preprint arXiv:2109.08472, 2021b.
  • Wang et al. [2019] Qinyi Wang, Yexin Zhang, Junsong Yuan, and Yilong Lu. Space-time event clouds for gesture recognition: From rgb cameras to event cameras. In WACV, 2019.
  • Weng et al. [2021] Wenming Weng, Yueyi Zhang, and Zhiwei Xiong. Event-based video reconstruction using transformer. In ICCV, 2021.
  • Wortsman et al. [2022] Mitchell Wortsman, Gabriel Ilharco, Jong Wook Kim, Mike Li, Simon Kornblith, Rebecca Roelofs, Raphael Gontijo Lopes, Hannaneh Hajishirzi, Ali Farhadi, Hongseok Namkoong, et al. Robust fine-tuning of zero-shot models. In CVPR, 2022.
  • Wu et al. [2023] Ziyi Wu, Nikita Dvornik, Klaus Greff, Thomas Kipf, and Animesh Garg. Slotformer: Unsupervised visual dynamics simulation with object-centric models. In ICLR, 2023.
  • Xiong et al. [2020] Ruibin Xiong, Yunchang Yang, Di He, Kai Zheng, Shuxin Zheng, Chen Xing, Huishuai Zhang, Yanyan Lan, Liwei Wang, and Tieyan Liu. On layer normalization in the transformer architecture. In ICML, 2020.
  • Xue et al. [2023] Le Xue, Mingfei Gao, Chen Xing, Roberto Martín-Martín, Jiajun Wu, Caiming Xiong, Ran Xu, Juan Carlos Niebles, and Silvio Savarese. Ulip: Learning a unified representation of language, images, and point clouds for 3d understanding. In CVPR, 2023.
  • Yao et al. [2023] Hantao Yao, Rui Zhang, and Changsheng Xu. Visual-language prompt tuning with knowledge-guided context optimization. In CVPR, 2023.
  • Yin et al. [2022] Dacheng Yin, Xuanchi Ren, Chong Luo, Yuwang Wang, Zhiwei Xiong, and Wenjun Zeng. Retriever: Learning content-style representation as a token-level bipartite graph. In ICLR, 2022.
  • Zaheer et al. [2017] Manzil Zaheer, Satwik Kottur, Siamak Ravanbakhsh, Barnabas Poczos, Russ R Salakhutdinov, and Alexander J Smola. Deep sets. NeurIPS, 2017.
  • Zanardi et al. [2019a] Alessandro Zanardi, Andreas Aumiller, Julian Zilly, Andrea Censi, and Emilio Frazzoli. Cross-modal learning filters for rgb-neuromorphic wormhole learning. RSS, 2019a.
  • Zanardi et al. [2019b] Alessandro Zanardi, Julian Zilly, Andreas Aumiller, Andrea Censi, and Emilio Frazzoli. Wormhole learning. In ICRA, 2019b.
  • Zhang et al. [2022a] Renrui Zhang, Rongyao Fang, Wei Zhang, Peng Gao, Kunchang Li, Jifeng Dai, Yu Qiao, and Hongsheng Li. Tip-adapter: Training-free clip-adapter for better vision-language modeling. In ECCV, 2022a.
  • Zhang et al. [2022b] Renrui Zhang, Ziyu Guo, Wei Zhang, Kunchang Li, Xupeng Miao, Bin Cui, Yu Qiao, Peng Gao, and Hongsheng Li. Pointclip: Point cloud understanding by clip. In CVPR, 2022b.
  • Zhou et al. [2022a] Chong Zhou, Chen Change Loy, and Bo Dai. Extract free dense labels from clip. In ECCV, 2022a.
  • Zhou et al. [2023] Jiazhou Zhou, Xu Zheng, Yuanhuiyi Lyu, and Lin Wang. E-clip: Towards label-efficient event-based open-world understanding by clip. arXiv preprint arXiv:2308.03135, 2023.
  • Zhou et al. [2022b] Kaiyang Zhou, Jingkang Yang, Chen Change Loy, and Ziwei Liu. Conditional prompt learning for vision-language models. In CVPR, 2022b.
  • Zhou et al. [2022c] Kaiyang Zhou, Jingkang Yang, Chen Change Loy, and Ziwei Liu. Learning to prompt for vision-language models. IJCV, 2022c.
  • Zhou et al. [2022d] Xingyi Zhou, Rohit Girdhar, Armand Joulin, Philipp Krähenbühl, and Ishan Misra. Detecting twenty-thousand classes using image-level supervision. In ECCV, 2022d.
  • Zhu and Yuan [2018] Alex Zihao Zhu and Liangzhe Yuan. Ev-flownet: Self-supervised optical flow estimation for event-based cameras. In RSS, 2018.
  • Zhu et al. [2021] Alex Zihao Zhu, Ziyun Wang, Kaung Khant, and Kostas Daniilidis. Eventgan: Leveraging large scale image datasets for event cameras. In ICCP, 2021.
  • Zhu et al. [2023] Xiangyang Zhu, Renrui Zhang, Bowei He, Ziyao Zeng, Shanghang Zhang, and Peng Gao. Pointclip v2: Adapting clip for powerful 3d open-world learning. In ICCV, 2023.

Appendix A Additional Related Work

Deep Learning for Event-based Classification. Depending on the utilization of the sparsity and asynchronous nature of event data, existing event-based classifiers can be mainly categorized into two classes [13], namely, synchronous and asynchronous methods. Synchronous models aggregate events to a grid-based representation, and then use standard modules such as Convolutional Neural Networks (CNNs) to process it [1, 41, 58, 16, 17, 27, 32]. Significant efforts have been made to achieve efficient and expressive event-to-frame conversion [6, 41, 1]. Recently, EST [16] has achieved state-of-the-art results in classification with an end-to-end learnable event-to-frame conversion pipeline. As a remedy for robustness in the presence of noise, DiST [32] proposes to suppress noisy events leveraging their spatio-temporal relationships, which is proved effective under camera motion and lighting variations.

On the other hand, asynchronous networks [37, 46, 2, 36, 42, 4, 39, 56] have been developed to address the computational latency inherent in grid-based methods, which directly apply Spiking Neural Networks or Graph Neural Networks to raw event inputs. However, these methods still consistently underperform synchronous methods across datasets [47, 58, 32]. As our primary goal is to achieve high accuracy instead of efficiency, we adopt the former category of methods as our baselines in the experiments.

Bridging Frame-based and Event-based Vision. Inspired by the success of classical computer vision, several works have introduced techniques from frame-based vision to process event data. Some papers focus on reconstructing natural images from events, and then apply conventional deep models on the converted frames [53, 54, 57, 60, 68, 5]. However, they introduce computational overhead which is at odds with event cameras’ low-latency nature. Another line of work tries to simulate event data from existing RGB image datasets, where ground-truth annotations can be automatically obtained from labeled frames [44, 52, 28, 17, 86]. The drawback is the large Sim2Real gap of the synthesized events such as unrealistic camera motions and the absence of sensor noises. The most relevant works to ours are methods that transfer knowledge learned from RGB images to event-based models [27, 65, 61, 76, 77, 43]. However, they either require paired recordings of image and event data, or massive labels in the image domain. In this work, we utilize CLIP pre-trained on RGB image-text pairs for data-efficient event-based classification. Our method converts events into frames via simple counting, and directly applies CLIP for zero-shot classification. We can further boost its performance via few-shot fine-tuning, without the need for paired RGB images or large amounts of labels.

CLIP-based Few-Shot Transfer Learning. Transfer learning aims to leverage models trained on large-scale datasets to help learning on data-scarce tasks. In the field of event-based object recognition, existing grid-based methods have also utilized models pre-trained on RGB images from ImageNet as their backbones to improve performance via fine-tuning [16, 32, 41, 1]. Trained on large-scale image-text pairs, CLIP [50] has shown great potential in learning transferable representations for various downstream tasks. To further enhance the few-shot accuracy of CLIP, one line of work [83, 82, 73] proposes to insert learnable text tokens to perform task-specific prompt tuning. CLIP-Adapter [14], Tip-Adapter [78], and WiSE-FT [69] instead learn lightweight adapters over CLIP features.

In addition to 2D image classification, CLIP has also been extended to 2D detection [23, 84], segmentation [51, 80], and video analysis [66, 30]. Our work is inspired by PointCLIP [79, 87], which projects point clouds to multi-view images and performs zero-shot and few-shot shape recognition with CLIP. Different from PointCLIP, events only capture the boundary information of objects compared to the complete surfaces presented in point clouds. Also, we design a Transformer-based adapter for event temporal information fusion, while PointCLIP simply uses an MLP since their multi-view projections follow a fixed order.

Appendix B Details on N-ImageNet Robustness Variants

Here, we provide more information about the robustness variants of N-ImageNet [32] test sets. The original training and testing set (it is actually the validation set, but we call it test set for simplicity) of N-ImageNet are both captured with a Samsung DVS Gen3 [59] event camera moving around the screen under the same environmental conditions. To test event-based classifiers’ robustness against variations in the data capture process, the authors create 9 variants of the same test set. Variants 1-5 change the camera motions used to trigger events, where different moving directions, frequencies, and amplitudes of the camera trajectory are employed. Variants 6-9 alternate the lighting conditions of the environment, such as extremely low or high illuminations. Overall, these variations cause a large degradation in the performance of existing event-based classifiers. See Appendix D.3 for visualizations of some data variants.

Appendix C Implementation Details

In our few-shot experiments, all models including baselines and EventCLIP are trained on the same subset of data. All reported results are averaged over three runs, and we empirically find that the performance variation is small.

Baselines. EST [16] is the state-of-the-art method on N-ImageNet which utilizes learnable kernels to convert raw events into grid-based representations. Event Histogram [41] converts the event counts into a two-channel image grouped by their polarity. Sorted Time Surface [1] adopts the sorted indices of event timestamps to ensure durability against camera speed changes. DiST [32] is specifically designed to improve the robustness against event camera noise and corruptions, which filters noise with local statistics. It achieves state-of-the-art results on the robustness benchmark of N-ImageNet variants.

We adopt the online official implementation of EST111https://github.com/uzh-rpg/rpg_event_representation_learning [16] and DiST222https://github.com/82magnolia/n_imagenet [32]. The implementation of Event Histogram and Sorted Time Surface are also adopted from the DiST codebase. We re-train all the models with their default settings on each dataset, but decrease the learning rate and number of training epochs when observing severe overfitting. ResNet34 [24] pre-trained on the RGB version of ImageNet [9] is adopted as their backbones, and fine-tuned jointly under the few-shot setting. As shown in the experiments, we tried DiST with pre-trained ResNet50 and ResNet101 backbones, but did not observe clear improvement in accuracy. We also tried pre-trained ViT [10] models. Even with state-of-the-art data-efficient training strategy [62], the joint fine-tuning suffers from severe overfitting, leading to results even worse than ResNet backbones. We hypothesize that this is because vision transformers are data-hungry, and thus are not suitable for the few-shot learning setting. To evaluate the model ensemble performance on the robustness variants of N-ImageNet, we directly use the pre-trained weights from the official release.

Data Augmentation. Following DiST [32], we use random jittering, random horizontal flip along the spatial dimension, and random reverse along the temporal dimension as data augmentations. We tried other event augmentations such as random event dropping, and random cropping over both spatial and temporal dimensions, but did not observe clear improvement. Since we convert events to 3-channel RGB frames, EventCLIP additionally benefits from the well-studied RGB image augmentation literature. We apply RandAugment [7] 333We use the implementation from torchvision to the converted event frames during training. For each loaded event stream, we apply the same set of operations to all frames converted from this data. RandAugment consistently improves the performance of EventCLIP on N-Caltech and N-ImageNet, while bringing little gain on N-Cars. Finally, the resulting frames are resized and center cropped to 224 ×\times 224 following CLIP.

Few-shot EventCLIP. We adopt the pre-trained weights of CLIP from their official online release444https://github.com/openai/CLIP. For EventCLIP with the Transformer-based visual adapter, we stack 2 standard Transformer encoder modules [64], with a token size equal to 256256 and 4 heads. We choose the pre-LN Transformer variant [71] as it leads to more stable training and less overfitting. To ensure the permutation-equivariant property, we do not apply positional encoding to the Transformer input following Yin et al. [74] and Wu et al. [70]. For the MLP-based visual adapter baseline, we adopt the best-performing setting from PointCLIP [79], which concatenates image features F={𝒇i}i=1MF=\{\bm{f}_{i}\}_{i=1}^{M} to extract a global feature, and fuses with per-frame features via residual connections. For the text adapter, we treat the text features W={𝒘i}i=1KW=\{\bm{w}_{i}\}_{i=1}^{K} as the weight of the fully-connected layer in a classifier, and update it via gradient descent. When applying the visual adapter only, we set the residual ratio α\alpha to 0.50.5, while we use α=0.8\alpha=0.8 to further alleviate overfitting when training two adapters jointly. On N-ImageNet, we always use α=0.95\alpha=0.95 as we observe severe overfitting.

We train all models with the Adam [34] optimizer for 100 epochs on N-Caltech and N-ImageNet, and 50 epochs on N-Cars. We use a batch size of 32 on N-Caltech and N-Cars, and 128 on N-ImageNet. When the number of training data is smaller than 32, e.g., N-Cars under the 10-shot learning setting only has 20 samples, we set the batch size as the number of data available. On N-Caltech and N-Cars, when applying the visual and text adapter separately, we set the peak learning rate as 2×1042\times 10^{-4} and 1×1031\times 10^{-3} for them, respectively. When training them jointly, we set the peak learning rate as 2×1042\times 10^{-4}. Besides, we divide the learning rate by a factor of 10 when training on N-ImageNet to present overfitting. We also adopt a linear learning rate warmup schedule during the first 5%\% of training steps, and decay the learning rate to 0 in a cosine schedule. We do not use any weight decay or gradient clipping as we did not find them useful in preliminary experiments.

Fine-tuning EventCLIP. We fine-tune the image encoder of CLIP, while keeping the text encoder frozen. We use the same hyper-parameters and training settings as few-shot EventCLIP, except that we use a 10×\times smaller learning rate on CLIP’s image encoder. When fine-tuned on the entire dataset, we train shorter for 50 epochs. Besides tuning all the model parameters, we also tried other parameter-efficient fine-tuning methods such as only tuning the bias terms, only tuning LayerNorm, and LoRA [26], but observed worse performance compared to naive fine-tuning.

Robust event classification with EventCLIP. We average the class logits of EventCLIP with class logits predicted by baseline event-based classifiers. We search for the weight factors to balance the two terms on the normal validation set, and fix it for all the remaining robustness variants.

Learning from unlabeled data with EventCLIP. There are two settings in this task: fully unsupervised learning where no labels are available, and semi-supervised learning where we have a few labeled samples per class. In the unsupervised setting, we use the zero-shot EventCLIP to generate pseudo labels. In the semi-supervised setting, we first train EventCLIP on the labeled data similar to few-shot learning, and then use it to generate pseudo labels. We choose horizontal flip and temporal flip as TTA methods. Given an unlabeled event \mathcal{E}, we first generate four versions of it {i}i=14\{\mathcal{E}_{i}\}_{i=1}^{4} by applying the augmentations combinatorially. Then, we run EventCLIP to predict class probabilities {𝒑iK}i=14\{\bm{p}_{i}\in\mathbb{R}^{K}\}_{i=1}^{4} (KK is the number of classes), and get the class labels {ci=argmaxc𝒑i}i=14\{c_{i}=\operatorname*{\arg\!\max}_{c}\bm{p}_{i}\}_{i=1}^{4}. We discard examples with an inconsistent {ci}i=14\{c_{i}\}_{i=1}^{4}. To further enhance the label quality, we select examples with confidence scores higher than τ\tau, i.e., max(𝒑i)>τ,i{1,2,3,4}\max(\bm{p}_{i})>\tau,\forall\ i\in\{1,2,3,4\}. Finally, we take the top-kk most confident examples from each category to form the pseudo-labeled training set. After obtaining the training data, we follow few-shot EventCLIP to train our model on it with the joint feature adapter.

For hyper-parameters, we choose k=30k=30 in both settings. For τ\tau, we need to use a very high value of 0.9990.999 in unsupervised learning, as the zero-shot EventCLIP is over-confident in its predictions. After few-shot adaptation, EventCLIP is better calibrated, and we can use a lower value of 0.50.5. The model performance is insensitive to it here.

Appendix D Additional Experimental Results

Refer to caption
Refer to caption
Figure 5: Ablation on event time window size NN (×103\times 10^{3}) on N-Caltech and N-ImageNet. Entry with is the optimal setting.

D.1 Ablation Study on Event Time Window Size NN

Robust event-to-frame conversion is an active research field in event-based vision. In this paper, we adopt the simple event histogram representation as it already gives good performance. To resist camera and object motion changes, we convert every NN event into one frame. The optimal NN varies across datasets, as they are often captured by event cameras with different resolutions. We study the effect of NN on few-shot classification accuracy in Fig. 5. Overall, EventCLIP is not sensitive to NN within a reasonable range, as the accuracy difference is smaller than 2%\% in most cases. As a future direction, one can explore better event representations to further improve EventCLIP’s performance.

Refer to caption
Refer to caption
Figure 6: Ablation on feature adapters on N-Cars and N-Caltech. Trans. stands for the Transformer-based visual adapter. CoOp means the text adapter based on prompt tuning.

D.2 Ablation Study on Feature Adapters

We perform ablation studies on the feature adapters on N-Cars and N-Caltech. On N-Cars, we test an additional text adapter with prompt tuning (dubbed as CoOp [83]), which learns context vectors in text prompts instead of using pre-defined templates. Fig. 6 presents the few-shot accuracy of different adapters. CoOp adapter is consistently worse than other adapters, while consuming much more resources as it requires backpropagation through CLIP’s text encoder. Similar to N-Caltech, the joint feature adapter performs better in the low-shot scenarios, and the gap becomes negligible when more data are provided. Overall, we recommend users choose the text feature adapter as a starting point, since it achieves competitive performance, while requiring less computation and hyper-parameter tuning. Instead, the joint adapter can achieve the best performance after tuning.

D.3 Qualitative Results on Robust Classification

In Fig. 8, we visualize data from the N-ImageNet test set and its robustness variants. The original event image presents the best visual quality with both sharp object boundaries and small background noises. This is because the original data are captured under a regular camera trajectory (square) and a small moving displacement. In contrast, with an irregular moving trajectory (Variant 2, only horizontal movement), some boundaries are missing due to small image gradients along the moving direction. With a larger moving amplitude (Variant 3, 5), the event images show severe motion blur, and there are lots of background noisy events. For lighting changes, both too-low (Variant 6, 7) and too-high (Variant 9) illuminations result in distorted object boundaries. Overall, these data variations cause significant train-test discrepancy, leading to a large performance drop in event-based classifiers trained solely on event camera datasets. On the contrary, CLIP is trained on Internet-scale data covering diverse environmental conditions, which greatly improves the robustness of EventCLIP. Indeed, we successfully classify all variants of this data, while DiST fails on Variant 5, 6, and 7.

We show another example in Fig. 9. The event images under different camera motions follow similar distortions. However, under the low-light condition (Variant 6, 7), the chairs almost disappear, making the recognition task problematic. As a result, neither EventCLIP nor baselines is able to predict the correct category.

Labeled Data per Class 1 3 5 10 20
Few-shot only 30.54 32.64 32.92 37.04 39.58
Few-shot + Pseudo-labels 35.98 37.04 37.53 40.30 42.02
Table 8: Semi-supervised learning results. We report the accuracy of EventCLIP trained only on few-shot labeled data vs. few-shot labeled data plus generated pseudo-labels.

D.4 Semi-Supervised Learning with EventCLIP

We conduct semi-supervised learning on N-ImageNet (Mini), where a few labeled data are available per class, and we pseudo-label the remaining unlabeled data. As shown in Tab. 8, training on generated pseudo-labels consistently improves the accuracy compared to using only labeled data, showing a promising direction of leveraging raw data.

Event-to-Frame (CPU) Model Forward (GPU)
Time (ms) 6.76 8.73
Table 9: Speed analysis of EventCLIP on N-ImageNet. The event camera resolution is 640×\times480. We report the runtime (ms) of EventCLIP’s two main components.

D.5 Speed Analysis of EventCLIP

Low imaging latency is an important property of event cameras. We test the inference time of our EventCLIP pipeline, which consists of two components: event-to-frame conversion (CPU) and model forward pass (GPU). All speeds are measured on a Linux desktop with AMD Ryzen 9 5950X CPU (16-Core) and NVIDIA GeForce 3090 GPU. We report the runtime on N-ImageNet which has the highest frame resolution of 640×\times480 in Tab. 9. Due to the use of simple event counting, our event-to-frame conversion time is 6.76 ms. Adding to the 8.73 ms model forward time, EventCLIP is able to run at more than 50 FPS, rendering it suitable for real-time applications with event cameras.

Refer to caption
Refer to caption
Figure 7: Comparison of few-shot accuracy between EventCLIP and baselines initialized from weights pre-trained on N-ImageNet. Our method still achieves more data-efficient learning on N-Cars, while performing competitively on N-Caltech.

D.6 N-ImageNet Pre-trained Baselines

In our main experiments, all baselines employ backbones pre-trained on the RGB images from ImageNet. Here, we evaluate a more challenging scenario, where the backbones are initialized from weights pre-trained on the large-scale N-ImageNet dataset. Fig. 7 compares the few-shot accuracy of EventCLIP and the baselines. N-ImageNet pre-training provides substantial domain-specific knowledge for event-based classification, and thus greatly improves their performance. Still, EventCLIP is able to outperform the baselines with a sizeable margin when the number of data per category is small (e.g., 10-shot), and achieve competitive results with more training data. This demonstrates the effective knowledge transfer process of our approach.

Refer to caption
(a) RGB image
Refer to caption
(b) Original
Refer to caption
(c) Variant 2
Refer to caption
(d) Variant 2
Refer to caption
(e) Variant 5
Refer to caption
(f) Variant 6
Refer to caption
(g) Variant 7
Refer to caption
(h) Variant 9
Refer to caption
(a) RGB image
Refer to caption
(b) Original
Refer to caption
(c) Variant 2
Refer to caption
(d) Variant 2
Refer to caption
(e) Variant 5
Refer to caption
(f) Variant 6
Refer to caption
(g) Variant 7
Refer to caption
(h) Variant 9
Figure 8: An N-ImageNet [32] sample from the “monarch butterfly” category. We show (a) the source RGB image from ImageNet [9] used to generate the events for reference. We plot (b) the event histograms of this data under the normal capture setting and (c-h) several robustness variants. Variant 2, 3, and 5 introduce changes in the camera trajectory, while Variant 6, 7, and 9 use different lighting during data capture. Left: we apply the red-blue color map for better visual quality. Right: the actual inputs to EventCLIP with the gray-scale color map.
Refer to caption
(a) RGB image
Refer to caption
(b) Original
Refer to caption
(c) Variant 2
Refer to caption
(d) Variant 2
Refer to caption
(e) Variant 5
Refer to caption
(f) Variant 6
Refer to caption
(g) Variant 7
Refer to caption
(h) Variant 9
Refer to caption
(i) RGB image
Refer to caption
(j) Original
Refer to caption
(k) Variant 2
Refer to caption
(l) Variant 2
Refer to caption
(m) Variant 5
Refer to caption
(n) Variant 6
Refer to caption
(o) Variant 7
Refer to caption
(p) Variant 9
Figure 9: An N-ImageNet sample from the “folding chair” category. We again plot the red-blue colorized images at left and the gray-scale input to EventCLIP at right. Due to extreme low light, events under Variant 6 do not contain any object information, making classification impossible.
Refer to caption
(a) N-ImageNet sample
Refer to caption
(b) N-ImageNet sample
Refer to caption
(c) Gen1 [8] detection data
Figure 10: Transfer Detic [84] to perform instance segmentation on (a), (b) synthetic data, and (c) real-world detection data.
Dataset N-Caltech
Shots 0 1 3 5 10 20
Visual-MLP 69.67 73.52 77.43 78.55 82.69 83.28
Visual-Trans. 69.67 76.51 81.44 82.99 85.21 85.73
Text 69.67 77.35 81.58 82.81 84.89 86.33
Joint 69.67 77.89 82.48 83.19 85.62 86.41
Table 10: Full zero-shot and few-shot classification accuracy (%\%) of EventCLIP on N-Caltech.
Dataset N-Cars
Shots 0 10 30 50 100 200
Visual-Trans. 82.28 84.55 86.62 87.36 89.51 90.33
Text 82.28 84.37 87.06 87.20 89.40 90.34
Joint 82.28 84.77 87.57 88.44 89.49 90.29
Table 11: Full zero-shot and few-shot classification accuracy (%\%) of EventCLIP on N-Cars.
Dataset N-ImageNet
Shots 0 1 3 5 10 20
Visual-Trans. 20.78 23.36 24.45 25.20 26.64 28.45
Text 20.78 24.04 25.47 26.55 28.24 30.25
Joint 20.78 24.25 25.84 26.96 28.63 30.57
Table 12: Full zero-shot and few-shot classification accuracy (%\%) of EventCLIP on N-ImageNet.

D.7 Full Numerical Results

In the main paper, we plot EventCLIP’s zero-shot and few-shot classification accuracy in Fig.  3 and Fig.  4. To ease future comparison, we report all those numbers in Tab. 10, Tab. 11, and Tab. 12.

Appendix E Limitations and Future Works

In this paper, we mainly focus on the event-based object recognition problem. It is still unclear how to utilize large pre-trained models for other event camera tasks, such as detection [8, 48] and segmentation [65, 61]. We conduct preliminary experiments to transfer an open-world instance segmentation model Detic [84] to event data. Since Detic also builds upon CLIP, it is able to detect objects from N-ImageNet samples as shown in Fig. 10 (a), (b). However, objects present more complex motions in real-world captured events, degrading the converted event frames’ visual quality drastically. Still, Detic is able to detect some objects as shown in Fig. 10 (c). But it also misses objects with sparse edges such as the motorbikes and trucks. Besides, it predicts some weird classes such as bananas and scissors due to the large domain gap. Therefore, it is worth studying better event-to-frame conversion or model adaptation methods to better leverage the pre-trained vision foundation models.

Another direction is to directly learn a joint embedding space of events and texts. Due to a lack of event-text dataset, we can leverage RGB images as the intermediate representation, as done in some recent work [22, 72]. We can first leverage the large-scale RGB video datasets such as CO3D [55] to simulate event data. Then, we train an event encoder to align the extracted features with a pre-trained image encoder such as CLIP.