| | |
| | |
Stat |
Members: 3667 Articles: 2'599'751 Articles rated: 2609
09 February 2025 |
|
| | | |
|
Article overview
| |
|
Ego-Only: Egocentric Action Detection without Exocentric Pretraining | Huiyu Wang
; Mitesh Kumar Singh
; Lorenzo Torresani
; | Date: |
3 Jan 2023 | Abstract: | We present Ego-Only, the first training pipeline that enables
state-of-the-art action detection on egocentric (first-person) videos without
any form of exocentric (third-person) pretraining. Previous approaches found
that egocentric models cannot be trained effectively from scratch and that
exocentric representations transfer well to first-person videos. In this paper
we revisit these two observations. Motivated by the large content and
appearance gap separating the two domains, we propose a strategy that enables
effective training of egocentric models without exocentric pretraining. Our
Ego-Only pipeline is simple. It trains the video representation with a masked
autoencoder finetuned for temporal segmentation. The learned features are then
fed to an off-the-shelf temporal action localization method to detect actions.
We evaluate our approach on two established egocentric video datasets: Ego4D
and EPIC-Kitchens-100. On Ego4D, our Ego-Only is on-par with exocentric
pretraining methods that use an order of magnitude more labels. On
EPIC-Kitchens-100, our Ego-Only even outperforms exocentric pretraining (by
2.1% on verbs and by 1.8% on nouns), setting a new state-of-the-art. | Source: | arXiv, 2301.01380 | Services: | Forum | Review | PDF | Favorites |
|
|
No review found.
Did you like this article?
Note: answers to reviews or questions about the article must be posted in the forum section.
Authors are not allowed to review their own article. They can use the forum section.
|
| |
|
|
|