Publication | Open Access
DeCAF: A Deep Convolutional Activation Feature for Generic Visual Recognition
3.6K
Citations
38
References
2013
Year
Convolutional Neural NetworkEngineeringMachine LearningGeneric Visual RecognitionImage ClassificationImage AnalysisData SciencePattern RecognitionFixed FeatureVideo TransformerVision RecognitionDeep Convolutional NetworkMachine VisionFeature LearningVision Language ModelComputer ScienceDeep LearningComputer VisionObject RecognitionDeep Convolutional Features
Generic visual tasks often differ from the original training set and lack sufficient labeled data, making conventional deep architecture adaptation difficult. The study evaluates whether deep convolutional activation features can be repurposed for generic visual tasks and releases DeCAF to enable such experimentation. The authors investigate and visualize semantic clustering of these features across scene recognition, domain adaptation, and fine‑grained tasks, and provide an open‑source implementation with network parameters. They demonstrate that features from various network levels form a fixed representation that significantly outperforms state‑of‑the‑art on several vision challenges.
We evaluate whether features extracted from the activation of a deep convolutional network trained in a fully supervised fashion on a large, fixed set of object recognition tasks can be re-purposed to novel generic tasks. Our generic tasks may differ significantly from the originally trained tasks and there may be insufficient labeled or unlabeled data to conventionally train or adapt a deep architecture to the new tasks. We investigate and visualize the semantic clustering of deep convolutional features with respect to a variety of such tasks, including scene recognition, domain adaptation, and fine-grained recognition challenges. We compare the efficacy of relying on various network levels to define a fixed feature, and report novel results that significantly outperform the state-of-the-art on several important vision challenges. We are releasing DeCAF, an open-source implementation of these deep convolutional activation features, along with all associated network parameters to enable vision researchers to be able to conduct experimentation with deep representations across a range of visual concept learning paradigms.
| Year | Citations | |
|---|---|---|
2009 | 60.2K | |
1998 | 56.5K | |
2008 | 35.7K | |
2005 | 31.6K | |
2006 | 20.5K | |
1989 | 11.6K | |
2009 | 10K | |
2012 | 6.6K | |
2001 | 6.4K | |
1997 | 6.1K |
Page 1
Page 1