Publication | Closed Access
Temporally Distributed Networks for Fast Video Semantic Segmentation
200
Citations
46
References
2020
Year
Unknown Venue
Deep CnnConvolutional Neural NetworkMachine VisionImage AnalysisData ScienceMachine LearningPattern RecognitionEngineeringInherent Temporal ContinuityVideo HallucinationComputer ScienceVideo UnderstandingPresent TdnetVideo TransformerDeep LearningVideo RetrievalVideo InterpretationComputer Vision
We present TDNet, a temporally distributed network designed for fast and accurate video semantic segmentation. We observe that features extracted from a certain high-level layer of a deep CNN can be approximated by composing features extracted from several shallower sub-networks. Leveraging the inherent temporal continuity in videos, we distribute these sub-networks over sequential frames. Therefore, at each time step, we only need to perform a lightweight computation to extract a sub-features group from a single sub-network. The full features used for segmentation are then recomposed by application of a novel attention propagation module that compensates for geometry deformation between frames. A grouped knowledge distillation loss is also introduced to further improve the representation power at both full and sub-feature levels. Experiments on Cityscapes, CamVid, and NYUD-v2 demonstrate that our method achieves state-of-the-art accuracy with significantly faster speed and lower latency.
| Year | Citations | |
|---|---|---|
Page 1
Page 1