Longshortnet: Exploring temporal and semantic features fusion in streaming perception

C Li, ZQ Cheng, JY He, P Li, B Luo… - ICASSP 2023-2023 …, 2023 - ieeexplore.ieee.org
C Li, ZQ Cheng, JY He, P Li, B Luo, H Chen, Y Geng, JP Lan, X Xie
ICASSP 2023-2023 IEEE International Conference on Acoustics …, 2023ieeexplore.ieee.org
Streaming perception is a fundamental task in autonomous driving that requires a careful
balance between the latency and accuracy of the autopilot system. However, current
methods for streaming perception are limited as they rely only on the current and adjacent
two frames to learn movement patterns, which restricts their ability to model complex scenes,
often leading to poor detection results. To address this limitation, we propose LongShortNet,
a novel dual-path network that captures long-term temporal motion and integrates it with …
Streaming perception is a fundamental task in autonomous driving that requires a careful balance between the latency and accuracy of the autopilot system. However, current methods for streaming perception are limited as they rely only on the current and adjacent two frames to learn movement patterns, which restricts their ability to model complex scenes, often leading to poor detection results. To address this limitation, we propose LongShortNet, a novel dual-path network that captures long-term temporal motion and integrates it with short-term spatial semantics for real-time perception. Our proposed LongShortNet is notable as it is the first work to extend long-term temporal modeling to streaming perception, enabling spatiotemporal feature fusion. We evaluate LongShortNet on the challenging Argoverse-HD dataset and demonstrate that it outperforms existing state-of-the-art methods with almost no additional computational cost. 1
ieeexplore.ieee.org
以上显示的是最相近的搜索结果。 查看全部搜索结果