YOLIC: An efficient method for object localization and classification on edge devices

K Su, Y Tomioka, Q Zhao, Y Liu - Image and Vision Computing, 2024 - Elsevier
Image and Vision Computing, 2024Elsevier
In the realm of Tiny AI, we introduce “You Only Look at Interested Cells”(YOLIC), an efficient
method for object localization and classification on edge devices. Through seamlessly
blending the strengths of semantic segmentation and object detection, YOLIC provides
improved computational efficiency and precision compared to traditional methods. By
adopting Cells of Interest for classification instead of individual pixels, YOLIC encapsulates
relevant information, reduces computational load, and enables rough object shape …
Abstract
In the realm of Tiny AI, we introduce “You Only Look at Interested Cells” (YOLIC), an efficient method for object localization and classification on edge devices. Through seamlessly blending the strengths of semantic segmentation and object detection, YOLIC provides improved computational efficiency and precision compared to traditional methods. By adopting Cells of Interest for classification instead of individual pixels, YOLIC encapsulates relevant information, reduces computational load, and enables rough object shape inference. Importantly, the need for bounding box regression is obviated, as YOLIC capitalizes on the predetermined cell configuration that provides information about potential object location, size, and shape. To tackle the issue of single-label classification limitations, a multi-label classification approach is applied to each cell for effectively recognizing overlapping or closely situated objects. This paper presents extensive experiments on multiple datasets to demonstrate that YOLIC achieves detection performance comparable to the state-of-the-art YOLO algorithms while surpassing in speed, exceeding 30fps on a Raspberry Pi 4B CPU.
Elsevier
以上显示的是最相近的搜索结果。 查看全部搜索结果