Abstract:
Zero-shot object detection is an emerging approach that aims to detect unseen object categories by leveraging the knowledge from known categories. In this study, we propo...Show MoreMetadata
Abstract:
Zero-shot object detection is an emerging approach that aims to detect unseen object categories by leveraging the knowledge from known categories. In this study, we propose a novel zero-shot object detection framework based on the popular YOLO (You Only Look Once) architecture. Our approach addresses the challenge of detecting objects that have not been seen during training by learning a mapping between the visual features of unseen categories and their textual descriptions. We introduce a semantic embedding space that captures the relationship between visual and textual representations and use it to generate visual features for the unlabeled object instances. By aligning these visual features with the textual descriptions of unseen categories, we enable the detection of previously unseen object classes. Our experimental results demonstrate the effectiveness of our approach, achieving competitive performance for both unseen and seen object categories on kaggle datasets. Moreover, our framework's incremental learning capability allows for the continuous incorporation of new unseen object categories without the need for retraining the entire system. The proposed zero-shot object detection method opens up new opportunities for various applications, such as autonomous vehicles, surveillance systems, and robotics, where the ability to recognize novel objects is crucial with the high level of Accuracy-98.2%, Precision-97.9%, Recall-97.8%, F1 score-96.8% using You Only Look Once (YOLO) in machine learning algorithm.
Published in: 2024 IEEE International Conference on Information Technology, Electronics and Intelligent Communication Systems (ICITEICS)
Date of Conference: 28-29 June 2024
Date Added to IEEE Xplore: 22 August 2024
ISBN Information: