| 英文摘要 |
This paper focuses on addressing the challenges of real-time inference under computational resource and power limitations, especially on embedded edge devices. Optimizing both training and inference speeds efficiently in these contexts has become a significant research topic. The YOLO (You Only Look Once) algorithm, which processes images using a single pass through a convolutional neural network (CNN), achieves high performance and detection rates. Compared to two-stage object detection models like R-CNN and Faster R-CNN, YOLO significantly increases recognition speed. In this context, the research aims to develop an intelligent surveillance system by integrating deep learning techniques on the FPGA-KV260 platform. The KV260, as a high-performance FPGA platform, maximizes the potential of deep learning algorithms, enhancing computational efficiency and reducing power consumption. This experiment aims to achieve the following objectives using the YOLOv4 algorithm: lGenerate YOLOv4 Weights File:By training the YOLOv4 model with an appropriate dataset, a weights file required for the project will be generated. Creating a custom training dataset ensures higher accuracy compared to open-source options and aligns anchor boxes with project-specific needs, minimizing unwanted detection results. lStream Video to KV260 for Real-Time Detection:A webcam will stream video data to the FPGA-KV260 platform, where the YOLOv4 model will perform facial detection, with results displayed on a computer. The experimental results indicate that the proposed method can achieve over 95% accuracy under favorable lighting conditions. |