| 研究生: |
盧柏翰 LU, POHAN |
|---|---|
| 論文名稱: |
面向多場景的小物件偵測之高效率 YOLOv11 輕量化模型 An Efficient Lightweight YOLOv11 Model for Small Object Detection in Multi-Scene Environments |
| 指導教授: |
戴顯權
Tai, Shen-Chuan |
| 學位類別: |
碩士 Master |
| 系所名稱: |
電機資訊學院 - 電機工程學系 Department of Electrical Engineering |
| 論文出版年: | 2026 |
| 畢業學年度: | 114 |
| 語文別: | 英文 |
| 論文頁數: | 91 |
| 中文關鍵詞: | 小物件偵測 、輕量化模型 、多尺度偵測架構 、GhostConv 、YOLO 、多場景學習 |
| 外文關鍵詞: | Small Object Detection, Lightweight Model, Multi-scale Detection Architecture, GhostConv, YOLO, Multi-scene Learning |
| ORCID: | 0009-0002-9151-181X |
| ResearchGate: | Deep Learning |
| 相關次數: | 點閱:52 下載:0 |
| 分享至: |
| 查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報 |
在多場景環境中進行小物件偵測,常因目標尺度有限與視覺條件複雜而面臨挑戰,而實際部署情境亦對模型的體積與計算效率提出嚴格限制。本論文提出一種以 YOLOv11 為基礎的輕量化、部署導向小物件偵測框架,適用於多元場景下的實務應用。本研究以 YOLOv11-nano 為基準模型,透過效率導向的架構調整策略,包括通道重新配置(Channel Reallocation)、基於 GhostConv 的特徵融合,以及加入 P2 分支的多尺度偵測設計,以在不增加網路深度或引入重量級模組的前提下,保留高解析度空間特徵。實驗結果顯示,於 TACO、PlastOPol 與 VisDrone 等資料集上,所提出之方法能在顯著降低模型體積與參數量的情況下,維持良好的小物件偵測能力。整體而言,本方法在模型體積約降低40%、參數量控制於1.5M的條件下,提供一個具備實際部署可行性的多場景小物件偵測解決方案。
Small object detection in multi-scene environments is challenging due to limited object scale and complex visual conditions, while practical deployment requires compact and efficient detection models. This thesis presents a lightweight, deployment-oriented object detection framework based on YOLOv11 for small-object detection across diverse scenes. Built on the YOLOv11-nano baseline, the proposed method applies efficiency-driven architectural refinements, including channel reallocation, GhostConv-based feature fusion, and an extended multi-scale design with an additional P2 branch to preserve high-resolution spatial information, without increasing network depth or introducing heavyweight modules. Experiments on the TACO, PlastOPol, and VisDrone datasets show that the proposed framework maintains effective small-object detection capability under significantly reduced model size and parameter budgets. With approximately 40% reduction in model size and a parameter count of 1.5M, the proposed method offers a practical solution for deployment-oriented small-object detection in multi-scene environments.
[1] W. Wei, Y. Cheng, J. He, and X. Zhu, “A review of small object detection based on deep learning,” Neural Computing and Applications, vol. 36, no. 12, pp. 6283–6303, 2024.
[2] R. Varghese and M. Sambath, “Yolov8: A novel object detection algorithm with enhanced performance and robustness,” in 2024 International conference on advances in data engineering and intelligent computing systems (ADICS). IEEE, 2024, pp. 1–6.
[3] C.-Y. Wang, I.-H. Yeh, and H.-Y. Mark Liao, “Yolov9: Learning what you want to learn using programmable gradient information,” in European conference on computer vision. Springer, 2024, pp. 1–21.
[4] A. Wang, H. Chen, L. Liu, K. Chen, Z. Lin, J. Han et al., “Yolov10: Real-time end-to-end object detection,” Advances in Neural Information Processing Systems, vol. 37, pp. 107 984–108 011, 2024.
[5] R. Khanam and M. Hussain, “Yolov11: An overview of the key architectural enhancements,” arXiv preprint arXiv:2410.17725, 2024.
[6] Y. Tian, Q. Ye, and D. Doermann, “Yolov12: Attention-centric real-time object detectors,” arXiv preprint arXiv:2502.12524, 2025.
[7] X. Xu, Q. Li, J. Pan, X. Lu, H. Wei, M. Sun, and H. Zhang, “Esod-yolo: an enhanced efficient small object detection framework for aerial images,” Computing, vol. 107, no. 2, p. 54, 2025.
[8] M. Chao, C. Peng, L. Yun, C. Zhang, H. Wang, and Z. Chen, “A lightweight small object detection model for uav images based on deep semantic integration,” Scientific Reports, vol. 15, no. 1, p. 31888, 2025.
[9] M. Hu, Z. Li, J. Yu, X. Wan, H. Tan, and Z. Lin, “Efficient-lightweight yolo: improving small object detection in yolo for aerial images,” Sensors, vol. 23, no. 14, p. 6423, 2023.
[10] B. Du, Y. Huang, J. Chen, and D. Huang, “Adaptive sparse convolutional networks with global context enhancement for faster object detection on drone images,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2023, pp. 13 435–13 444.
[11] C. Zhang and J. Yang, “Emsd-detr: efficient small object detection for uav aerial images based on enhanced rt-detr model,” The Journal of Supercomputing, vol. 81, no. 9, pp. 1–33, 2025.
[12] D. Chen and L. Zhang, “Sl-yolo: A stronger and lighter drone target detection model,” arXiv preprint arXiv:2411.11477, 2024.
[13] H.-K. Jung, “Yolo-drone: An efficient object detection approach using the ghosthead network for drone images,” arXiv preprint arXiv:2511.10905, 2025.
[14] G. Sun and F. Zhang, “Drse-yolo: Efficient and lightweight architecture for accurate waste detection,” IET Image Processing, vol. 19, no. 1, p. e70022, 2025.
[15] K. Han, Y. Wang, Q. Tian, J. Guo, C. Xu, and C. Xu, “Ghostnet: More features from cheap operations,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2020, arXiv:1911.11907.
[16] X. Chen, C. Yang, J. Mo, Y. Sun, H. Karmouni, Y. Jiang, and Z. Zheng, “Cspnext: A new efficient token hybrid backbone,” Engineering Applications of Artificial Intelligence, vol. 132, p. 107886, 2024.
[17] X. Li et al., “Slim-neck by gsconv: A better design paradigm of feature fusion for object detectors,” arXiv preprint arXiv:2206.02424, 2022.
[18] Z. Ge, S. Liu, F. Wang, Z. Li, and J. Sun, “Yolox: Exceeding yolo series in 2021,” arXiv preprint arXiv:2107.08430, 2021.
[19] P. Proenc¸a and P. Sim˜oes, “Taco: Trash annotations in context for litter detection,” CVPR Workshops, 2020.
[20] P. Consortium, “Plastopol: A dataset for plastic pollution detection in coastal environments,” 2022. [Online]. Available: https://research.wur.nl/en/datasets/ plastopol-a-dataset-for-litter-detection/
[21] D. Du, P. Zhu, L. Wen, X. Bian, H. Lin, Q. Hu, T. Peng, J. Zheng, X. Wang, Y. Zhang et al., “Visdrone-det2019: The vision meets drone object detection in image challenge results,” in Proceedings of the IEEE/CVF international conference on computer vision workshops, 2019, pp. 0–0.