Edge AI for Automotive Vulnerable Road User Safety: Deployable Detection via Knowledge Distillation
Akshay Karjol, Darrin M. Hanna
TLDR
This paper uses knowledge distillation to enable accurate, deployable Vulnerable Road User detection on edge AI for automotive safety, overcoming INT8 quantization issues.
Key contributions
- Developed a knowledge distillation framework for compact YOLOv8-S from YOLOv8-L for VRU detection.
- Achieved 3.9x model compression while preserving INT8 quantization robustness for edge deployment.
- KD student maintained accuracy (-5.6% mAP) where the teacher failed (-23% mAP) under INT8 quantization.
- KD transferred precision calibration, reducing false alarms by 44% and surpassing teacher's FP32 precision.
Why it matters
Deploying accurate VRU detection on edge AI for automotive safety is challenging due to model size and quantization. This work shows knowledge distillation is essential to overcome these hurdles, enabling robust, high-precision detection on resource-constrained hardware. It makes safety-critical AI deployable in real-world automotive applications.
Original Abstract
Deploying accurate object detection for Vulnerable Road User (VRU) safety on edge hardware requires balancing model capacity against computational constraints. Large models achieve high accuracy but fail under INT8 quantization required for edge deployment, while small models sacrifice detection performance. This paper presents a knowledge distillation (KD) framework that trains a compact YOLOv8-S student (11.2M parameters) to mimic a YOLOv8-L teacher (43.7M parameters), achieving 3.9x compression while preserving quantization robustness. We evaluate on full-scale BDD100K (70K training images) with Post-Training Quantization to INT8. The teacher suffers catastrophic degradation under INT8 (-23% mAP), while the KD student retains accuracy (-5.6% mAP). Analysis reveals that KD transfers precision calibration rather than raw detection capacity: the KD student achieves 0.748 precision versus 0.653 for direct training at INT8, a 14.5% gain at equivalent recall, reducing false alarms by 44% versus the collapsed teacher. At INT8, the KD student exceeds the teacher's FP32 precision (0.748 vs. 0.718) in a model 3.9x smaller. These findings establish knowledge distillation as a requirement for deploying accurate, safety-critical VRU detection on edge hardware.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.