TY - JOUR
T1 - Audio-Based Aircraft Detection System for Safe RPAS BVLOS Operations
AU - Mariscal-Harana, Jorge
AU - Alarcón, Víctor
AU - González, Fidel
AU - Calvente, Juan José
AU - Pérez-Grau, Francisco Javier
AU - Viguria, Antidio
AU - Ollero, Aníbal
PY - 2020/12/5
Y1 - 2020/12/5
N2 - For the Remotely Piloted Aircraft Systems (RPAS) market to continue its current growth rate, cost-effective ‘Detect and Avoid’ systems that enable safe beyond visual line of sight (BVLOS) operations are critical. We propose an audio-based ‘Detect and Avoid’ system, composed of microphones and an embedded computer, which performs real-time inferences using a sound event detection (SED) deep learning model. Two state-of-the-art SED models, YAMNet and VGGish, are fine-tuned using our dataset of aircraft sounds and their performances are compared for a wide range of configurations. YAMNet, whose MobileNet architecture is designed for embedded applications, outperformed VGGish both in terms of aircraft detection and computational performance. YAMNet’s optimal configuration, with >70% true positive rate and precision, results from combining data augmentation and undersampling with the highest available inference frequency (i.e., 10 Hz). While our proposed ‘Detect and Avoid’ system already allows the detection of small aircraft from sound in real time, additional testing using multiple aircraft types is required. Finally, a larger training dataset, sensor fusion, or remote computations on cloud-based services could further improve system performance.
AB - For the Remotely Piloted Aircraft Systems (RPAS) market to continue its current growth rate, cost-effective ‘Detect and Avoid’ systems that enable safe beyond visual line of sight (BVLOS) operations are critical. We propose an audio-based ‘Detect and Avoid’ system, composed of microphones and an embedded computer, which performs real-time inferences using a sound event detection (SED) deep learning model. Two state-of-the-art SED models, YAMNet and VGGish, are fine-tuned using our dataset of aircraft sounds and their performances are compared for a wide range of configurations. YAMNet, whose MobileNet architecture is designed for embedded applications, outperformed VGGish both in terms of aircraft detection and computational performance. YAMNet’s optimal configuration, with >70% true positive rate and precision, results from combining data augmentation and undersampling with the highest available inference frequency (i.e., 10 Hz). While our proposed ‘Detect and Avoid’ system already allows the detection of small aircraft from sound in real time, additional testing using multiple aircraft types is required. Finally, a larger training dataset, sensor fusion, or remote computations on cloud-based services could further improve system performance.
KW - deep learning
KW - sound event detection
KW - convolutional neural networks
KW - audio processing
KW - embedded systems
UR - http://www.scopus.com/inward/record.url?scp=85097445971&partnerID=8YFLogxK
U2 - 10.3390/electronics9122076
DO - 10.3390/electronics9122076
M3 - Article
SN - 2079-9292
VL - 9
SP - 1
EP - 13
JO - Electronics
JF - Electronics
IS - 12
M1 - 2076
ER -