Abstract
Automated optical inspection of FAST is realized by exploiting advances in drone technology and deep-learning techniques. The AI-powered drone-based automated inspection is time-efficient and reliable, which guarantees the stable operation of FAST.
The Five-hundred-meter Aperture Spherical radio Telescope (FAST), also known as the “China Sky Eye”, is the world’s largest single-dish radio telescope1. Its optical geometry is outlined in Fig. 1a. The reflector is a partial sphere of radius R = 300 m. The planar partial spherical cap of the reflector has a diameter of 519.6 m, 1.7 times larger than that of the former largest radio telescope. The large reflecting surface makes FAST the world’s most sensitive radio telescope. It was used by astronomers to observe, for the first time, fast radio bursts in the Milky Way and to identify more than 500 new pulsars, four times the total number of pulsars identified by other telescopes worldwide. More interesting and exotic objects may yet be discovered using FAST.
However, each coin has two sides. A larger reflecting surface is more prone to external damage due to environmental factors. The FAST reflector comprises a total of 4450 spliced trilateral panels, made of aluminum with uniform perforations to reduce weight and wind impact. Falling objects (e.g., during extreme events such as rockfalls, severe windstorms, and hailstorms) may cause severe dents and holes in the panels. Such defects adversely impact the study of small-wavelength radio waves, which demands a perfect dish surface. Any irregularity in the parabola scatters these small waves away from the focus, causing information loss.
The rapid detection of surface defects for timely repair is hence critical for maintaining the normal operation of FAST. This is traditionally done by direct visual inspection. Skilled inspectors climb up the reflector and visually examine the entire surface, searching for and replacing any panels showing dents and holes. However, this procedure has several limitations. Firstly, there is danger involved in accessing hard-to-reach places high above the ground. Secondly, it is labor- and time-consuming to scrutinize all the thousands of panels. Thirdly, the procedure relies heavily on the inspectors’ expertize and is prone to human-based errors and inconsistencies.
The remedy to the shortcomings of manual inspection at FAST is automated inspection. A recent publication in Light: Advanced Manufacturing by Li et al.2 made the first step towards automating the inspection of FAST by integrating deep-learning techniques with drone technology. As a first step, the research team integrated deep-learning techniques with the use of drones to automatically detect defects on the reflector surface. Specifically, they began by manually controlling a drone equipped with a high-resolution RGB camera to fly over the surface along a predetermined route (Fig. 1b). During the flight, the camera captured and recorded videos of the surface condition. One benefit of the advanced flight stability of drones is that the recorded videos can capture much information on surface details. Moreover, thanks to the GPS device and the RTK module onboard the drone platform, every video frame can be tagged with the corresponding drone location with centimeter-level accuracy. The physical locations of the panels that appear in each frame can thus be determined.
Previous works involving defect detection using aerial imagery were primarily designed to detect large defects and were not reliable for detecting very small defects3,4,5,6,7,8. In contrast, this work aims to inspect the large surface of FAST from on high. The surface defects in drone imagery exhibit large-scale variation and high inter-class similarity (Fig. 1c). To tackle the above challenges, the research team introduced a simple yet effective cross-fusion operation for deep detectors, which aggregates multi-level features in a point-wise selective manner to help detect defects of various scales and types (Fig. 1d). The cross-fusion method is lightweight and computationally efficient, particularly valuable features for onboard drone applications. Future work will implement the algorithm on embedded hardware platforms to process captured videos onboard the drone, and to make the inspection system more autonomous and robust.
References
Nan, R. D. et al. The five-hundred-meter aperture spherical radio telescope (FAST) project. Int. J. Mod. Phys. D. 20, 989–1024 (2011).
Li, J. N. et al. Automated optical inspection of FAST’s reflector surface using drones and computer vision. Light Adv. Manuf. 4, 1 (2023).
Zhu, J. Q. et al. Pavement distress detection using convolutional neural networks with images captured via UAV. Autom. Constr. 133, 103991 (2022).
Zhang, C. B., Chang, C. C. & Jamshidi, M. Concrete bridge surface damage detection using a single-stage detector. Comput.-Aided Civ. Infrastruct. Eng. 35, 389–409 (2020).
Du, F. J., Jiao, S. J. & Chu, K. L. Application research of bridge damage detection based on the improved lightweight convolutional neural network model. Appl. Sci. 12, 6225 (2022).
Liu, C. Y. et al. Insulator faults detection in aerial images from high-voltage transmission lines based on deep learning model. Appl. Sci. 11, 4647 (2021).
Liu, J. J. et al. An improved method based on deep learning for insulator fault detection in diverse aerial images. Energies 14, 4365 (2021).
Di Tommaso, A. et al. A multi-stage model based on YOLOv3 for defect detection in PV panels based on IR and visible imaging by unmanned aerial vehicle. Renew. Energy 193, 941–962 (2022).
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Wang, L. AI-powered drone-based automated inspection of FAST. Light Sci Appl 12, 63 (2023). https://doi.org/10.1038/s41377-023-01094-6
Published:
DOI: https://doi.org/10.1038/s41377-023-01094-6