Abstract
This study investigated the use of computer vision–based algorithms for detecting needle-type instrument readings in the cockpit of an aerial vehicle. A flight data recorder plays a crucial role in aviation safety investigations and flight operation review. In practice, not all aerial vehicles are equipped with flight data recorders, and this poses problems in the retrieval of instrument readings during investigations. Installing a lightweight recorder such as a camera in the cockpit to record the instrument panel is a solution to the mentioned problem. Although recorded flight data can be retrieved through human inspection, computer vision–based algorithms enable more rapid and efficient detection. Accordingly, this study developed two computer vision–based algorithms operated in both of the grayscale color space and the value layer of the hue–saturation–value color space. Performance of the four combinations is then compared and the best combination of algorithm along with operation space is suggested. The airspeed meter of a Bell 206 helicopter was selected to test the proposed detection algorithm in this study. GPS data and human inspection results were used as references. Herein, experimental results are presented, performance of algorithms is discussed, and conclusions are provided. This study contributes to aviation safety investigations and flight operation review involving aerial vehicles that are not equipped with flight data recorders.
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-022-13226-y/MediaObjects/11042_2022_13226_Fig1_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-022-13226-y/MediaObjects/11042_2022_13226_Fig2_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-022-13226-y/MediaObjects/11042_2022_13226_Fig3_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-022-13226-y/MediaObjects/11042_2022_13226_Fig4_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-022-13226-y/MediaObjects/11042_2022_13226_Fig5_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-022-13226-y/MediaObjects/11042_2022_13226_Fig6_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-022-13226-y/MediaObjects/11042_2022_13226_Fig7_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-022-13226-y/MediaObjects/11042_2022_13226_Fig8_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-022-13226-y/MediaObjects/11042_2022_13226_Fig9_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-022-13226-y/MediaObjects/11042_2022_13226_Fig10_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-022-13226-y/MediaObjects/11042_2022_13226_Fig11_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-022-13226-y/MediaObjects/11042_2022_13226_Fig12_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-022-13226-y/MediaObjects/11042_2022_13226_Fig13_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-022-13226-y/MediaObjects/11042_2022_13226_Fig14_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-022-13226-y/MediaObjects/11042_2022_13226_Fig15_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-022-13226-y/MediaObjects/11042_2022_13226_Fig16_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-022-13226-y/MediaObjects/11042_2022_13226_Fig17_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-022-13226-y/MediaObjects/11042_2022_13226_Fig18_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-022-13226-y/MediaObjects/11042_2022_13226_Fig19_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-022-13226-y/MediaObjects/11042_2022_13226_Fig20_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-022-13226-y/MediaObjects/11042_2022_13226_Fig21_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-022-13226-y/MediaObjects/11042_2022_13226_Fig22_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-022-13226-y/MediaObjects/11042_2022_13226_Fig23_HTML.png)
Similar content being viewed by others
References
CAA (2018) Civil aviation act of the Republic of China (Taiwan). https://law.moj.gov.tw/ENG/LawClass/LawAll.aspx?pcode=K0090001. Accessed 2 Jun 2019
Chi J, Liu L, Liu J, Jiang Z, Zhang G (2015) Machine vision based automatic detection method of indicating values of a pointer gauge. Math Probl Eng 2015:1–19. https://doi.org/10.1155/2015/283629
Chmielińska J, Jakubowski J (2016) Application of matlab for automatic reading of analogue measuring instruments. Przegla̧d Elektrotechniczny 92(11):95–98
Dong S, Wang P, Abbas K (2021) A survey on deep learning and its applications. Comput Sci Rev 40:100379. https://doi.org/10.1016/j.cosrev.2021.100379
Ezatzadeh S, Keyvanpour MR (2019) Vifa: an analytical framework for vision-based fall detection in a surveillance environment. Multimed Tools Appl 78 (18):25515–25537. https://doi.org/10.1007/s11042-019-7720-3
FAA (2021) FAA Order Jo 7110.65Z - air traffice control. https://www.faa.gov/air_traffic/publications/atpubs/atc_html/chap3_section_11.html. Accessed 25 Jul 2021
Gonzales RC, Woods RE (2002) Digital image processing. Prentice Hall, Englewood Cliffs
Hsiao FY, Lang CN (2016) Real-time target determination and tracking with an airborne video system. J Aeronaut Astronaut Aviat 48(1):11–19
Khan H, Rasool G, Bouaynaya NC, Johnson CC (2019) Rotorcraft flight information inference from cockpit videos using deep learning. In: Vertical flight society’s 75th annual forum and technology display, Philadelphia, United States
Medjram S, Babahenini MC, Taleb-Ahmed A, et al. (2018) Automatic hand detection in color images based on skin region verification. Multimed Tools Appl 77. https://doi.org/10.1007/s11042-017-4995-0
Pratt WK (1991) Digital image processing. Wiley, New York
Sablatnig R, Kropatsch WG (1994) Automatic reading of analog display instruments. In: Proceedings of 12th international conference on pattern recognition, vol 1, pp 794–797
Shin S, Hwang I (2016) Helicopter flight information inference from cockpit video data using dbscan clustering. In: The 16th AIAA aviation technology, integration, and operations conference, Washington, D.C., Paper No.: AIAA-2016-4078
Smith A (1978) Color gamut transform pairs. ACM Siggraph Computer Graphics 12:12–19. https://doi.org/10.1145/800248.807361
TTSB (2017) Safety recommendation to the aviation accidence of emerald pacific airlines b-31118 helicopter. In: Safety recommendation, Taiwan transportation safety board, report No.: ASC-ASR-18-10-012. https://www.ttsb.gov.tw/1133/1154/1155/1160/20795/post
Vision-1000 (2020) Website of vision 1000 in appareo. https://www.appareo.com/aviation/flight-data-monitoring/vision-1000/. Accessed 20 May 2020
Xuand Tian Fang L, Gao X (2015) An automatic recognition method of pointer instrument based on improved hough transform. p 96752T. https://doi.org/10.1117/12.2202805. Accessed 2 Jun 2019
Ye X, **e D, Tao S (2013) Automatic value identification of pointer-type pressure gauge based on machine vision. J Comput 8. https://doi.org/10.4304/jcp.8.5.1309--1314
Zhou Y (2016) Research on identification method of pointer type instrument in substation. Int J Sci 3(6):125–130
Acknowledgements
The authors thank the Taiwan Transportation Safety Board (previously known as the Aviation Safety Council) and the Emerald Pacific Airlines of Taiwan for providing cockpit videos used for analysis.
Funding
This research was funded by the Taiwan Ministry of Science and Technology under grant number MOST 107-2221-E-032-027.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of Interests
The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendices
Appendix A: Abbreviation
The following abbreviations are used in this manuscript:
- CVR:
-
Cockpit voice recorder
- DBSCAN:
-
Density-based spatial clustering of applications with noise
- FDR:
-
Flight data recorder
- GPS:
-
Global Positioning System
- HSV:
-
Hue-saturation-value
- ICAO:
-
International Civil Aviation Organization
- RGB:
-
Red-green-blue
- STD:
-
Standard deviation
- TTSB:
-
Taiwan Transportation Safety Board
Appendix B: Pseudo code
1.1 B.1 Transformation of color space
![figure a](http://media.springernature.com/lw685/springer-static/image/art%3A10.1007%2Fs11042-022-13226-y/MediaObjects/11042_2022_13226_Figa_HTML.png)
1.2 B.2 Binarization and edge detection
Suppose that the thresholds for binarization are \((T_{\min \limits },T_{\max \limits })\). The edge detection employs Matlab function edge() directly. The pesudo code is presented as follows:
![figure b](http://media.springernature.com/lw685/springer-static/image/art%3A10.1007%2Fs11042-022-13226-y/MediaObjects/11042_2022_13226_Figb_HTML.png)
1.3 B.3 Build of needle mask
![figure c](http://media.springernature.com/lw685/springer-static/image/art%3A10.1007%2Fs11042-022-13226-y/MediaObjects/11042_2022_13226_Figc_HTML.png)
![figure d](http://media.springernature.com/lw685/springer-static/image/art%3A10.1007%2Fs11042-022-13226-y/MediaObjects/11042_2022_13226_Figd_HTML.png)
1.4 B.4 Needle angle detection using mask
![figure e](http://media.springernature.com/lw685/springer-static/image/art%3A10.1007%2Fs11042-022-13226-y/MediaObjects/11042_2022_13226_Fige_HTML.png)
1.5 B.5 Identification of needle from background
The biggest object in the binary image is identified by Matlab function bwconncomp(). Suppose that the image is B.jpg. The Matlab code is presented as follows:
function [NDL,biggest] = NDL_idf(Bunit8) Bunit8 = impread('B.jpg'); Bdble = double(Bunit8); Bnml = Bdble/255; Bcnd = bwconncomp(Bnml); neighborhood. index = 1:Bcnd.NumObjects; numPixels = cellfun(@numel,B.PixelIdxList); Identify the index of the biggest object [biggest,idx] = max(numPixels); index(idx) = []; Bnml(Bcnd.PixelIdxList{index(i)}) = 0; objects end NDL = Bnml;
1.6 B.6: Needle angle detection using image moment
![figure f](http://media.springernature.com/lw685/springer-static/image/art%3A10.1007%2Fs11042-022-13226-y/MediaObjects/11042_2022_13226_Figf_HTML.png)
![figure g](http://media.springernature.com/lw685/springer-static/image/art%3A10.1007%2Fs11042-022-13226-y/MediaObjects/11042_2022_13226_Figg_HTML.png)
Rights and permissions
About this article
Cite this article
Hsiao, FY., Chang, FY., Vida, P. et al. Reading detection of needle-type instrument in a noisy environment using computer vision-based algorithms. Multimed Tools Appl 82, 1749–1782 (2023). https://doi.org/10.1007/s11042-022-13226-y
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11042-022-13226-y