Next Article in Journal
Digital Finance and Advanced Manufacturing Industry Development in China: A Coupling Coordination Analysis
Next Article in Special Issue
Travel-Energy-Based Timetable Optimization in Urban Subway Systems
Previous Article in Journal
Study on the Effect of Non-Linear Ventilation Pipe Composite Measures on the Slope Permafrost Subgrade
Previous Article in Special Issue
A Time-Space Network-Based Optimization Method for Scheduling Depot Drivers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automatic Obstacle Detection Method for the Train Based on Deep Learning

1
School of Electronic and Information Engineering, Beijing Jiaotong University, Beijing 100044, China
2
School of Computer and Information Technology, Beijing Jiaotong University, Beijing 100044, China
*
Author to whom correspondence should be addressed.
Sustainability 2023, 15(2), 1184; https://doi.org/10.3390/su15021184
Submission received: 30 November 2022 / Revised: 25 December 2022 / Accepted: 4 January 2023 / Published: 9 January 2023

Abstract

:
Automatic obstacle detection is of great significance for improving the safety of train operation. However, the existing autonomous operation of trains mainly depends on the signaling control system and lacks the extra equipment to perceive the environment. To further enhance the efficiency and safety of the widely deployed fully automatic operation (FAO) systems of the train, this study proposes an intelligent obstacle detection system based on deep learning. It collects perceptual information from industrial cameras and light detection and ranging (LiDAR), and mainly implements the functionality including rail region detection, obstacle detection, and visual–LiDAR fusion. Specifically, the first two parts adopt deep convolutional neural network (CNN) algorithms for semantic segmentation and object detection to pixel-wisely identify the rail track area ahead and detect the potential obstacles on the rail track, respectively. The visual–LiDAR fusion part integrates the visual data with the LiDAR data to achieve environmental perception for all weather conditions. It can also determine the geometric relationship between the rail track and obstacles to decide whether to trigger a warning alarm. Experimental results show that the system proposed in this study has strong performance and robustness. The system perception rate (precision) is 99.994% and the recall rate reaches 100%. The system, applied to the metro Hong Kong Tsuen Wan line, effectively improves the safety of urban rail train operation.

1. Introduction

With the rapid development of rail transit, fully automated operations systems step into a stage of rapid development [1,2]. The standard “Railway applications Automated urban guided transport Safety requirements Part I: General” (GB/T 32588.1-2016) [3] issued in 2016 clearly states that in driverless train operation (DTO) and unattended train operation (UTO) operation modes, there will be no operating staff in the front-end cabin of the train. The control system is responsible for performing rail area detection and preventing the risk of collision with obstacles within the line boundaries [4,5]. Therefore, the operational safety of fully automatic operation (FAO) systems requires the support of the automatic environmental perception of trains.
The obstacles which may appear on the rail lines, such as pedestrians, falling rocks, and intruders, would break the safety conditions of rail line closure. Thus, detecting these obstacles and reducing the probability of collision accidents are the goals of the intelligent train detection system [6].
With the maturity of urban rail transit technology, intelligent train detection systems have become an important development direction. The intelligent train detection system covers modern data communication, automatic control, highly reliable positioning, and other advanced technologies, and represents the highest level of rail transit modernization. First, the intelligent train detection system is not affected by the subjective factors of drivers; it is of great significance to improve the safety of vehicle operation [7,8,9]. Second, compared with the traditional manual driving train, the intelligent train detection system will surely improve the operation efficiency. It has strong flexibility and adaptability in the supply and configuration of transportation services, which effectively ensures the punctuality and comfort of train operation. It also helps to enhance the service quality of the transportation system.
At present, there are many schemes of intelligent detection systems applied in the traffic field. These mainly include the detection system scheme based on computer vision, radar (millimeter-wave radar), and visual–LiDAR (Light Detection and Ranging) fusion. The detection scheme based on computer vision collects the images from the moving direction of the train through cameras [10]. This method only requires the configuration of the camera sensor and does not require complicated calibration [11,12]. This solution is economical, simple to deploy, and can be employed in a dark environment. However, the camera is greatly affected by the external environment (such as light, rain, and snow). The detection scheme based on radar mainly transmits and recovers the signal through the radar and compares the received signal frequency with its frequency to obtain the relative distance and speed between the vehicle and the obstacle [13]. Radar has the characteristics of small volume, light weight, and high spatial resolution. Meanwhile, the millimeter wave guide seeker has a strong ability to penetrate fog, smoke, and dust. However, this solution requires the deployment of equipment beside the railway, which increases the difficulty of implementation and maintenance. The detection scheme based on visual–LiDAR fusion integrates the information of visual and LiDAR [14]. It can make up for the deficiencies of a single vision or LiDAR detection schemes. The scheme is less affected by environmental factors, so it has high reliability.
In this paper, the key methods of the intelligent train detection system are studied. We integrate visual and LiDAR information, which realizes pixel-wise identification of the rail track area and obstacles in front of the running train. The FAO system can leverage the perceptual information to warn the control center or directly output a brake. It ensures that the train can accurately identify the obstacles ahead and enhances the safety of train operation.
The general contributions of this work are as follows:
(1)
The convolutional neural networks are leveraged to realize the computer vision tasks under the urban railway scenario, i.e., rail track area detection and obstacle detection;
(2)
A visual and LiDAR information fusion method is proposed. It achieves a better performance for obstacle detection by implementing the two aforementioned tasks in real-time;
(3)
A method to judge the relationship between rail track and obstacles is proposed to improve the accuracy of early warning.

2. Literature Review

2.1. Rail Track Region Detection

The existing rail track region detection algorithms are mainly divided into two categories. One is to extract the rail region by feature extraction operator, the other is to detect the rail region based on deep learning method.
Qi et al. [15] propose to use gradient direction histogram features to build integral images and use the area growth algorithm to detect the rail area. Based on the prior knowledge, Wu et al. [16] preliminarily determine the characteristics of the rail in the lower part of the image and then detect the rail area through the angle calibration measurement algorithm. This algorithm shows strong robustness to both color and light. Researchers [17,18] used a multi-threshold algorithm to segment rail tracks in a complex environment. They used the pixel tracking algorithm to extract the feature points of the rail track and select the appropriate curve model to construct the rail equation.
However, the method based on the feature extraction operator is relatively fixed to a specific scenario; it is difficult to apply one method in all railway environments. This is because the method based on the feature extraction operator is usually rule-based, while in the actual environment, the rail track is changeable, and the single rule is insufficient to describe the rail track in detail. Furthermore, with the rapid development of deep learning in recent years [19,20], the convolutional neural network has been applied in many fields [21,22]. We find that the convolutional neural network has a strong feature extraction capability for the whole image. Segmentation networks such as SegNet [23] and ERFNet [24] show good performance in image segmentation. Chaurasia et al. [25] propose the ENet network structure by reducing the size of the feature map in advance and reducing the size of the decoder, which improved the running speed of the model. ROMERA et al. [24,26] designed the ERFNet network to improve the segmentation accuracy while keeping the model with low computational complexity. However, these methods are aimed at semantic segmentation of automotive autonomous driving, few studies focus on the semantic segmentation of rail track. Therefore, this study focuses on the semantic segmentation of rail track based on deep learning.

2.2. Obstacle Detection

In the research of obstacle detection technology, one may leverage the efficient existing methods based on deep learning models for object detection.
Object detectors based on CNN, such as Faster RCnn [22], SSD [27], and YOLO [28], can recognize and locate objects in images. Faster RCnn generates a large number of bounding boxes and then performs image recognition on each of them. YOLO directly predicts category scores and box offsets in multiple categories within one step, with a fixed set of default bounding boxes of different sizes at each position of the feature map with different shapes. SSD is a combination of Faster RCNN and YOLO; the regressor-based model is adopted to directly return the category and location of objects in the network, and the region-based concept is also used. In the process of detection, many candidate regions are used as ROI.
Based on the Caffe framework, Garnett [29] uses Google net as the backbone network and SSD and StixelNet as two network branches for pedestrian position detection and conventional obstacle detection, respectively. The networks use the labels generated by LiDAR for training and realize the detection and classification of three-dimensional information of objects through the network.
Aycard et al. [30] use the original data of LiDAR and stereo vision system as input to detect the surrounding obstacles in the visual sensing system and LiDAR sensing system, respectively, and obtain the list of moving objects. Premebida et al. [31] propose a pedestrian and vehicle detection system using cameras and LiDAR. The system uses a traditional Adaboost classifier and GMM classifier to classify ROI (region of interest) data and obtain detection results; however, the performance of the classifier has limited accuracy. Fan et al. [32] applied a 3D point cloud camera for object detection and distance measurement, but the algorithm is complex and demonstrates low real-time performance and the equipment is expensive. Garcia et al. [33] designed a detection architecture in which the vehicle is equipped with visual–LiDAR and LiDAR for demanding different distances. The method combines detection-level fusion with track-level fusion which improves the robustness and detection accuracy.
Current detection methods can only detect common obstacles within the detection range of sensors. However, the railway scene often involves unpredictable obstacles. At the same time, the current methods lack the understanding of the railway scene. Only the obstacles within the rail track limit will affect the operation of the current train, and the obstacles within the near rail track limit do not need the warning. Aiming at these shortcomings, in this paper, we propose an intelligent train detection system. The proposed system not only detects the obstacle, but also detects the rail track. The results of obstacle detection and rail track detection are further combined in order to judge whether the detected obstacle affects the train running.

2.3. Visual–LiDAR Fusion

The algorithms of visual–LiDAR fusion can be divided into pre-fusion and post-fusion:
(1)
Pre-fusion includes data fusion and feature fusion. The main method is to combine the features of the original image data and point cloud data, and then detect and locate them. Rovid et al. [34] proposed a neural network method to fuse the image and LIDAR point cloud to form a dense point cloud and estimate the position of the target based on the point cloud. However, the pre-fusion method consumes a lot of resources and is insufficient to realize the on-board real-time application;
(2)
Post-fusion is also called decision fusion. Firstly, the data obtained by the two sensors are detected, and the target detection results are fused and located. Sualeh et al. [35] first use the network to detect the image and point cloud, respectively, and find the two corners and the third corner of the diagonal of the target detection bounding box according to the three-dimensional point cloud so as to update the shape and position of the target.
Most of the existing rail track obstacle detection method relies on a single sensor; few of them are based on multi-sensor fusion. To detect the objects in front of the train quickly and obtain the shape and position of the target accurately. We use post-fusion technology to detect the shape and position of the obstacles.

3. Method

This section first introduces the workflow of the proposed method, and then each functional part will be introduced in detail.

3.1. Workflow

The main function of the intelligent train detection system is to detect the obstacles in the rail track and give early warning. The algorithm needs to provide information about the number, type, and distance of obstacles. To realize the sensing function, the intelligent train detection system in this paper includes the following equipment as the input: one short focal camera is used to detect near obstacles, and one long focal camera is used to detect long-range obstacles. Besides, a LiDAR is used to measure the distance of the obstacle.
The main components of the intelligent train detection system consist of rail region identification, obstacle detection, vision and LiDAR fusion, and spatial relationship judgment. The rail region identification module detects the train running area and provides rail region and rail track direction information; the obstacle detection module detects common obstacles in rail transit, such as trains and pedestrians, and provides location and classification information; the vision and LiDAR Fusion module detects the unknown obstacles in the rail region (moreover, it can output the distance information of obstacles which are detected by the obstacle detection module); the spatial relationship judgment module can judge whether the obstacle needs warning according to the rail track direction and the position of the obstacle. The workflow of the intelligent train detection system is shown in Figure 1.

3.2. Rail Region Identification

One core function of the intelligent train detection system is rail track identification, which is to extract the rail area from the input image. Accurate and reliable rail region identification can not only provide accurate rail limits [36,37], but also accurate regions of interest [38,39] for obstacle detection. This makes obstacle detection more accurate and efficient. However, the traditional image processing algorithm is difficult to adapt to all scenes because of the complex environment of the rail region in the operation environment and the interlacing of various scenes such as curves and straights. Besides, in recent years, artificial intelligence, especially deep learning, has achieved rapid development and has been widely used in various fields. We find that the semantic segmentation algorithm has a strong ability to extract image features. Therefore we try to apply semantic segmentation to rail region recognition which mainly include a cascading downsampling layer (Cascading downsampling layer: Downsampling is performed by convolution and pooling), a convolution layer, a dilated convolution layer (Dilated convolution layer: Convolution with skip connections), a deconvolution layer (Deconvolution layer: The inverse operation of convolution which increases the resolution of the feature map), and a Softmax layer (Softmax layer: Used to achieve pixel-level classification). The image segment network structure is shown in Figure 2.
The rail region recognition network mainly consists of the encoding section and decoding section. The encoding section is mainly used to analyze and extract the features of the rail region, which is composed of the convolution layer and the pooling layer. In addition, the encoding section also contains a dilated convolution layer. The convolution layer is mainly used to extract the rail region features of each layer. The pooling layer is mainly used to reduce the resolution of the image. On the one hand, the subsequent convolutional layer can have a larger receptive field. On the other hand, the computational amount of subsequent convolutional operations can be reduced to ensure the real-time performance of the algorithm. Dilated convolution extracts the features of different scales by applying different dilated convolution coefficients. These features are combined to achieve feature extraction of different scales without changing the size of the image feature map. The decoding section is followed by the encoding section. The decoding section mainly consists of a series of upsampling layers. The semantic information of the features extracted from the encoding section is restored to the original size to realize the pixel-level image understanding. In the last part of the network is the Softmax layer, which is mainly used to classify each pixel in the image to achieve pixel-wise rail track segmentation. The specific structure of the rail region detection network is shown in Table 1.
The segmentation results are shown in Figure 3. Green pixels are the track area.

3.3. Obstacle Detection

In obstacle detection part, the system uses a convolutional neural network to detect obstacles in the rail region. Depthwise separable convolution and residual network structure are used. We combine the output of the multi-layer feature layer to implement the detection, which includes bounding box regression and obstacle classification.
Considering the real-time requirements of the algorithm for practical train applications, the forward obstacle detection network was designed based on the fast deep learning objects detection network SSD [27]. The network includes three main parts: the first part is the feature extraction network, which extracts the features of the input rail environment image through multi-layer neural networks; the second part is the extra convolution layer, which aims to obtain the feature maps with a smaller scale; the third part is the network prediction layer, which is used to predict the target position of the forward obstacle. The obstacle detection network structure is shown in Figure 4. As the main part of the train forward obstacle detection framework, the feature extraction network consumes most of the computing resources of the network. At the same time, the feature extraction of the rail environment image is also determined by this part. In the forward obstacle detection network, the residual connection is used in the design of the network [40]. The residual connection can eliminate the gradient vanishing. Meanwhile, depthwise separable convolution is applied in the forward obstacle detection network [20]. Depthwise separable convolution can reduce the computation of the network while maintaining the accuracy of the network. In addition, multiscale feature map fusion is applied in the design of a forward obstacle detection network. Multi-scale feature fusion combines feature maps with different resolutions so that the rich detail information in the lower layer of the network can be fused with the rich feature information in the upper layer. Therefore, the network has a stronger feature extraction ability.
The goal of setting extra layers in the train forward obstacle detection framework is to further compress the features extracted by the feature extraction layer to obtain more global features. Thus, it can be used for the subsequent detection of large-size obstacles.
After completing the design of the extra layer, the network prediction layer should be designed. The network prediction layer selects several layers from different feature graphs for decision-making. Each pixel in the feature map of the prediction layer can map its position to the original graph and corresponds to an obstacle detection bounding box of fixed size in the feature map. To detect obstacles of different sizes, six sizes of feature maps are selected for prediction. The category of the target bounding box at each position and the offset of the initial bounding box were obtained. The final size and position of the obstacle detection bounding box were obtained by combining the location information of the initial target bounding box set.
The a priori bounding box used to predict trains gradually increases from low-level feature maps to high-level feature maps. The formula used to predict the a priori bounding box of layer k is shown in Equation (1):
S k = S min + S max S min m 1 ( k 1 ) ,   k [ 1 , m ]
where Sk is the predicted bounding box the size of which is between the Smin and the Smax, Smin is the minimum scale for prediction, Smax is the maximum scale for prediction, and m is the number of feature maps for prediction. In addition, we added three additional layers at the end of the train identification network to improve the accuracy of train detection. The specific structure of the additional layers is shown in Table 2.
The obstacle detection effect is shown from Figure 5, Figure 6 and Figure 7. It can be seen that the proposed method has a good effect on a variety of scenarios such as daylight, night, tunnels, and elevated tracks.

3.4. Visual–LiDAR Fusion

The intelligent train detection system further combines the information of images and LiDAR sensors to achieve a better understanding of the environment. Firstly, the system unifies the visual and LiDAR data through combining the spatiotemporal information. Multi-level vision and LiDAR fusion algorithms are used to obtain the distance and class information of obstacles. Combined with prior knowledge, the results are checked again to obtain the final perceptual outputs.
Single visual information is difficult to achieve accurate obstacle distance measurement. In addition, visual information is susceptible to changes in light and other factors. In addition, a single LiDAR is difficult to determine the type of obstacles, and it cannot accurately identify the boundaries of the tracks. The intelligent train detection system integrates visual information with LiDAR information to realize accurate all-weather obstacle perception.
(1)
Multi-sensor spatiotemporal fusion
The visual module of the intelligent train detection system is mainly used for rail area and train recognition. At the same time, the LiDAR module preprocesses the LiDAR point cloud data and filters out the data with a large height of the point cloud. After the preprocessing of point cloud data, the spatial LiDAR points are mapped to the image through data association. Thus, sparse image depth information can be obtained.
(2)
Abnormal obstacle screening
In the actual train operation environment, there are some objects which rarely appear, but the occurrence of them will have an impact on the train operation. Therefore, we need to detect these abnormal obstacles. “Abnormal obstacle” refers to an object with a small probability. Due to insufficient samples, these objects are difficult to be learned by the deep learning method. After obtaining the depth information of the image, the rail region LiDAR points are extracted by mapping of the rail region image. The rail surface fitting of the LiDAR point cloud is carried out by the RANSAC [41] algorithm to obtain the space equation of the rail surface. Based on the space equation of the rail surface, the point clouds above the rail plane are clustered to identify the obstacles within the rail boundaries.
(3)
Distance measurement
According to the mapping relationship between the image and LiDAR, the LiDAR point corresponding to the obstacle detected by the image is determined. The average distance of the LiDAR point set to the obstacle is calculated as the distance of the obstacle. The fusion effect of the image and LiDAR is shown in Figure 8.

3.5. Spatial Relationship Judgment

During the actual train operation, only the obstacles on the train track will affect the train operation. On the contrary, the obstacles outside the rail track do not affect the train operation. For the intelligent train detection system, only the obstacles that have an impact on the normal operation of the train ahead need to be warned. The current algorithm only detects the obstacle and rail region but does not judge the relationship between the obstacle and the current train. Therefore, it is necessary to judge the relationship between the current rail and obstacles. The workflow of spatial relation judgment is shown in Figure 9.
Step1: Use the chain code tracking algorithm to extract the contours of the rail region [42] to obtain the outer contour of the rail region;
Step2: Select the set of key points on the left and right sides of the contour of the rail region as the left and right boundaries;
Step3: Determine whether there is a turnout core point according to the ordinate change of the outer contour set. If the ordinate of the outer contour set decreases and then increases, this point is the turnout core point. If there are turnout cores, proceed to 3.1; otherwise, proceed to 3.2;
Step3.1: Use the quadratic function to fit the left and right contour sets, respectively, and get two rail curves;
Step3.2. Use the quadratic function to fit the midpoints of the left and right contours to obtain the rail center curve;
Step4: If the rail curve intersects the lower box line of the obstacle bounding box, the obstacle is considered to be on the current rail track; otherwise, the obstacle is on the adjacent rail track.

4. Experiment Results

The experiment was conducted on Hong Kong Tsuen Wan line. There are 16 stations along the whole of the Tsuen Wan Line, with a total length of about 16 km. Among them, 4 stations of Tsuen Wan Station, Kui Hing Station, Kui Fong station, and Li Jing Station are above-ground stations, and the other 12 stations are underground. The operation line includes the ground section and tunnel section. The smallest radius of the line is 310 m and the maximum grade of the line is 3%. It is a typical rail transit line, which can cover most scenes encountered by train operation.
To evaluate the effectiveness of the proposed method, the test lasted nine months and involved the installation of Train Intelligent Detection System (TIDS) equipment on five trains of the Tsuen Wan Line as well as the corresponding on-time running tests, with a cumulative mileage of more than 200,000 km and a total of more than 50,000 km of data collected after installation. The video with resolution of 1280 × 720 are collected. All experiments were trained based on Titan Xp GPU, Pytorch 1.1.0, CUDA 9.0, and CUDNN 7.1.
For a straight track, TIDS can achieve more than 240 m detect range. The image detection effect is shown in Figure 10. For a grade 3% maximum ramp, TIDS can achieve no less than 100 m detection. The detection effect is shown in Figure 11.
In the case of the sharpest curve of the Tsuen Wan Line (R = 312 m), the system can achieve a detection distance of no less than 70 m. The image detection effect is shown in Figure 12.
We sampled more than 100 h of data for statistical analysis of false alarm rate (FAR), miss detection rate (MDR), perception, and recall of train obstacle detection system results. The results of this analysis are shown in Table 3.
Table 4 shows the analysis results and the detection indicators of this system. The false alarm rate is 0.008% and the miss detection rate is 0. It shows that our method has higher precision. Although there are still some false alarms in our methods at this stage, as a perception system, false alarms do not cause security problems. On the contrary, it is miss detection that can cause the safety problems, and the miss detection rate is 0 in this study. This shows that our method can meet the requirements of practical application.
At the same time, to prove the effectiveness of the fusion method proposed in this paper, we used vision and LiDAR for train detection, respectively. For the proposed fusion method, we can achieve accurate forward obstacle detection and distance measurement. The forward obstacle detection abilities of different methods are shown in Table 5.
The “×” in Table 5 indicates that the obstacle cannot be detected. The results show that our fusion method can achieve better obstacle detection than the single sensor method. This is because we know the pixel of the forward obstacle from the camera and its corresponding point from LiDAR; thus, we can achieve accurate forward obstacle detection and distance measurement.

5. Conclusions

In this paper, an intelligent train detection system is proposed and applied to the urban rail scenario. The application of different focal cameras improves the perceived range. The fusion of LiDAR supplements the distance information which is missing from the image and realizes the detection of unknown small obstacles. In addition, the proposed system makes full use of existing deep learning-based object detection and semantic segmentation algorithms to extract regions of interest based on subway scene features. Finally, the spatial judgment algorithm meets the perception requirements of the rail scenario. Experiments show that the system can detect the obstacles on the rail track reliably and robustly.
As the active obstacle detection technology is increasingly widely accepted by the market, the key technologies and implementation methods of the TIDS system must be paid more attention. In the future, this technology will develop in the direction of high RAMS. When the FAO line is operating without a driver on duty, the highly reliable SIL4 TIDS will serve as the “eyes” of the train to further detect foreign objects intruding into the boundary in real time and report to the signal system [43,44]. The existing methods mainly focus on on-board obstacle detection; however, the braking distance of the train is too long, and the autonomous perceived distance of the train cannot meet the braking requirements in curves and other scenarios. Therefore, in the future, it is necessary to further develop vehicle–road collaborative perception and improve obstacle detection distance. In addition, a convolutional neural network is used in this study; neural networks have uncertainty, especially in a dynamic environment [45,46]. In future studies, we will estimate the uncertainty to improve the interpretability of the model.

Author Contributions

Conceptualization, Q.Z. and F.Y.; methodology, Q.Z.; software, R.W.; validation, W.S.; formal analysis, R.W.; investigation, R.W.; resources, G.L.; data curation, F.Y.; writing—original draft preparation, Q.Z.; writing—review and editing, Q.Z.; visualization, G.L.; supervision, F.Y.; project administration, F.Y.; funding acquisition, F.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Guangxi Key Research and Development Program, grant number AB22035008 and the Basic Research Funds for Central Universities, grant number 2022JBZY024.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data is not available for commercial reasons at this stage.

Acknowledgments

The authors would like to thank the insightful and constructive comments from anonymous reviewers.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhu, L.; Li, Y.; Yu, F.R.; Ning, B.; Tang, T.; Wang, X. Cross-layer defense methods for jamming-resistant CBTC systems. IEEE Trans. Intell. Transp. Syst. 2020, 22, 7266–7278. [Google Scholar] [CrossRef]
  2. Feng, J.; Liu, L.; Pei, Q.; Li, K. Min-max cost optimization for efficient hierarchical federated learning in wireless edge networks. IEEE Trans. Parallel Distrib. Syst. 2021, 33, 2687–2700. [Google Scholar] [CrossRef]
  3. BS EN 62267:2009; Railway Applications. Automated Urban Guided Transport (AUGT). Safety Requirements. British Standards Institution: London, UK, 2009.
  4. Li, Y.; Zhu, L.; Wang, H.; Yu, F.R.; Liu, S. A cross-layer defense scheme for edge intelligence-enabled CBTC systems against MitM attacks. IEEE Trans. Intell. Transp. Syst. 2020, 22, 2286–2298. [Google Scholar] [CrossRef]
  5. Liu, L.; Feng, J.; Pei, Q.; Chen, C.; Ming, Y.; Shang, B.; Dong, M. Blockchain-enabled secure data sharing scheme in mobile-edge computing: An asynchronous advantage actor–critic learning approach. IEEE Internet Things J. 2020, 8, 2342–2353. [Google Scholar] [CrossRef]
  6. Zhang, J. Technical analysis of automatic train obstacle detection device for urban rail transit. Tunn. Rail Transit 2021, S01, 4. [Google Scholar]
  7. Yang, Z.; Cheung, V.; Gao, C.; Zhang, Q. Train Intelligent Detection System based on Convolutional Neural Network. In Proceedings of the International Conference on Applied Human Factors and Ergonomics, Washington, DC, USA, 24–28 July 2019; pp. 157–165. [Google Scholar]
  8. Wang, Y.; Cong, J.; Wang, P.; Liu, X.; Tang, H. A Data-Fusion Approach for Speed Estimation and Location Calibration of a Metro Train Based on LowCost Sensors in Smartphones. IEEE Sens. J. 2019, 19, 10744–10752. [Google Scholar] [CrossRef]
  9. Dong, H.; Ning, B.; Cai, B.; Hou, Z. Automatic train control system development and simulation for high-speed railways. IEEE Circuits Syst. Mag. 2010, 10, 6–18. [Google Scholar] [CrossRef]
  10. Yan, L. Research on Urban Rail Transit Green Energy Electric Intelligent Unmanned Technology. Journal of Physics: Conference Series. 2019, 1168, 032117. [Google Scholar] [CrossRef]
  11. Lüy, M.; Çam, E.; Ulamış, F.; Uzun, I.; Akın, S.İ. Initial results of testing a multilayer laser scanner in a collision avoidance system for light rail vehicles. Appl. Sci. 2018, 8, 475. [Google Scholar] [CrossRef] [Green Version]
  12. Komalachitra, S.; Bhargavi, S.; Devika, R.; Kumuthavalli, S.; Narayani, M.S. Railway obstacle detection and power storage using image processing and hybrid technology. In Proceedings of the 2017 IEEE International Conference on Electrical, Instrumentation and Communication Engineering (ICEICE), Karur, India, 27–28 April 2017; pp. 1–3. [Google Scholar]
  13. Wu, R.; Jin, Z.; Zhong, T.; Dai, H. Vehicle Ranging System Based on Millimeter Wave Radar. Automob. Pract. Technol. 2019, 2, 33–35. [Google Scholar]
  14. Zhang, Q.; Yang, F.; Zhang, B. Research on The Application of Intelligent Train Obstacle Detection System in The Automatic Operation of Beijing New Airport Line. Railw. Locomot. Veh. 2019, 39, 114–118. [Google Scholar]
  15. Qi, Z.; Tian, Y.; Shi, Y. Efficient railway tracks detection and turnouts recognition method using HOG features. Neural Comput. Appl. 2013, 23, 245–254. [Google Scholar] [CrossRef]
  16. Wu, H.; Siu, W.C. Real time railway extraction by angle alignment measure. In Proceedings of the IEEE International Conference on Image Processing, Quebec City, QC, Canada, 27–30 September 2015. [Google Scholar]
  17. Xiang, Q.; Zhang, Z.; Chen, Q.; Wu, C.; Wang, Y. Video-based adaptive railway recognition in complex scene. In Proceedings of the 2016 International Conference on Audio, Language and Image Processing (ICALIP), Shanghai, China, 11–12 July 2016. [Google Scholar]
  18. Qiang, X. Research on the Boundary Recognition and Real-Time Dynamic Tracking System of Modern Trams; Soochow University: Taipei, Taiwan, 2016. [Google Scholar]
  19. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef] [Green Version]
  20. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  21. Li, X.; Liu, Z.; Luo, P.; Change, L.C.; Tang, X. Not all pixels are equal: Difficulty-aware semantic segmentation via deep layer cascade. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3193–3202. [Google Scholar]
  22. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 28, 91–99. [Google Scholar] [CrossRef] [Green Version]
  23. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
  24. Romera, E.; Alvarez, J.M.; Bergasa, L.M.; Arroyo, R. ERFNet: Efficient Residual Factorized ConvNet for Real-Time Semantic Segmentation. IEEE Trans. Intell. Transp. Syst. 2017, 1, 1–10. [Google Scholar] [CrossRef]
  25. Paszke, A.; Chaurasia, A.; Kim, S.; Culurciello, E. Enet: A deep neural network architecture for real-time semantic segmentation. arXiv 2016, arXiv:1606.02147. [Google Scholar]
  26. Shi, Y. Research on Road Scene Semantic Segmentation Based on Deep Neural Network; Southwest Jiaotong University: Chengdu, China, 2018. [Google Scholar]
  27. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar]
  28. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  29. Garnett, N.; Silberstein, S.; Oron, S.; Fetaya, E.; Verner, U.; Ayash, A.; Goldner, V.; Cohen, R.; Horn, K.; Levi, D. Real-time categorybased and general obstacle detection for autonomous driving. In Proceedings of the 2017 IEEE International Conference on Computer Vision Workshops, Venice, Italy, 22–29 October 2017; pp. 198–205. [Google Scholar]
  30. Aycard, O.; Baig, Q.; Bota, S.; Nashashibi, F.; Nedevschi, S.; Pantilie, C.; Parent, M.; Resende, P.; Vu, T.-D. Intersection safety using LiDAR and stereo vision sensors. In Proceedings of the 2011 IEEE Intelligent Vehicles Symposium (IV), Baden-Baden, Germany, 5–9 June 2011; pp. 863–869. [Google Scholar]
  31. Premebida, C.; Monteiro, G.; Nunes, U.; Peixoto, P. A LiDAR and vision-based approach for pedestrian and vehicle detection and tracking. In Proceedings of the 2007 IEEE Intelligent Transportation Systems Conference, Bellevue, WA, USA, 30 September–3 October 2007; pp. 1044–1049. [Google Scholar]
  32. Fan, Z.; Liu, H.; He, J.; Jiang, S.; Du, X. PointFPN: A Frustum-based Feature Pyramid Network for 3D Object Detection. In Proceedings of the 2020 IEEE 32nd International Conference on Tools with Artificial Intelligence (ICTAI), Baltimore, MD, USA, 9–11 November 2020; pp. 1129–1136. [Google Scholar]
  33. Chavez-Garcia, R.O.; Aycard, O. Multiple sensor fusion and classification for moving object detection and tracking. IEEE Trans. Intell. Transp. Syst. 2015, 17, 525–534. [Google Scholar] [CrossRef] [Green Version]
  34. Rovid, A.; Remeli, V.; Szalay, Z. Raw fusion of camera and sparse LiDAR for detecting distant objects. Automatisierungstechnik 2020, 68, 337–346. [Google Scholar] [CrossRef]
  35. Sualeh, M.; Kim, G.W. Visual-LiDAR Based 3D Object Detection and Tracking for Embedded Systems. IEEE Access 2020, 8, 156285–156298. [Google Scholar] [CrossRef]
  36. Nassu, B.T.; Ukai, M. A vision-based approach for rail extraction and its application in a camera pan–tilt control system. IEEE Trans. Intell. Transp. Syst. 2012, 13, 1763–1771. [Google Scholar] [CrossRef]
  37. Ruder, M.; Mohler, N.; Ahmed, F. An obstacle detection system for automated trains. In Proceedings of the IEEE IV 2003 Intelligent Vehicles Symposium, Columbus, OH, USA, 9–11 June 2003; pp. 180–185. [Google Scholar]
  38. Selver, M.A.; Zoral, E.Y.; Belenlioglu, B.; Dogan, S. Predictive modeling for monocular vision based rail track extraction. In Proceedings of the 2017 10th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI), Shanghai, China, 14–16 October 2017; pp. 1–6. [Google Scholar]
  39. Wang, Z.; Wu, X.; Yu, G.; Li, M. Efficient Rail Area Detection Using Convolutional Neural Network. IEEE Access 2018, 6, 77656–77664. [Google Scholar] [CrossRef]
  40. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–2 July 2016; pp. 770–778. [Google Scholar]
  41. Fischler, M.A.; Bolles, R.C. Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
  42. Freeman, H. Boundary encoding and processing. Pict. Process. Psychopictorics 1970, 241–266. [Google Scholar]
  43. Zhu, L.; Liang, H.; Wang, H.; Ning, B.; Tang, T. Joint Security and Train Control Design in Blockchain-Empowered CBTC System. IEEE Internet Things J. 2021, 9, 8119–8129. [Google Scholar] [CrossRef]
  44. Feng, J.; Zhang, W.; Pei, Q.; Wu, J.; Lin, X. Heterogeneous Computation and Resource Allocation for Wireless Powered Federated Edge Learning Systems. IEEE Trans. Commun. 2022, 70, 3220–3233. [Google Scholar] [CrossRef]
  45. Tutsoy, O.; Barkana, D.E.; Tugal, H. Design of a completely model free adaptive control in the presence of parametric, non-parametric uncertainties and random control signal delay. ISA Trans. 2018, 76, 67–77. [Google Scholar] [CrossRef]
  46. Tutsoy, O. CPG based RL algorithm learns to control of a humanoid robot leg. Int. J. Robot. Autom. 2015, 30, 1–7. [Google Scholar] [CrossRef]
Figure 1. The workflow of the intelligent train detection system.
Figure 1. The workflow of the intelligent train detection system.
Sustainability 15 01184 g001
Figure 2. Architecture of image segment network.
Figure 2. Architecture of image segment network.
Sustainability 15 01184 g002
Figure 3. Rail area segmentation results.
Figure 3. Rail area segmentation results.
Sustainability 15 01184 g003
Figure 4. Image detection network structure.
Figure 4. Image detection network structure.
Sustainability 15 01184 g004
Figure 5. Detection results of straight road, curves and multi-turnout areas.
Figure 5. Detection results of straight road, curves and multi-turnout areas.
Sustainability 15 01184 g005
Figure 6. Detection results of night and tunnel conditions.
Figure 6. Detection results of night and tunnel conditions.
Sustainability 15 01184 g006
Figure 7. Detection results of bad weather.
Figure 7. Detection results of bad weather.
Sustainability 15 01184 g007
Figure 8. Illustration of image and LiDAR fusion.
Figure 8. Illustration of image and LiDAR fusion.
Sustainability 15 01184 g008
Figure 9. Workflow of spatial relation judgment.
Figure 9. Workflow of spatial relation judgment.
Sustainability 15 01184 g009
Figure 10. Straight track detection.
Figure 10. Straight track detection.
Sustainability 15 01184 g010
Figure 11. Ramp detection.
Figure 11. Ramp detection.
Sustainability 15 01184 g011
Figure 12. Curve (R = 312 m), maximum visible distance 73 m.
Figure 12. Curve (R = 312 m), maximum visible distance 73 m.
Sustainability 15 01184 g012
Table 1. Rail region detection network.
Table 1. Rail region detection network.
LevelTypeOutputFeature Map Size
1Cascading downsampling16240 × 180
2–3Convolution64240 × 180
4Cascading downsampling 128120 × 90
5–6Convolution64120 × 90
7Cascading downsampling 25660 × 45
8–9Convolution25660 × 45
10–11Dilated convolution (rate = 2)25660 × 45
12–13Dilated convolution (rate = 4)25660 × 45
14Deconvolution128120 × 90
15–16Convolution128120 × 90
17Deconvolution64240 × 180
18–19Convolution64240 × 180
20Deconvolution16480 × 360
21–22Convolution16480 × 360
23Softmax2480 × 360
Table 2. Structure of additional layers.
Table 2. Structure of additional layers.
LevelConvolution KernelOutput
Conv6_11 × 1 × 1287 × 7 × 128
Conv6_23 × 3 × 2564 × 4 × 256
Conv7_11 × 1 × 1284 × 4 × 128
Conv7_23 × 3 × 2562 × 2 × 256
Conv8_11 × 1 × 1282 × 2 × 128
Conv8_22 × 2 × 2561 × 1 × 256
Table 3. Analysis of experimental results of intelligent train detection system.
Table 3. Analysis of experimental results of intelligent train detection system.
The System TimeResults
Effective running seconds (s) 363,268
False alarm time(s) 30
Miss detection time(s) 0
Table 4. Indicator of intelligent train detection system.
Table 4. Indicator of intelligent train detection system.
IndicatorsResults
False alarm rate0.008%
Miss detection rate0
Table 5. Forward obstacle detection ability.
Table 5. Forward obstacle detection ability.
Obstacle Distance (m)Metrical Error
Vision MethodLiDAR MethodFusion Method
60−4%−1.83%−0.5%
802.75%1%0.75%
1003.7%×1.1%
1203.5%×−0.25%
1405.5%×1.57%
1605.1%×2.69%
1806.28%×−1.44%
2006.8%×1.45%
2209.82%×−1.4%
2407.59%×1.46%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, Q.; Yan, F.; Song, W.; Wang, R.; Li, G. Automatic Obstacle Detection Method for the Train Based on Deep Learning. Sustainability 2023, 15, 1184. https://doi.org/10.3390/su15021184

AMA Style

Zhang Q, Yan F, Song W, Wang R, Li G. Automatic Obstacle Detection Method for the Train Based on Deep Learning. Sustainability. 2023; 15(2):1184. https://doi.org/10.3390/su15021184

Chicago/Turabian Style

Zhang, Qiang, Fei Yan, Weina Song, Rui Wang, and Gen Li. 2023. "Automatic Obstacle Detection Method for the Train Based on Deep Learning" Sustainability 15, no. 2: 1184. https://doi.org/10.3390/su15021184

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop