Skip to main content
Log in

FCN and LSTM Based Computer Vision System for Recognition of Vehicle Type, License Plate Number, and Registration Country

  • Published:
Automatic Control and Computer Sciences Aims and scope Submit manuscript

Abstract

We propose an advanced Automatic number-plate recognition (ANPR) system, which not only recognizes the number and the issuing state, but also the type and location of the vehicle in the input image. The system is based on a combination of existing methods, modifications to neural network architectures and improvements in the training process. The proposed system uses machine-learning approach and consists of three main parts: segmentation of input image by Fully Convolutional Network for localization of license plate and determination of vehicle type; recognition of the characters of the localized plate by a Maxout CNN and LSTM; determination of the state that has issued the license plate by a CNN. The training of these neural network models is accomplished using a manually labeled custom dataset, which is expanded with data augmented techniques. The resulting system is capable of localizing and classifying multiple types of vehicles (including motorcycles and emergency vehicles) as well as their license plates. The achieved precision of the localization is 99.5%. The whole number recognition accuracy is 96.7% and character level recognition accuracy is 98.8%. The determination of issuing state is precise in 92.8% cases.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. Patel, C., Shah, D., and Patel, A., Automatic number plate recognition system (ANPR): A survey, Int. J. Comput. Appl., 2013, vol. 69, no. 9, pp. 21–33.

    Google Scholar 

  2. Everingham, M., Eslami, S.A., Van Gool, L., Williams, C.K., Winn, J., and Zisserman, A., The Pascal visual object classes challenge: A retrospective, Int. J. Comput. Vision, 2015, vol. 111, no. 1, pp. 98–136.

    Google Scholar 

  3. Russakovsky, O., Deng, J., Su, H., Krause, J., Satheesh, S., Ma, S., Huang, Z., Karpathy, A., Khosla, A., Bernstein, M., Berg, A.C., and Fei-Fei, Li, ImageNet Large Scale Visual Recognition Challenge, Int. J. Comput. Vision, 2015, vol. 115, no. 3, pp. 211–252.

    Article  MathSciNet  Google Scholar 

  4. Lin, T.-Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., Dollár, P., Zitnick, C.L., Microsoft COCO: Common objects in context, European Conference on Computer Vision, 2014, pp. 740–755.

    Google Scholar 

  5. Sermanet, P., Eigen, D., Zhang, X., Mathieu, M., Fergus, R., and LeCun, Y., Overfeat: Integrated recognition, localization and detection using convolutional networks, arXiv:1312.6229, 2013.

    Google Scholar 

  6. Simonyan, K. and Zisserman, A., Very deep convolutional networks for large-scale image recognition, arXiv:1409.1556, 2014.

    Google Scholar 

  7. He, K., Zhang, X., Ren, S., and Sun, J., Deep residual learning for image recognition, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 770–778.

    Google Scholar 

  8. Ren, S., He, K., Girshick, R., and Sun, J., Faster R-CNN: Towards real-time object detection with region proposal networks, Adv. Neural Inf. Process. Syst., 2015, pp. 91–99.

    Google Scholar 

  9. Redmon, J., and Farhadi, A., YOLO9000: Better, faster, stronger, arXiv:1612.08242, 2016.

    Google Scholar 

  10. Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.Y., and Berg, A.C., SSD: Single shot multibox detector, European Conference on Computer Vision, 2016, pp. 21–37.

    Google Scholar 

  11. Grangier, D., Bottou, L., and Collobert, R., Deep convolutional networks for scene parsing, ICML 2009 Deep Learning Workshop, 2009, vol. 3.

  12. Farabet, C., Couprie, C., Najman, L., and LeCun, Y., Scene parsing with multiscale feature learning, purity trees, and optimal covers, arXiv:1202.2160, 2012.

    Google Scholar 

  13. Badrinarayanan, V., Handa, A., and Cipolla, R., Segnet: A deep convolutional encoder-decoder architecture for robust semantic pixel-wise labeling, arXiv:1505.07293, 2015.

    Google Scholar 

  14. Paszke, A., Chaurasia, A., Kim, S., and Culurciello, E., Enet: A deep neural network architecture for real-time semantic segmentation, arXiv:1606.02147, 2016.

    Google Scholar 

  15. Shelhamer, E., Long, J., and Darrel, T., Fully convolutional networks for semantic segmentation, IEEE Trans. Pattern Anal. Mach. Intell., 2017, vol. 39, no. 4, pp. 640–651.

    Article  Google Scholar 

  16. Li, H. and Shen, C., Reading car license plates using deep convolutional neural networks and LSTMs, arXiv:1601.05610, 2016.

    Google Scholar 

  17. Hannun, A., Case, C., Casper, J., Catanzaro, B., Diamos, G., Elsen, E., Prenger, R., Satheesh, S., Sengupta, S., Coates, A., and Ng, A.Y., Deep speech: Scaling up end-to-end speech recognition, arXiv:1412.5567, 2014.

    Google Scholar 

  18. Xu, K., Ba, J., Kiros, R., Cho, K., Courville, A., Salakhudinov, R., Zemel, R., and Bengio, Y., Show, attend and tell: Neural image caption generation with visual attention, International Conference on Machine Learning, 2015, pp. 2048–2057.

    Google Scholar 

  19. He, P., Huang, W., Quao, Y., Chen, C.L., and Tang, X., Reading scene text in deep convolutional sequences, Thirtieth AAAI Conference on Artificial Intelligence, 2016.

    Google Scholar 

  20. Goodfellow, I.J., Warde-Farley, D., Mirza, M., Courville, A., and Bengio, Y., Maxout networks, arXiv:1302.4389, 2013.

    Google Scholar 

  21. Jaderberg, M., Vedaldi, A., and Zisserman, A., Deep features for text spotting, European Conference on Computer Vision, 2014, pp. 512–528.

    Google Scholar 

  22. Hochreiter, S. and Schmidhuber, J., Long short-term memory, Neural Comput., 1997, vol. 9, no. 8, pp. 1735–1780.

    Article  Google Scholar 

  23. Dorbe, N., Kadikis, R., and Nesenbergs, K., Vehicle type and license plate localisation and segmentation using FCN and LSTM, New Challenges of Economic and Business Development—2017: Digital Economy, 2017, pp. 132–140.

    Google Scholar 

  24. Gers, F.A. and Schmidhuber, J., Recurrent nets that time and count, Neural Networks, 2000. IJCNN 2000, Proceedings of the IEEE-INNS-ENNS International Joint Conference on, 2000, vol. 3, pp. 189–194.

    Article  Google Scholar 

  25. Graves, A. Fernández, S., Gomez, F., and Schmidhuber, J., Connectionist temporal classification: Labelling unsegmented sequence data with recurrent neural networks, Proceedings of the 23rd International Conference on Machine Learning, 2006, pp. 369–376.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Nauris Dorbe.

Additional information

The article was translated by the authors.

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Dorbe, N., Jaundalders, A., Kadikis, R. et al. FCN and LSTM Based Computer Vision System for Recognition of Vehicle Type, License Plate Number, and Registration Country. Aut. Control Comp. Sci. 52, 146–154 (2018). https://doi.org/10.3103/S0146411618020104

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.3103/S0146411618020104

Keywords

Navigation