Abstract
This paper addresses the challenge of reading low contrast text on tyre sidewall images of vehicles in motion. It presents first of its kind, a full scale industrial system which can read tyre codes when installed along driveways such as at gas stations or parking lots with vehicles driving under 10 mph. Tyre circularity is first detected using a circular Hough transform with dynamic radius detection. The detected tyre arches are then unwarped into rectangular patches. A cascade of convolutional neural network (CNN) classifiers is then applied for text recognition. Firstly, a novel proposal generator for the code localization is introduced by integrating convolutional layers producing HOG-like (Histogram of Oriented Gradients) features into a CNN. The proposals are then filtered using a deep network. After the code is localized, character detection and recognition are carried out using two separate deep CNNs. The results (accuracy, repeatability and efficiency) are impressive and show promise for the intended application.
| Original language | English |
|---|---|
| Pages (from-to) | 1264 - 1275 |
| Number of pages | 12 |
| Journal | IEEE Transactions on Intelligent Transportation Systems |
| Volume | 22 |
| Issue number | 2 |
| Early online date | 24 Jan 2020 |
| DOIs | |
| Publication status | Published - 1 Feb 2021 |
| Event | International Conference on Automation Science and Engineering - Vancouver, Canada Duration: 22 Aug 2019 → 26 Aug 2019 Conference number: 15 http://case2019.hust.edu.cn/ |
Keywords
- Intelligent vehicles
- deep learning
- computer vision
- tyre (tire) sidewall
- Optical Character Recognition (OCR)