摘要
Purpose: This paper aims to focus on lane detection of unmanned mobile robots. For the mobile robot, it is undesirable to spend lots of time detecting the lane. So quickly detecting the lane in a complex environment such as poor illumination and shadows becomes a challenge. Design/methodology/approach: A new learning framework based on an integration of extreme learning machine (ELM) and an inception structure named multiscale ELM is proposed, making full use of the advantages that ELM has faster convergence and convolutional neural network could extract local features in different scales. The proposed architecture is divided into two main components: self-taught feature extraction by ELM with the convolution layer and bottom-up information classification based on the feature constraint. To overcome the disadvantages of poor performance under complex conditions such as shadows and illumination, this paper mainly solves four problems: local features learning: replaced the fully connected layer, the convolutional layer is used to extract local features; feature extraction in different scales: the integration of ELM and inception structure improves the parameters learning speed, but it also achieves spatial interactivity in different scales; and the validity of the training database: a method how to find a training data set is proposed. Findings: Experimental results on various data sets reveal that the proposed algorithm effectively improves performance under complex conditions. In the actual environment, experimental results tested by the robot platform named BIT-NAZA show that the proposed algorithm achieves better performance and reliability. Originality/value: This research can provide a theoretical and engineering basis for lane detection on unmanned robots.
源语言 | 英语 |
---|---|
页(从-至) | 361-371 |
页数 | 11 |
期刊 | Assembly Automation |
卷 | 42 |
期 | 3 |
DOI | |
出版状态 | 已出版 - 24 5月 2022 |