TY - JOUR
T1 - Self-supervised learning to visually detect terrain surfaces for autonomous robots operating in forested terrain
AU - Zhou, Shengyan
AU - Xi, Junqiang
AU - McDaniel, Matthew W.
AU - Nishihata, Takayuki
AU - Salesses, Phil
AU - Iagnemma, Karl
PY - 2012/3
Y1 - 2012/3
N2 - Autonomous robotic navigation in forested environments is difficult because of the highly variable appearance and geometric properties of the terrain. In most navigation systems, researchers assume a priori knowledge of the terrain appearance properties, geometric properties, or both. In forest environments, vegetation such as trees, shrubs, and bushes has appearance and geometric properties that vary with change of seasons, vegetation age, and vegetation species. In addition, in forested environments the terrain surface is often rough, sloped, and/or covered with a surface layer of grass, vegetation, or snow. The complexity of the forest environment presents difficult challenges for autonomous navigation systems. In this paper, a self-supervised sensing approach is introduced that attempts to robustly identify a drivable terrain surface for robots operating in forested terrain. The sensing system employs both LIDAR and vision sensor data. There are three main stages in the system: feature learning, feature training, and terrain prediction. In the feature learning stage, 3D range points from LIDAR are analyzed to obtain an estimate of the ground surface location. In the feature training stage, the ground surface estimate is used to train a visual classifier to discriminate between ground and nonground regions of the image. In the prediction stage, the ground surface location can be estimated at high frequency solely from vision sensor data.
AB - Autonomous robotic navigation in forested environments is difficult because of the highly variable appearance and geometric properties of the terrain. In most navigation systems, researchers assume a priori knowledge of the terrain appearance properties, geometric properties, or both. In forest environments, vegetation such as trees, shrubs, and bushes has appearance and geometric properties that vary with change of seasons, vegetation age, and vegetation species. In addition, in forested environments the terrain surface is often rough, sloped, and/or covered with a surface layer of grass, vegetation, or snow. The complexity of the forest environment presents difficult challenges for autonomous navigation systems. In this paper, a self-supervised sensing approach is introduced that attempts to robustly identify a drivable terrain surface for robots operating in forested terrain. The sensing system employs both LIDAR and vision sensor data. There are three main stages in the system: feature learning, feature training, and terrain prediction. In the feature learning stage, 3D range points from LIDAR are analyzed to obtain an estimate of the ground surface location. In the feature training stage, the ground surface estimate is used to train a visual classifier to discriminate between ground and nonground regions of the image. In the prediction stage, the ground surface location can be estimated at high frequency solely from vision sensor data.
UR - http://www.scopus.com/inward/record.url?scp=84857251326&partnerID=8YFLogxK
U2 - 10.1002/rob.21417
DO - 10.1002/rob.21417
M3 - Article
AN - SCOPUS:84857251326
SN - 1556-4959
VL - 29
SP - 277
EP - 297
JO - Journal of Field Robotics
JF - Journal of Field Robotics
IS - 2
ER -