Abstract
Space exploration has drawn increasing attention to space control technology. For debris removal missions and on-orbit servicing, accurate pose estimation of a noncooperative target is critical. This article introduces the satellite relative pose estimation network (SR-Net) two-stage training method for a noncooperative target via RGB images. As the first stage in regressing the 3D translation, we combined the detection and translation regression modules into a single model. SR-Net decouples the translation and rotation information in stage two by utilizing classification instead of regression, using the detected picture as input and fitting a rotation by minimizing the weighted least squares. Furthermore, a large-scale dataset for 6-DoF pose estimation is introduced, which can be utilized as a benchmark for various state-of-the-art monocular vision-based 6-DoF pose estimation methods. Ablation studies are used to verify the effectiveness and scalability of each module. SR-Net can be added to a baseline model as a separate module to improve the 6-DoF pose estimation accuracy for noncooperative targets. The results are extremely encouraging since they show that using only vision data, it is feasible to accurately estimate the 6-DoF pose of a noncooperative target.
Original language | English |
---|---|
Pages (from-to) | 31557-31573 |
Number of pages | 17 |
Journal | Multimedia Tools and Applications |
Volume | 82 |
Issue number | 20 |
DOIs | |
Publication status | Published - Aug 2023 |
Keywords
- CNN
- Noncooperative target
- Object detection
- Pose estimation
- Weighted least squares