Lotti, Alessandro
(2026)
Deep learning for spacecraft pose estimation: from algorithms to hardware-in-the-loop testbed, [Dissertation thesis], Alma Mater Studiorum Università di Bologna.
Dottorato di ricerca in
Scienze e tecnologie aerospaziali, 37 Ciclo. DOI 10.48676/unibo/amsdottorato/12562.
Documenti full-text disponibili:
Abstract
The rapid growth of the New Space economy is increasing demand for in-orbit servicing and active debris removal missions that rely on rendezvous and proximity operations. In this context, relative pose estimation between a chaser and its target is a key enabling technology, achievable with various sensors (e.g., cameras, LIDAR). Monocular vision is attractive for its minimal system budget footprint. However, the complexity of space imagery and onboard deployment constraints still hinder its adoption in space. Monocular satellite pose estimation methods, especially when paired with deep learning for image processing, still suffer from domain gap, high computational cost, and limited uncertainty quantification. This work investigates these issues through an extensive set of experiments. First, a domain-adversarial, Vision Transformer-based pipeline is introduced to reduce the performance gap between synthetic and real imagery. This approach ranked fourth and fifth in ESA’s second Satellite Pose Estimation Challenge. Architectural and data-level strategies, such as generative models and hybrid neural network architectures, are then investigated to further mitigate domain gap while lowering computational cost. A parametric study is conducted to scale image processing on Edge TPUs, showing real-time performance with minimal energy consumption. Additionally, Sparse Variational Gaussian Processes (SVGPs) are explored to provide uncertainty estimates from neural network feature maps, showing a moderate-to-strong correlation between predicted uncertainty and actual errors on synthetic images. For unknown targets, the pipeline must begin with an inspection phase. To this aim, lightweight localization models are evaluated on previously unseen spacecraft at far range, followed by a lightweight deep-learning-based feature matcher that drives a frame-to-frame pose estimation pipeline to recover camera motion. Results highlight current dataset limitations for this task. Finally, the development, software integration, and calibration of a HIL testbed at the University of Bologna are presented, together with a dataset generation pipeline featuring a nearly symmetric target.
Abstract
The rapid growth of the New Space economy is increasing demand for in-orbit servicing and active debris removal missions that rely on rendezvous and proximity operations. In this context, relative pose estimation between a chaser and its target is a key enabling technology, achievable with various sensors (e.g., cameras, LIDAR). Monocular vision is attractive for its minimal system budget footprint. However, the complexity of space imagery and onboard deployment constraints still hinder its adoption in space. Monocular satellite pose estimation methods, especially when paired with deep learning for image processing, still suffer from domain gap, high computational cost, and limited uncertainty quantification. This work investigates these issues through an extensive set of experiments. First, a domain-adversarial, Vision Transformer-based pipeline is introduced to reduce the performance gap between synthetic and real imagery. This approach ranked fourth and fifth in ESA’s second Satellite Pose Estimation Challenge. Architectural and data-level strategies, such as generative models and hybrid neural network architectures, are then investigated to further mitigate domain gap while lowering computational cost. A parametric study is conducted to scale image processing on Edge TPUs, showing real-time performance with minimal energy consumption. Additionally, Sparse Variational Gaussian Processes (SVGPs) are explored to provide uncertainty estimates from neural network feature maps, showing a moderate-to-strong correlation between predicted uncertainty and actual errors on synthetic images. For unknown targets, the pipeline must begin with an inspection phase. To this aim, lightweight localization models are evaluated on previously unseen spacecraft at far range, followed by a lightweight deep-learning-based feature matcher that drives a frame-to-frame pose estimation pipeline to recover camera motion. Results highlight current dataset limitations for this task. Finally, the development, software integration, and calibration of a HIL testbed at the University of Bologna are presented, together with a dataset generation pipeline featuring a nearly symmetric target.
Tipologia del documento
Tesi di dottorato
Autore
Lotti, Alessandro
Supervisore
Co-supervisore
Dottorato di ricerca
Ciclo
37
Coordinatore
Settore disciplinare
Settore concorsuale
Parole chiave
Spacecraft pose estimation
Vision-based relative navigation
Monocular vision
Domain gap
Embedded deep learning
Robotic hardware-in-the-loop (HIL) testbed
DOI
10.48676/unibo/amsdottorato/12562
Data di discussione
21 Gennaio 2026
URI
Altri metadati
Tipologia del documento
Tesi di dottorato
Autore
Lotti, Alessandro
Supervisore
Co-supervisore
Dottorato di ricerca
Ciclo
37
Coordinatore
Settore disciplinare
Settore concorsuale
Parole chiave
Spacecraft pose estimation
Vision-based relative navigation
Monocular vision
Domain gap
Embedded deep learning
Robotic hardware-in-the-loop (HIL) testbed
DOI
10.48676/unibo/amsdottorato/12562
Data di discussione
21 Gennaio 2026
URI
Statistica sui download
Gestione del documento: