基于事件相机与神经形态硬件的高效星载航天器姿态估计 / Efficient Onboard Spacecraft Pose Estimation with Event Cameras and Neuromorphic Hardware
1️⃣ 一句话总结
这篇论文提出了一种结合事件相机和神经形态处理器的航天器姿态估计新方法,能够在太空极端光照和高速运动条件下实现低延迟、低功耗的实时感知,为未来自主空间任务提供了实用解决方案。
Reliable relative pose estimation is a key enabler for autonomous rendezvous and proximity operations, yet space imagery is notoriously challenging due to extreme illumination, high contrast, and fast target motion. Event cameras provide asynchronous, change-driven measurements that can remain informative when frame-based imagery saturates or blurs, while neuromorphic processors can exploit sparse activations for low-latency, energy-efficient inferences. This paper presents a spacecraft 6-DoF pose-estimation pipeline that couples event-based vision with the BrainChip Akida neuromorphic processor. Using the SPADES dataset, we train compact MobileNet-style keypoint regression networks on lightweight event-frame representations, apply quantization-aware training (8/4-bit), and convert the models to Akida-compatible spiking neural networks. We benchmark three event representations and demonstrate real-time, low-power inference on Akida V1 hardware. We additionally design a heatmap-based model targeting Akida V2 and evaluate it on Akida Cloud, yielding improved pose accuracy. To our knowledge, this is the first end-to-end demonstration of spacecraft pose estimation running on Akida hardware, highlighting a practical route to low-latency, low-power perception for future autonomous space missions.
基于事件相机与神经形态硬件的高效星载航天器姿态估计 / Efficient Onboard Spacecraft Pose Estimation with Event Cameras and Neuromorphic Hardware
这篇论文提出了一种结合事件相机和神经形态处理器的航天器姿态估计新方法,能够在太空极端光照和高速运动条件下实现低延迟、低功耗的实时感知,为未来自主空间任务提供了实用解决方案。
源自 arXiv: 2604.04117