#4399 Lunar Rover AI-Assistance :: REASON v3 White Paper
#4399 Lunar Rover AI-Assisted Navigation :: REASON v3 White Paper
[REASON v3 Whitepaper 4.pdf]
#4399 Provide an executive summary of the uploaded white paper.
The white paper "REASON - Real-time Executive Autonomy System with On-board Neural Reasoning," authored by Ivan Delgado, P.E. in March 2026, proposes a novel research framework for LLM-driven autonomous decision-making in space assets and autonomous vehicles
Core Objective
REASON aims to overcome the limitations of traditional rule-based autonomy—such as "single-signal blindness" and an inability to handle unanticipated scenarios—by integrating Large Language Models (LLMs) capable of reasoning about complex temporal trends and mission context
Architectural Framework
The system utilizes a formally partitioned four-layer architecture to balance advanced reasoning with safety:
Layer 1 (Perception): Processes raw sensor data into structured summaries
. Layer 2 (LLM Multi-Agent Pipeline): Employs parallel domain-expert agents (Terrain, Power, Mobility, Thermal, Comms, and Crew) that report to a Synthesis Agent
. This layer runs on high-performance edge AI hardware (ASUS Ascent GX10) . Layer 3 (Safety Enforcement): A deterministic C++ "hardware guardian" that enforces inviolable hard limits regardless of LLM recommendations
. Layer 4 (Control): Executes wheel and actuator commands at high frequency (100Hz)
.
Key Innovations
Authority Chain: Enforces a strict "Recommend, Approve, Validate, Execute" workflow that maps to NASA mission operations
. Latency Management: A 12-second full deliberation cycle is mitigated by a 200ms "Scout" fast-interrupt path for sudden hazards and continuous monitoring by Layer 3
. Byzantine Fault Detection: The LLM's ability to cross-reference multiple sensors (e.g., comparing wheel current to reported slip) allows it to detect inconsistent or corrupted sensor data
. Certification Pathway: By isolating non-deterministic LLM behavior in Layer 2 and keeping safety-critical constraints in a traditional, deterministic C++ Layer 3, the system offers a viable route for NASA flight software certification
.
Safety and Development
REASON requires frozen model weights during missions to ensure reproducibility and auditability, while allowing for adaptation through in-context learning
Is the LLM expected to handle vision inputs?
No, the Large Language Model (LLM) in the REASON framework is not expected to handle raw vision or image inputs directly
Instead, the architecture utilizes a structured abstraction process where vision data is processed by deterministic systems before reaching the LLM pipeline
Layer 1 (Perception) Processing: Raw sensor data, including LiDAR point clouds and stereo camera imagery, is processed by C++ modules
. Structured Summaries: These modules distill visual information into "structured summaries" containing semantic world models and uncertainty bounds
. Semantic Reasoning: The LLM receives text-based health classifications (e.g., NOMINAL, CRITICAL), trend data, and human-readable alert strings rather than raw pixels
. Domain-Specific Fields: For the Terrain Agent specifically, visual data is converted into numeric fields such as slope, crater proximity, and rock density for the LLM to analyze
.
By using this approach, the system ensures the LLM reasons about interpreted domain states—like "entrapment risk"—rather than attempting to process raw voltages or image frames, which helps manage latency and reliability
Describe the size and capability of the LLM likely to be required
The REASON framework utilizes a specific high-efficiency model designed for agentic workflows in resource-constrained environments like a lunar rover.
LLM Size and Architecture
Primary Model: The framework specifies Nemotron-3 Nano 30B.
Architecture: It uses a 30B-A3B hybrid Mixture-of-Experts (MoE) design. While it has 30 billion total parameters, only 3 billion are active during any single inference task.
Memory Footprint: At NVFP4 quantization, the model requires approximately 15GB of VRAM. A 4-bit version typically needs about 24GB of RAM for broader use-cases.
Context Window: The model natively supports a context length of up to 256,000 tokens, which is used to process mission history and multi-sensor snapshots.
Required Capabilities
The LLM must possess specific "agentic" capabilities to function within the REASON pipeline:
On-board Neural Reasoning: The model must reason from "physical first principles" to handle novel terrain and unanticipated hazards.
Tool Calling: It requires integrated support for tool calling to interact with the rover's various subsystems and generate executable JSON action packages.
Multimodal Synthesis: While the current whitepaper focuses on text-based structured summaries, the specified Nemotron Nano Omni variant is capable of unifying video, audio, and text into a single reasoning stream.
Temporal Awareness: The model must be able to track trends (e.g., a wheel slowly losing traction) over a deliberation window rather than just reacting to instantaneous snapshots.
Cross-Domain Expert Depth: The model is expected to carry deep knowledge in specific domains such as terramechanics, power systems engineering, and orbital mechanics to act as a domain-expert advisor.
Compute Hardware
To run these capabilities in real-time, the framework specifies the ASUS Ascent GX10 AI Compute Module:
GPU: NVIDIA Blackwell (integrated GB10) capable of up to 1 petaFLOP of AI performance using FP4.
Memory: 128GB of LPDDR5x coherent unified system memory to support parallel execution of the six subsystem agents.
From <https://gemini.google.com/app/> Google Gemini 3.0 (Fast)
Comments
Post a Comment