Posts

#4399 Lunar Rover AI-Assistance :: REASON v3 White Paper

#4399 Lunar Rover AI-Assisted Navigation :: REASON v3 White Paper [REASON v3 Whitepaper 4.pdf] #4399  Provide an executive summary of the uploaded white paper. The white paper "REASON - Real-time Executive Autonomy System with On-board Neural Reasoning," authored by Ivan Delgado, P.E. in March 2026, proposes a novel research framework for LLM-driven autonomous decision-making in space assets and autonomous vehicles . Core Objective REASON aims to overcome the limitations of traditional rule-based autonomy—such as "single-signal blindness" and an inability to handle unanticipated scenarios—by integrating Large Language Models (LLMs) capable of reasoning about complex temporal trends and mission context . Architectural Framework The system utilizes a formally partitioned four-layer architecture to balance advanced reasoning with safety: Layer 1 (Perception): Processes raw sensor data into structured summaries . Layer 2 (LLM Multi-Agent Pipeline): Employs para...

#4398 Nvidia DGX Spart Tokens/Second for GPT-OSS-20b, Granite 4 30B, Nemotron 3 Nano 30B and Gemma 4 31B

#4398 Nvidia DGX Spart Tokens/Second for GPT-OSS-20b, Granite 4 30B, Nemotron 3 Nano 30B and Gemma 4 31B #4398  Describe the tokens/second inferencing on the Nvidia DGX Spark for GPT-OSS-20B, Granite 4 30B, NVIDIA Nemotron-3-Nano-30B-A3B, and Gemma 4 31B Based on technical specifications for the NVIDIA DGX Spark (featuring the GB10 Grace Blackwell Superchip with 128GB of unified memory) and the specific model architectures you mentioned, here is the estimated tokens-per-second (tps) performance. These figures assume the use of NVFP4 (4-bit) quantization, which is the native optimization target for the Blackwell architecture. DGX Spark Inferencing Performance (Estimated) Model Parameters (Total/Active) Estimated Tokens/Sec Key Optimization GPT-OSS-20B 20.9B / 3.6B (MoE) 140 – 165 Top-4 expert routing; optimized for edge. Granite 4.1 30B 30.7B (Dense) 65 – 80 Dense architecture results in higher compute per token. Nemotron-3-Nano-30B-A3B 30.1B / 3.5B (Hybrid) 180 – 210 Mamba-2/Mo...