<RETURN_TO_BASE

NVIDIA ThinkAct Revolutionizes Robot Control with Vision-Language-Action Reasoning

NVIDIA's ThinkAct framework introduces a dual-system approach combining vision-language reasoning with reinforced visual latent planning, significantly improving robot manipulation and planning in complex environments.

Bridging High-Level Reasoning and Robot Control

NVIDIA and National Taiwan University researchers have introduced ThinkAct, a novel framework that advances embodied AI by integrating vision, language, and action through reinforced visual latent planning. Unlike traditional vision-language-action (VLA) models that directly map inputs to actions, ThinkAct separates reasoning and control, enabling more effective long-term planning and adaptability in complex environments.

Dual-System Architecture

ThinkAct employs two interconnected components:

  • Reasoning Multimodal LLM (MLLM): This model performs detailed, stepwise reasoning on visual scenes and language instructions, producing a visual plan latent that captures high-level intents and planning contexts.
  • Action Model: A Transformer-based policy that executes the robot's actions based on the visual plan latent.

This asynchronous design allows the reasoning module to generate plans at a slower pace, while the action module carries out fine-grained controls rapidly.

Reinforced Visual Latent Planning

A key innovation is the use of reinforcement learning with action-aligned visual rewards:

  • Goal Reward: Aligns predicted start and end positions with demonstration trajectories to ensure goal completion.
  • Trajectory Reward: Uses dynamic time warping (DTW) to regularize predicted trajectories against expert demonstrations.

These rewards are combined with format correctness scores to encourage accurate and physically plausible plans.

Training Pipeline

The training process involves:

  • Supervised Fine-Tuning (SFT): Teaching trajectory prediction, reasoning, and formatting using annotated data.
  • Reinforced Fine-Tuning: Enhancing reasoning quality via Group Relative Policy Optimization (GRPO) to maximize visual rewards.
  • Action Adaptation: Training the action policy through imitation learning guided by the frozen LLM's latent plans.

Performance on Benchmarks

ThinkAct outperforms existing methods on various benchmarks:

  • SimplerEnv: Gains 11–17% improvement over strong baselines, particularly excelling in long-horizon and diverse tasks.
  • LIBERO: Achieves 84.4% success rate, leading in spatial, object, goal, and long-horizon challenges.

Embodied Reasoning and Adaptation

On tasks like EgoPlan-Bench2 and RoboVQA, ThinkAct demonstrates superior multi-step planning and question-answering capabilities. It supports few-shot adaptation, achieving notable success with as few as 10 demonstrations.

Self-Reflection and Correction

ThinkAct exhibits emergent behaviors such as failure detection (e.g., recognizing dropped objects) and automatic replanning to recover from errors.

Implementation Highlights

The framework uses the Qwen2.5-VL 7B MLLM backbone, vision encoders like DINOv2, text encoders like CLIP, and a Q-Former to connect reasoning outputs to the action policy. It has been tested extensively in both simulated and real-world settings, proving its scalability and robustness.

ThinkAct sets a new benchmark for embodied AI systems, enabling intelligent robots capable of thoughtful planning, real-time control, quick adaptation, and self-correction in dynamic environments.

🇷🇺

Сменить язык

Читать эту статью на русском

Переключить на Русский