Добавить новость


News in English


Новости сегодня

Новости от TheMoneytizer

Vision-language-action models are the next leap in autonomous robotics

GR00T N1 is an example of a vision-language-action model. Source: NVIDIA

Robotics has traditionally used modular pipelines. Perception, planning, and control sit in separate systems and connect through hand-tuned interfaces. This approach works for simple, well-defined tasks. It struggles when environments change or when robots must follow flexible instructions. Vision-language-action, or VLA, models offer a different path.

Systems such as Figure AI’s Helix, NVIDIA’s GR00T N1, and Google DeepMind’s RT-1, introduced last year, combine vision, language understanding, and motor control into a single model. These systems operate end-to-end and act directly on real robots.

This shift matters now because recent work shows practical, on-device deployments. These can reduce latency, improve dexterity, and allow faster task changes. VLAs point toward robots that understand natural instructions, carry out multi-step tasks, and move smoothly without fragile, hand-built pipelines.

Let’s look at how VLAs work, compare major approaches, and examine hardware, deployment, and safety considerations for commercial robotics teams.

What are vision-language-action models?

Vision-language-action models are unified AI systems that combine vision, language understanding, and action into one end-to-end model. VLAs take in images (or video) and language instructions, and produce continuous motor commands that drive a robot’s behavior in the physical world.

This approach differs from traditional robotics. Older systems split perception, planning, and control into separate modules. Engineers connect them with hand-built rules, which often fail in messy and flexible environments.

VLAs build on vision-language models (VLMs) by adding action. They do more than recognize scenes or answer questions. They decide how a robot should move, grasp, and manipulate objects.

Through joint training across vision, semantics, and motor behavior, VLAs learn shared representations that support flexible task execution. This foundation leads directly into the key VLA architectures that now drive rapid progress in autonomous robotics.



Key architectures drive vision-language-action progress

Several recent vision-language-action architectures show how this new paradigm moves from research into working robotic systems. Each takes a different path toward unifying perception, language, and action.

Helix – High-frequency dexterous control

Helix is a VLA model developed by Figure AI to control the full upper body of its humanoid robots. It targets arms, hands, torso, and fingers at high frequency.

Helix uses a dual-system design. A large vision-language backbone handles high-level reasoning and task understanding. A separate, fast visuomotor policy converts those internal representations into continuous control signals.

This split allows Helix to generalize across tasks while still meeting the real-time demands of dexterous manipulation in unstructured environments.

Helix architecture. Source: Figure AI

GR00T N1 – Open, generalist robotics foundation model

GR00T N1, introduced by NVIDIA, follows a foundation-model approach for robotics. It is trained offline on a mix of robot trajectories, human demonstration videos, and synthetic data. The goal is broad generalization across tasks and robot platforms.

NVIDIA has shown GR00T N1 running on real humanoid hardware, including bimanual manipulation. Like large language models (LLMs), it emphasizes pretraining once and adapting widely.

GR001 N1 model architecture. Source: NVIDIA

RT-2 – Scalable embodied AI

RT-2, from Google DeepMind, extends the Gemini 2.0 multimodal backbone into continuous action control. It demonstrates strong generalization to unseen objects and multi-step tasks. Recent on-device variants reduce latency and support offline operation.

Together, these approaches set the stage for how VLAs integrate with physical robot stacks.

RT-2 architecture. Source: Google DeepMind

How VLAs integrate with physical robot stacks

Vision-language-action models rely on rich, fused sensing. RGB and depth cameras, lidar, IMUs, and force/torque sensors feed multimodal encoders so the model sees geometry, texture, and contact states in real time.

Onboard compute shapes what’s possible. Real-time inference for multimodal transformers demands GPUs or specialized accelerators. Otherwise, latency kills safety and responsiveness.

That creates a trade-off: Run the VLA locally for low latency and offline operation, or use a hybrid cloud setup for heavier reasoning and model updates. RT-2’s on-device variant illustrates the local approach, reduces network delays, and enables faster reactions.

Next, we’ll examine practical deployment challenges and considerations that commercial teams must face when adopting VLAs.

Practical deployment challenges and considerations

While VLAs promise transformative abilities, real deployment still faces hard challenges.

Real-world robustness

Real-world robustness remains a major hurdle. VLAs can be brittle when lighting changes, scenes are cluttered, or sensors report noisy data. Ensuring reliable behavior in varied settings demands extensive testing and safety assurance.

Hardware limits—heat, power draw, and communication bandwidth—can further constrain performance on mobile robots.

Efficiency and model size

Efficiency and model size also matter. Large VLA models strain onboard resources. Emerging work on smaller, efficient variants (e.g., research into compact VLA models) shows that leaner architectures can still deliver meaningful control for specific tasks.

Benchmarking and standards

Benchmarking and standards are nascent. Conferences like ICLR see a surge of VLA research, but the field lacks widely accepted benchmarks and test suites for fair evaluation across both simulation and real robots.

Where VLA research and industry are headed

Looking ahead, vision-language-action research shows clear momentum. The next wave focuses on deeper multimodal and embodied AI systems that move beyond today’s designs.

One major shift appears in architecture. Researchers now explore diffusion-based and hybrid models instead of purely autoregressive policies. These approaches generate action sequences more efficiently and align reasoning with control, which improves generalization across tasks.

Another trend centers on embodied cognition. New models connect continuous perception with time-aware action planning and intermediate reasoning. This helps robots understand context over longer horizons and complete multi-step tasks more reliably.

The ecosystem also expands quickly. Open frameworks and shared datasets, such as community-driven efforts like LeRobot, make experimentation easier and encourage collaboration across labs and companies. Together, these trends point toward VLAs that scale better, adapt faster, and see wider adoption in commercial robotics.

A practical step toward truly autonomous robots

Vision-language-action models mark a clear break from older, modular robotics pipelines. They connect perception, language understanding, and control in a single system, which allows robots to interpret instructions and act with far more flexibility.

For commercial robotics teams, this shift opens the door to natural-language interfaces, stronger generalization across tasks, and robots that operate more naturally in human spaces.

I see VLAs as a practical step toward machines that truly understand what to do and how to do it. Success, however, depends on thoughtful adoption that balances ambitious capabilities with hardware limits, safety requirements, and real-world deployment constraints.

About the author

Pratik Shinde is a content and SEO Expert at Omdena and a full-stack digital marketer with over six years of experience driving organic growth for SaaS, AI, and technology brands. He takes a holistic approach to marketing by combining SEO, content strategy, paid acquisition, and AI-powered automation to deliver measurable business outcomes.

Previously, Shinde has led high-impact SEO and link-building initiatives for multiple global SaaS companies, helping them grow authority, traffic, and conversions across competitive markets.

The post Vision-language-action models are the next leap in autonomous robotics appeared first on The Robot Report.

Читайте на сайте


Smi24.net — ежеминутные новости с ежедневным архивом. Только у нас — все главные новости дня без политической цензуры. Абсолютно все точки зрения, трезвая аналитика, цивилизованные споры и обсуждения без взаимных обвинений и оскорблений. Помните, что не у всех точка зрения совпадает с Вашей. Уважайте мнение других, даже если Вы отстаиваете свой взгляд и свою позицию. Мы не навязываем Вам своё видение, мы даём Вам срез событий дня без цензуры и без купюр. Новости, какие они есть —онлайн с поминутным архивом по всем городам и регионам России, Украины, Белоруссии и Абхазии. Smi24.net — живые новости в живом эфире! Быстрый поиск от Smi24.net — это не только возможность первым узнать, но и преимущество сообщить срочные новости мгновенно на любом языке мира и быть услышанным тут же. В любую минуту Вы можете добавить свою новость - здесь.




Новости от наших партнёров в Вашем городе

Ria.city
Музыкальные новости
Новости России
Экология в России и мире
Спорт в России и мире
Moscow.media










Топ новостей на этот час

Rss.plus