Showing posts with label Robotics. Show all posts
Showing posts with label Robotics. Show all posts

2.8.25

Computing Changes How We Think—But Creativity, Not Just GPUs, Will Decide AI’s Next Decade

 In a wide-ranging Bloomberg interview, Dr. Wang Jian (founder of Alibaba Cloud) makes a forceful case that the era of AI “toy problems” is over. I agree. The last two years moved us from brittle demos to systems that reliably draft code, analyze documents, and support human decision-making. His analogy that more compute is like upgrading from a bicycle to a rocket is compelling: when the cost and scale of computation change, the feasible solution space—and our mental models—change with it.

Where I especially align is his view that markets are not just places to sell, but living testbeds where technology matures under real constraints. This resonates with best practices in ML ops: no benchmark, however well chosen, substitutes for deployment feedback. China’s dense competitive landscape, as he notes, creates short iteration loops—startups push features, rivals answer, users vote—accelerating collective learning. In ML terms, it’s a virtuous cycle of data, gradient steps, and evaluation at production scale.

I also appreciate his skepticism about tidy labels like AI → AGI → ASI. In practice, capability is a continuum: larger context windows, better tool use, richer memory, and planning—these blur categorical boundaries. Treating progress as increasing capability across tasks avoids false thresholds and keeps builders focused on measurable gains.

That said, I diverge on several points.

First, Dr. Wang downplays compute as a long-term bottleneck. I’m not fully convinced. While creativity and product insight absolutely dominate value creation, frontier training remains capital- and energy-intensive. Export controls, supply chain variability, and power availability still shape who can train or serve the most advanced models. For many labs, clever data curation and distillation help—but they don’t erase the physics and economics of scaling laws.

Second, on robotics, he frames AI as a new “engine” for an existing vehicle. Conceptually useful—but today’s embodied intelligence also requires tight integration across perception, control, simulation, and safety, not just swapping motors. Progress is real (foundation models for vision and language transfer surprisingly well), yet reliable grasping, long-horizon autonomy, and recovery from edge cases remain research frontiers. The “AI engine” metaphor risks underestimating those system-level challenges.

Third, the notion that no current advantage forms a durable moat is directionally optimistic and healthy for competition; still, moats can emerge from datasets with verified provenance, reinforcement-learning pipelines at scale, distribution, and compliance. Even if individual components commoditize, the orchestration (agents, tools, retrieval, evals, and workflow integration) can compound into real defensibility.

Finally, I agree with his emphasis that creativity is the scarcest input. Where I’d extend the argument is execution discipline: teams need evaluation harnesses, safety checks, and shipping cadences so creativity feeds a measurable loop. In other words, pair inspired ideas with ruthless metrics.

The upshot: Dr. Wang’s thesis—compute reshapes thinking, markets mature tech, creativity drives breakthroughs—captures much of what’s powering AI right now. My caveats don’t negate his vision; they refine it. The winners will be those who marry inventive product design with pragmatic engineering and acknowledge that, even in a marathon, hardware, data, and distribution still set the course.

4.6.25

SmolVLA: Hugging Face's Compact Vision-Language-Action Model for Affordable Robotics

 Hugging Face has introduced SmolVLA, a compact and efficient Vision-Language-Action (VLA) model designed to democratize robotics by enabling robust performance on consumer-grade hardware. With only 450 million parameters, SmolVLA achieves competitive results compared to larger models, thanks to its training on diverse, community-contributed datasets.

Bridging the Gap in Robotics AI

While large-scale Vision-Language Models (VLMs) have propelled advancements in AI, their application in robotics has been limited due to high computational demands and reliance on proprietary datasets. SmolVLA addresses these challenges by offering:

  • Compact Architecture: A 450M-parameter model that balances performance and efficiency.

  • Community-Driven Training Data: Utilization of 487 high-quality datasets from the LeRobot community, encompassing approximately 10 million frames.

  • Open-Source Accessibility: Availability of model weights and training data under the Apache 2.0 license, fostering transparency and collaboration.

Innovative Training and Annotation Techniques

To enhance the quality of training data, the team employed the Qwen2.5-VL-3B-Instruct model to generate concise, action-oriented task descriptions, replacing vague or missing annotations. This approach ensured consistent and informative labels across the diverse datasets.

Performance and Efficiency

SmolVLA demonstrates impressive capabilities:

  • Improved Success Rates: Pretraining on community datasets increased task success on the SO100 benchmark from 51.7% to 78.3%.

  • Asynchronous Inference: Decoupling perception and action prediction from execution allows for faster response times and higher task throughput.

  • Resource-Efficient Deployment: Designed for training on a single GPU and deployment on CPUs or consumer-grade GPUs, making advanced robotics more accessible.

Getting Started with SmolVLA

Developers and researchers can access SmolVLA through the Hugging Face Hub:

By offering a compact, efficient, and open-source VLA model, SmolVLA paves the way for broader participation in robotics research and development, fostering innovation and collaboration in the field.

22.5.25

NVIDIA Launches Cosmos-Reason1: Pioneering AI Models for Physical Common Sense and Embodied Reasoning

 NVIDIA has unveiled Cosmos-Reason1, a groundbreaking suite of AI models aimed at advancing physical common sense and embodied reasoning in real-world environments. This release marks a significant step towards developing AI systems capable of understanding and interacting with the physical world in a human-like manner.

Understanding Cosmos-Reason1

Cosmos-Reason1 comprises multimodal large language models (LLMs) trained to interpret and reason about physical environments. These models are designed to process both textual and visual data, enabling them to make informed decisions based on real-world contexts. By integrating physical common sense and embodied reasoning, Cosmos-Reason1 aims to bridge the gap between AI and human-like understanding of the physical world. 

Key Features

  • Multimodal Processing: Cosmos-Reason1 models can analyze and interpret both language and visual inputs, allowing for a comprehensive understanding of complex environments.

  • Physical Common Sense Ontology: The models are built upon a hierarchical ontology that encapsulates knowledge about space, time, and fundamental physics, providing a structured framework for physical reasoning. 

  • Embodied Reasoning Capabilities: Cosmos-Reason1 is equipped to simulate and predict physical interactions, enabling AI to perform tasks that require an understanding of cause and effect in the physical world.

  • Benchmarking and Evaluation: NVIDIA has developed comprehensive benchmarks to assess the models' performance in physical common sense and embodied reasoning tasks, ensuring their reliability and effectiveness. 

Applications and Impact

The introduction of Cosmos-Reason1 holds significant implications for various industries:

  • Robotics: Enhancing robots' ability to navigate and interact with dynamic environments. 

  • Autonomous Vehicles: Improving decision-making processes in self-driving cars by providing a better understanding of physical surroundings.

  • Healthcare: Assisting in the development of AI systems that can comprehend and respond to physical cues in medical settings.

  • Manufacturing: Optimizing automation processes by enabling machines to adapt to changes in physical environments.

Access and Licensing

NVIDIA has made Cosmos-Reason1 available under the NVIDIA Open Model License, promoting transparency and collaboration within the AI community. Developers and researchers can access the models and related resources through the following platforms:



 OpenAI has released GPT-OSS , a pair of open-weight language models designed for strong reasoning and agentic workflows— gpt-oss-120b and ...