We've all seen those impressive videos online of humanoid robots running across parking lots, jumping onto platforms, or completing obstacle courses with astounding precision. These clips showcase amazing feats of engineering and suggest a future filled with these agile mechanical machines.
But the reality is that dependable robots for unpredictable real-world tasks are still far from being commercially viable. Mastering balance and locomotion, for example, is one big hurdle.
However, the real challenge lies in reasoning and manipulation. As we've already seen, today's robots can excel at one task under controlled conditions, such as stacking boxes or pushing a table, but they fail to transfer that knowledge to new and unfamiliar contexts, like carrying a tray through a crowded cafeteria or navigating through a moving crowd while balancing something in its arms.
This gap has ignited a global race to build generalized robotic intelligence. And in China, the tech startup X Square Robot has entered the race with a breakthrough called Wall-OSS; it's an open-source small foundational model for embodied intelligence.
Wall-OSS is designed to act as an "embodied brain." It can generalize across multiple robot forms, training fine motor control for humanoid hands, as well as upper body movements, while also enabling continuous reasoning in different environments. Rather than programming a robot into pre-designed scripts, Wall-OSS allows it to adapt on-the-go, responding to non-routine conditions such as changing light, moving objects, or shifting instructions.
At its core, Wall-OSS is trained on billions of Vision-Language-Action samples. This includes real-world robotic datasets paired with generative video, simulating conditions robots are most likely to face in people's homes, public places, or establishments. By blending perception, instruction, and movement, the model strengthens both generalization and robustness.
The architecture itself uses a shared attention layer with task-routed Feed-Forward Networks. This means each type of input such as vision, language, or motor signals gets its own specialized processing pathway, which is then combined into an integrated response. This way, the robot doesn't simply "see" a box and "hears" the word "carry," but combines the two pathways to determine intent, which is to "pick up the box from the floor and place it on the table among other objects."
This lets robots coordinate high-level reasoning (such as sequencing steps) with motor control (like adjusting finger grip), all while planning several moves ahead through Chain of Thought (CoT) reasoning. The latter means it can simulate multiple steps before executing an action. So instead of reacting in isolation, the model generates an internal plan, similar to rehearsing its steps. And by doing so, it reduces trial-and-error when it starts to move.
The significance of Wall-OSS is its release as open source. For the first time, a foundation model for embodied AI will be available on platforms like GitHub and Hugging Face, ready for labs or startups to experiment with. Developers can embed Wall-OSS into their own robot designs, retrain it on specific tasks, or contribute to expanding its dataset.
This shift mirrors earlier moments in AI where open frameworks democratized innovation. And by opening access, X Square Robot is betting that community-driven development will accelerate progress faster than any single company could achieve alone.
To demonstrate the power of its new framework, X Square Robot has paired Wall-OSS with its latest robot, Quanta X2. Unlike earlier prototypes, Quanta X2 integrates a dexterous hand with 62 degrees of freedom across the body, a 7-degree-of-freedom arm, and even self-rotating tool clamping for cleaning attachments. It can perform teleoperated gestures, communicate through expressive movements, and, of course, execute household chores.
The Quanta line is positioned as a bridge between research and real-world application. By running Wall-OSS, Quanta X2 showcases how open-source embodied AI can extend beyond labs into homes, service industries, and industrial settings.
X Square Robot also recently announced approximately $100 million in fresh capital. This, the company says, will help expand developer access, grow training datasets, and accelerate the rollout of additional humanoid models.
The immediate promise includes robotic assistants in homes, hotel staff automation, and industrial handling. In reality, Wall-OSS could allow robots to adapt to tasks that vary day-to-day. They could be cleaning tables today, and welcoming guests the next. By sharing the codebase openly, X Square Robot is also encouraging use in unexpected domains, from eldercare to disaster relief.
The release of Wall-OSS marks a shift in robotics from something guarded and exclusive to something open and collaborative. The real question now is not if robots can be taught real-world intelligence, but in what life-changing ways can we use this resource to improve our everyday lives.