The bottleneck for inference right now isn't just raw FLOPS or even memory bandwidth—it's the compiler stack. The graveyard of AI hardware startups is filled with chips that beat NVIDIA on specs but couldn't run a standard PyTorch graph without segfaulting or requiring six months of manual kernel tuning.
Until I see a dev board and a working graph compiler that accepts ONNX out of the box, this is just a very expensive CGI render.
That seems like not much compared to the hundreds of billions of dollars US companies currently invest into their AI stack? OpenAI pays thousands of engineers and researchers full time.
Edit: It kind of looks like there's no silicon anywhere near production yet. Probably vaporware.
Also, the 3D graphic of their chip on a circuit board is missing some obvious support pieces, so it's clearly not from a CAD model.
Lots of chip startups start as this kind of vaporware, but very few of them obfuscate their chip timelines and anticipated release dates this much. 5 years is a bit long to tapeout, but not unreasonable.
This seems indicative enough for me, give or take a quarter or two probably, from the latest news post on their website:
> VSORA is now preparing for full-scale deployment, with development boards, reference designs, and servers expected in early 2026.
https://vsora.com/vsora-announces-tape-out-of-game-changing-...
Seems they have partners as well, who describe working together with a Taiwanese company as well.
You never know, guess they could have gotten others to fall for their illusions too, it's not unheard of. But considering how long time something like this takes to bring to market, that they have dev-boards ready is months rather than years at least gives me enough to wait until then to judge them too harshly.
So far, they just talk about it.
It sounds nice, but how much is it?