This specific GPU is… Kind of a mixed bag. It’s supposed to be built on a 6nm process, and the G100 is, according to Lisuan, the first domestic chip to genuinely rival the NVIDIA RTX 4060 in raw performance, delivering 24 TFLOPS of FP32 compute. It even introduced support for Windows on ARM, a feature even major Western competitors had not fully prioritized.
It appears to fall short of its marketing promises, though. An alleged Geekbench OpenCL listing revealed the G100 achieving a score of only 15,524, a performance tier that effectively ties it with the GeForce GTX 660 Ti, a card released in 2012. This places the “next-gen” Chinese GPU on par with 13-year-old hardware, making it one of the lowest-scoring entries in the modern database. The leaked specifications further muddied the waters, showing the device operating with only 32 Compute Units, a bafflingly low 300 MHz clock speed, and a virtually unusable 256 MB of video memory. We’ll likely see more benchmarks as the GPU makes its way to the hands of customers.
These “anemic” figures might represent an engineering sample failing to report correctly due to immature drivers—a theory supported by the test bed’s configuration of a Ryzen 7 8700G on Windows 10. But still, if true, the underlying silicon may still be fundamentally incapable of reaching the promised RTX 4060 performance targets, regardless of the actual specifications that are being reported.


This is why i played around with mcp over the holidays. The fact its a standard to allow an ai to talk to an api is kinda cool. And nothing is stopping you from making the api do some ai call in itself.
Personally, i find the tech behind ai’s, and even llm’s, super interesting but companies are just fucking it up and pushing it way ti fucking hard and in ways its not meant to be -_-
Thanks for the info and ill have to look into those non-llm ai’s :)