Why do AI chip startups make shitty NPUs that don't support floating point calculations?
>hey just quantize your shit model and hope it works
Why is it so hard to find a NPU with large memory and float32 support and high power efficiency?
>26 TOPs, 2.5W
>no floating point support
>no HBM, use ram from host you bitch
Is this really the best we've got in 2023? Why don't google sell TPUv4 cards?