I'd be interested in seeing benchmark data. The speed seemed pretty good in those examples.
Are there any Intel GPUs with a lot of vRAM that someone could recommend that would work with this?
Any performance benchmark against 'llamafile'[0] or others?
Would be nice if this came with scripts which could launch the examples on compatible GPUs on cloud providers (rather than trying to guess?). Would anyone else be interested in that? Considering putting it together.
Hm, no major cloud provider offers intel gpus.
[dead]
Looking forward to reviewing!
The company that did 4-cores-forever, has the opportunity to redeem itself, in its next consumer GPU release, by disrupting the "8-16GB VRAM forever" that AMD and Nvidia have been imposing on us for a decade. It would be poetic to see 32-48GB at a non-eye-watering price point.
Intel definitely seems to be doing all the right things on software support.