It is a mixture of experts model so it will run on a computer with a lot of RAM and a GPU.
Alternately, on an M3 Ultra Mac Studio with 256GB of unified memory, you can run a 4bit quant of GLM-4.6 at about 20 tokens/second. That compares to about 40 t/s for a 6bit quant of MiniMax M2. I am not sure how fast these will run if you have a Mac Studio 512GB that can load the unquantized versions of the models.
There was an image of a chart [^1] in the email I received announcing this. It is perhaps one of the worst charts I’ve seen in a while.
[1]: https://service.campaigndelivery.cn/resources/templateImages...
Huggingface[0] says that this is a 358B model. What kind of hardware is necessary to run it?
[0]: https://huggingface.co/zai-org/GLM-4.7
It is a mixture of experts model so it will run on a computer with a lot of RAM and a GPU.
Alternately, on an M3 Ultra Mac Studio with 256GB of unified memory, you can run a 4bit quant of GLM-4.6 at about 20 tokens/second. That compares to about 40 t/s for a 6bit quant of MiniMax M2. I am not sure how fast these will run if you have a Mac Studio 512GB that can load the unquantized versions of the models.
https://lifearchitect.ai/models-table/
I'm completely blown away by ZAI GLM 4.7.
Great performance for coding after I snatched a pretty good deal 50%+20%+10%(with bonus link) off.
60x Claude Code Pro Performance for Max Plan for the almost the same price. Unbelievable
Anyone cares to subscribe here is a link:
https://z.ai/subscribe?ic=OUCO7ISEDB
Looks like it's available on the coding plan as well.