APU for LLM Inference?

What about a cluster of M4 Minis? (or Mac Studios)

1 Like