this post was submitted on 05 Mar 2025
10 points (91.7% liked)

LocalLLaMA

2660 readers
2 users here now

Community to discuss about LLaMA, the large language model created by Meta AI.

This is intended to be a replacement for r/LocalLLaMA on Reddit.

founded 2 years ago
MODERATORS
 

Thinking about a new Mac, my MPB M1 2020 16 GB can only handle about 8B models and is slow.

Since I looked it up I might as well shared the LLM-related specs:
Memory bandwidth
M4 Pro (Mac Mini): 273GB/s M4 Max (Mac Studio): 410 GB/s

Cores cpu / gpu
M4 pro 14 / 20
M4 Max 16 / 40

Cores & memory bandwidth is of course important, but with the Mini I could have 64 GB ram instead of 36 (within my budget that is fixed for tax reasons).

Feels like the Mini with more memory would be better. What do you think?

top 6 comments
sorted by: hot top controversial new old
[–] [email protected] 1 points 3 days ago (1 children)

Have you considered the framework desktop? Uses amd strix halo. The same cost for the Mac mini (2k?) gets you around 90gb vram, out of a 128gb unified ram configuration.

[–] [email protected] 1 points 3 days ago (1 children)

Interesting, lots of "bang for the buck". I'll check it out

[–] [email protected] 1 points 3 days ago

Yup! They even had a demo clustering 5 of them to run deep seek proper

[–] [email protected] 1 points 4 days ago (1 children)

Depends on what model you want to run?

[–] [email protected] 1 points 3 days ago (1 children)

of course, I haven't looked at models >9B for now. So I have to decide if I want to run larger models quickly or even larger models quickly-but-not-as-quick-as-on-a- Mac-Studio.

Or I could just spend the money on API credits :D

[–] [email protected] 2 points 3 days ago

Use API credits. 64GB can barely run a 70B model. I have a MacBook Pro M3 Max with 128GB and can run those and even slightly bigger models. But results are underwhelming. Not bought for LLM only, but if I would have, I would be disappointed.