this post was submitted on 05 Mar 2025
10 points (91.7% liked)

LocalLLaMA

2661 readers
2 users here now

Community to discuss about LLaMA, the large language model created by Meta AI.

This is intended to be a replacement for r/LocalLLaMA on Reddit.

founded 2 years ago
MODERATORS
 

Thinking about a new Mac, my MPB M1 2020 16 GB can only handle about 8B models and is slow.

Since I looked it up I might as well shared the LLM-related specs:
Memory bandwidth
M4 Pro (Mac Mini): 273GB/s M4 Max (Mac Studio): 410 GB/s

Cores cpu / gpu
M4 pro 14 / 20
M4 Max 16 / 40

Cores & memory bandwidth is of course important, but with the Mini I could have 64 GB ram instead of 36 (within my budget that is fixed for tax reasons).

Feels like the Mini with more memory would be better. What do you think?

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 1 points 4 days ago (1 children)

Depends on what model you want to run?

[–] [email protected] 1 points 4 days ago (1 children)

of course, I haven't looked at models >9B for now. So I have to decide if I want to run larger models quickly or even larger models quickly-but-not-as-quick-as-on-a- Mac-Studio.

Or I could just spend the money on API credits :D

[–] [email protected] 2 points 4 days ago

Use API credits. 64GB can barely run a 70B model. I have a MacBook Pro M3 Max with 128GB and can run those and even slightly bigger models. But results are underwhelming. Not bought for LLM only, but if I would have, I would be disappointed.