this post was submitted on 30 Jul 2024
1200 points (98.0% liked)

linuxmemes

20688 readers
1071 users here now

I use Arch btw


Sister communities:

Community rules

  1. Follow the site-wide rules and code of conduct
  2. Be civil
  3. Post Linux-related content
  4. No recent reposts

Please report posts and comments that break these rules!

founded 1 year ago
MODERATORS
1200
submitted 1 month ago* (last edited 1 month ago) by [email protected] to c/[email protected]
 

List of icons/services suggested:

  • Calibre
  • Jitsi
  • Kiwix
  • Monero (Node)
  • Nextcloud
  • Pihole
  • Ollama (Should at least be able to run tiny-llama 1.1B)
  • Open Media Vault
  • Syncthing
  • VLC Media Player Media Server
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 2 points 1 month ago (5 children)

Heres a tip, most software has the models default context size set at 512, 2048, or 4092. Part of what makes llama 3.1 so special is that it was trained with 128k context so bump that up to 131072 in the settings so it isnt recalculating context every few minutes…

Some caveats, this massively increases memory usage (unless you quantize the cache with FA) and it also massively slows down CPU generation once the context gets long.

TBH you just need to not keep a long chat history unless you need it,.

[–] [email protected] 1 points 1 month ago* (last edited 1 month ago) (4 children)

Thank you thats useful to know. In your opinion what context size is the sweet spot for llama 3.1 8B and similar models?

[–] [email protected] 1 points 1 month ago (2 children)

4 core i7, 16gb RAM and no GPU yet

Honestly as small as you can manage.

Again, you will get much better speeds out of "extreme" MoE models like deepseek chat lite: https://huggingface.co/YorkieOH10/DeepSeek-V2-Lite-Chat-Q4_K_M-GGUF/tree/main

Another thing I'd recommend is running kobold.cpp instead of ollama if you want to get into the nitty gritty of llms. Its more customizable and (ultimately) faster on more hardware.

[–] [email protected] 1 points 1 month ago* (last edited 1 month ago) (1 children)

Thats good info for low spec laptops. Thanks for the software recommendation. Need to do some more research on the model you suggested. I think you confused me for the other guy though. Im currently working with a six core ryzen 2600 CPU and a RX 580 GPU. edit- no worries we are good it was still great info for the thinkpad users!

[–] [email protected] 1 points 1 month ago

8GB or 4GB?

Yeah you should get kobold.cpp's rocm fork working if you can manage it, otherwise use their vulkan build.

llama 8b at shorter context is probably good for your machine, as it can fit on the 8GB GPU at shorter context, or at least be partially offloaded if its a 4GB one.

I wouldn't recommend deepseek for your machine. It's a better fit for older CPUs, as it's not as smart as llama 8B, and its bigger than llama 8B, but it just runs super fast because its an MoE.

load more comments (1 replies)
load more comments (1 replies)