L_Acacia

joined 1 year ago
[–] [email protected] 2 points 4 weeks ago

Mistral modèles don't have much filter don't worry lmao

[–] [email protected] 3 points 1 month ago (4 children)

They is no chance they are the one training it. It costs hundreds of millions to get a descent model. Seems like they will be using mistral, who have scrapped pretty much 100% of the web to use as training data.

[–] [email protected] 3 points 1 month ago

Buying second hand 3090/7090xtx will be cheaper for better performances if you are not building the rest of the machine.

[–] [email protected] 2 points 1 month ago

You are limited by bandwidth not compute with llm, so accelerator won't change the interferance tp/s

[–] [email protected] 4 points 1 month ago

Too be fair to Gemini, even though it is worse than Claude and Gpt. The weird answer were caused by bad engineering and not by bad model training. They were forcing the incorporattion off the Google search results even though the base model would most likely have gotten it right.

[–] [email protected] 5 points 3 months ago (2 children)

The training doesn't use csam, 0% chance big tech would use that in their dataset. The models are somewhat able to link concept like red and car, even if it had never seen a red car before.

[–] [email protected] 2 points 3 months ago (2 children)

Google uses their own chip for AI

[–] [email protected] 3 points 3 months ago

They know the tech is not good enough, they just dont care and want to maximise profit.

[–] [email protected] 4 points 6 months ago

Whatsapp is europe's iMessage

[–] [email protected] 4 points 6 months ago

You can take a look at exllama and llama.cpp source code on github if you want to see how it is implemented.

[–] [email protected] 5 points 6 months ago

If you have good enough hardware, this is a rabbithole you could explore. https://github.com/oobabooga/text-generation-webui/

[–] [email protected] 3 points 6 months ago

Around 48gb of VRAM if you want to run it in 4bits

view more: next ›