L_Acacia

joined 1 year ago
[–] [email protected] 2 points 3 days ago (1 children)

zen integrates every upstream change a few hours after release, it is built as a set of patch on top of firefox just to make that easy

[–] [email protected] 2 points 1 week ago

they released a search engine where the model reads the first link before trying to answer your request

[–] [email protected] 2 points 3 weeks ago

llama.cpp works on windows too (or any os for that matter), though linux will vive you better performances

[–] [email protected] 6 points 2 months ago

Revolt tries to be a discord clone/replacement and suffer from some of the same issues. Matrix happens to have a lot of feature in common, but is focused on privacy and security at its core.

[–] [email protected] 2 points 3 months ago

Mistral modèles don't have much filter don't worry lmao

[–] [email protected] 3 points 3 months ago (4 children)

They is no chance they are the one training it. It costs hundreds of millions to get a descent model. Seems like they will be using mistral, who have scrapped pretty much 100% of the web to use as training data.

[–] [email protected] 3 points 4 months ago

Buying second hand 3090/7090xtx will be cheaper for better performances if you are not building the rest of the machine.

[–] [email protected] 2 points 4 months ago

You are limited by bandwidth not compute with llm, so accelerator won't change the interferance tp/s

[–] [email protected] 6 points 4 months ago

I use similar feature on discord quite extensively (custom emote/sticker) and i don't feel they are just a novelty. Allows us to have inside joke / custom reaction to specific event and I really miss it when trying out open source alternatives.

[–] [email protected] 4 points 4 months ago

Too be fair to Gemini, even though it is worse than Claude and Gpt. The weird answer were caused by bad engineering and not by bad model training. They were forcing the incorporattion off the Google search results even though the base model would most likely have gotten it right.

[–] [email protected] 5 points 5 months ago (2 children)

The training doesn't use csam, 0% chance big tech would use that in their dataset. The models are somewhat able to link concept like red and car, even if it had never seen a red car before.

[–] [email protected] 3 points 5 months ago* (last edited 5 months ago)

The models used are not trained on CP. The models weight are distributed freely and anybody can train a LORA on his computer. Its already too late to ban open weight models.

view more: next ›