this post was submitted on 23 Apr 2025
624 points (92.6% liked)
196
5125 readers
1727 users here now
Be sure to follow the rule before you head out.
Rule: You must post before you leave.
Other rules
Behavior rules:
- No bigotry (transphobia, racism, etc…)
- No genocide denial
- No support for authoritarian behaviour (incl. Tankies)
- No namecalling
- Accounts from lemmygrad.ml, threads.net, or hexbear.net are held to higher standards
- Other things seen as cleary bad
Posting rules:
- No AI generated content (DALL-E etc…)
- No advertisements
- No gore / violence
- Mutual aid posts are not allowed
NSFW: NSFW content is permitted but it must be tagged and have content warnings. Anything that doesn't adhere to this will be removed. Content warnings should be added like: [penis], [explicit description of sex]. Non-sexualized breasts of any gender are not considered inappropriate and therefore do not need to be blurred/tagged.
Also, when sharing art (comics etc.) please credit the creators.
If you have any questions, feel free to contact us on our matrix channel or email.
Other 196's:
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
While the order of magnitude is correct, running the bigger models is closer to playing a AAA game on 8 computers at the same time.
Yeah I did forget to consider that a lot of the web hosted models have a whole array of "experts" - Sub-LLMs that help fill in specialized information that a more generalized LLM wouldn't have. Not a problem for someone running an AI model on their home computer but something that likely happens most times that you're querying an AI online.
That's also true, though it's important to remember that the "experts" aren't experts in the classical sense. Say you have a word made up of 3 tokens, it's possible that each token is routed to a different expert. It's just a model architecture.