this post was submitted on 23 Apr 2025
624 points (92.6% liked)

196

5125 readers
1727 users here now

Be sure to follow the rule before you head out.


Rule: You must post before you leave.



Other rules

Behavior rules:

Posting rules:

NSFW: NSFW content is permitted but it must be tagged and have content warnings. Anything that doesn't adhere to this will be removed. Content warnings should be added like: [penis], [explicit description of sex]. Non-sexualized breasts of any gender are not considered inappropriate and therefore do not need to be blurred/tagged.

Also, when sharing art (comics etc.) please credit the creators.

If you have any questions, feel free to contact us on our matrix channel or email.

Other 196's:

founded 2 years ago
MODERATORS
 
you are viewing a single comment's thread
view the rest of the comments
[–] FooBarrington@lemmy.world 4 points 6 days ago (1 children)

While the order of magnitude is correct, running the bigger models is closer to playing a AAA game on 8 computers at the same time.

[–] infinitesunrise@slrpnk.net 2 points 6 days ago (1 children)

Yeah I did forget to consider that a lot of the web hosted models have a whole array of "experts" - Sub-LLMs that help fill in specialized information that a more generalized LLM wouldn't have. Not a problem for someone running an AI model on their home computer but something that likely happens most times that you're querying an AI online.

[–] FooBarrington@lemmy.world 2 points 5 days ago

That's also true, though it's important to remember that the "experts" aren't experts in the classical sense. Say you have a word made up of 3 tokens, it's possible that each token is routed to a different expert. It's just a model architecture.