this post was submitted on 23 Nov 2024
136 points (85.1% liked)

Technology

59578 readers
2908 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

I'm usually the one saying "AI is already as good as it's gonna get, for a long while."

This article, in contrast, is quotes from folks making the next AI generation - saying the same.

you are viewing a single comment's thread
view the rest of the comments
[–] MajorHavoc@programming.dev 9 points 6 hours ago (2 children)

Are you asserting that chatbots are so fundamentally different from LLMs that "oh shit we can't just throw more CPU and data at this anymore" doesn't apply to roughly the same degree?

[–] makyo@lemmy.world 3 points 3 hours ago (1 children)

I feel like people are using those terms pretty well interchangeably lately anyway

[–] Greg@lemmy.ca -1 points 2 hours ago

People that don't understand those terms are using them interchangeably

[–] Greg@lemmy.ca -3 points 2 hours ago (1 children)

Yes of course I'm asserting that. While the performance of LLMs may be plateauing, the cost, context window, and efficiency is still getting much better. When you chat with a modern chat bot it's not just sending your input to an LLM like the first public version of ChatGPT. Nowadays a single chat bot response may require many LLM requests along with other techniques to mitigate the deficiencies of LLMs. Just ask the free version of ChatGPT a question that requires some calculation and you'll have a better understanding of what's going on and the direction of the industry.

[–] MajorHavoc@programming.dev 2 points 1 hour ago (1 children)

I think you're agreeing, just in a rude and condescending way.

There's a lot of ways left to improve, but they're not as simple as just throwing more data and CPU at the problem, anymore.

[–] Greg@lemmy.ca 1 points 30 minutes ago

I'm sorry if I'm coming across as condescending, that's not my intent. It's never been as simple as just throwing more days and CPU at the problem. There were algorithmic challenges for every LLM evolution. There are still lots of potential improvements using the existing training data. But even if there wasn't, we'll still see loads of improvements in chat bots because of other techniques.