this post was submitted on 17 Jul 2024
125 points (95.0% liked)

Technology

59673 readers
3179 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] rottingleaf@lemmy.world 1 points 4 months ago (1 children)

Information theory, entropy in Markovian processes. Read up on these buzzwords to see why.

[–] sunbeam60@lemmy.one -2 points 4 months ago (1 children)

I think I know enough about these concepts to know that there isn’t any conclusive proof, observed in output or system state, to establish consensus that human speech output is generated differently to how LLMs generate output. If you have links to any papers that claim otherwise, I’ll be happy to read them.

[–] rottingleaf@lemmy.world -1 points 4 months ago (1 children)

What? Humans, ahem, collect entropy every moment of their existence.

[–] sunbeam60@lemmy.one 2 points 4 months ago (1 children)

I mean I have an opinion too; what I’m seeking is evidence.

[–] rottingleaf@lemmy.world 0 points 4 months ago (1 children)

Evidence for what?

I've just diagonally read a google link where the described way humans work with language appears for me to be very similar to GPT in rough strokes. Only human brain does a lot more than language. Hence the comparisons to the mechanical Turk.

Also Russell's teapot.

[–] sunbeam60@lemmy.one 2 points 4 months ago (1 children)

I’m not saying humans and LLMs generate language the same way.

I’m not saying humans and LLMs don’t generate language the same way.

I’m saying I don’t know and I haven’t seen clear data/evidence/papers/science to lean one way or the other.

A lot of people seem to believe humans and LLMs don’t generate language the same way. I’m challenging that belief in the absence of data/evidence/papers/science.

[–] rottingleaf@lemmy.world 0 points 4 months ago (1 children)

Like going out and meeting a dino - 50% yes, 50% no. It's a joke.

Russell's teapot again.

[–] JackGreenEarth@lemm.ee 1 points 4 months ago (1 children)

You're actually incorrect in regards to Russell's teapot in this instance. The correct approach is to admit to yourself and others you don't know. Not to assume a negative became you can't prove a positive, if you can't prove the negative either.

[–] rottingleaf@lemmy.world 1 points 4 months ago

I know I don't know, but this is a continuous system and the probability of something being in one particular state is infinitely small ; the probability of it being in certain range of that particular state is, ahem, not, but with the amount of moving things in LLMs and in human brains there are most likely quite a few radical differences between laws describing them.

Why am I incorrect? You can't disprove that there isn't that teapot flying at a certain orbit as well. Or you can, but not for all such statements.

What would be the criterion for saying that yes, human brain works with language just in the same way as LLMs do? What would be "same"? Logic exists inside defined constraints in the continuous world.

Unless you define what would prove something, you can't disprove it, but it's also not a scientific hypothesis. That's Popper's criterion.