this post was submitted on 08 Jun 2025
771 points (95.7% liked)

Technology

71143 readers
2960 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
 

LOOK MAA I AM ON FRONT PAGE

you are viewing a single comment's thread
view the rest of the comments
[–] sp3ctr4l@lemmy.dbzer0.com 17 points 1 day ago* (last edited 1 day ago) (1 children)

This has been known for years, this is the default assumption of how these models work.

You would have to prove that some kind of actual reasoning capacity has arisen as... some kind of emergent complexity phenomenon.... not the other way around.

Corpos have just marketed/gaslit us/themselves so hard that they apparently forgot this.

[–] riskable@programming.dev 5 points 1 day ago (1 children)

Define, "reasoning". For decades software developers have been writing code with conditionals. That's "reasoning."

LLMs are "reasoning"... They're just not doing human-like reasoning.

[–] sp3ctr4l@lemmy.dbzer0.com 5 points 1 day ago* (last edited 1 day ago)

Howabout uh...

The ability to take a previously given set of knowledge, experiences and concepts, and combine or synthesize them in a consistent, non contradictory manner, to generate hitherto unrealized knowledge, or concepts, and then also be able to verify that those new knowledge and concepts are actually new, and actually valid, or at least be able to propose how one could test whether or not they are valid.

Arguably this is or involves meta-cognition, but that is what I would say... is the difference between what we typically think of as 'machine reasoning', and 'human reasoning'.

Now I will grant you that a large amount of humans essentially cannot do this, they suck at introspecting and maintaining logical consistency, that they are just told 'this is how things work', and they never question that untill decades later and their lives force them to address, or dismiss their own internally inconsisten beliefs.

But I would also say that this means they are bad at 'human reasoning'.

Basically, my definition of 'human reasoning' is perhaps more accurately described as 'critical thinking'.