As a software developer who took an elective in neural networks - when people call LLMs stochastic parrots, that's not criticism of their results.

It's literally a description of how they work.

The so-called training data is used to build a huge database of words and the probability of them fitting together.

Stochastic because the whole thing is statistics.
Parrot because the answer is just repeating the most probable word combinations from its training dataset.

Calling an LLM a stochastic parrot is lile calling a car a motorised vehicle with wheels. It doesn't say anything about cars being good or bad. It does, however, take away the magic. So if you feel a need to defend AI when you hear the term stochastic parrot, consider that you may have elevated them to a god-like status, and that's why you go on the defense when the magic is dispelled.

Dieser Beitrag wurde bearbeitet. (3 Tage her)
Als Antwort auf Leeloo

pretty sure that's a fallacy, kinda like "a sculpture is just stone, therefore it can't be beautiful", or "a cell is just a bunch of proteins, therefore it cannot be a living creature".

Now, I'm not saying a huge database of probabilities can be intelligent (I hope it can't), just that I think a better argument is needed why in the case of a database of probabilities, what it's made of prevents it from being intelligent.

Dieser Beitrag wurde bearbeitet. (2 Tage her)
Als Antwort auf Wolf480pl

@wolf480pl
You would have to redefine intelligence for asking whether a list of numbers is intelligent to even make sense.

And your comparison is completely off. Beauty is not a property of the sculpture, it's, as they say, "in the eye pf the beholder". Some people find curves beautiful. Can a stone have curves? Yes, of course. Others may find sharp edges beautiful. Can a stone have sharp edges? Again, yes.

I suggest you consider once again whether you are elevating "AI" to a god-like status.

Als Antwort auf Wolf480pl

@wolf480pl
The effect that you are noticing is because the writers of the training material were intelligence. You are seeing the reflection of their intelligence in the output of the LLM: Here is output from an LLM that describes what an LLM is, and what it is not: johntinker.substack.com/p/misu…
Als Antwort auf James Wood

@mudri @lmorchard it’s not inductive at all though. It’s just parroting the patterns it sees in its training data. If it wasn’t common to see exchanges like that, the response would be utter nonsense.

People misunderstand what “training” is. It’s modeling the input. Humans develop the rules for how to model that input. Emergent properties of that process can easily *seem* like thinking or reason, but it’s an illusion.

Als Antwort auf Les Orchard

@lmorchard @mudri Be careful not to conflate the actual language model with its user interface. Whatever was sent to or received from the LLM went through the chatbot layer. Or possibly was handled by thd chatbot layer without ever touching the LLM. We don't know because the whole system is opaque.

This casual experiment may not be telling you what you think it's telling you. :)

Als Antwort auf Tobias Ernst

@tobifant Whilst we obviously can't show if humans have a soul, we can absolutely show that humans have e.g. abstracted concept frameworks that are not solely based on averages of language statistics. I understand what an "owl" is, for example, in a way separate to the numerical relationships between the word "owl" and other words. That is a really fundamental information processing difference and allows me to construct *novel* understandings of that concept in ways that an LLM couldn't.
Als Antwort auf Tobias Ernst

A LLM is not able to reason. It can fool you into believing it is intelligent and self aware, where in fact it just parrots the patterns it has stored. These patterns are however very human-like as they are the result of training on texts written by actual humans.

The fun part starts now where the entire internet got flooded by #ai generated content. All of this will be the training set for the next generation of LLM's. What could possibly go wrong?
@leeloo

Dieser Beitrag wurde bearbeitet. (1 Tag her)
Als Antwort auf Tobias Ernst

@Tobias Ernst @Leeloo We are already way past that point, although it isn't distributed evenly. One of the reason is that LLMs are machine learning applications, and machine learning is extremely effective at reaching its stated goals, the problems being to define those goals, and that they are hidden as a trade secret by the major LLM companies.

But it isn't difficult to figure out that these companies favor outputs that looks and sounds as human as possible, in order to exploit our innate tendency to seek humanity in looks and sounds, including language.

Als Antwort auf Leeloo

I myself like calling LLMs "glorified autocomplete". Or "Т9 на максималках" in Russian.

It's surprising just how defensive some people get when I say that even when they agree with my definition. They keep believing that just give this thing more parameters and something magical, something more than sum of its parts will emerge, any moment now, just one more model generation, just one more order of magnitude, I promise.

Als Antwort auf Gregory

@grishka
The fun part is that the next generation will have the current state of the internet as its training set. An internet that is flooded by #ai generated content.

The biggest issue those ai companies face at the moment is how to only ingest human generated content and filter out as much as possible of all of the ai generated crap that is out there.

Good luck with that.
@leeloo

Dieser Beitrag wurde bearbeitet. (1 Tag her)
Als Antwort auf Leeloo

As a side note, I sometimes worry about how much parroting happens in academia among humans even before/without LLMs, where people repeat things without understanding what they’re talking about. I guess at least for students, it sometimes is about learning to talk the talk, and then gradually developing more understanding and genuine thinking around topics. At least we humans are capable of developing that understanding if we bother to try.
Dieser Beitrag wurde bearbeitet. (1 Tag her)
Als Antwort auf Troed Sångberg

@troed
No, this is not just not true, it's absurdly not true.

Most of human thought isn't even language-based, let alone being representable as some kind of token generation. Most human thought is based on platforms that evolved long before language, that are demonstrably more capable than large language models at reasoning about the real world, since other entities that share these platforms are able to demonstrate quite sophisticated reasoning without involving language.

Als Antwort auf Resuna

@resuna At no point am I stating that LLMs are exactly like human brains.

blog.troed.se/posts/the-delta-…