zompist wrote: ↑Wed Jan 29, 2025 5:14 am
I think you're just redefining "prediction" here to make it sound like intelligence. It's not, nor does it imply a model.
Then I invite you to provide your own definition.
Flip a coin; I predict that 50% of the time it'll come up heads. […] Note that you don't need an intelligent system at all to make these predictions. You can in fact make predictions, with the precise statistical distribution, by using another 3d6-- three dead pieces of plastic.
I’ll note that this and the other coin and dice examples are all
Markov processes, so comments on them apply to these situations similarly.
On these ‘predictions’: I shouldn’t really use scare quotes because after all they
are predictions, but they’re very limited ones. For me it’s not just about the accuracy of the predictions, but the fact that an intelligence can act as a predictor for many different systems. (More on this below.)
People, especially AI boosters, love to compare them to people. But we shouldn't be throwing out science just because we found a cool toy. Science needs to make the minimum number of assumptions possible to explain a phenomenon. You can't just say "It does human-like things, so it's just like a human."
The basic premise I work on is that ‘simulating intelligence’ and ‘being intelligent’ are one and the same thing. Something which can simulate intelligence sufficiently well must itself be intelligent. Trying to argue otherwise leads to incoherencies like the Chinese Room.
For this reason, I basically disagree with the last sentence here. I
would say that something has human-level intelligence if it can do human-like things.
Mind you, I do agree that merely doing ‘human-like things’ isn’t enough to be considered ‘just like a human’. There are degrees of human-like-ness, and a text generator should only be considered ‘just like a human’ when it does indeed communicate ‘just like a human’. You seem to think this is a low bar; it really is not.
Do Markov chains write in a human-like way? Well… no. They get basic syntax right (which probably says something important about syntax), but not much more. Consider the example sentence you quote in the last post: it can barely stay on a single topic from the beginning to the end! Rather it drifts gradually from queens to crowns to rain to illusion to screaming, in a way which is not surprising at all considering that it can only look at the last two words. And if you make the context any longer, they just start repeating sentences verbatim from the source.
Do LLMs write in a human-like way? They’re much closer, yes. But they still make mistakes that no human would ever make, like the infamous glue pizza. There’s more subtle things too: for instance, it’s well-reported that they regularly veer between being weirdly agreeable and weirdly disagreeable, again in a non-human-like way. And, most damning of all, they are unable to truly
learn after the training period: as soon as something goes out of the context window it’s forgotten. Current LLMs compensate by making the context window enormous, but that’s just working around the problem, not solving it.
To put it as neutrally as possible, it's an open question what they are. I don't think they're human; as I've said before, if you really believe they are you should be demanding that Sam Altman be arrested as a slaveowner. But-- just as if you were investigating grey parrots-- you can't just assume they work like human brains, or assume that "intelligence" is a unitary concept that applies to humans, birds, octopuses, and collections of node weights.
I don’t assume this at all. As described above, ‘intelligence’ for me is a function of observable behaviour; it says nothing about how they work internally.
What kind of observable behaviour, then, is ‘intelligence’? I do very strongly think it comes down to the ability of one system to predict the behaviour of another system. For LLMs (and also animals!), the other system of choice is invariably humans, and the prediction task is ‘behave like a human would if confronted by the same problem’. And this is a strong requirement, because humans can in turn do
many different kinds of prediction task! We can do things all the way from catching a thrown ball to modelling other humans to solving Fermat’s Last Theorem. Not every human can do every thing, of course, but the range is clearly broader than anything else we currently know of.
It’s interesting to contemplate how these criteria could be applied to an alien species, where we have no expectation of human-like-ness in the first place. But, well, you’ve read Stanisław Lem; such a question may well be impossible to answer. We are forced to take humanity as our reference point because we have no other.
(As for Sam Altman being a slave-holder, I think that depends on
consciousness and
suffering more than intelligence. Those two properties do indeed depend on internals, which makes them much more difficult to talk about. For the record, I do not believe for a second that LLMs are conscious; whatever machinery yields their output, I don’t think that it’s the right sort for consciousness.)