Re: Soshul meedja.
Posted: Thu Mar 16, 2023 3:50 am
Me too! Probably the more Croftian version of it, but I'm not too picky.
Me too! Probably the more Croftian version of it, but I'm not too picky.
Hmm, what’s that?chris_notts wrote: ↑Thu Mar 16, 2023 3:50 amMe too! Probably the more Croftian version of it, but I'm not too picky.
I was thinking of this book:bradrn wrote: ↑Thu Mar 16, 2023 3:55 amHmm, what’s that?chris_notts wrote: ↑Thu Mar 16, 2023 3:50 amMe too! Probably the more Croftian version of it, but I'm not too picky.
Eh, Markov generators are just as much of a challenge. They're far simpler, are obviously not "intelligent", and yet are almost creepy in their ability to reproduce the basics of language. I agree with Brad that something like Construction Grammar is more able to handle learning from a corpus than UG. But then I think UG is about the worst of Chomsky's ideas.chris_notts wrote: ↑Thu Mar 16, 2023 3:03 amI want to see linguistics grapple with what it means, if anything, for an LLM to be able to produce fluent, grammatical English just by finding usage patterns, without directly encoding anything built in as complex as UG at all.Ares Land wrote: ↑Thu Mar 16, 2023 2:54 am Looking at it another way: artificial intelligence sheds light on how the brain works. Language centers in the human brain probably resemble ML language models somewhat. Without even going into ML, there's probably something of a Markov chain there.
(When I'm very tired, I certainly tend to pick the next statistically likely word instead of what I actually meant to say )
Arguably the machines in the Matrix found a way to do just that. "Preserve humanity" is pretty vague.bradrn wrote: ↑Wed Mar 15, 2023 11:01 pm This isn’t the argument. The argument is that a superintelligent AI just has to be fixated on any goal. If ‘preserve humanity‘ isn’t its priority, then there’s a good chance it won’t preserve humanity.
(The obvious answer to that is that then we should figure out how to make ‘preserve humanity’ a priority for it; that’s basically what the AI alignment people go on about.)
True, but isn't this just a further argument against the Chomskian school? They're the ones trying to separate the core of language itself from the rest of cognition and the embedded nature of the speaker. If you accept that the purpose and use of language is strongly intertwined with every aspect of its design and that there's no hard boundaries between grammar, lexicon, pragmatics etc. then surely formalism and the Chomskian programme is dead.zompist wrote: ↑Thu Mar 16, 2023 4:06 am Though we don't have UG as Chomsky envisions it, we do have a hundred million years of evolutionary history as animals functioning in the world. We didn't evolve to randomly generate plausible-sounding text; we did evolve to see and interact with things in the world.
(And yes, I'm aware that you could add visual data, mechanical sensors, motor devices, etc, etc. But then it's not just a language model, is it?)
Not really, no. Output like this:zompist wrote: ↑Thu Mar 16, 2023 4:06 amEh, Markov generators are just as much of a challenge. They're far simpler, are obviously not "intelligent", and yet are almost creepy in their ability to reproduce the basics of language.chris_notts wrote: ↑Thu Mar 16, 2023 3:03 amI want to see linguistics grapple with what it means, if anything, for an LLM to be able to produce fluent, grammatical English just by finding usage patterns, without directly encoding anything built in as complex as UG at all.Ares Land wrote: ↑Thu Mar 16, 2023 2:54 am Looking at it another way: artificial intelligence sheds light on how the brain works. Language centers in the human brain probably resemble ML language models somewhat. Without even going into ML, there's probably something of a Markov chain there.
(When I'm very tired, I certainly tend to pick the next statistically likely word instead of what I actually meant to say )
viii--the queen's croquet ground near our breath.…simply cannot be compared to what current LLMs can produce. In general, the latter seems indistinguishable (or at least very nearly so) from human utterances — there is a reason universities are panicking about ChatGPT-produced essays but never worried about Markov chains!
this moment, the white rabbit cried alice opened the fall never heard it, i wonder?"
said, quiet thing sat down in another moment she went straight on, to herself.
Of course! But there is a chance it could go some way towards explaining language.But, take your LLM to a cocktail party and ask it to describe the room, and analyze the social relationships in it. Uh oh! Maybe a purely linguistic corpus does not fully explain human functioning.
I totally agree that the philosophy behind Chomsky's UG is wrong... I'm for cognitive linguistics too. But when you refute the weakest and silliest part of a theory, you do not disprove the whole theory. There are cognitivist versions of formalism too-- Lakoff's work, for instance.chris_notts wrote: ↑Thu Mar 16, 2023 4:16 amTrue, but isn't this just a further argument against the Chomskian school? They're the ones trying to separate the core of language itself from the rest of cognition and the embedded nature of the speaker. If you accept that the purpose and use of language is strongly intertwined with every aspect of its design and that there's no hard boundaries between grammar, lexicon, pragmatics etc. then surely formalism and the Chomskian programme is dead.zompist wrote: ↑Thu Mar 16, 2023 4:06 am Though we don't have UG as Chomsky envisions it, we do have a hundred million years of evolutionary history as animals functioning in the world. We didn't evolve to randomly generate plausible-sounding text; we did evolve to see and interact with things in the world.
(And yes, I'm aware that you could add visual data, mechanical sensors, motor devices, etc, etc. But then it's not just a language model, is it?)
You say this, but it hasn't stopped him and his ideas dominating much of the field for decades, despite the fact that everything we know about how evolution and cognition work would suggest he was talking rubbish. It's depressing how insular and anti-empirical much of linguistics has managed to be.Moose-tache wrote: ↑Thu Mar 16, 2023 4:30 am chris_notts: "True, but isn't this just a further argument against the Chomskian school?"
We don't need more arguments against Chomsky. His school of linguistics is absurd on its face and always has been.
This is potentially a good counterargument to using LLMs as evidence in the debate about the nature of language.But the fact that AI can create lucid bullshit doesn't do anything to further disprove his ramblings, since an AI generates lucid bullshit by applying superhuman amounts of computing power to the problem. You might as well say the room of infinite monkeys and infinite typewriters disproves this or that linguistic theory.
Didn’t notice this comment earlier. Let me post this in response (and as a plain interesting piece of work in and of itself): PaLM-E: An Embodied Multimodal Language Model.
Did you seriously stop reading before you got to level 3 generators? Yes, amazingly, ChatGPT does better than looking at two-word sequences in a 25,000 word text.bradrn wrote: ↑Thu Mar 16, 2023 4:29 am Not really, no. Output like this:…simply cannot be compared to what current LLMs can produce. In general, the latter seems indistinguishable (or at least very nearly so) from human utterances — there is a reason universities are panicking about ChatGPT-produced essays but never worried about Markov chains!
Chomsky's ideas on UG are about 1% of his work on linguistics and 0% of the useful part. Don't talk rubbish yourself.chris_notts wrote: ↑Thu Mar 16, 2023 4:44 amYou say this, but it hasn't stopped him and his ideas dominating much of the field for decades, despite the fact that everything we know about how evolution and cognition work would suggest he was talking rubbish. It's depressing how insular and anti-empirical much of linguistics has managed to be.Moose-tache wrote: ↑Thu Mar 16, 2023 4:30 am chris_notts: "True, but isn't this just a further argument against the Chomskian school?"
We don't need more arguments against Chomsky. His school of linguistics is absurd on its face and always has been.
Er, yes, it appears I did. Sorry.zompist wrote: ↑Thu Mar 16, 2023 4:53 amDid you seriously stop reading before you got to level 3 generators?bradrn wrote: ↑Thu Mar 16, 2023 4:29 am Not really, no. Output like this:…simply cannot be compared to what current LLMs can produce. In general, the latter seems indistinguishable (or at least very nearly so) from human utterances — there is a reason universities are panicking about ChatGPT-produced essays but never worried about Markov chains!
I’d actually be quite interested in seeing this. Has anyone attempted it? I feel doubtful it would produce anything near the level of ChatGPT, but it’s hard to know until I’ve seen it.How well would a higher-level Markov generator do if it was trained on 300 bllion words?
I do think ChatGPT is a remarkable achievement, but by this time it's also remarkably overhyped.
When you have a mega-mondo database of English, then you can do some absolutely stunning statistical feats. You can do them better with LLM but you can do them with Markov too.
It's a foundational idea! By accepting it, and then focusing most of your area's research on a small number of big well known languages (often the ones the author speaks) while making strong universality assumptions, you systematically bias your output as a field. What positive contribution of Chomsky's do you think outweighs the damage?zompist wrote: ↑Thu Mar 16, 2023 4:56 amChomsky's ideas on UG are about 1% of his work on linguistics and 0% of the useful part. Don't talk rubbish yourself.chris_notts wrote: ↑Thu Mar 16, 2023 4:44 amYou say this, but it hasn't stopped him and his ideas dominating much of the field for decades, despite the fact that everything we know about how evolution and cognition work would suggest he was talking rubbish. It's depressing how insular and anti-empirical much of linguistics has managed to be.Moose-tache wrote: ↑Thu Mar 16, 2023 4:30 am chris_notts: "True, but isn't this just a further argument against the Chomskian school?"
We don't need more arguments against Chomsky. His school of linguistics is absurd on its face and always has been.
To mention but one, as I recall he was one of the first people to popularise syntax as a legitimate and interesting area of study.chris_notts wrote: ↑Thu Mar 16, 2023 5:03 am What positive contribution of Chomsky's do you think outweighs the damage?
but that's the point, we indeed *don't* know how the mind works. we do not know what consciousness is made out of. my lil bro works in that field, and he knows we don't know. ask other neuroscientists, they also don't know. the consciousness research community gets excited at things like "look, i made this equation to calculate the complexity of the EEG signals from the brain, and they correlate to how conscious you are as you fall asleep or something". we have no idea what consciousness is made out of. they don't even have a good definition for consciousness. now, ofc, that don't mean we can't have computers get consciousness... but it also doesn't mean we can. what even is consciousness? how would you know if gpt-9 is feeling feelings or just producing the kinds of language that it thinks are consistent with feeling feelings?Moose-tache wrote: ↑Wed Mar 15, 2023 11:12 pm Multiple pages into to omnidirectional epistemological slap fight, and my original conclusion remains undisloged: If you think the logical functions performed by a computer cannot be a form of consciousness, you do not know what consciousness is made out of.
How can I know if you are feeling feelings or just producing the kinds of language that are consistent with feeling feelings?
No, it really isn't. If you read Syntactic Structures or even Aspects, there's nothing about it there. Generative grammar started as an attempt to formalize the rules of syntax, and it was explicitly agnostic about whether those rules were even in the brain or not. Chomsky used to relegate all such details to "performance", insisting he was only studying "competence." A syntactic theory was an attempt to describe how the syntax of a language worked; Chomsky's methods were easily applied and extended for 70 years now without anyone having to believe in UG.
This is an old charge against Chomsky; it had some validity in 1965, but it's nonsense today. When you throw an army of grad students at the world for 70 years, the field is no longer limited to the languages Chomsky knew (and let's not have any nonsense about "just English"; his dissertation is on Hebrew and he knows French fluently).By accepting it, and then focusing most of your area's research on a small number of big well known languages (often the ones the author speaks)
Demolishing Skinner's extremely reductive approach.What positive contribution of Chomsky's do you think outweighs the damage?