Re: Syntax random
Posted: Mon Jun 29, 2020 8:34 pm
I had to wait a good few years to get mine! And even then I had to ask for them.
I had to wait a good few years to get mine! And even then I had to ask for them.
Haha, thanks for the traditional (and I guess archaic by now?) welcome!Ser wrote: ↑Mon Jun 29, 2020 8:02 pmThank you for that!priscianic wrote: ↑Mon Jun 29, 2020 12:14 amThere actually is somewhat of a literature within a broadly Chomskyan framework on this kind of phenomenon (you might see it called "matrix que", since it's a que appearing in a matrix/main clause); for instance, Etxepare (2007), Demonte and Fernández Soriano (2014), and Corr (2016), to name just a few. I suspect the Corr might be especially interesting to conlangers, since it looks at these kinds of "matrix complementizers" in a (micro)comparative perspective, looking at variation in different varieties of Ibero-Romance.
By the way, there is a long tradition here... dating to 15 years ago or so... of welcoming people with images of some pickles and tea... I guess that tradition is getting well-worn and tired now, seeing that no one has yet done it. (!)
Another by the way, priscianic: have you read the final two books Priscian's grammar of Latin discussing syntax?
Really? Consider the following sequence of sentence:Tropylium wrote: ↑Mon Jun 29, 2020 9:21 pm (Plus, yes, I insist that languages are necessarily finite entities. They have a finite number of speakers who utter (or think, or parse) a finite number of words-per-minute with finite hearing and articulation fidelity over finite lifespans. Any theory that generates infinitely many "possible sentences" is infinitely wrong.)
I'm not sure I understand this argument. The whole idea about ungrammaticality judgments is to find the limits on what is possible. If you exhaustively know what's possible, then you also exhaustively know what's impossible, and likewise if you exhaustively know what's impossible, then you also exhaustively know what's possible (assuming that you can do some sort of excluded middle inference for possibility, such that there are no in-between states that are neither possible nor impossible).Tropylium wrote: ↑Mon Jun 29, 2020 9:21 pm It appears to me that most arguments in theoretical syntax have the form "sentence X is bad/ungrammatical, therefore we shall explain this happening because…".
Where does this come from? This seems to me to be unlike anything else in linguistics. Nobody thinks it needs a particular explanation if any given language such as English lacks any perhaps possible feature like a word fiffen, a locative case, a series of ejectives, a single lexeme expressing 'sea turtle', or a distinct grocery store clerk sociolect. Instead what we seek to explain / model are the features that do exist.
I think Tropylium’s issue is that we don’t seem to do this with any other part of language. If we see a corpus of a dead language, and the corpus lacks, say, the consonant /w/, we don’t feel any burning need to justify this. If a language seems to have no perfect aspect, then we don’t feel the need to exhaustively test every combination of words until we find it (or not). So why is it any different with syntax?priscianic wrote: ↑Mon Jun 29, 2020 10:00 pmI'm not sure I understand this argument. The whole idea about ungrammaticality judgments is to find the limits on what is possible. If you exhaustively know what's possible, then you also exhaustively know what's impossible, and likewise if you exhaustively know what's impossible, then you also exhaustively know what's possible (assuming that you can do some sort of excluded middle inference for possibility, such that there are no in-between states that are neither possible nor impossible).Tropylium wrote: ↑Mon Jun 29, 2020 9:21 pm It appears to me that most arguments in theoretical syntax have the form "sentence X is bad/ungrammatical, therefore we shall explain this happening because…".
Where does this come from? This seems to me to be unlike anything else in linguistics. Nobody thinks it needs a particular explanation if any given language such as English lacks any perhaps possible feature like a word fiffen, a locative case, a series of ejectives, a single lexeme expressing 'sea turtle', or a distinct grocery store clerk sociolect. Instead what we seek to explain / model are the features that do exist.
This issue comes up in trying to study dead languages. You might look at the attested corpora, and come up with a particular explanation for the data you see. Then you think about what kinds of predictions that theory makes, and you note that it predicts that a certain class of complicated sentences should be grammatical. However, you don't find any such sentences in the corpus, given their complexity, and thus you're at an impasse: you don't know for sure whether those sentences are grammatical or not, and thus whether your theory makes the right predictions or not. If you were working with speakers of a living language, then in this situation you would construct some sentences and contexts and elicit acceptability judgments.
And of course, the most interesting cases are those where one theory predicts that a certain set of sentences should be grammatical, but in actuality they aren't. And so you point out those cases, say "the correct theory needs to account for these judgments, and the old theory does not predict them", and then either revise the old theory or come up with a new one.
So I'm not sure I understand the problem with using negative data (i.e. judgments of unacceptability).
You won't ask questions like that until you have a theory that makes predictions, and you want to test those predictions. But once you have a theory that makes predictions, you might start asking seemingly mundane and uninteresting questions like that.bradrn wrote: ↑Mon Jun 29, 2020 10:31 pmI think Tropylium’s issue is that we don’t seem to do this with any other part of language. If we see a corpus of a dead language, and the corpus lacks, say, the consonant /w/, we don’t feel any burning need to justify this. If a language seems to have no perfect aspect, then we don’t feel the need to exhaustively test every combination of words until we find it (or not). So why is it any different with syntax?priscianic wrote: ↑Mon Jun 29, 2020 10:00 pmI'm not sure I understand this argument. The whole idea about ungrammaticality judgments is to find the limits on what is possible. If you exhaustively know what's possible, then you also exhaustively know what's impossible, and likewise if you exhaustively know what's impossible, then you also exhaustively know what's possible (assuming that you can do some sort of excluded middle inference for possibility, such that there are no in-between states that are neither possible nor impossible).Tropylium wrote: ↑Mon Jun 29, 2020 9:21 pm It appears to me that most arguments in theoretical syntax have the form "sentence X is bad/ungrammatical, therefore we shall explain this happening because…".
Where does this come from? This seems to me to be unlike anything else in linguistics. Nobody thinks it needs a particular explanation if any given language such as English lacks any perhaps possible feature like a word fiffen, a locative case, a series of ejectives, a single lexeme expressing 'sea turtle', or a distinct grocery store clerk sociolect. Instead what we seek to explain / model are the features that do exist.
This issue comes up in trying to study dead languages. You might look at the attested corpora, and come up with a particular explanation for the data you see. Then you think about what kinds of predictions that theory makes, and you note that it predicts that a certain class of complicated sentences should be grammatical. However, you don't find any such sentences in the corpus, given their complexity, and thus you're at an impasse: you don't know for sure whether those sentences are grammatical or not, and thus whether your theory makes the right predictions or not. If you were working with speakers of a living language, then in this situation you would construct some sentences and contexts and elicit acceptability judgments.
And of course, the most interesting cases are those where one theory predicts that a certain set of sentences should be grammatical, but in actuality they aren't. And so you point out those cases, say "the correct theory needs to account for these judgments, and the old theory does not predict them", and then either revise the old theory or come up with a new one.
So I'm not sure I understand the problem with using negative data (i.e. judgments of unacceptability).
(My gut feeling is that there’s something wrong with this argument, but I’m not entirely sure what it is…)
First, we do make negative arguments in other parts of linguistics. We have the entire idea of phonotactics, which is precisely defining what phonemic sequences are and are not possible. We define things like deponent verbs, which have parts of their paradigm "missing." Historical linguistics absolutely talks about losing features; if a proto-language has ejectives and a daughter does not, we ask what happened.Tropylium wrote: ↑Mon Jun 29, 2020 9:21 pm It appears to me that most arguments in theoretical syntax have the form "sentence X is bad/ungrammatical, therefore we shall explain this happening because…".
Where does this come from? This seems to me to be unlike anything else in linguistics. Nobody thinks it needs a particular explanation if any given language such as English lacks any perhaps possible feature like a word fiffen, a locative case, a series of ejectives, a single lexeme expressing 'sea turtle', or a distinct grocery store clerk sociolect. Instead what we seek to explain / model are the features that do exist.
I agree with you that infinite capacity is not actually theoretically required. But your statement that it's actually wrong is just as dogmatic and unsupported as Chomsky's.(Plus, yes, I insist that languages are necessarily finite entities. They have a finite number of speakers who utter (or think, or parse) a finite number of words-per-minute with finite hearing and articulation fidelity over finite lifespans. Any theory that generates infinitely many "possible sentences" is infinitely wrong.)
How would you go about generating a grammar of a programming language, e.g. C?Tropylium wrote: ↑Mon Jun 29, 2020 9:21 pm Plus, yes, I insist that languages are necessarily finite entities. They have a finite number of speakers who utter (or think, or parse) a finite number of words-per-minute with finite hearing and articulation fidelity over finite lifespans. Any theory that generates infinitely many "possible sentences" is infinitely wrong.
No, not all of these sentences are grammatical. I take "grammaticality" to mean that a real, living human will actually come to understand a sentence. Yet at minimum, if we repeat a sentence such as this for one thousand years — then no human cannot in principle even take in the entire sentence, much less process it and understand it to hold a particular meaning.bradrn wrote: ↑Mon Jun 29, 2020 9:32 pmConsider the following sequence of sentence:
I know that.
I know that you know that.
I know that you know that I know that.
I know that you know that I know that you know that.
I know that you know that I know that you know that I know that.
… ad infinitum
All these sentences are acceptable and grammatical, and there are an infinite number of them, so a theory to explain them must be able to generate infinitely many possible sentences.
This is a non-problem: I wouldn't. Programming languages are not natural languages and they do not have any mental grammars. No human natively writes C.
Granted. The difference I'm pointing at is that syntax starts with enormously permissive theories and seeks to falsify parts of them. Lexicology, phonology, semantics etc. generally don't: they start out by assuming nothing unwarranted and only add items to the model when they find evidence that they exist. The latter seems to me like the proper amount of epistemic caution.priscianic wrote: ↑Mon Jun 29, 2020 10:00 pmIf you exhaustively know what's possible, then you also exhaustively know what's impossible, and likewise if you exhaustively know what's impossible, then you also exhaustively know what's possible (assuming that …)
Why should we have any synchronic theories particular individual languages at all? Why is not not good enough to document what actually exists?priscianic wrote:This issue comes up in trying to study dead languages (…) you don't know for sure whether those sentences are grammatical or not, and thus whether your theory makes the right predictions or not.
I do not grant the background assumptions here: I have never seen any reason to assume a phenomenon of "movement" to exist at all.priscianic wrote:But people are asking why you get island effects with movement/displacement, because, given that in general you can have unbounded long-distance (i.e. cross-clausal) movement, you might expect this to be true in all cases. But it isn't!
I was offering you a simpler task.
They quickly enter this territory when it comes to phonetic syllable or word structure. I can remember coming upon a phonological description of Thai syllables which explained that why several words I knew to exist did not exist.Tropylium wrote: ↑Wed Jul 01, 2020 3:37 pm No phonological inventory makes any predictions, and no dictionary does either. We still have theories of phonology or semantics, but they do not make any a priori predictions about what features any one particular language should have — only about what feature some languages could have.
It's interesting to learn that German, Swedish and Sanskrit don't exist. Mind you, there are some old claims around that Sanskrit is a 'priestly fraud'.
You're being deeply uncharitable here. This is not what Tropylium is saying. He's saying that a set of compounds do not all suddenly spring into being without being uttered; the first one is coined, and then another based on that model, and then another based on some subset of those two, and so on, potentially by different speakers each time. Words do not exist independently of people who use them.
Again, this is just as dogmatic a claim as Chomsky's. Also, just as unnecessary.
And so what? It's quite possible and unremarkable for humans to create an artifact, or a text, that no one human can understand fully. It's barely possible to read the entire Talmud and its commentaries, but it takes half a lifetime. The Tibetan Buddhist canon is far longer in number of volumes at least (I'm not sure how long each is). No one could read everything in Wikipedia. Outside of texts, no one person could (say) master the entire corpus of law in a single state, or the entire multi-million-line code base for a very large scale project, or everything that's part of the standard model in physics. (Feynman did his damndest, but that was 40 years ago.)Yet at minimum, if we repeat a sentence such as this for one thousand years — then no human cannot in principle even take in the entire sentence, much less process it and understand it to hold a particular meaning.
I agree. This is a big complaint of mine about syntactic theories — they tend to start with an assumption and then seek to explain all language in terms of that assumption, when it really should be the other way around.Tropylium wrote: ↑Wed Jul 01, 2020 3:37 pmGranted. The difference I'm pointing at is that syntax starts with enormously permissive theories and seeks to falsify parts of them. Lexicology, phonology, semantics etc. generally don't: they start out by assuming nothing unwarranted and only add items to the model when they find evidence that they exist. The latter seems to me like the proper amount of epistemic caution.priscianic wrote: ↑Mon Jun 29, 2020 10:00 pmIf you exhaustively know what's possible, then you also exhaustively know what's impossible, and likewise if you exhaustively know what's impossible, then you also exhaustively know what's possible (assuming that …)
Yes, but it also removes all risk of explaining anything. There is an essential difference between word-formation and syntax: speakers of natural languages generally do not create new words as they talk (onomatopoeia and productive compounding excepted), but do continually produce novel sentences which have never been heard before. And listeners can clearly understand these novel sentences, so there must be some rule somewhere specifying how these sentences get interpreted. At the same time, some sentences — e.g. ‘on cat the sitting see mat I the’ — are clearly unintelligible, and get rejected as nonsense, so clearly there must be some rule which they are violating. The question syntax asks is: what, exactly, are these rules? And how precise can we make them? By contrast, if I am reading your post correctly, your approach seems to say ‘some sentences get produced, others don’t, no need to complicate the situation any further’ — an approach which is correct in its own way, but fails to have any predictive power.I do not grant the background assumptions here: I have never seen any reason to assume a phenomenon of "movement" to exist at all.priscianic wrote:But people are asking why you get island effects with movement/displacement, because, given that in general you can have unbounded long-distance (i.e. cross-clausal) movement, you might expect this to be true in all cases. But it isn't!
In terms of purely ground-level observable facts, what is being claimed in cases like these seems to be that e.g. sentences A B C, A B D and E C A B exist, but E D A B does not. Such a fact seems to be as much of a non-problem to me as the fact that the English lexicon has the words cat, bat and copycat but not ˣcopybat. Compounds are/have been created independently one at a time, only when needed. At no point in history has there occurred a mass Cartesian product of some sets of nouns bringing n·m compounds into some language "immediately".
Adopting the same approach in syntax would immediately remove all risk of ever overgenerating anything.
To be honest, I'm not quite seeing what you're seeing here. The kinds of complaints I usually see about formal/theoretical/generative syntax are things (strawmen) like: it analyzes everything "as if it's English", and that it can't account for the diversity of human language—in other words, that it's too restrictive, not that it's too permissive. Even bradrn, in their reply to this point, says "they tend to start with an assumption and then seek to explain all language in terms of that assumption, when it really should be the other way around" (which doesn't seem to be actually agreeing with your point here, unless I'm misreading things? It seems like bradrn is complaining about syntactic theories being too restrictive, rather than too permissive? Maybe I'm misunderstanding something).Tropylium wrote: ↑Wed Jul 01, 2020 3:37 pmGranted. The difference I'm pointing at is that syntax starts with enormously permissive theories and seeks to falsify parts of them. Lexicology, phonology, semantics etc. generally don't: they start out by assuming nothing unwarranted and only add items to the model when they find evidence that they exist. The latter seems to me like the proper amount of epistemic caution.priscianic wrote: ↑Mon Jun 29, 2020 10:00 pmIf you exhaustively know what's possible, then you also exhaustively know what's impossible, and likewise if you exhaustively know what's impossible, then you also exhaustively know what's possible (assuming that …)
Am I understanding you right here? Are you trying to make the normative claim that we shouldn't study dead languages with the tools of synchronic linguistics? One, you're gonna need to do a lot more convincing to get me to accept that, and two, I don't see how it's relevant to my point about how negative evidence is useful for theory-building.Tropylium wrote: ↑Wed Jul 01, 2020 3:37 pmWhy should we have any synchronic theories particular individual languages at all? Why is not not good enough to document what actually exists?priscianic wrote:This issue comes up in trying to study dead languages (…) you don't know for sure whether those sentences are grammatical or not, and thus whether your theory makes the right predictions or not.
I don't know what kinds of phonology or semantics you're familiar with, but it's certainly not the kind of phonology or semantics I'm familiar with. Phonology is not just phoneme inventories (some phonologists don't even believe in existence of phonemes, much less in the existence of phoneme inventories), and semantics is not just about dictionaries and lexical semantics (plenty of semanticists basically just gloss over lexical semantics entirely).Tropylium wrote: ↑Wed Jul 01, 2020 3:37 pm No phonological inventory makes any predictions, and no dictionary does either. We still have theories of phonology or semantics, but they do not make any a priori predictions about what features any one particular language should have — only about what feature some languages could have.
The background assumptions aren't the point here—you can replace it with whatever floats your boat. The point is that sometimes you might expect certain things to be grammatical (for whatever reason), and they aren't—and that naturally leads you to wonder why. If you don't ever experience these kinds of expectations and questions, I'm not sure what I can say. To me, questions like "why is the world one way, and not the other?" are natural and interesting questions that curious people ask, the kinds of questions that drive scientific inquiry. I'm not sure I can personally grok how someone wouldn't have these same kinds of questions about human language, and I think that's a core part of why I'm so confused by your responses here, because they seem to be trying to deny that these kinds of questions are interesting and worth asking.Tropylium wrote: ↑Wed Jul 01, 2020 3:37 pmI do not grant the background assumptions here: I have never seen any reason to assume a phenomenon of "movement" to exist at all.priscianic wrote:But people are asking why you get island effects with movement/displacement, because, given that in general you can have unbounded long-distance (i.e. cross-clausal) movement, you might expect this to be true in all cases. But it isn't!
Good catch in pointing out that my complaint contradicts Tropylium’s complaint; I have no idea how I didn’t notice that when I wrote it.priscianic wrote: ↑Thu Jul 02, 2020 3:10 amTo be honest, I'm not quite seeing what you're seeing here. The kinds of complaints I usually see about formal/theoretical/generative syntax are things (strawmen) like: it analyzes everything "as if it's English", and that it can't account for the diversity of human language—in other words, that it's too restrictive, not that it's too permissive. Even bradrn, in their reply to this point, says "they tend to start with an assumption and then seek to explain all language in terms of that assumption, when it really should be the other way around" (which doesn't seem to be actually agreeing with your point here, unless I'm misreading things? It seems like bradrn is complaining about syntactic theories being too restrictive, rather than too permissive? Maybe I'm misunderstanding something).Tropylium wrote: ↑Wed Jul 01, 2020 3:37 pmGranted. The difference I'm pointing at is that syntax starts with enormously permissive theories and seeks to falsify parts of them. Lexicology, phonology, semantics etc. generally don't: they start out by assuming nothing unwarranted and only add items to the model when they find evidence that they exist. The latter seems to me like the proper amount of epistemic caution.priscianic wrote: ↑Mon Jun 29, 2020 10:00 pmIf you exhaustively know what's possible, then you also exhaustively know what's impossible, and likewise if you exhaustively know what's impossible, then you also exhaustively know what's possible (assuming that …)
I don't see how what you are saying about words differs from sentences in these languages, except in degree. Also, where compound verbs are freely formed in highly inflected languages, you automatically get an immense number of derivative words. Such forms generally exist independently of whether someone has ever uttered them. Now, English is not so free in generating compound words from nouns and similar parts of speech as the three languages I mentioned. While 42-legged was English before I independently coined it this morning, it is debatable as to how many words it constitutes. When it comes to the long Sanskrit compounds that serve the same purpose as clauses in other languages, I don't think the term 'coining' is appropriate.KathTheDragon wrote: ↑Wed Jul 01, 2020 7:50 pmYou're being deeply uncharitable here. This is not what Tropylium is saying. He's saying that a set of compounds do not all suddenly spring into being without being uttered; the first one is coined, and then another based on that model, and then another based on some subset of those two, and so on, potentially by different speakers each time. Words do not exist independently of people who use them.
I do not think this follows. If "copybat" is rejected as not being a word, this doesn't mean that there has to exist a rule that it violates — it merely means that it is not an existing word. It very well could exist, if there was a referent for it.
Well, I do pare this back a little bit. I am aware that many novel sentences can be easily analyzed as having an already known structure that's simply dressed up with different choices of lexemes, and these analyses will predict many other similar tamely novel ("lexically novel") sentences. "Nameless blue nodes sit stringently" and all that pop punk.bradrn wrote:By contrast, if I am reading your post correctly, your approach seems to say ‘some sentences get produced, others don’t, no need to complicate the situation any further’ — an approach which is correct in its own way, but fails to have any predictive power.
No, I don't mean that there's anything bad with this if you already have ended up with an expectation… The original observation I made is that this seems to be the only common workflow for bringing syntactic theories in better accord with languages as they exist in reality. And I mean not just the question, but also the answer, which invariably is something roughly "here's a constraint that prevents this from being grammatical".priscianic wrote:The point is that sometimes you might expect certain things to be grammatical (for whatever reason), and they aren't—and that naturally leads you to wonder why.
"Why" is always a good question, but it is fundamentally a question about diachrony. "Everything is the way it is because it got that way."priscianic wrote:To me, questions like "why is the world one way, and not the other?" are natural and interesting questions that curious people ask, the kinds of questions that drive scientific inquiry.
Not for any sense of "exist" that is supposed to be on the same level as forms that are uttered.
Explaining novel sentences is one of the things generative grammars do. Another thing I see them doing though is trying to reduce sentences as being generated from other sentences or archetypes or operations (rather than simply taken from memory) even when they aren't novel at all.Richard W wrote:My feeling is that novel sentences are generated on the model of other sentences. Generative grammars are an approximation to this process.