Generative linguistics is pretty isolated these days... chugging along but using a set of methodologies and premises that are pretty far removed from the rest of academia. Generative linguistics generally had trouble in two key areas 1) first language acquisition (where increasingly ornate innate machinery is needed to explain how kids arrive at the 'right' grammar) and 2) language processing, where the structural representations were never great at predicting processing difficulty, nor implementable in a machine. Consequently, these two fields have long been in psychology (and the reason I'm a psych graduate student, of all things!), and draw heavily on statistics cognitive science, NLP and in the latter case, information theory.
Chomsky is brilliant, no doubt, and critically showed the world how much latent structure there is in language. However, he is IMHO pretty wrong in thinking that this latent structure can't be learned (e.g. inferring a probabilistic context-free grammar).
One thing to be said for generative linguistics is that a whole lot of super-interesting phenomena have only been characterized within specific generative frameworks--language documentation and psycholinguistic research can only progress with a formal system in which language can be rigorously characterized. So one of the huge projects in the future will be translating the 50+ years of research into the formalisms in the modern computer science / statistical NLP / Bayesian cognitive science / learning-centric developmental psychology stack.
>a set of methodologies and premises that are pretty far removed from the rest of academia.
The rest of academia has agreed on a set of methodologies and premises? No wonder we feel left out!
>increasingly ornate innate machinery is needed to explain how kids arrive at the 'right' grammar
This is too vague to respond to. But theories do tend to get more complex as we learn more, since there's more data to account for.
>where the structural representations were never great at predicting processing difficulty, nor implementable in a machine.
Which structural representations assumed by (say) GB theory are not implementable in a machine?
>he is IMHO pretty wrong in thinking that this latent structure can't be learned (e.g. inferring a probabilistic context-free grammar).
Chomsky always agreed that it was possible to learn context-free grammars via statistical methods. That's why he placed such a great emphasis in the 60s on showing that context-free grammars are not a suitable model for natural language.
Your last paragraph is fairly astonishing, insofar as it admits that generative linguistics has obtained lots of interesting results which cannot be characterized in "modern" terms. That sounds like a pretty strong indication that generative linguistics has got something right!
> Which structural representations assumed by (say) GB theory are not implementable in a machine?
Most of the individual constraints proposed in the myriad papers on GB and Minimalism are probably implementable by machine. But no one in Chomskyan generative syntax seems interested in explicitly spelling out the full set of principles that would underly a large-coverage grammar--except maybe Steedman's Minimalist Grammar formalism, which is ignored by most people who call themselves "syntacticians". In contrast, the HPSG and LFG communities have attempted to provide large-scale grammars and a lot of NLP work has used them in a serious way, but those communities are no longer very active.
Is there a work which lays out modern Minimalist generative syntax in full formal detail, and shows how this formal system handles a very large range of different syntactic phenomena? Such that it would be possible to produce a large hand-parsed corpus? It seems like this is what would be needed for generative syntax to have relevance outside linguistics departments. If it exists and I'm just unaware of it, I'd be glad to hear about it!
I'm not sure why you're asking this about Minimalism specifically. There are already wide coverage parsers based on various generative frameworks (e.g. HPSG, LFG, CCG).
It's also a bit odd to suggest that any framework for which there isn't a wide coverage parser lacks any relevance outside linguistics departments. I know of lots of examples of cross-disciplinary work involving generative linguistics, but most of it doesn't relate to parsing at all. I'd say that wide coverage parsers are actually a pretty niche interest, which is one reason why people don't tend to work on them much.
>However, he is IMHO pretty wrong in thinking that this latent structure can't be learned (e.g. inferring a probabilistic context-free grammar).
Please correct me if I'm wrong but your representation of Chomsky's thesis seems to be the opposite of what it is.
Chomsky says that the ability to learn a language is innate - you don't need to go to school to do it, nor do you need to learn grammar rules.
You say that Chomsky says that the "latent structure in language" cant be learned and if my representation is correct, then yours is way off and seems partisan.
> Chomsky is brilliant, no doubt, and critically showed the world how much latent structure there is in language. However, he is IMHO pretty wrong in thinking that this latent structure can't be learned (e.g. inferring a probabilistic context-free grammar).
Presumably you have a critique of his poverty of stimulus argument and not merely a blanket dismissal?
The simple critique of the poverty of the stimulus argument is that there's no poverty of the stimulus: children receive ample, rich linguistic input as well as useful feedback. Children receive a lot more negative evidence than Chomsky suggests. https://en.wikipedia.org/wiki/Poverty_of_the_stimulus#Agains... Is Chomsky's argument for the poverty of the stimulus based on empirical observations of child language acquisition?
Chomsky is brilliant, no doubt, and critically showed the world how much latent structure there is in language. However, he is IMHO pretty wrong in thinking that this latent structure can't be learned (e.g. inferring a probabilistic context-free grammar).
One thing to be said for generative linguistics is that a whole lot of super-interesting phenomena have only been characterized within specific generative frameworks--language documentation and psycholinguistic research can only progress with a formal system in which language can be rigorously characterized. So one of the huge projects in the future will be translating the 50+ years of research into the formalisms in the modern computer science / statistical NLP / Bayesian cognitive science / learning-centric developmental psychology stack.