[ 3 / biz / cgl / ck / diy / fa / ic / jp / lit / sci / vr / vt ] [ index / top / reports ] [ become a patron ] [ status ]
2023-11: Warosu is now out of extended maintenance.

/sci/ - Science & Math


View post   

File: 31 KB, 544x408, sentence-structure.png [View same] [iqdb] [saucenao] [google]
6305914 No.6305914 [Reply] [Original]

/sci/, please take me back to when I was never taught this and teach me the math behind grammar and sentence structure

>> No.6305930

>>>/lit/

>> No.6305939

>>>/lit/

>> No.6305942

I bet you're a filthy prescriptivist.
Fuck off to >>>/lit/

>> No.6305944

>syntax trees

Ahahaha, have fun with scrambling languages with that approach, chumps.

>> No.6305977

The easiest way to model a grammar is through a context-free grammar: https://en.wikipedia.org/wiki/Context_free_grammars

Note that this only allow for very simple grammars and ususally has problems with gender/number concordance, although that's not a huge problem if you plan on using english only.

For more advanced grammars that allow for some context, I sugest that you pick a book on Natural Language Processing (and in Automata Theory, if you don't feel confortable with regular expressions/context free grammars).

>> No.6306390
File: 1.34 MB, 458x296, chomsky shrug.gif [View same] [iqdb] [saucenao] [google]
6306390

ha ha ha

>> No.6306409

>>6305914
Parse trees are good for a specific parse, but it's often times possible to parse a sentence in lots of different ways (draw lots of different trees) all with different meanings. Mainly, thee take-away I got out of parse trees while learning NLP is that languages are such shit that NLP is more or less impossible except through probabilistic means.

>> No.6306442

>>6305942
>parse tree
>prescriptivism
Do you even linguistics, bro?

>> No.6306449

Because human languages are terrible at following rules.
Seriously terrible.
Just like they're terrible at not being ambiguous.
Probably a symptom of them evolving slowly over time rather than being something that was designed exactly

>> No.6306467

>>6306409
>Mainly, thee take-away I got out of parse trees while learning NLP is that languages are such shit that NLP is more or less impossible except through probabilistic means.
You say that like probabilistic means are bad.

>> No.6306482

>>6305930
>>6305939
Fuck off back to your consciousness and 0.999... threads. Wikipedia:
>Linguistics is the scientific[1] study of language.[2]

>>6305914
>teach me the math behind grammar and sentence structure
https://en.wikipedia.org/wiki/Context_free_grammars#Formal_definitions

>> No.6306550

>>6306449
What about Esperanto?

>> No.6306562
File: 52 KB, 500x283, image.jpg [View same] [iqdb] [saucenao] [google]
6306562

>>6306550
Which countries uses it? That was an interesting idea but in practice it became as useful as elfic.

>> No.6306568

>>6306550
Esperanto's arguably WORSE.

http://www.xibalba.demon.co.uk/jbr/ranto/

>> No.6306587

>>6305944
The details of scrambling languages actually have a lot of interesting information to contribute to linguistics, and need to be explored more deeply. Turns out that when you get into the details of it, scrambling generally has certain restrictions, and these things can be played with to tease out really neat stuff.

>> No.6306606
File: 9 KB, 355x253, Head-subj-avm.png [View same] [iqdb] [saucenao] [google]
6306606

>>6305944
This is just one of the reasons why HPSG's are a God-tier formalism.

>> No.6306611

>>6305914
It is difficult to get into modern linguistics because many of the arguments are spread out over many different articles at different times.

For contemporary Chomskyan linguistics the reader Minimalist Syntax: Essential Readings has most of the main modern parts of generative grammar, along with The Minimalist Program itself.
Ed Stabler has written quite a bit about about programming them/putting them in logic.

Heim and Kratzner is a good book for getting into semantics and a sense of how it relates to the syntax. Also the work of Montague. There are many many theories in linguistics, but these have a lot of the technology used in "mainstream" syntax and semantics.

Aravind Joshi has been working on another computational approach TAG. Coming from type theory and Lambek calculus are other grammars related more directly to mathematics. These overlap with Chomsky grammars a lot. These come out of the work of Grishin, Lambek, Ajdukiewicz, Steedman, Chris Barker, Gentzen, Moortgat and others, which are in many ways related to the CS work of Curry and the Church lambda-calculus, as well as model theory, and categorial logic.

These are only a handful of related theories, and there are many others. Many probabilistic theories make use of parts of some of these structures in many ways, while stripping away other aspects to be taken care of by n-grams or other kinds of "intelligent" pattern-seeking put to probabilistic parameters. Almost no theory of syntax and semantics totally avoids the tools found in what I've mentioned, though there's a lot more because it is a very new and diverse field. I am also just restricting to what you have asked, which is for syntax/semantics.

>> No.6306618

>>6306611
>minimalist program
>not a bunch of bullshit that's only taken seriously because of Chomsky's earlier work

>semantics
>not highly speculative

I bet you believe in null constituents too

>> No.6306627

>>6306618
I'm not saying what I think is or isn't speculative, or even that the minimalist program is right*, but these books have some of the earlier work, and regularly reference back to them. I recommended them because they contain a lot of current stuff in context of old stuff.

(*I don't. But it has a very interesting relationship with many other theories of grammar that I think are compelling. All of this is the beginning of what will hopefully be a much broader theory, which will probably take at least a generation or two of revision.)

The other computational theories** cover some more abstract principles, and show relationships to lots of other theories of grammar.

(**These are related to very general mathematical statements, and most elementary formalisms of grammar.)

Null constituents might be a problem of notation..

>> No.6306629

Are there any constructed languages that are structured to have almost no ambiguity?

>> No.6306641

learn a case-system language with no word order. it's fun. Hungarian is probably the easiest, but I suggest Russian for the number of speakers.

>> No.6306680

>>6306629
http://attempto.ifi.uzh.ch/site/
http://en.wikipedia.org/wiki/Attempto_Controlled_English
>Attempto Controlled English (ACE) is a controlled natural language, i.e. a subset of standard English with a restricted syntax and a restricted semantics described by a small set of construction and interpretation rules.

>> No.6306728

>>6306627
>Null constituents might be a problem of notation..
It's a problem with conceptualisation. Syntacticians are overstepping their ground into semantics. They do it with wh-movement too.

>> No.6306984

>>6306467
It means that at best we can translate from one language to another probabilistically but we can never derive any real meaning from it.

>> No.6306999

doubleplusungood

>> No.6307006
File: 18 KB, 267x273, I don't think so tim.jpg [View same] [iqdb] [saucenao] [google]
6307006

>>6306984
>mfw

>>6306629
http://en.wikipedia.org/wiki/Lojban

>> No.6307982

>>6305914
>please take me back to when I was never taught this
why.avi

>> No.6308169

>>6307006
CS/NLP nerds are as dogmatic about their hopeless "formalism" as us "rule-based" linguists. Give up the ghost man, and let's go forward into this century.

>> No.6308175

>>6308169
also, probability models are essentially "meaningless" I don't know how you get around this; run statistics on some very surface distinct languages like English and Yoruba or one of the Kichaga languages and tell me how much of their semantics system you've learned.

>> No.6308791

bumpan

>> No.6309941 [DELETED] 

bump

>> No.6310081 [DELETED] 

bump

>> No.6310216

bump

>> No.6310223

>>6306409
You basically get a fuzzy set of correct meanings which make sense without context, then use context to narrow it down.

Then you realize that the most correct meaning won't always be correct for any finitely complex parser, which is why even humans make errors.

>> No.6310225

>>6306550
It was deliberately designed and therefore doesn't count as a natural language.

Many NLP-parsing machines use a modified version of Esperanto as an intermediary language.

>> No.6310314 [DELETED] 

bump

>> No.6310439 [DELETED] 

bump

>> No.6310516 [DELETED] 

bump

>> No.6310640

bump

>> No.6311955 [DELETED] 

.

>> No.6312064

>not using Combinatory categorial grammars
>2014

>> No.6312123

>>6306629
http://www.ithkuil.net/
With the added advantage of a script that looks like it was made by The Predator.

>> No.6312568

>>6312064
if ccg's are better, why are cfg's more common?

>> No.6312583

>>6306629
yes
http://en.wikipedia.org/wiki/Mathematics

>> No.6312581

>>6306611
Nobody sensible believes in Chomsky anymore though. GG as a whole has gone off the deep end somewhere in the late 80s.

>> No.6312585

>>6306641
Russian isn't a word order free language.

>> No.6312761

>>6306606
Enjoy your slow-as-shit parses!

>> No.6312976

>>6306984
>It means that at best we can translate from one language to another probabilistically but we can never derive any real meaning from it.
How so? Meaning may be probabilistic too.

>> No.6313017

>>6312064
Yes! (And I'm the 'Chomskyan'). Although generalizations found in categorial logic/MTLG have been more interesting to me recently, and fix a lot of the shortcomings of CCG in the most natural way (they also fit neatly in with MGs).

>>6312568
Doesn't know about equivalence proofs.

Also, again not that I wholeheartedly believe minimalism, but MG is just a MCFG, and most of the Chomsky-bashing I've read completely mischaracterizes modern syntax. None of it is that far fetched or even distinct from what almost all linguists use to some extent. There are gentzen presentations of MG which emulate it completely and are nearly identical to any mainstream formalism you can cook up.

>> No.6313044

>>6312568
To be more clear, cfg's are more common cause those are the straightforward scraps that trickled down to CS people. many linguistic phenomena seem to go beyond CFGs, and 'typed' combinators as in CCG, or 'multiple' dimensions for the CFG to work out over as in MCFGs are two of the most natural ways to expand CFGs. Depending on choice of combinators, CCGs usually push a CFG a little into the "mildly sensitive" region, while every MG is equivalent to some MCFG. Basically, the reason is CFGs are simpler so CS people like them, but linguistics realized it is better to move past them than try to brute force them to work, which may just be impossible.

>> No.6313062

>>6305944
>hasn't seen a syntax paper from the last 15 years
scrambling, at least not of the slavic and japanese flavors, seems to be unrestricted by any reasonable definition. In fact, the limitations come out from a very rich interaction of minimality principles (or particular mixing properties, etc.)

>> No.6313065

>>6313062
*does NOT seem to be unrestricted

>> No.6313148

>>6305914
I allays had trouble understanding grammar. Am I retarded?

>> No.6314026

>>6313065
>does NOT seem to be unrestricted
I'm going to need some citation on this.

>>6313148
You are retarded even among retards. People with William's syndrome, who average an IQ of roughly 70, have no problem speaking, and speaking fluently at that! You, sir, have no business here and I encourage you to promptly leave. Good day!

>> No.6314045

>>6306611
>Aravind Joshi has been working on another computational approach TAG. Coming from type theory and Lambek calculus are other grammars related more directly to mathematics. These overlap with Chomsky grammars a lot. These come out of the work of Grishin, Lambek, Ajdukiewicz, Steedman, Chris Barker, Gentzen, Moortgat and others, which are in many ways related to the CS work of Curry and the Church lambda-calculus, as well as model theory, and categorial logic.
Anon, no. Type theory is everywhere. Why here too?

>> No.6314047

>>6305914
>implying that isn't actually useful

>> No.6314049

>>6306984
>humans can do it
>machines can't
>humans are special

>> No.6314654

>>6314045
What's wrong with type theory?

>> No.6314666

>>6314654
Nothing, it's actually cool as fuck. I'm just waiting for the coming of typist type II.

>> No.6314813

>>6305914
... but I don't believe it.

>> No.6314816

>>6314026
Hm. I tried to dig a few up, but it turns out many of the papers which we've been reading in lab are not out yet, since it's a relatively recent research topic. The "seminal papers" which most work references are Saito 1992 and Saito 2006. There are a number of binding and quantificational (and I think some hyperraising? I don't study Japanese) restrictions. Recent work has gone into making the A/A' distinction come out of other principles, but they seem to not have the same (non-)restrictions as each other anyway.

>>6314045
>>6314654
I'm also confused. I'm not trying to convince anyone of type theory; it's a fact that it simply directly inspired some work in syntax and semantics which I listed there, seeing as OP was asking about math approaches to grammar. "MTLG" stands for multimodal type logical grammar - it straightforwardly is type logic as a grammar, and people like Oehrle and Moortgat have been working on it and relating it to the CCGs of Steedman.

>> No.6314932

>>6314813
>not believing in phrase structure grammars
>2014

>> No.6316165
File: 247 KB, 250x196, tell me more.gif [View same] [iqdb] [saucenao] [google]
6316165

>>6314816
>we've been reading in lab are not out yet
Who's "we?" What have you been reading?

>> No.6316474 [DELETED] 

bump

>> No.6316529

>mfw pseudoscience philistines are still trying to model natural language without any insight from the systems and evolutionary neuroscience of language

>> No.6316634

>>6316529
>pseudoscience philistines
Depends on the subfield in linguistics.

>still trying to model natural language
That's about right. Phrase structure grammars seem pretty close, but not quite perfect.

> without any insight from the systems and evolutionary neuroscience of language
0/10, apply yourself.

>> No.6316827

>>6316165
Would risk anon-iminity. I do something too specific in too small a field, and you could google me very easily. I'll say that some of it was Boskovic's more recent/inchoate work, which is very interesting. My advisor is a first generation Chomsky student (though I'm not Chomskyan, exactly) at one of the 10 major US ling schools.

>>6316529
0/10 We read 50+ pages of evolutionary science for my syntax course week one (not a ton, but a lot considering it's not course content), not to mention how much neuro/psych even us formalists have to take (1-2 years min + colloquia).

>> No.6316916

>>6305914

shit, we had a test on this today

>> No.6316936

>>6316827
> I do something too specific in too small a field, and you could google me very easily.
is this a may may, or do you just copy and paste the exact same this every time?

>> No.6318544 [DELETED] 

bump

>> No.6318996

bump