Constraint grammarConstraint grammar (CG) is a methodological paradigm for natural language processing (NLP). Linguist-written, context-dependent rules are compiled into a grammar that assigns grammatical tags ("readings") to words or other tokens in running text. Typical tags address lemmatisation (lexeme or base form), inflexion, derivation, syntactic function, dependency, valency, case roles, semantic type etc. Each rule either adds, removes, selects or replaces a tag or a set of grammatical tags in a given sentence context. Context conditions can be linked to any tag or tag set of any word anywhere in the sentence, either locally (defined distances) or globally (undefined distances). Context conditions in the same rule may be linked, i.e. conditioned upon each other, negated, or blocked by interfering words or tags. Typical CGs consist of thousands of rules, that are applied set-wise in progressive steps, covering ever more advanced levels of analysis. Within each level, safe rules are used before heuristic rules, and no rule is allowed to remove the last reading of a given kind, thus providing a high degree of robustness. The CG concept was launched by Fred Karlsson in 1990 (Karlsson 1990; Karlsson et al., eds, 1995), and CG taggers and parsers have since been written for a large variety of languages, routinely achieving accuracy F-scores for part of speech (word class) of over 99%.[1] A number of syntactic CG systems have reported F-scores of around 95% for syntactic function labels. CG systems can be used to create full syntactic trees in other formalisms by adding small, non-terminal based phrase structure grammars or dependency grammars, and a number of Treebank projects have used CG for automatic annotation. CG methodology has also been used in a number of language technology applications, such as spell checkers and machine translation systems. Rule syntax and formatA Constraint Grammar parser expects as its input a stream of morphologically analysed tokens, typically produced by a Finite-state transducer-based analyser (common ones are the Xerox tools twolc/lexc/xfst, HFST or Apertium's lttoolbox). Each token may be ambiguous and have many readings, the surface form with all its readings is called a cohort. Below is a possible example analysis of ", and X was like “" in the input format expected by VISL CG-3: "<,>" "," cm "<and>" "and" conj "<X>" "X" num pl "X" noun prop "<was>" "be" verb past p1 sg "be" verb past p3 sg "<like>" "like" adj "like" subj "like" pr "like" verb inf "like" verb pres "like" verb imp "<“>" "“" lquot This snippet shows 5 cohorts, each with one or more readings. The surface wordforms are in Below is an example rule (again in VISL CG-3 format) for picking the third person reading of "was" (by removing the first person reading), given that there is no first person pronoun to the left: REMOVE (verb p1) IF (0C (verb)) (NEGATE *-1 (prn p1)) ; Here In CG-3, rules may also be given names, e.g. A rule can also pick a single reading if we're sure that all the other readings must be wrong given the constraints: SELECT:quoting ("like" subj) IF (-1 ("<was>")) (1 (lquot) OR (":")) ; In this rule, we see that we can refer to both wordforms and baseforms in tagsets (they are treated just like any other tag, and a reading will always match its wordform). Here the second constraint uses LIST prequote = lquot ":" ; SELECT:quoting ("like" subj) IF (-1 ("<was>")) (1 prequote) ; An equivalent definition would be After running the above rules, we should end up with this: "<,>" "," cm "<and>" "and" conj "<X>" "X" num pl "X" noun prop "<was>" "be" verb past p3 sg "<like>" "like" subj "<“>" "“" lquot If we used
LIST nominal = noun prn ; ADD (@SUBJ) IF (NEGATE *-1 nominal) (0C (prop)) (1C finiteverb) ; This is called a "mapping rule", and we may end up with multiple such mapping tags per cohort, in which case we can disambiguate using the same SELECT/REMOVE rules. ImplementationsCG-1The first CG implementation was CGP by Fred Karlsson in the early 1990s. It was purely LISP-based, and the syntax was based on LISP s-expressions (Karlsson 1990). CG-2Pasi Tapanainen's CG-2 implementation mdis[2] removed some of the parentheses in the grammar format and was implemented in C++, interpreting the grammar as a Finite State Transducer for speed. CG-2 was later reimplemented (with a non-FST method) by the VISL group at Syddansk Universitet as the open source VISL CG [1], keeping the same format as Tapanainen's closed-source mdis.
CG-3![]() ![]() The VISL project later turned into VISL CG-3, which brought further changes and additions to the grammar format, e.g.:
There is also simple IDE for CG-3 developed by VISL,[2] which provides syntax highlighting and lets you see input and output and possible errors as you write your grammar. There is also an Emacs mode Unlike the Tapanainen implementation, the VISL implementations do not use finite state transducers. Rules are ordered within sections, which gives more predictability when writing grammars, but at the cost of slower parsing and the possibility of endless loops. There have been experimental open-source FST-based reimplementations of CG-2 that for small grammars reach the speed of VISL CG-3, if not mdis.[3] List of systems
References
External links
|
Portal di Ensiklopedia Dunia