Dipanjandas.com

An Affinity Based Greedy Approach towards
Chunking for Indian Languages
Dipanjan Das, Monojit Choudhury, Sudeshna Sarkar and Anupam Basu
Department of Computer Science and Engineering Indian Institute of Technology, Kharagpur email: [email protected], {monojit, sudeshna, anupam}@cse.iitkgp.ernet.in Abstract
Abney (1991) viewed chunks to be connected sub-graphs of the parse tree of a sentence. For example, The young girl was sitting on a bench.
language text. Chunking for Indian lan-guages require a novel approach because of the relatively unrestricted order ofwords within a word group. A compu- [The young girl]/NP [was sitting]/VP [on a on valency theory and feature structures where NP, VP and PP are noun, verb and preposi- draws an analogy of chunk formationin free word order languages with the Several methods for chunking have been explored for English and other European languages since the 90’s. (Abney, 1991) introduced a context-free unavailability of large annotated corpora grammar based approach for chunking of English forces one to adopt a statistical approach focussed on statistical techniques. Machine learning based approaches eliminate the need of manual rule creation and is language independent. (Ramshaw and Marcus, 1995) proposed a transformation based learning paradigm for text chunking, whichachieved approximately 92% precision and recallrates. This work involved text corpora as training Introduction
examples where the chunks were marked manually.
Other works report more than 90% accuracy in Chunking or local word grouping is an intermedi- identifying English noun, verb, prepositional, ad- ate step towards parsing and follows part of speech jective and adverbial phrases (Buccholz et al, 1999; tagging of raw text. Other than simplifying the pro- Zhang et al, 2002). The work of (Vilain and Day, cess of full parsing, chunking has a number of appli- 2000) in identifying chunks is worth mentioning cations in problems like speech processing and in- here because their system used transformation based formation retrieval. Chunking has been traditionally rules which would either be hand-crafted or trained.
defined as the process of forming groups of words The trained system performed with precision and based on local information (Bharati et al, 1995).
recall rates of 89% and 83% respectively, whereas they attained much lower precision and recall rates this description may have a nested structure as well.
in the hand-engineered system, and hence their The following sentence elucidates some chunks for observations justified the use of the learning based and context sensitive rules for part of speech tag sequences to chunk sentences. The rules in their work were also derived from training data.
wood[of] table[on] very beautiful flower-vase was The problem of chunking for Indian languages (There was a very beautiful flower-vase on takes a new dimension for a two reasons.
for most free-word order languages, the internalstructure of the chunks often have an unrestrictedorder, for which chunk identification becomes a In sentence 3, kAThera TebilaTira upara forms a challenging task in comparison to English text.
noun chunk (NN) which has two chunks within it - However, full parsing of free-word order languages, a noun chunk and a postposition (PPI). In the noun which is a hard task, becomes much easier to solve chunk khuba sundara phuladAni, there is an adjecti- when the intermediate step of identifying chunks val chunk (JJ) khuba sundara and a noun phuladAni is achieved. Second, there is a dearth of annotated nested inside. It should be noted that due to lack of strong word ordering constraints in Bengali, the number of sentences used for training an NLP chunks can be freely permuted in the above sentence system for languages such as English goes beyond to yield grammatically correct sentences having the a few millions. In the scenario of Indian languages, same meaning. However, words may not be per- such colossal amount of training data is hardly muted across the chunks. For example, the sentence sundara TebilaTira upara khuba kAThera phuladAni Chila.
beautiful table[of] very wood[of] flower- An approach towards chunking based on valency theory and feature structure unification has been described here. An analogy of unification of FSs of syntactic units has been drawn with the formationof complex structures from atoms and radicals. Inthe formalism reported the parse tree of a sentence has further been imagined to be similar to the syntax, but a different semantics. Thus informally, molecular structure of a substance. A chunker for a chunk or a word group is a string of words inside Bengali has been implemented using this formalism which no other word of the sentence can enter with a precision of around 96% and a recall of without changing the meaning of the sentence. The approximately 98%. In the next section we describe present paper assumes this definition to identify the nature of chunk formation in Bengali.
chunks for a particular input sentence.
The problem of dividing a Hindi sentence into Chunking in Free Word Order
word groups has been explored by (Bharati et al, Languages
1995). The output of the grouper served as an inputto a computational Paninian parser. However, in In this paper, chunking, especially with a view to- their work, they made a distinction between local wards free word order languages, is defined to be the word groups and phrases. They assert from a com- process of identifying phrases in a sentence that have putational point that the recognition of noun phrases an internal arrangement independent of the over- 1In this paper, Bengali script has been written in italicized all structure of the sentence. Chunks according to Roman fonts following the ITRANS notation.
and verb phrases is neither simple nor efficient.
Therefore, the scope of the grouper was limitedand much of the disambiguation task was left tothe parser.
in Hindi has been explored by (Ray et al, 2003)using a rewrite rule based approach.
technique gave only one possible grouping of aparticular sentence, did not consider sentence levelconstraints, which at times generated ill-formedgroups and the case of noun-noun groups has beencompletely ignored.
Formalism
Overview of the Framework
The present paper focuses on a formalism devel-oped using an analogy with formation of bonds in Figure 1 shows the framework proposed in this chemical compounds. The criteria that governs the work. The main components of the system are the series of bond formations is the minimization of POS tagger, the chunking system and a search space energy through which a stable state is reached. Two atomic units can form a bond if and only if they and tags each word with its POS category and other have a mutual attraction towards each other, which morphological attributes valid for that category.
is governed by the property called valency.During The grouper tries to formulate valid groups based bond formation of two structures, there is an amount on the POS tags and the morphological attributes of energy that is released and this brings the system as provided by the tagger. The first phase of the to a more stable state. Ultimately, when the system grouping is the feature structure instantiation of cannot release any more energy, the final structure each and every word of the sentence followed by the grouping algorithm. At each step of grouping,a search space pruner reduces the number ofredundant groups that may have been formed by the We can extend this concept to chunking (or pars- chunking system. The current work focusses only ing in general) as follows: The individual words of a sentence are the atoms, which have valencies rep-resentative of their syntactic and/or semantic prop-erties. The chunks can be viewed as semi-stable Valency and Feature Structures
molecular complexes formed out of a set of atoms,which in turn have its own valencies. A quantity Valency theory takes an approach towards an called the bond energy reflects the relative affinity analysis of sentences that focuses on the role of between two syntactic units. The higher the bond energy between two units, the higher is the prob- certain verbs allow only certain kinds of predicates ability that they will form a chunk. The complete some of which are compulsory while some others parse-tree of the sentence is analogous to the final may be optional. The arguments that a verb can chemical compound that is formed. The concepts of take are defined in terms of its valency.
valencies and feature structures have been used to introduced by French structuralist Lucien Tesniere realize a chunker in the lines of the aforementioned in dependency grammar formalism. Numerous the- analogy. The following subsections examine the de- oretical publications (Helbig, 1992; Welke, 1995) velopment of this formalism in detail.
have established it as arguably the most widely used model of complementation in languages like units to its left and right that is represented German, French, Romanian and Latin. Valency the- through the corresponding bond energy. Sec- oretic description of English has also been explored tion 3.3 provides further discussion on the theo- retical and practical issues concerning bond en- valency theory has focused on the syntactic and se- mantic valencies of verbs and occasionally of nouns.
Category is simply the POS category of the syn- tactic unit that the FS allows in the concerned A generalization of the valency theory has been proposed in this work, where instead of onlynouns or verbs, each syntactic unit of a sentence Attributes are the allowable attribute of the syntac- at every level of analysis has a certain number of tic category in the respective hand.
chunking, where only adjacent words or groups can N ewCategory - At every step of the chunking pro-
be attached to form larger groups, the units (word cess two feature structures unite to form a new or word groups) have a maximum of two valencies associated with them - the right and the left. The right valency defines the capability of the unit to N ewAttributes - Similarly, N ewAttributes are
form groups with the units that lie to its right in a the attributes of the FS obtained through uni- sentence. The left valency can be defined likewise.
However, the affinity of a unit to form groups also The FS corresponding to all the POS categories depends on the syntactic attributes of the adjacent and their different attributes is designed manually through linguistic analysis. However, in the pres- group with an adjective to its left, but never with ence of an annotated corpus, these templates can be of the adjacent units are also important. We use Bond Energy
feature structure (FS) to represent the syntacticaffinities of a syntactic unit in terms of the valencies and bond energies. Successful unification between [[kAT hera T ebilaT ira]NN [upara]P P I ]NN adjacent FS implies formation of a valid group with tion era, as in kAT hera, has an affinity or valency process releases an energy equal to the bond energy for a noun or noun group (NN) to its right. Simi- associated with that particular valency.
larly, a noun in the oblique form like T ebilaT iracan form a NN chunk with a post-position like The basic contents of the FS are shown in upara to its right. Therefore, the sequence of words kAT here T ebilera upara can have two possible morphological attributes, the FS also contains [kAT hera [T ebilaT ira [upara]P P I ]NN ]NN (left) hand specifies the constraints on the units to and the aforementioned one, out of which only the the right (left) of an object that can be grouped latter is syntactically valid. The concept of bond en- with it. The constraints are specified as a list of ergy helps us solve this particular problem. Suppose, 5-tuples <BondEnergy, Category, Attributes, the bond energy between NN and PPI is smaller than that between NN-possesive and NN. Then in our ap- elements of the 5-tuples can be described as follows: proach, the group between NN-pssesive and NN willform before NN and PPI, thus leading to the correctgrouping. Similarly, an adjective can connect to a BondEnergy - The formalism assumes a variable
noun to its right and a qualifier to its left. However, affinity for a particular syntactic unit towards in a sentence, when an adjective is followed by a Figure 2: Contents of a Feature Structure noun and preceded by a qualifier, the correct group- elements of BE. We are interested in a patial ing requires adjective to be grouped with the quali- ordering between the bond energies. Note that fier first, and then the new adjectival group is joined we need to compare BEij only with BEki, for with the noun. This is illustrated by the chunk all ks other than j, beacause an ambiguity oc- curs only when a particular category can form in example 3. Therefore, we can conclude that the bonds with the units both to its right and left.
bond energy between qualifier and adjective is more We call such elements as comparable pairs.
than the bond energy between an adjective and a Thorough linguistic analysis has led us to 47 noun.Thus, bond energy helps us decide which way unique values for the bond energies, based on which the bond will form (or in other words, units will be all possible ambiguous pairs can be compared. Note grouped) whenever there are multiple possibilities.
that it is not important to know what are these 47 val- Let there be N POS categories from C1 to CN .
ues, only the ordering is important. In other words, Let Ci · lef tBE(Cj) be the bond energy between these 47 distinct values have been chosen arbitrar- Ci and Cj when Cj is to the left of Ci.
ily from the set of natural numbers. It should be rightBE(Cj) can be defined likewise. We make mentioned here that a given sentence can be inher- ently ambiguous and lead to multiple valid group- ings. Since, we do not preclude the use of the same i · lef tBE(Cj ) is 0 if Ci cannot be grouped bond energy value for a comparable pair, multiple grouping emerge as a natural possibility. We illus- • Ci · lef tBE(Cj) is equal to Cj · rightBE(Ci).
trate this with the following examples.
• We need to estimate only N ×N bond energies, that can be stored in a matrix BE, such that ij = Ci · lef tBE(Cj ) = Cj · rightBE(Ci) • It is not necessary to assign exact values to the 1. The list of FSs is stored in an array F .
2. Initially, for every pair of FSs in the sentence, the bond energy between them is calculated andstored in a max-heap B.
3. The arrays F and B are stored in a global stack Thus, in order to retain the possibility of both the groupings, we must assign the same bond energiesto the pairs NN-possesive and NN, and NN and VN 4. The stack is popped to get a partial grouped list of FSs - denoted by B and F for this run.
The Grouping Algorithm
5. The maximum bond energy bmax is extracted from B and the two corresponding FSs x and y The grouping algorithm of the formalism proceeds by first instantiating the feature structures for eachand every word and the punctuation marks in a sen- 6. If only one such pair is present, then the bond follows a divide and conquer approach that eases the complexity of full parsing. The following sub- of the hands of either x or y. Let the FS after subsections concentrate on the details of the algo- unification be z. z is inserted into the array F .
Instantiation of Feature Structures
7. The BondEnergies between z and the FSs to After an input sentence sentence is fed to the POS its right and left which are denoted by bz tagger, the sentence comes as input to the chunking system with the POS tags and morphological 8. At step 5, if there are multiple values of b attributes attached with each of the words and for each of the values, Steps 6 and 7 are re- sentinels. The tagged sentence is fed into the initial instantiator. The steps of instantiation are: stored in the global stack S. The control flowstarts from Step 5 again.
1. Each word of the sentence is extracted one after tial grouped structure is stored in S and control 2. A rule file which assists in the process of in- stantiation consists of the template of FS for 10. If at Step 5, the value of bmax is zero, the a POS category and its attibutes. The present system uses inflection as the only syntactic at-tribute for chunking purposes.
3. For each word, a new instance for the FS is cre- Handling Conjunction
ated from the template, which is selected from A thumbrule for handling conjunctions and other the rule file based on the POS tag and the at- conditionals is applied before storing the configu- tribute information associated with the word.
ration B and F in the stack S in Step 11/12. The The Grouping Process
After the instantiation of the feature structures of 1. For all the FSs in F , it is checked whether each syntactic unit of the sentence, the grouping pro- cess starts a greedy process and proceeds in the fol- Attributesv>,<P OSw, Attributesw> exist in the list of tuples stored in a file for handling output from the chunker: hugalI/NN [nadIra [dui tIre]/NN]/NN kolakAtA/NN [ga.De uTheChe.]/VF 2. For all such sets of 3 FSs, the 3 FSs are united However, note that the correct grouping is [hugalI to form a single FS x with new category P OSx nadIra]/NN.poss [dui tIre]/NN]/NN [kolakAtA]/NN and attributes Attributesx, both of which is 3. After unification, the new FS x is inserted into Implementation and Results
the list F , and the bond energies of x withthe FSs to its right and left are inserted into B A chunker for Bengali has been implemented based maintaining the decreasing order of the array.
on the aforementioned framework. It depends on An Example
a morphological analyzer (MA) and a POS tagger,both of which have been developed in-house and Below we illustrate the process of chunking with an have accuracies of 95% and 91% respectively. The salient features of the chunker for Bengali are men- Input sentence:
hugalI nadIra dui tIre kolakAtA ga.De uTheChe.
Hoogly river’s two bank Kolkata built up • The chunker uses 39 tags and 47 distinct values Kolkata has grown up on the either banks of river Hoogly.
After MA and POS-tagging and FS instantia-
• The chunks identified by the system are nested.
Their categories are noun-chunks (NN), finite verb chunks (VF), non-finite verb chunks (VN), adverbial chunks (ADV) and adjectival chunks (JJ). The punctuation marks other than ”,” form • The templates of the feature structures are spec- ified manually in a rule file. There are 75 such templates for the different POSs and their re-spective attributes.
The process of bonding: There are two possible
bonds that can be formed at the first step - that
• Multiple word groupings for the same sentence between dui - tIre and ga.De - uTheChe.
former has the higher bond energy 2 and thereforeis chosen. The resulting category is an NN (noun • The chunker cannot identify multiword ex- pressions, named entities and clausal conjunc- tions of conditional constructs .It cannot handle nadIra [dui tIre]/NN kolakAtA ga.De cases where semantic issues are involved.
Since, the accuracy of the chunker heavily de- pends on the accuracies of the pre-processing steps At this point again two bonds can form, both of such as MA and POS tagging, evaluation of the which have the same bond energy. Since the bonds system as such will not reflect the accuracy of the are non-overlapping, it does not matter which chunking algorithm. Therefore, in order to estimate one is chosen first. Finally, we get the following the accuracy of the chunking algorithm rather than 2The meaning of the symbols are - NP: proper noun, poss: the system as a whole, the present chunker has been possesive, JQ: quantifying adjective, NN: noun, V: Verb, nf: tested on a POS-annotated corpus of about 30,000 non-finite, f: finite, CAT: category, attr: attribute, be: bond en-ergy words. The corpus has been developed in-house by manually annotating a subset of the Bengali mono- Acknowledgement
lingual corpus provided by CIIL. A few conven-tions have been kept in mind while evaluating the This work has been partially supported by Media accuracy of the algorithm. Since it produces word groups which may be nested, the outermost chunkhave been considered during evaluation. Therefore,for a particular sentence, the chunker produces a References
set of non-overlapping chunks which may have a S. Abney. 1991 Parsing by Chunks. Principle Based correct or an incorrect structure. The average pre- Parsing. Kluwer Acedemic Publishers.
cision and recall rates are 96.12% and 98.03% re- D. Allerton. 1982. Valency and the English Verb. Lon- spectively. A cent percent accuracy in chunking is not observed since the system cannot identify multi- A. Bharati, V. Chaitanya, and R. Sangal. 1995. Natural word expressions and named entities. Most of the Language Processing: A Paninian Perspective. Pren- incorrect chunks are due to the occurrences of these two phenomena in natural language text.
S. Buchholz, J. Veenstra, and W. Daelemans. 1999. Cas- caded Grammatical Relation Assignment. Proceed- Conclusion
In this paper, a framework has been described forchunking of free word order languages based on a G. Helbig. 1992. Probleme der Valenzund Kasustheorie. generalization of the valency theory implementedin form of feature structures.
T. Herbst. 1999. English Valency Structures - A first strong analogy between formation of word groups sketch. Technical report EESE 2/99. and formation of larger chemical structures from bond energy between syntactic units provided a lot CoNLL-2000 and LLL-2000. Lisbon, Portugal, 2000.
of representational ease in capturing the problem of L. A. Ramshaw, and M. P. Marcus. 1995. Text Chunk- ing using Tansformation-Based Learning. ACL ThirdWorkshop on Very Large Corpora The formalism can be extended to complete P. R. Ray, Harish, V., S. Sarkar, and A. Basu 2003 Part of parsing of free word order languages. This can be Speech Tagging and Local Word Grouping Techniques accomplished through the inclusion of semantic for Natural Language Parsing in Hindi. Proceedings of valencies and non-adjacent syntactic valencies, International Conference on Natural Language Pro-cessing (ICON 2003). Mysore, India.
which will be associated with a non-adjacent bondenergy value. The idea is similar to that of karaka or M. Vilain, and D. Day 2000. Phrase Parsing with Rule traditional valencies of verbs. Also, the system can Sequence Processors: an Application to the SharedCoNLL Task. Proceedings of CoNLL-2000 and LLL- be extended by starting to chunk the words without the intervention of the POS tagger. An untaggedsentence may serve as the input of the chunker K. Welke. 1995. Dependenz, Valenz und Konstituenz.
and the system may tag and chunk in one go to Ludwig M. Eichinger and Hans-Werner Eroms (eds.).
Dependenz und Valenz. make the chunker independent of the POS tagger.
A morphological analyzer would serve as the frontend to such a system. If semantic resources become T. Zhang, F. Damerau, and D. Johnson.
Chunking based on a Generalization of Winnow. Jour- available for Bengali and other free word order nal of Machine Learning Research, 2 (March).
languages, then the present system can be extendedto a full parser for such languages.

Source: http://www.dipanjandas.com/files/lwg.pdf

Alzheimer's disease overview nytimes.doc nov 2009

Alzheimer's Disease Overview Alzheimer's disease (AD), is one form of dementia that gradually gets worse over time. It affects memory, thinking, and behavior. Memory impairment, as well as problems with language, decision-making ability, judgment, and personality, are necessary features for the diagnosis. Reference from A.D.A.M. ALTERNATIVE NAMES Senile dementia - Alzheimer's type (SDAT)

nbcgib.uesc.br

Title Incremental Cost-Effectiveness (ICE) Statistical Inference from Two Unbiased SamplesAuthor Bob Obenchain <[email protected]>Maintainer Bob Obenchain <[email protected]>Description Given two unbiased samples of patient level data on cost and effectivenessfor a pair of treatments, make head-to-head treatment comparisons by (i) generating thebivariate bootstrap resampling distribution

Copyright ©2010-2018 Medical Science