<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A New Corpus Resource for Studies in the Syntactic Characteristics of Terminologies in Contemporary English</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Alex C. Fang</string-name>
          <email>1acfang@cityu.edu.hk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jing Cao</string-name>
          <email>cjing3@student.cityu.edu.hk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Yang Song</string-name>
          <email>songyang2@student.cityu.edu.hk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Dialogue Systems Group Department of Chinese, Translation and Linguistics City University of Hong Kong Tat Chee Avenue</institution>
          ,
          <addr-line>KowloonHong Kong SAR</addr-line>
          ,
          <country country="CN">PR China</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>In this paper, we present a new corpus resource that has been constructed specially for the study of the syntactic characteristics of terminologies. The corpus is based on the British component of the International Corpus of English (ICE-GB), comprising four parallel subject domains from two text categories (i.e. academic vs. popular prose) with a total of about 200,000 running word tokens. The resource is richly annotated at lexical, grammatical, syntactic, and terminological levels. It is also parameterized according to both text categories and subject domains. The corpus resource is expected to contribute towards a linguistically motivated description of terms and their internal structures. It is also expected to provide an analytical framework for the study of relations between terminological use and text categories as well as subject domains.</p>
      </abstract>
      <kwd-group>
        <kwd>syntactic tree</kwd>
        <kwd>treebank</kwd>
        <kwd>syntactic function</kwd>
        <kwd>terminology</kwd>
        <kwd>ICE-GB</kwd>
        <kwd>noun phrase</kwd>
        <kwd>term annotation</kwd>
        <kwd>corpus</kwd>
        <kwd>syntax</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Automatic term recognition (ATR) and extraction have been a challenging task
and encouraged rigorous efforts of researchers from a wide range of backgrounds
and disciplines. Nevertheless, past work on terminological extraction tends to focus
on specific subject domains, and mainly in the field of biochemistry and medicine
such as Ananiadou et al. 2000, Nenadic et al. 2005, Aubin and Hamon 2006, and
Ville-Ometz et al. 2007, to name just a few. Some work on other domains such as
computing (e.g. Eumeridou et al. 2004; L’Homme 2002; Nakagawa and Mori 2003),
economy (e.g. Rodriguez et al. 2007), and legislation (e.g. Ha et al. 2008; Kit and Liu
2008). Those studies are domain specific in a good sense that they concentrate on
domain-specific issues like domain knowledge and associated knowledge expressions
on the lexical level. Yet they are domain limited in an undesirable sense, which leads
to difficulty in evaluating the performance and interoperability of the existing term
recognition systems across a set of different domains. Additionally, it remains an
issue how such systems will adapt to new domains.</p>
      <p>Another noticeable issue is that, among the linguistic features employed in ATR
systems, syntactic features have been mainly observed at the phrasal level, and
seldom from the perspective of syntactic structures at a clausal level. Grammatical
patterns, such as ‘noun’, ‘noun + noun’, ‘adjective + noun’, ‘noun + preposition +
noun’, have been integrated with statistic measurements to determine the termhood
(e.g. Frantzi et al. 2000; Pazienza et al. 2005). Eumeridou et al. (2004) go beyond the
grammatical patterns and examine how term occurrences correlate the argument
structure of verbs across three domains chosen from the British National Corpus.
Their findings show an uneven distribution of terms in different argument structures1,
and they also notice the influence that different domains have upon term occurrences.
Although the study focuses on the verbal syntax only, it does indicate that syntactic
features of terminological entities warrant a worthwhile research topic and that text
categories such as registerial types and subject domains should also be a parameter to
consider. It is reasonable to believe that further improvement of ATR systems can be
achieved by exploring deeper, linguistically motivated analysis of the relation
between terminologies and linguistic parameters.</p>
      <p>The main focus of this paper is to present a new corpus resource that has been
constructed specially for the study of the syntactic characteristics of terminologies.
Existing term-annotated corpora are typically domain-specific, such as GENIA (Ohta
et al. 2002), and typically used as a resource for statistical training. The new corpus
resource is different in that it is built on general domains and is richly annotated for
syntactic information, especially for detailed annotation of the syntactic categories
and their functions within the clause complex that is often dependent on verb
subcategorization. The corpus is based on the British component of the International
Corpus of English (ICE-GB), comprising four parallel subject domains from two text
categories (i.e. academic vs. popular prose) with a total of about 200,000 running
word tokens. The resource is richly annotated at lexical, grammatical, syntactic, and
terminological levels. It is also parameterized according to both text categories and
subject domains. The tree bank is expected to contribute towards a linguistically
motivated description of terms and their associated syntactic structures. It will also
provide an analytical framework for the study of relations between terminological use
and text types as well as subject domains. The richly annotated trees will facilitate
studies in the linguistic relations of terms for the purpose of ontology construction.</p>
      <p>In the rest of this paper, we will first of all describe the construction of the corpus,
including the selection of the corpus material, the annotation schemes for grammar
and syntax, and an inter-annotator analysis of the manual annotation of terms. We
shall then report some of our initial empirical observations of the syntactic
characteristics of noun phrases (NP) that are terminological entities as opposed to
generic NPs across different types and domains. For this purpose, we will describe the
distribution of general NPs in terms of text categories and subject domains. We will
1 In lexical semantic terms, argument structure refers to the semantic type of the verb and its related
elements such as agent and theme. The same term is also loosely used in syntax to refer to the
subcategorisation, or valency structure or complémentation type of verbs.
then describe the distribution of terminological NPs according to the same parameters,
focusing on their syntactic functions in the tree structure.
2
2.1</p>
    </sec>
    <sec id="sec-2">
      <title>Corpus Construction</title>
      <sec id="sec-2-1">
        <title>Corpus resource for term annotation</title>
        <p>Our on-going research attempts to extend the previous studies by exploring the
syntactic characteristics of terminological entities across different text types and
subject domains in contemporary English. To achieve our objectives, the British
component of the International Corpus of English (ICE-GB; Greenbaum 1996) was
chosen as a basis for the following reasons: First, it is encoded for a variety of text
categories and subject domains. Secondly, it is already grammatically tagged,
syntactically parsed and manually validated. Finally and most importantly, it is
annotated with a rich set of linguistically motivated syntactic relations that will
maximally enhance our intended study. The following sections will first describe the
resource created from the ICE-GB and introduce its part-of-speech (POS) and
syntactic annotations.
2.1.1</p>
        <sec id="sec-2-1-1">
          <title>Creation of a sub-corpus</title>
          <p>The British component of the International Corpus of English (ICE-GB) is a
onemillion-word corpus comprising both spoken and written British English from the
1990s (Greenbaum 1996; Fang 2007). The spoken section represents 60% of the total
size of the corpus with 300 sample texts. The written section accounts for 40% of the
corpus with 200 texts. Each component text has about 2,000 word tokens. Table 1
summarizes the text categories in the ICE-GB together with the number of component
texts.
Given the purpose of our study, texts from the category of informational writing
constitute a suitable source of texts, which is further divided into three sub-categories:
academic writing, popular writing and press news reports. Two contrastive text types,
i.e., academic writing and popular writing, were chosen. The two text types cover four
parallel subject domains comprising ten texts each. Table 2 presents the composition
of the sub-corpus created from ICE-GB.
As can be seen from Table 2, the sub-corpus comprises 80 texts similar in size with a
total number of 193,206 word tokens.
2.1.2</p>
        </sec>
        <sec id="sec-2-1-2">
          <title>Tree annotations in the ICE-GB</title>
          <p>All the texts in ICE-GB are richly annotated grammatically and syntactically
(Fang 1996, 2000, 2006, 2007). When the 80 texts from ICE-GB were selected to
create the sub-corpus, a treebank was effectively created that comprises 8,306
syntactic trees.</p>
          <p>PU CL(main,montr,pass,pres)
SU NP()
DT DTP()
DTCE ART(def) {The}
NPHD N(com,plu) {fibres}
NPPO PP()
P PREP(ge) {of}
PC NP()</p>
          <p>NPHD N(com,sing) {group B}
VB VP(montr,pres,pass)
OP AUX(pass,pres) {are}
MVB V(montr,edp) {found}
A PP()
P PREP(ge) {in}
PC NP()
DT DTP()
DTCE ART(def) {the}
NPPR AJP(attru)
AJHD ADJ(ge) {autonomic}
NPPR AJP(attru)
AJHD ADJ(ge) {nervous}</p>
          <p>NPHD N(com,sing) {system}</p>
          <p>PUNC PUNC(per) {.}</p>
          <p>As noted in Figure 1 above, the tree structure is richly annotated with fine-grained
grammatical and syntactic information. At the grammatical level, words are coded
with part-of-speech (POS) tags that include a head tag (such as nouns, verb, and
adjectives) with a set of attributes indicating the subcategorizations of the head tag.
For instance, the verb found enclosed within a pair of curly brackets is tagged as
V(montr,edp), namely, a mono-transitive verb in past participial form. As another
example, {The} is assigned a label ART(def), meaning it is a definite article, and
{fibres} is a common noun in its plural form. Syntactically, each node comprises
two labels: one representing its syntactic category (such as noun phrase and adjective
phrase) and the other the syntactic function. Take the node SU NP() as an example,
which indicates that it is a noun phrase (NP) functioning as the subject (SU) of the
clause. The same NP comprises a determiner (DT), the head (NPHD) and a
postmodifier (NPPO). The definite article The constitutes the central determiner (DTCE), a
daughter node of DT. See Appendix for a complete list of all the parsing symbols.
With such a system of syntactic categories and their associated syntactic functions,
the corpus forms a valuable testbed according to which grammatical relations of
various kinds can be investigated. The syntactic framework will also form an
informative context within which terms and term relations can be usefully examined.
2.2</p>
        </sec>
      </sec>
      <sec id="sec-2-2">
        <title>Term annotation</title>
        <p>Term annotation was carried out manually during a period of four months, and has
gone through the following procedures:
• Training of the annotators: The training session helps the annotators get
familiar with the special format of the target texts, which are parsed and
represented in a form exemplified in Fig. 1.
• Analysis of inter-annotator agreement: This step was taken to establish the
consistency and therefore the quality of the annotations by the three different
annotators given the same text, and a higher statistic agreement will
demonstrate the confidence of the manual annotation.
• Actual annotation: With an annotation guideline, annotators mark up the
terms with the help of dictionaries, online dictionaries and term banks.
• Manual examination of terminological annotations.</p>
        <p>In the remaining of this section, we shall first describe the annotation guideline and
then report the results from the inter-annotator agreement test. The basic statistics of
the terminologically annotated corpus resource will be presented in Section 3.
2.2.1</p>
        <sec id="sec-2-2-1">
          <title>Annotation guideline</title>
          <p>Before describing the guideline, we first introduce the operational definition of
terminological entities. To our understanding, terms by definition primarily
correspond to noun-phrase (NP) groups and thus consist of words that are single
nouns or complex noun phrases (Kageura et al. 2004; Nakagawa 2001; Nakagawa and
Mori 2003). Following Eumeridou et al. (2004), we also consider terms in a
pragmatic sense. Take text w2a-031 for example. The text is about “blind shaft
drilling” under the domain of technology. In addition to terms in technology and
engineering, we may also mark up terminological entities from related domains such
as environment. Given such a definition, a working guideline for annotation was made:</p>
          <p>Among the NPs, proper names of places, countries, organizations or
institutes are excluded from the current study, and therefore, will not be
annotated.</p>
          <p>Variant terms will be annotated.</p>
          <p>o Singular and plural forms of a term will both be regarded as terms
in case some termbanks only collect singular form of a term.
o When an N1+N2 compound is a term, the sequence N2 + of + N1 will
also be treated as a term.</p>
          <p>o Variant spellings of the same term will be accepted.</p>
          <p>With nested terms, we only mark up the longest part as a multi-word term.
Terms are marked with ‘&lt;’ at the beginning and ‘&gt;’ at the end in the tree
diagram, and the resulting NP is described by an additional attribute ‘term’.
See Figure 2.</p>
          <p>PU CL(main,montr,pass,pres)
SU NP(term)
DT DTP()
DTCE ART(def) {The}
NPHD N(com,plu) {&lt;fibres&gt;}
NPPO PP()
P PREP(ge) {of}</p>
          <p>PC NP()
NPHD N(com,sing) {group B}</p>
          <p>A PP()
P PREP(ge) {in}
PC NP(term)
DT DTP()
DTCE ART(def) {the}
NPPR AJP(attru)
AJHD ADJ(ge) {&lt;autonomic}
NPPR AJP(attru)
AJHD ADJ(ge) {nervous}</p>
          <p>NPHD N(com,sing) {system&gt;}</p>
          <p>Three annotators were trained to mark up terms. All the three annotators are
university students majoring in linguistics. Among them, two are undergraduates who
have been admitted to postgraduate study and one is a PhD candidate. To measure the
inter-annotator agreement, two texts were taken from the pre-selected sub-corpus
from ICE-GB, with a total number of about 4,000 words. During the annotation stage,
the annotators were allowed to refer to the guideline or other sources such as online
termbanks and dictionaries, in addition to their linguistic knowledge. They were not
allowed to confer with each other over the annotation.</p>
          <p>We then compared the annotations among the three annotators by using F score,
which is considered to be a standard measure to determine the inter-annotator
agreement (Corbett et al. 2007) and has been commonly used in previous studies (see,
for example, Demetriou and Gaizauskas 2003, Morgan et al. 2004, Vlachos and
Gasperin 2006 and Kolarik 2008). Therefore, the inter-annotator agreement was
computed pair-wise using a measure defined in (1):
(1)
where M1 and M2 are the number of markable terms in a given text marked up by
Annotators 1 and 2 respectively, and C is the total number of times both annotators
agree on a markable term in that same text. To calculate the F score, the total number
of terms marked by annotators A, B, and C were counted respectively. Next, all of
the exact matches were found and counted. For an exact match, the left and right
boundaries had to match entirely.</p>
          <p>In this section, we present some initial descriptive statistics and chart the
distribution of NP constructions across different text categories and domains. We will
first explain how we retrieve the syntactic functions of NPs according to the tree
structure, followed by a description of the basic statistics of NP constructions in the
corpus. We shall then present the preliminary observations of the syntactic features of
NPs that are marked as terms.</p>
          <p>A general description of NP constructions by category and domain
As explained in Section 2.1.2, every NP is assigned a function label and additional
attributes if necessary. To count the frequency of NP constructions in trees is
straightforward in most cases except for two scenarios, where the functions are
labeled as CJ (conjoin; see Fig. 3) and DEFUNC (appositive NP that does not perform
any syntactic function; see Fig. 4). In Fig. 3, the direct object NP is described by the
attribute coordn, indicating the presence of a coordinated construction whose
conjoins are marked as CJ. In such a scenario, a CJ will inherit the function of its
mother node and be counted as a separate OD NP. Therefore, instead of counting one
OD and two CJ functions, we count two OD functions for the NPs in Fig 3. Similarly,
NPs with DEFUNC labels are also relocated and assigned the function label of the
governing NP. See Fig. 4 for an example, where DEFUNC NP is treated as SU NP. In this
particular case, instead of one DEFUNC and one SU, two SU functions are counted.</p>
          <p>OD NP(coordn)
CJ NP()
DT DTP()
DTCE ART(def) {the}
NPHD N(com,plu) {gods}
COOR CONJUNC(coord) {and}
CJ NP()</p>
          <p>NPHD N(com,plu) {customs}</p>
          <p>With this treatment of conjoin and appositive NPs, NP constructions in all the
eight subject domains were retrieved and summarized in Table 4. As can be observed
in Table 4, there is an uneven distribution of 17 different functions of NPs across
domains. In general, NPs seem to occur most frequently at the position of PC in all the
domains, followed by SU and OD. Nevertheless, when we examine the functions by
category and domain, we notice more interesting patterns. First, NPs in domains of
academic writing tend to occur less frequently at the position of SU than those in their
counterparts of popular writing. Second, domains in academic writing are more likely
to have a comparatively higher occurrence of PC as a syntactic function than their
counterparts in popular writing. They also tend to have fewer occurrences of OD.
3.2</p>
          <p>A statistical description of term-NP constructions</p>
          <p>When examining the distribution of term-NPs, we also related the CJ and DEFUNC
functions to their mother nodes. Accordingly, the actual distribution of term-NPs
across difference categories and domains were calculated and presented in Table 5.</p>
        </sec>
        <sec id="sec-2-2-2">
          <title>Function</title>
          <p>A
AJPR
AVPR
CO
CS
DT
ELE
FOC
NPPO
NPPR
OD
OI
PC
SU
Total</p>
          <p>Interesting features emerge from the initial frequency count. First, academic
writing tends to have more terms than popular writing in both parameters (i.e.
category and domain). In a broad sense, the total number of terms in academic writing
is higher than that of popular writing. From the perspective of subject domains,
individual domains belonging to academic writing tend to have more terms than their
counterparts in popular writing. Such a result suggests that formal writing tends to
contain more term candidates than informal writing. Second, science domains (i.e.
NAT and TEC) tend to contain more terms than arts domains (i.e. HUM and SOC). It
can be also noticed that the number of terms in AHUM is higher than that of ATEC,
and it is understandable since AHUM has the highest number of NPs among the
domains in academic writing. Third, across the eight domains term-NPs seem to
appear most frequently at the position of PC, followed by SU and OD. Fourth, it would
be easy to make a contrastive study on certain syntactic functions across the eight
domains. For example, terms are more likely to occur at the position of A in ATEC
when compared with the other seven domains, and they are more likely to appear at
the position CS in AHUM when examined across domains. Such information can be
taken as a flexible value in assigning weights to syntactic functions in accordance
with particular domains in ATR.</p>
          <p>It is worth mentioning that syntactic labels at the phrasal level can be further
classified at the clausal level. For example, a considerable number of NPs occur at the
position of PC, which should be related to its mother node, namely PP, whose
functions could be analyzed differently as A PP and NPPO PP, revealing further
variations of use across the eight categories.
4</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Conclusion</title>
      <p>In this paper, we presented a new corpus resource that has been constructed
specially for the study of the syntactic characteristics of terminologies for a
linguistically motivated description of terms and their internal structures. The corpus
is based on the British component of the International Corpus of English, comprising
four parallel subject domains from two text categories (i.e. academic vs. popular
prose) with a total of about 200,000 running word tokens. It is richly annotated at
lexical, grammatical, syntactic, and terminological levels. It is parameterized
according to both text categories and subject domains. We first described the
construction of the corpus, including the selection of the corpus material, the
annotation schemes for grammar and syntax, and an inter-annotator analysis of the
annotation of terms. We then described the corpus resource by reporting some of our
initial empirical observations of NP constructions and term-NP constructions.
Interesting patterns were observed in terms of syntactic distribution of NPs and
termNPs across different categories and domains. In particular, term-NPs show observable
difference across different categories and domains. In other words, the corpus
resource can provide an analytical framework for the study of relations between
terminological use and text types as well as subject domains.</p>
      <p>Acknowlegement
The work described in this paper was supported partially by research grants (Nos
7002190, 7200120 and 7002387) from City University of Hong Kong. The authors
would like to thank the reviewers for their valuable comments and suggestions.
ANANIADOU, S., ALBERT S. &amp; SCHUHMANN D. (2000). Evaluation of Automatic Term
Recognition of Nuclear Receptors from MEDLINE. Genome Informatics 11. p. 450-451.
AUBIN, S. &amp; HAMON T. (2006). Improving Term Extraction with Terminological Resources. In
FinTAL 2006, LNAI 4139. p. 380-387.</p>
      <p>CORBETT, P., BATCHELOR, C. &amp; TEUFEL, S. (2007). Annotation of Chemical Named Entities. In
Proceedings of BioNLP 2007: Biological, translational, and clinical language processing. p.
57-64.
DEMETRIOU, G. &amp; GAIZAUSKAS, R. (2003). Corpus resources for development and evaluation of
a biological text mining system. In Proceedings of the Intelligent Systems for Molecular
Biology (ISMB) Workshop on Text Mining (BioLINK2003).</p>
      <p>EUMERIDOU, E., NKWENTI-AZEH, B. &amp; MCNAUGHT, J. (2004). An Analysis of Verb
Subcategorization Frames in Three Special Language Corpora with View towards Automatic
Term Recognition. Computers and the Humanities, 38. p. 37-60.</p>
      <p>FANG, A.C. (1996). AUTASYS: Automatic Tagging and Cross-Tagset Mapping. In S.
GREENBAUM Ed. p. 110-124.</p>
      <p>FANG, A.C. (2000). From Cases to Rules and Vice Versa: Robust Practical Parsing with
Analogy. In Proceedings of the Sixth International Workshop on Parsing Technologies. p.
7788.</p>
      <p>FANG, A.C. (2006). Evaluating the Performance of the Survey Parser with the NIST Scheme. In
LNCS 3878: Computational Linguistics and Intelligent Text Processing, ed. by A. Gelbukh.
Berlin Heidelberg: Springer-Verlag. P. 168-179.</p>
      <p>FANG, A.C. (2007). English Corpora and Automated Grammatical Analysis. Beijing: The
Commercial Pess.</p>
      <p>FRANTZI, K., ANANIADOU, S. &amp; MIMA, H. (2000). Automatic recognition of multi-word terms:
the C-value/NC-value method. International Journal of Digital Libraries, 3(2). p.117-132.
GREENBAUM, S. (1996). Comparing English World: The International Corpus of English.
Oxford: Oxford University Press.</p>
      <p>HA, L. A., MITKOV, R. &amp; CORPAS, G. (2008). Mutual Terminology Extraction Using a Statistical
Framework. In Procesamiento del lenguaje natural. N. 41. p. 107-112.</p>
      <p>KIT, C. &amp; LIU, X. (2008). Measuring Mono-word Termhood by Rank Difference via Corpus
Comparison. Terminology 14:2. p. 204-229.</p>
      <p>KOLÁ IK, C. &amp; KLINGER, R. (2008). Chemical Names: Terminological Resources and Corpora
Annotation. LREC 2008 Workshop: Building and Evaluating Resources for Biomedical Text. p.
51-58.</p>
      <p>L'HOMME, M.C. (2002). What can Verbs and Adjectives Tell us about Terms? In Proceedings
of Terminology and Knowledge Engineering (TKE 2002). p. 65-70. France.</p>
      <p>MORGAN, A.-A., HIRSCHMAN, L., COLOSIMO, M., YEH, A.-S. &amp; COLOMBE, J.-B. (2004). Gene
Name Identification and Normalization using a Model Organism Database. Journal of
Biomedical Informatics, 37(6). p. 398-410.</p>
      <p>NAKAGAWA, H. &amp; MORI, T. (2003). Automatic term recognition based on statistics of
compound nouns and their components. Terminology, 9 (2). p. 201-219.</p>
      <p>NENADIC, G., SPASIC, I., &amp; ANANIADOU, S. (2005). Mining Biomedical Abstracts: What’s in a
Term? In K.-Y. Su, J. Tsujii, J.-H. Lee and O. Kwong (Eds.) Natural Language Processing –
IJCNLP 2004 First International Joint Conference, Lecture Notes in Computer Science vol.
3248. p. 797-806.</p>
      <p>OHTA, T., TATEISI, Y., MIMA, H. &amp; TSUJII, J. (2002). GENIA Corpus: An Annotated Research
Abstract Corpus in Molecular Biology Domain. In Proceedings of the Human Language
Technology Conference, San Diego, USA.</p>
      <p>PAZIENZA, M.T., PENNACCHIOTTI, M. &amp; ZANZOTTO, F.M. (2005). Terminology Extraction: An
Analysis of Linguistic and Statistical Approaches. StudFuzz 185. p. 255-279.
RODRÍGUEZ, B., MARIO, F., NOYA, E. D., OTERO, P.G., MARTÍNEZ, M.L., MATO, E. M.M., ROJO,
G. SANTALLA DEL RÍO, M.P., &amp; DOCÍO, S.S. (2007). A Corpus and Lexical Resources for
Multiword Terminology Extraction in the Field of Economy in a Minority Language. In Proceedings
of the 3rd Language and Technology Conference. p. 359-363.</p>
      <p>VILLE-OMETZ, F., ROYAUTÉ, J. &amp; ZASADZINSKI, A. (2007). Enhancing automatic recognition
and extraction of term variants with linguistic features. Terminology 13 (1). p. 35-59.
A
ADJ
ADV
AJHD
AJP
AJPO
AJPR
ANTIT
ART
AUX
AVB
AVHD
AVP
AVPO
AVPR
CF
CJ
CL
CLEFTIT
CLOP
CO
CONJUNC
CONNEC
COOR
CS
CT
DEFUNC
DISMK
DISP
DT
DTCE
DTDE
DTP
DTPE
DTPO
DTPR
DTPS
ELE
EXOP
EXTHERE
FOC
FRM
GENF
GENM
IMPOP
VLACHOS, A. &amp; GASPERIN, C. (2006). Bootstrapping and Evaluating Named Entity Recognition
in the Biomedical Domain. In Proceedings of BioNLP in HLT-NAACL. p. 138-145.
Appendix: A Complete List of Parsing Symbols</p>
      <p>INDET
INTOP
INVOP
LIM
LK
MVB
N
NADJ
NONCL
NOOD
NOSU
NP
NPHD
NPPO
NPPR
NUM
OD
OI
OP
P
PARA
PC
PMOD
PP
PRED
PREDG
PREP
PROD
PROFM
PRON
PRSU
PRTCL
PS
PU
PUNC
REACT
SBMO
SU
SUB
SUBHD
SUBP
TO
V
VB
VP</p>
      <p>Indetermined
Interrogative operator
Inversion operator
Limitor
Linker
Main verb
Noun
Nominal adjective
Non-clause
Notional object
Notional subject
Noun phrase
Noun phrase head
Noun phrase postmodifier
Noun phrase premodifier
Numeral
Direct object
Indirect object
Operator
Prepositional
Paratactic
Prepositional complement
Preposition premodifier
Prepositiional phrase
Predicate
Predicate group
Preposition
Provisional object
Pro-nominal form
Pronoun
Provisional subject
Particle
Stranded preposition
Parsing unit
Punctuation
Reactional signal
Subordinator phrase premodifier
Subject
Subordinator
Subordinator phrase head
Subordinator phrase
Infinitive to
Verb
Verbal
Verb phrase</p>
    </sec>
  </body>
  <back>
    <ref-list />
  </back>
</article>