<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>AMRITA_CEN @ FIRE 2015: Extracting Entities for Social Media Texts in Indian Languages</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Anand Kumar M</string-name>
          <email>m_anandkumar@cb.amrita.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Shriya Se</string-name>
          <email>shriyaseshadrik.r@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Soman KP</string-name>
          <email>kp_soman@amrita.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>CEN, Amrita Vishwa Vidyapeetham</institution>
          ,
          <addr-line>Coimbatore</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
      </contrib-group>
      <fpage>85</fpage>
      <lpage>88</lpage>
      <abstract>
        <p>This contemporary work is done as a slice of the shared task on Entity Extraction from Social Media Text Indian Languages in Forum for Information Retrieval and Evaluation (FIRE2015). Nowadays people are extensively using social media platforms like Face book, Twitter, etc, to exchange their thoughts. The twitter messages are growing rapidly and their style and short nature present a new challenge in language technology field. This extensive amount of textual data is also increases the interest in Information Extraction (IE) on such textual data. Named entity extraction is one of the essential tasks in Information Extraction, aims to extract and classify entities from text. The performance of the present standard language processing tools is severely affected on Tweet messages. Hence, different improvised and nonimprovised algorithms are necessary for extracting these entities from the informal text. This paper deals with the extracting the Named Entities from twitter messages of four Indian Languages. The extraction of the Named entity relies mainly on the domain specific features and conventional features. A well known supervised algorithm, Support Vector Machine (SVM) is used to extracting the entities. • Theory of computation~Support vector machines • Computing methodologies~Natural language processing • Information systems~Information extraction • Human-centered computing~Social tagging systems</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. INTRODUCTION</title>
      <p>
        Named Entity Recognition (NER) is a task of extracting the
named entities like people, organization and location etc. from a
text. NER is an important task in Natural Language Processing
(NLP) system for question answering and information extraction
[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Twitter, a micro blogging platform is an important source for
continuously updating the information. Twitter messages are 140
character length short texts which are noisy, informal nature and
seldom contain an informative content. For extracting these
information, many algorithms are proposed and in practice too. In
general, tweets are based on various products or about different
persons. Identifying the named entities in tweets is a difficult task
for two reasons [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. The first one is tweet contains the large
number of distinct named entity types (except location and
person), so it is difficult to find an adequate size of annotated
examples for each entity type. Another reason is due to the lack of
sufficient context, it’s challenging to determine the entity type.
Entity extraction task is important for companies to detect their
crisis and to response in an earlier stage. By doing entity
extraction in twitter data, a company can solve the crisis and can
make earlier response which leads to improve their profit. The
performance of the present standard language processing tools is
severely affected on Tweet messages. So adapting the domain and
applying that too for Indian Language is a challenging task.
Normalizing the Tweet messages is also not advisable. Hence an
improvised machine learning technique is needed for better
accuracy. Few researchers treated entity extraction as two level
tasks, which include segmentation and disambiguation.
Segmentation is used to find out the entity boundaries and
disambiguation assist to identify the entity type. There is no prior
works on entity extraction for Twitter messages in Indian
languages. Widely machine learning approaches are followed for
entity recognition. Machine learning algorithm, SVM has been
successfully applied to Indian language processing tasks which
includes NER , sense disambiguation [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], morphological analyzer
[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ][
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] and clause boundary identification [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Entities are
extracted and linked with DBpedia source using machine learning
algorithms [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. MaxEnt based Hybrid NER system is tried for
Indian languages [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. SVM and CRF based entity recognition for
Indian languages have been done with rich feature set [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ].
We converted the given format into sequence labeling format
using BIO encoding for representing entities and uses a well
known machine learning algorithm, Support vector machine for
classification. We include POS tag, nearby context, orthographic
and language specific features for all the four languages. In
addition to that we used Brown clusters for English language.
      </p>
      <p>The rest of this paper is composed of four sections. In
Section 2, we discuss about the dataset descriptions and explained
the size of training data and testing data. The methodology and the
features are explained in Section 3. We report our experiment
results in Section 4 and conclude the paper with future directions
in Section 5.</p>
    </sec>
    <sec id="sec-2">
      <title>2. DATA SET DESCRIPTIONS</title>
      <p>
        Given the four Indian languages tweets as an input, the task is to
extract the named entities and identify its type. The task
organizers released the Twitter dataset and its entities for four
languages namely Tamil, English, Hindi and Malayalam. There
are 22 types of entities present in the training dataset in which the
majority of entities from person, location and organization. The
size of the training and testing data is shown in the Table.1 in
which the most of the tweet contains URL and leading topics are
politics and cinema. The major issue in the dataset is that the size
of the training data is lesser than the size of the test data (except
for Malayalam) which leads to a drop in the accuracy level of the
system. We have converted the tweets into the conventional BIO
format and treated them as a sequence labeling problem. Table.1
also describes the token counts of training and testing data of all
the four languages. Count of primary entities in the training
dataset is shown in the Table.2.
divided into small chunks called as tokens. Training and testing
tweets are tokenized. Given named entities and tokenized training
tweets are merged to create BIO format in one token-per line
fashion. The BIO formatted sequence data is then tagged using
POS tagger of the respective languages. For Tamil Language in
house Twitter POS tagger and Amrita POS tagger [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] is used.
Hindi NLTK POS tagger is applied for Hindi tweets and for
Malayalam, Amrita POS tagger is used and for English the tokens
is tagged using English Twitter POS tagger developed by Gimbel
et.al.[
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. Features are illustrated in Table.4 and explained in
Subsection 3.1. Features are extracted for each language and train
the system with support vector machine based classifier. We
trained our system using the SVM based tool “SvmLight” [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ].
Finally, the BIO format tokens are converted into the given
annotation format.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. SYSTEM DESCRIPTIONS</title>
      <p>This section describes our system submitted in Entity extraction
from twitter messages task in FIRE 2015. In the proposed Twitter
NER system, feature extraction plays an important role, as the
accuracy of the system rely on the extracted features. Generally,
for text mining and information extraction, preprocessing is
necessary and it is mandatory for the twitter dataset.
Preprocessing includes Normalization and Tokenization.
Generally the sentences are divided into tokens based on the white
spaces. These tokens are further normalized where superficial
variations are removed. But, normalization of Twitter messages
aims to standardize the non-standard words, spellings variations,
expand informal abbreviations (e.g., tmrw for tomorrow), and
standardize phonetic substitutions. Here we avoided stemming
and Twitter data normalization. Case folding is appropriate for
English language where case variations are exist but it is not
required for Indian Languages where no such variation are exists.
The methodology of the proposed system is depicted in the
Figure.2. The training dataset contains two files which are raw
tweets and type annotated extracted entities. The tweet file
consists of “Tweet ID”, “UserID” and Tweets. The entity file
consists of “Tweet ID”, “UserID”, Entity Type, Entity, Starting
Index and length. We have combined these files and converted
into the conventional BIO formatted text in which B-XXX tag
indicates the beginning word of an entity type XXX and I-XXX is
used for the following chunks of an entity. The tag other than
named entity is tagged as O. In tokenization the tweets are further</p>
      <p>Word Features: The current word is used as feature.
Additionally, to provide context information, the previous and
next words of a current word are also used as features (w-1, w0,
w+1).</p>
      <p>Named Entities: Named Entity tag of the previous word, next
words are used as features (t-1, t+1).</p>
      <sec id="sec-3-1">
        <title>Features</title>
        <sec id="sec-3-1-1">
          <title>Word Features</title>
        </sec>
        <sec id="sec-3-1-2">
          <title>Named entities</title>
        </sec>
        <sec id="sec-3-1-3">
          <title>POS tags</title>
        </sec>
        <sec id="sec-3-1-4">
          <title>Prefixes and Suffixes</title>
        </sec>
        <sec id="sec-3-1-5">
          <title>End features</title>
        </sec>
        <sec id="sec-3-1-6">
          <title>Hash, Numbers, Punctuation and Acronym with periods</title>
        </sec>
        <sec id="sec-3-1-7">
          <title>Language Specific features</title>
        </sec>
        <sec id="sec-3-1-8">
          <title>Length</title>
        </sec>
        <sec id="sec-3-1-9">
          <title>Position</title>
        </sec>
        <sec id="sec-3-1-10">
          <title>Brown Cluster(Only for English)</title>
        </sec>
        <sec id="sec-3-1-11">
          <title>Capital, All Capital, Contains Apostrophe(Only for English)</title>
          <p>Prefixes and Suffixes: Prefix and suffix information are helpful
to identify the named entities, particularly for unknown words.
The first and last n characters are used as features; characters are
ranging from 3 to 5 (P3, P4, P5 S3, S4, S5).</p>
          <p>End features: Here, we have used four different kind of features (
. , ! ?). If the token ends with dot or comma or exclamatory or
question mark, then the corresponding feature is set to 1.
Otherwise, it is set to 0. For example, for the token “help!” the
feature is “0010”.</p>
          <p>Non-alphabet features: If a token contains hash (#) or any digit
or punctuation or acronym with periods, then the corresponding
feature is set to 1. Otherwise, it is set to 0. For example, for the
token “B.J.P” the feature is “0011”.</p>
          <p>Other features: Length of the token and the position in current
tweet is considered as features for all the four languages.
Language Specific features are used only for Tamil and English.
For Tamil, we check whether the token contains ‘ரூ ’ or ‘ரூ .’
followed by digit and any digit followed by -ல் ,-இல் ,கு , க்ு
. For
English, we consider the digit followed by “st” “th”. In addition
to that for English language we have used brown cluster (size of
10) and orthographic features like Starts with capital, all capital
and contains apostrophe as features.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>3.2 Run Submissions</title>
      <p>We have participated in all the four Indian languages and
submitted our outputs to organizers.</p>
      <p>Run-1: All the four languages with aforesaid features and SVM
training without Validation.</p>
      <p>Run-2: Only Tamil system has been submitted. In feature
extraction, instead of standard POS tagger, we have used our own
Twitter POS tagger which is developed with small amount of
manually annotated examples.</p>
      <p>Run3: Only Tamil system has been submitted. Tamil Twitter
POS tagger has been used. In addition to that, in order to tune the
SVM parameters, we have taken first 1000 tokens (from BIO
format) as development set and train the system accordingly.</p>
    </sec>
    <sec id="sec-5">
      <title>4. RESULTS</title>
      <sec id="sec-5-1">
        <title>Tamil</title>
        <p>P
55.23
POS Tags: POS tags contain the essential information for
identifying and disambiguating named entities. The current
word’s part-of-speech tag, the previous and next words (p-1, p0,
p+1) part-of-speech tags are used as features.</p>
        <p>
          We have used the Twitter POS tagger developed by Gimbel
et.al.[
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] for English language. Twitter POS tagger does not exist
for Indian languages, so we have used the standard POS taggers
except for Tamil system (Run-2 and Run-3). Malayalam POS
tags are retrieved from the in house Malayalam POS tagger.
NLTK Hindi POS tagger is used for tagging the Hindi tweets. For
Tamil language (Run-1), we have used our standard Tamil POS
tagger [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] and Twitter POS tagger. We have developed a Twitter
POS tagger for Tamil with 400 manually annotated tweets
collected from SAIL-2015 neutral dataset [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ].
        </p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>5. CONCLUSIONS AND FUTURE SCOPE</title>
      <p>In this paper, we have presented an approach to identify and
extract the named entities on the Twitter messages in Indian
languages. Due to the huge volumes of rapidly increasing tweets,
text mining applications like entity extractions are immediately
required to accomplish language processing tasks in informal
short texts. This entity extraction is highly helpful in the field of
information extraction and trend analysis. Here we have extracted
the different types of named entities, such as person, location,
organization, etc. We have converted the training data into
sequence labeling format and we extracted rich language specific
features for training the SVM classifier. We would also like to
perform a through error analysis to understand the issues, so we
can easily address them in future. We want to invest time in
crafting the informal text features to improve Twitter NER for
Indian languages. Other directions includes, using Twitter POS
tagger for Indian languages, adding entity specific prefix suffix
features and word embeddings with unlabeled Twitter data.</p>
    </sec>
    <sec id="sec-7">
      <title>6. ACKNOWLEDGMENTS</title>
      <p>First and foremost, we would like to thank Forum of Information
Retrieval and Evaluation (FIRE 2015) organizers for organizing a
wonderful research evaluation workshop and giving opportunities
for researchers to present their work on Text Analytics and
Natural Language Processing (NLP). We also like to thank
Computational Linguistics Research Group (CLRG), AU-KBC
Research Centre, for organizing the Entity Extraction from Social
Media Text Indian Languages (ESM-IL) Task.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Toh</surname>
            , Zhiqiang, Bin Chen, and
            <given-names>Jian</given-names>
          </string-name>
          <string-name>
            <surname>Su</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <article-title>Improving twitter named entity recognition using word representations Shared tasks of the 2015 workshop on noisy user-generated text: Twitter lexical normalization and named entity recognition</article-title>
          .
          <source>ACL-IJCNLP</source>
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Ritter</surname>
            , Alan,
            <given-names>Sam</given-names>
          </string-name>
          <string-name>
            <surname>Clark</surname>
            , and
            <given-names>Oren</given-names>
          </string-name>
          <string-name>
            <surname>Etzioni</surname>
          </string-name>
          .
          <article-title>"Named entity recognition in tweets: an experimental study</article-title>
          .
          <source>" Proceedings of the Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics</source>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Anand</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Rajendran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            ,
            <surname>Soman</surname>
          </string-name>
          ,
          <string-name>
            <surname>K.P.</surname>
          </string-name>
          <article-title>Tamil word sense disambiguation using support vector machines with rich features (2014)</article-title>
          <source>International Journal of Applied Engineering Research</source>
          ,
          <volume>9</volume>
          (
          <issue>20</issue>
          ), pp.
          <fpage>7609</fpage>
          -
          <lpage>7620</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Anand</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Rajendran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            , &amp;
            <surname>Soman</surname>
          </string-name>
          ,
          <string-name>
            <surname>K. P.</surname>
          </string-name>
          <article-title>AMRITA@ FIRE-2014: Morpheme Extraction for Tamil using Machine Learning</article-title>
          .
          <source>Working notes in FIRE 2014 - MET Task</source>
          ,
          <year>2014</year>
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Anand</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Dhanalakshmi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            ,
            <surname>Soman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.P.</given-names>
            ,
            <surname>Rajendran</surname>
          </string-name>
          ,
          <string-name>
            <surname>S. Factored</surname>
          </string-name>
          <article-title>statistical machine translation system for English to Tamil language (</article-title>
          <year>2014</year>
          )
          <article-title>Pertanika Journal of Social Science</article-title>
          and Humanities,
          <volume>22</volume>
          (
          <issue>4</issue>
          ), pp.
          <fpage>1045</fpage>
          -
          <lpage>1061</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Dhivya</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dhanalakshmi</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Anand</surname>
            <given-names>Kumar</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Soman</surname>
          </string-name>
          ,
          <string-name>
            <surname>K.P.</surname>
          </string-name>
          <article-title>Clause boundary identification for Tamil language using dependency parsing (2012) Lecture Notes of the Institute for Computer Sciences</article-title>
          ,
          <source>Social-Informatics and Telecommunications Engineering</source>
          ,
          <volume>62</volume>
          LNICST, pp.
          <fpage>195</fpage>
          -
          <lpage>197</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Barathi</given-names>
            <surname>Ganesh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.B.</given-names>
            ,
            <surname>Abinaya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            ,
            <surname>Anand</surname>
          </string-name>
          <string-name>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Vinayakumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            ,
            <surname>Soman</surname>
          </string-name>
          ,
          <string-name>
            <surname>K.P.</surname>
          </string-name>
          <article-title>AMRITA - CEN@NEEL : Identification and linking of twitter entities</article-title>
          , (
          <year>2015</year>
          ) CEUR Workshop Proceedings,
          <volume>1395</volume>
          , pp.
          <fpage>64</fpage>
          -
          <lpage>65</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Saha</surname>
            ,
            <given-names>Sujan</given-names>
          </string-name>
          <string-name>
            <surname>Kumar</surname>
          </string-name>
          , et al.
          <article-title>"A hybrid approach for named entity recognition in indian languages</article-title>
          .
          <source>" Proceedings of the IJCNLP-08 Workshop on NER for South and South East Asian Languages</source>
          .
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Abinaya.N</given-names>
            ,
            <surname>Neethu</surname>
          </string-name>
          <string-name>
            <given-names>John</given-names>
            , Anand Kumar.M and
            <surname>Soman.K.P. Amrita@</surname>
          </string-name>
          fire-2014:
          <article-title>Named entity recognition for Indian languages</article-title>
          .
          <source>Working notes in FIRE 2014 - NER Task</source>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Abinaya</surname>
          </string-name>
          .N et.al,.
          <year>2014</year>
          .
          <article-title>AMRITA@FIRE-2014 Named entity recognition for Indian languages using rich features</article-title>
          .
          <source>In Proceedings of the 6th 2014 Forum on Information Retrieval Evaluation (FIRE '14)</source>
          , ACM.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Malarkodi</surname>
            ,
            <given-names>RK Rao</given-names>
          </string-name>
          <string-name>
            <surname>Pattabhi and Sobha Lalitha Devi</surname>
          </string-name>
          (
          <year>2012</year>
          ).
          <article-title>Tamil NER - Coping with Real Time Challenges</article-title>
          .
          <source>In Proceedings of the Workshop on Machine Translation and Parsing in Indian Languages, COLING</source>
          <year>2012</year>
          , Mumbai, India.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>Kevin</given-names>
            <surname>Gimpel</surname>
          </string-name>
          , et.al, ,
          <article-title>Part-of-speech tagging for Twitter: annotation, features, and experiments, Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies: short papers</article-title>
          ,
          <source>June</source>
          <volume>19</volume>
          -24,
          <year>2011</year>
          , Portland, Oregon
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>T.</given-names>
            <surname>Joachims</surname>
          </string-name>
          . Svmlight:
          <article-title>Support vector machine</article-title>
          .
          <year>1999</year>
          , http://svmlight. joachims. org/, University of Dortmund,
          <volume>19</volume>
          (
          <issue>4</issue>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>Braja</given-names>
            <surname>Gopal</surname>
          </string-name>
          <string-name>
            <given-names>Patra</given-names>
            ,
            <surname>Dipankar Das</surname>
          </string-name>
          ,
          <string-name>
            <surname>Amitava Das</surname>
            , and
            <given-names>Rajendra</given-names>
          </string-name>
          <string-name>
            <surname>Prasath</surname>
          </string-name>
          .
          <article-title>Shared Task on Sentiment Analysis in Indian Languages (SAIL) Tweets - An Overview. In the proceeding of the Mining Intelligence and Knowledge Exploration (MIKE-</article-title>
          <year>2015</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>