<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Thematically Related Words toward Creative Information Retrieval</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Eiko</forename><surname>Yamamoto</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Graduate School of Engineering</orgName>
								<orgName type="institution">Kobe University</orgName>
								<address>
									<addrLine>1-1 Rokkodai-cho, Nada-ku</addrLine>
									<postCode>657-8501</postCode>
									<settlement>Kobe</settlement>
									<region>Hyogo</region>
									<country key="JP">Japan</country>
								</address>
							</affiliation>
						</author>
						<author role="corresp">
							<persName><forename type="first">Hitoshi</forename><surname>Isahara</surname></persName>
							<email>isahara@nict.go.jp</email>
							<affiliation key="aff1">
								<orgName type="institution">National Institute of Information and Communications Technology</orgName>
								<address>
									<addrLine>3-5 Hikaridai, Seika-cho, Soraku-gun</addrLine>
									<postCode>619-0289</postCode>
									<settlement>Kyoto</settlement>
									<country key="JP">Japan</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Thematically Related Words toward Creative Information Retrieval</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">F2096CA3FACA40939F04C3F9A21A51B0</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T01:21+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Natural Language Processing</term>
					<term>retrieval support</term>
					<term>related words</term>
					<term>thematic relation</term>
					<term>taxonomical relation H5.2. INFORMATION INTERFACES AND PRESENTATION (e.g.</term>
					<term>HCI): User Interfaces -Natural Language; H.3.3. INFORMATON STORAGE and RETRIEVAL: Information Search and Retrieval</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>We introduce a mechanism that provides key words which can make human-computer interaction increase in the course of information retrieval, by using natural language processing technology and mathematic measure for calculating degree of inclusion. We show what type of words should be added to the current query, i.e. keywords which previously had been input, in order to make humancomputer interaction more creative. We try to extract related word sets from documents by employing casemarking particles derived from syntactic analysis. Then, we verify which kind of related words is more useful as an additional word for retrieval support.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>INTRODUCTION</head><p>Nowadays, we can access huge amount of text data available on the web. The increase of the data quantity causes a paradigm shift for web retrieval. Rhetorically speaking, we can take a walk among the huge text data. The web retrieval supports we need in this novel situation are neither simple query expansion nor our (or someone's) record of previously input keywords, but we need interfaces which interact with people in new ways. What is crucial for such interface is not constructions of interface, i.e. how each part of interface is arranged on the screen, but what information is presented to interact with users.</p><p>New ideas pop into one's head when he/she strolls in library, bookstore, or even around town. We need retrieval supports which enable us to expand such creativity. Making computer smarter to automatically extract "correct" retrieval result is one-side way of developing support systems for information retrieval. Seeing the advice provided to a user by computer, how the user achieves next retrieval is one of the most important viewpoints for the future intelligent user interface. We need a technology that enables computer to understand huge text data and make it possible to expand the users' way of thinking.</p><p>In this paper, we introduce a mechanism that provides key words which can make human-computer interaction (HCI) during the information retrieval increase, by using natural language processing technology and mathematic measure for calculating degree of inclusion. Concretely, we show what type of words should be added to the current query, i.e. keywords which previously had been input, in order to make HCI more creative.</p><p>relation <ref type="bibr">[9]</ref>. 1 The former is a relation representing the physical resemblance among objects, such as, "cow" and "animal," which is typically a semantic relation; the latter is a non-taxonomical relation among objects through a thematic scene, such as "milk" and "cow" as recollected in the scene "milking a cow," which includes causal relation and entailment relation. Taxonomically related words are generally used to query expansion and it is comparatively easy to identify taxonomical relations from linguistic resources such as dictionaries and thesauri. On the other hand, it is difficult to identify thematic relations because they are rarely maintained in linguistic resources.</p><p>Most of the previous researches of information retrieval support tended to focus on the improvement of recall by the query expansion. However, our aim is to direct users to the informative information for them by query suggestion. Though some users sometimes do not realize their real intention of retrieval, we would like them to find their hidden needs via interaction in which system shows them suggestive terms.</p><p>In this paper, we try to extract related word sets from documents in Japanese by employing case-marking particles derived from syntactic analysis. Then, we compared the results retrieved with words related only taxonomically and those retrieved with words that included a word related non-taxonomically to the other words in order to verify what kind of relation makes humancomputer interaction more creative.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>WORD SET EXTRACTION METHOD</head><p>In order to derive word sets that direct users to obtain information, we applied the method based on the Complementary Similarity Measure (CSM) which can estimate inclusive relations between two vectors [10]. This measure was developed as a means of recognizing degraded machine-printed text <ref type="bibr">[5]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Estimating Inclusive Relation between Words</head><p>We first extract word pairs having an inclusive relation of the appearance patterns between the words by calculating the CSM values. An appearance pattern is expressed a kind of co-occurrence relation by an n-dimensional binary feature vector. Therefore, the dimension of each vector corresponds to a co-occurring word, a document, or a sentence. When V i = (v i1 , ..., v in ) is a vector for word w i and V j = (v j1 , ..., v jn ) is a vector for word w j , CSM(V i , V j ) is defined by the following formula: 1 The taxonomical relation which is, for example, provided by WordNet <ref type="bibr" target="#b0">[1]</ref> corresponds to "classical" relation by Morris and Hirst <ref type="bibr">[7]</ref>, and the thematic relation corresponds to "non-classical" relation.</p><p>CSM is an asymmetric measure because the denominator is asymmetric. Therefore, CSM(V i , V j ) usually differs from CSM(V j , V i ) exchanged between V i and V j . For example, when V i is 1110010111 and V j is 1000110110, parameters for CSM(V i , V j ) are a = 4, b = 3, c = 1, and d = 2, and CSM(V i , V j ) is greater than CSM(V j , V i ). According to the asymmetric feature, we can estimate whether the appearance pattern of w i includes the appearance pattern of w j . If w i is "animal" and w j is "tiger," CSM would estimate that "animal" is a hypernym of "tiger."</p><p>Extracted word pairs based on the appearance patterns are expressed by a tuple &lt;w i , w j &gt; which is a directed set of words. Tuple &lt;w i , w j &gt; represents that CSM(V i , V j ) is greater than CSM(V j , V i ) when words w i and w j have each appearance pattern represented by each binary vector V i and V j . We call w i the "left word" and w j the "right word."</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Constructing Related Word Sets</head><p>We next connected such word pairs with CSM values greater than a certain threshold and constructed word sets.</p><p>If we adopt simpler mechanism such as the co-occurrence frequency, which extracts only the co-occurrence relations between words, two tuples extracted from different sentences cannot be merged easily. A feature of our method is that because we use the CSM to calculate the degree of inclusion of appearance patterns between all combinations of words in whole collection of texts, we can connect word pairs consistently. That is to say, we can extract not only pairs of related words but also sets of related words. In other words, our CSM-based method is relevant not only to information within a sentence or a document, but also to information from a wider context. That is, once we obtain two tuples &lt;A, B&gt; and &lt;B, C&gt;, even though the tuples have been extracted from different sentences or documents, we can obtain word set {A, B, C} in order.</p><p>Suppose we have tuples &lt;A, B&gt;, &lt;B, C&gt;, &lt;Z, B&gt;, &lt;C, D&gt;, &lt;C, E&gt;, and &lt;C, F&gt;, which are word pairs having greater CSM values than the threshold (TH) in the order of their values. For example, let &lt;B, C&gt; be an initial word set {B, C}. We create a word set as follows.</p><p>1. We find the tuple with the greatest CSM value among the tuples in which the word at the tail of the current word setfor example, C in {B, C}is a left word, and connect the right word of the tuple to the tail of the current word set. In this example, word "D" is connected to {B, C} because &lt;C, D&gt; has the greatest</p><formula xml:id="formula_0">. ) 1 ( ) 1 ( , ) 1 ( , ) 1 ( , , ) )( ( ) , ( 1 1 1 1 ∑ ∑ ∑ ∑ = = = = − ⋅ − = ⋅ − = − ⋅ = ⋅ = + + − = n k jk ik n k jk ik n k jk ik n k jk ik j i v v d v v c v v b v v a d b c a bc ad V V CSM</formula><p>CSM value among the three tuples &lt;C, D&gt;, &lt;C, E&gt;, and &lt;C, F&gt;, making the current word set {B, C, D}.</p><p>2. This process is repeated until no tuples with a CSM value greater than TH can be chosen.</p><p>3. We find the tuple with the greatest CSM value among the tuples in which the word at the head of the current word setfor example, B in {B, C, D}is the right word, and connect the left word of the tuple to the head of the current word set. In this example, Word "A" is connected to the head of {B, C, D} because &lt;A, B&gt; has a CSM value greater than that of &lt;Z, B&gt;, making the current word set {A, B, C, D}.</p><p>4. This process is repeated until no tuples with a CSM value greater than TH can be chosen.</p><p>In this example, we obtained the word set {A, B, C, D} beginning with tuple &lt;B, C&gt; as the initial word set {B, C}.</p><p>In this way, we construct all word sets by beginning with each tuple, using tuples whose CSM values are greater than TH. Then from the word sets obtained, we remove word sets that are embedded in other word sets.</p><p>If we set TH to a low value, it is possible to obtain lengthy word sets. When the TH is too low, the number of tuples that must be considered becomes overwhelming and the reliability of the measurement decreases. Consequently, we experimentally set TH.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Extracting Word Sets with Thematic Relation</head><p>Finally, we use a thesaurus to extract word sets with a thematic relation. We remove word sets with taxonomical relations from whole word set we extracted, and get the leftover as word sets with thematic (at least nontaxonomical) relations. The heading words in a thesaurus are categorized to represent a taxonomical relationship. If a word set extracted by the CSM-based method demonstrates a taxonomical relation among the words, the words in the CSM-based word set are classified into one category in the thesaurus; that is, if an extracted word set agrees with the thesaurus, we can conclude that a taxonomical relation exists among the words. Through this approach, we remove those word sets with a taxonomical relation by examining the distribution of words in the categories. The rest of the word sets have a non-taxonomical relation -including a thematic relation -among the words. We then extract those word sets that do not agree with the thesaurus, having identified them as word sets with a thematic relation, that is, thematically related word sets.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>LINGUISTIC DATA</head><p>We extract word sets by utilizing inclusive relations of the appearance pattern between words based on a modifiee/modifier relationship in documents. The Japanese language has case-marking particles that indicate the semantic relation between two elements in a dependency relation, which is a kind of modifiee/modifier relationship.</p><p>For our experiment, we used such particles and extracted the data from the documents we gathered.</p><p>First, we parsed sentences with the KNP 2 . From the results, we collected dependency relations matching one of the following five patterns of case-marking particles. With A, B, P, Q, R, and S as nouns (including compound words); V as a verb; and &lt;X&gt; as a case-marking particle with its role in parentheses, the five patterns are as follows: From this set of dependency relations, we compiled the following types of experimental data 3 :</p><formula xml:id="formula_1">A &lt;no (of)&gt; B P &lt;wo (object)&gt; V Q &lt;ga (subject)&gt; V R &lt;ni (dative)&gt; V S &lt;ha (topic)&gt; V</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>NN-data based on co-occurrence between nouns.</head><p>For each sentence in our document collection, we gathered nouns followed by all five of the case-marking particles we used and nouns proceeded by &lt;no&gt;; that is, A, B, P, Q, R, and S. For the above sentence, we can gather Chloe, 2 A Japanese parser developed at Kyoto University.</p><p>3 Japanese case-marking particles define not deep semantics but rather surface syntactic relations between words/phrases; therefore, we utilized not semantic meanings between words, but classifications by casemarking particles. Therefore, the method proposed in this paper is applicable to other languages when a syntactic analyzer that classifies relations between elements, such as subject, direct object, and indirect object, exists for the language. For example, from the output of English parser, we can compile necessary linguistic data, such as Wo-data using collocations between verb and its direct object, Gadata from collocations between verb and its subject, Nidata from collocations between verb and its indirect object, and SO-data from collocations between subject and object of a verb.</p><p>Mike, Judy, bara, and hanataba. The number of data items equals the number of sentences in the documents.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>NV-data based on a dependency relation between noun and verb.</head><p>We gathered nouns P, Q, R, and S followed by each of the case-marking particles &lt;wo&gt;, &lt;ga&gt;, &lt;ni&gt;, and &lt;ha&gt; for each verb V. We named them Wo-data (with 20,234 gathered data items), Ga-data (15,924), Ni-data (14,215), and Ha-data (15,896), respectively. For the verb okutta in the above sentence, the Wo-data is hanataba, Ga-data is Mike, and so on.</p><p>The number of data items equals the number of kinds of verbs.</p><p>SO-data based on a collocation between subject and object. We gathered subject Q followed by the case-marking particle &lt;ga&gt; that depends on the same verb V as the object P for each object followed by the case-marking particle &lt;wo&gt;. For the above example, we can gather the subject Mike for the object hanataba because we have the dependency relations Mike &lt;ga&gt; okutta and hanataba &lt;wo&gt; okutta. The number of data items equals the number of kinds of objects, where each of them co-occurs with a subject in a sentence and depends on same verb as the subject (4,437).</p><p>When we represent experimental data with a binary vector, the vector corresponds to the appearance pattern of a noun. Parameters for calculating the CSM-value correspond to the number of dimensions in each situation. Figure <ref type="figure">1</ref> shows images of the appearance pattern expressed by the binary vector for each data item. The number of dimensions equals the number of data items for each experimental data. For NN-data, each dimension corresponds to a sentence. The element of the vector is 1 if the noun appears in the sentence and 0 if it does not. Similarly, for NV-data, each dimension corresponds to a verb. For SO-data, we represent the appearance pattern for each subject with a binary vector whose dimension corresponds to an object.</p><p>Therefore, if we calculate the CSM value between Vector A and Vector B, each of the parameters a, b, c, and d used for CSM formula explained above corresponds to the number of each of the following cases:</p><p>The number of dimensions which both Vector A and Vector B have 1 as each element.</p><p>The number of dimensions which Vector A has 1, though Vector B has 0.</p><p>The number of the dimensions which Vector B has 1, though Vector A has 0.</p><p>The number of the dimensions which both Vector A and Vector B have 0.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>EXPERIMENT</head><p>In our experiment, we used domain-specific documents in Japanese from the medical domain gathered from the Web pages of a medical school. The Japanese documents we used totaled 225,402 sentences (10,144 pages, 37MB).</p><p>In applying the CSM-based method, we represented experimental data for medical terms with a binary vector as explained above. We used descriptors in the 2005 Medical Subject Headings (MeSH) thesaurus <ref type="foot" target="#foot_0">4</ref> and translated them into Japanese. The number of terms in Japanese appearing in this experiment is 2,557. We constructed word sets consisting of these medical terms and chose the word sets consisting of three or more terms from them. Figures <ref type="figure">2 and  3</ref> show examples of word sets constructed with the CSMbased method. Note that we obtained word sets comprising Japanese medical terms that appear in the Japaneselanguage medical documents we used. For explanatory purposes, in the following part of this paper we use English terms obtained from the MeSH thesaurus.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>data -causation -depression -reduction -platelet count -bone marrow examination neonate -patent ductus arteriosus</head><p>-necrotizing enterocolitis secretion -gastric acid -gastric mucosa -duodenal ulcer skin -atopic dermatitis -herpes viruses -antiviral drugs fatigue -uterine muscle -pregnancy toxemia water -oxygen -hydrogen -hydrogen ion person -nicotiana -smoke -oxygen deficiencies Notice that some headings are classified into more than one category.</p><p>We examined the distribution of terms in the MeSH categories for each word set and extracted word sets that do not agree with the MeSH thesaurus as word sets with a thematic relation. Table <ref type="table" target="#tab_1">1</ref> shows the number of word sets that agree and that disagree with the MeSH thesaurus. By way of exception, for example, we obtained the word set "tree -forest -orangutan" from NN-data. "Tree" is classified into two categories "Organisms (B)" and "Technology and Food and Beverages (J)"; "forest" is classified into "J" and "orangutan" is classified into "B." In this case, we consider that a relation exists between "forest" and "orangutan" via "tree," and we treat this word set as being distributed in one category.</p><p>In Table <ref type="table" target="#tab_1">1</ref> we found that, for NN-data and NV-data, the ratio of CSM-based word sets that agreed with MeSH thesaurus was between 7.5% and 29.1%. Of the CSM-based word sets, Wo-data provided the highest agreement ratio.</p><p>The apparent reason for this is that the object case represented by the case-marking particle &lt;wo&gt; restricts nouns more stringently than do the others. Also, comparing the results of NN-data and NV-data, we found that the word sets extracted from NV-data agreed with the MeSH thesaurus to a greater degree than did those extracted from NN-data. This suggests that we obtained more word sets having taxonomical relations among words from NV-data than from NN-data.</p><p>SO-data is based on a collocation between subject and object; that is, the word sets obtained comprise subjects followed by the case-marking particle &lt;ga&gt; that depend on the same verb as the object for each object followed by the case-marking particle &lt;wo&gt;. For example, when we have "ningen (person) &lt;ga&gt; hon (book) &lt;wo&gt; yomu (read)," which means, "a person reads a book," and "nezumi (mouse) &lt;ga&gt; hon (book) &lt;wo&gt; kajiru (gnaw)," which means, "a mouse gnaws a book," we estimate the relation between the words ningen and nezumi with CSM. Therefore, we can surmise that the information we obtain from this data will not agree with a general thesaurus because we do not limit the verbs that subjects and objects depend on. Actually, the word sets we obtained from SO-data agreed little with the MeSH thesaurus.</p><p>Figure <ref type="figure">4</ref> shows examples of taxonomically related word sets, which agree with the MeSH thesaurus, that is, in which all the composing terms in a word set are classified into one category. The symbol in brackets represents the type of data from which each word set was obtained. As the result, we obtained the rest as word sets with a thematic relation, that is, thematically related word sets, which are 847 word sets.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>VERIFICATION</head><p>In verifying the capability of our word sets to retrieve Web pages, we examined whether our word sets could help limit the search results to more informative Web pages with Google as a search engine. To do this, in our obtained word sets with a thematic relation, we used 294 word sets in which one of the terms is classified into one category and the rest are classified into another category. Figure <ref type="figure">5</ref> shows examples of the word sets.  We used the terms that composed such word sets as the key words to input into the search engine and retrieved Web pages. We created three types of search terms from a word set. Suppose the word set is {X 1 , …, X n , Y}, where X i is classified into one category and Y is classified into another. The first type uses all terms except the one classified into a category different from the others: {X 1 , …, X n }, removing Y. The second type uses all terms except the one in the same category as the rest: {X 1 , …, X k-1 , X k+1 , …, X n } removing X k and Y. In our verification, we removed the term X k with the highest or lowest frequency among X i .</p><p>The third type uses terms in Type 2 and Y, i.e., terms in another category: {X 1 , …, X k-1 , X k+1 , …, X n , Y}. When we consider Type 2 as base key words, Type 1 is a set of key words with the addition of one term having the highest or lowest frequency among the terms in the same category; i.e., the additional term X k has a feature related to frequency and is taxonomically related to other terms. Type 3 is a set of key words with the addition of one term in a category different from those of the other component terms; i.e., the additional term Y seems to be thematically related to other terms.</p><p>The retrieval results are shown in Figures <ref type="figure">6 and 7</ref> including the results for each the highest frequency and the lowest frequency. The horizontal axis is the number of pages retrieved with Type 2 and the vertical axis is the number of pages retrieved with Type 1 or Type 3 that a certain term X k or Y is added to Type 2. The circles show the results with Type 1 and the crosses show the results with Type 3. The diagonal line in the graph shows that adding one term to Type 2 does not affect the number of Web pages retrieved.</p><p>As shown in Figure <ref type="figure">6</ref>, most crosses fall further below the line. This graph indicates that adding a search term related non-taxonomically tends to make a bigger difference than adding a term related taxonomically and with high frequency. This means that adding a term related nontaxonomically to key words is crucial to retrieving informative pages, i.e., such terms are informative terms themselves.   decreasing the number of Web pages retrieved. This means that, in comparison to the high frequency terms, which might not be so informative in themselves, the terms in the other category -related non-taxonomically -are effective for retrieving useful Web pages.</p><p>Constantly, in Figure <ref type="figure">7</ref>, most circles fall further below the line. This indicates that adding a term related taxonomically and with low frequency tends to make a bigger difference than does adding a term with high frequency. Certainly, additional terms with low frequency would be informative terms, even though they are related taxonomically, because they may be rare terms on the Internet. Thus, the taxonomically related terms with low frequencies are quantitatively effective for information retrieval as the nontaxonomically related terms.</p><p>Table <ref type="table">3</ref> shows the number of cases in which term in different category decreases the number of hit pages more than low frequency term. In comparing these numbers, we also found that the additional term with low frequency helped to reduce the number of Web pages retrieved, making no effort to determine the kind of relation the term had with the other terms. Thus, the terms with low frequencies are quantitatively effective when used for retrieval.</p><p>However, if we consider contents of the results retrieved with Type 1 and Type 3, it is clear that big differences exist between them. For example, consider "latency perioderythrocyte -hepatic cell" obtained from SO-data in Figure <ref type="figure">3</ref>. "Latency period" is classified into a category different from the other terms and "hepatic cell" has the lowest frequency in this word set. When we used all the three terms, we obtained pages related to "malaria" at the top of the results and the title of the top page was "What is malaria?" in Japanese. With "latency period" and "erythrocyte," we again obtained the same page at the top, although it was not at the top when we used "erythrocyte" and "hepatic cell" which have a taxonomical relation.</p><p>As we showed above, the terms with thematic relations with other search terms are effective at directing users to informative pages. Quantitatively, terms with a high frequency are not effective at reducing the number of pages retrieved; qualitatively, low frequency terms may not effective to direct users to informative pages.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>CONCLUSION</head><p>We introduced a mechanism that provides key words which can make human-computer interaction (HCI) increase, by using natural language processing technology and mathematic measure for calculating degree of inclusion. We showed what type of words should be added to the current query, i.e. keywords which previously had been input, in order to make HCI more creative.</p><p>We extracted related word sets from documents by employing case-marking particles derived from syntactic analysis. Then, we verified which kind of related word is more useful as an additional word for retrieval support.</p><p>That is, we found the additional term which is thematically related to other terms is effective at retrieving informative pages by comparing the results retrieved with words related only taxonomically and those retrieved with words that included a word related non-taxonomically to the other words. This suggests that words with a thematic relation can be useful to make the HCI more active.</p><p>As for the future directions of this work, one of most crucial issues is evaluation. We will evaluate the effectiveness of our method from human-centered viewpoints, possibly by human judgement.</p><p>In the future, we can understand the contents of huge text data with higher natural language processing technology and develop a system which makes it possible to expand the users' ways of thinking.</p><p>5. Hagita, N. and Sawaki, M. Robust recognition of degraded machine-printed characters using complementary similarity measure and error-correction learning. In Proc. SPIE -The International Society for Optical <ref type="bibr">Engineering, 2442</ref><ref type="bibr">Engineering, , (1995))</ref> </p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 2 .Figure 1 .</head><label>21</label><figDesc>Figure 2. Examples of word sets extracted from NN-data</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 4 .Figure 3 .</head><label>43</label><figDesc>Figure 4. Examples of taxonomically related word sets</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 5 .Figure 7 .Figure 6 .</head><label>576</label><figDesc>Figure 5. Examples of word sets used to verify</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 1 . The numbers of word sets that agreed/disagreed with the MeSH thesaurus.</head><label>1</label><figDesc></figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 2</head><label>2</label><figDesc></figDesc><table><row><cell>shows the number of cases in which</cell></row><row><cell>term in different category decreases the number of hit pages</cell></row><row><cell>more than high frequency term. Then, we found that most</cell></row><row><cell>of the additional terms with high frequency contributed less</cell></row><row><cell>than additional terms related non-taxonomically to</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 2 . Number of cases in which the term in different category decreases the number of Web pages retrieved more than the high frequency term.Table 3 . Number of cases in which the term in different category decreases the number of Web pages retrieved more than the low frequency term.</head><label>23</label><figDesc></figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head></head><label></label><figDesc>, 236-244. 6. Hearst, M. A. Automatic acquisition of hyponyms from large text corpora, In Proc. Coling 92, (1992), 539-545. 7. Morris, J. and Hirst, G. Non-classical lexical semantic relations. Workshop on Computational Lexical In Proc. Human Language Technology Conference of the NAACL, (2004). 8. Pantel, P. and Pennacchiotti, M. Espresso: Leveraging generic patterns for automatically harvesting semantic relations In Proceedings of ACL 2006, (2006), 113-120. 9. Wisniewski, E. J. and Bassok, M. What makes a man similar to a tie? Cognitive Psychology, 39, (1999), 208-238. 10. Yamamoto, E., Kanzaki, K., and Isahara, H. Extraction of hierarchies based on inclusion of co-occurring words with frequency information. In Proc. IJCAI2005, (2005), 1166-1172.</figDesc><table><row><cell>Type of Data Semantics,</cell><cell>NN</cell><cell></cell><cell>NV</cell><cell></cell></row><row><cell></cell><cell></cell><cell cols="3">Wo Ga Ni Ha</cell></row><row><cell>Number of word sets for verification</cell><cell cols="4">175 43 23 13 26</cell></row><row><cell>Number of cases in</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>which Type 3 defeated</cell><cell>61</cell><cell>18</cell><cell>7</cell><cell>6 13</cell></row><row><cell>Type 1 in retrieval</cell><cell></cell><cell></cell><cell></cell><cell></cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_0">The U.S. National Library of Medicine created, maintains, and provides the Medical Subject Headings (MeSH ® ) thesaurus.</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title level="m" type="main">WordNet: An electronic lexical database</title>
		<author>
			<persName><forename type="first">C</forename><surname>Fellbaum</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1998">1998</date>
			<publisher>The MIT Press</publisher>
			<pubPlace>Cambridge, Mass</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">The distribution inclusion hypotheses and lexical entailment</title>
		<author>
			<persName><forename type="first">M</forename><surname>Geffet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Dagan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. ACL 2005</title>
				<meeting>ACL 2005</meeting>
		<imprint>
			<date type="published" when="2005">2005</date>
			<biblScope unit="page" from="107" to="114" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Automatic detection of causal relations for question answering</title>
		<author>
			<persName><forename type="first">R</forename><surname>Girju</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. ACL Workshop on Multilingual summarization and question answering</title>
				<meeting>ACL Workshop on Multilingual summarization and question answering</meeting>
		<imprint>
			<date type="published" when="2003">2003</date>
			<biblScope unit="page" from="76" to="114" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Automatic discovery of part-whole relations</title>
		<author>
			<persName><forename type="first">R</forename><surname>Girju</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Badulescu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Moldovan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computational Linguistics</title>
		<imprint>
			<biblScope unit="volume">32</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="83" to="135" />
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
