=Paper=
{{Paper
|id=None
|storemode=property
|title=Models for Automatic Retrieval of Health Information on the Web
|pdfUrl=https://ceur-ws.org/Vol-729/paper5.pdf
|volume=Vol-729
}}
==Models for Automatic Retrieval of Health Information on the Web==
Models for automatic retrieval of health information on the
Web
Ana Marilza Pernas1.2 Jonas Bulegon Gassen2 José Palazzo M. de Oliveira2
1 2 2
Centro de Desenvolvimento Instituto de Informática Instituto de Informática
Tecnológico, Universidade Federal de Universidade Federal do Rio Grande Universidade Federal do Rio Grande
Pelotas, RS, Brasil do Sul, Porto Alegre, Brasil do Sul, Porto Alegre, Brasil
ana.pernas@inf.ufrgs.br jbgassen@inf.ufrgs.br palazzo@inf.ufrgs.br
ABSTRACT be: sponsors, if exists and who are; the subject of the page; the
Evaluate the data available in Web pages is necessary to allow ways of contact with the responsible by the web page; the web
the suggestion of adequate content to a specific public, for such, page readability; if the content is continuously updated (fresh-
an idea is to develop an automatic detection mechanism of the ness).
quality present in Web pages content. Due to its quickly and not To achieve this objective there are organizations that grant cer-
standardized growth, the content freely available in the Web has tificates of quality in form of quality-stamps to ensure the quali-
reached very large proportions, becoming extremely difficult to ty of the information presented on a website. That is the case of
automatic manage this large mass of data. One of the main rea- Web Mèdica Acreditada1 (WMA), an organization of the Medi-
son for this complexity is the not appliance of standards for its cal Association of Barcelona that offers ―a quality programme
construction, which should lead to a standardized access. This addressed to medical websites. Through a voluntarily certifica-
article describes the development of models and techniques to tion process, websites that follow that programme meet a set of
locate, standardize and extract the content in web pages asso- quality criteria, making possible a trustworthy virtual communi-
ciated with health issues. After that, the objective is to provide ty on the Internet for general public, patients and health profes-
an appropriate content to evaluate the quality of a web page sionals‖ [1]. Other examples of organizations which grant quali-
according to specific metrics. ty seals to websites are the Internet Content Rating Association
(ICRA) [2] and the Internet Quality Agency (IQUA) [3].
Categories and Subject Descriptors Even with this kind of organization to ensure the quality of
H.3.3 [Information Search and Retrieval]: Information filter- health information on the Web by a formal request, the best
ing; Retrieval models; Selection process. alternative would be to orchestrate this practice performing an
J.3 [Life and Medical Sciences]: Health; Medical information automated analysis of the content presented in health web pages.
systems. However, given the size of the Web nowadays this task would
be really complex and hard, as an example the indexing of the
Web content by the Yahoo! indicates that ―the production of the
General Terms Web Map, an index of the World Wide Web that is a critical
Management, Design, Experimentation, Standardization, Verifi- component of search {takes} (75 hours elapsed time, 500 tera-
cation. bytes of MapReduce intermediate data, 300 terabytes total out-
put)‖ using larger clusters of 3500 nodes [19]. In a post pub-
Keywords lished in July 2008 in the Google Blog the Web size was eva-
Location and extraction models, standardization, quality metrics. luated as already exceeding 1 trillion of URLs (Uniform Re-
source Locator) [4].
1. INTRODUCTION To enable automatic retrieval of data from web pages about
One of the biggest challenges in Web systems is to automatically health, some models and techniques are presented here to sup-
deal with the large amount of data available in the Web as a port: (i) location of health web-data (ii) standardization and (iii)
large distributed system. As a consequence of its rapid not stan- (ii) extraction of this data, based on pre-established criteria.
dardized growing, the Web reached a huge proportion that be- Conscious of the complexity of dealing to the entire Web, the
came extremely difficult to be efficiently controlled, being the location, standardization and extraction models presented here
automatic management of data very complex. are applied to specific search engines.
The automatic web-data management is especially desired when The remainder of this article is structured as follows. Section 2
the subject is the quality evaluation of a given web page for its explores some related works. Section 3 describes the general
subsequent recommendation to a specific public. In the case of vision of this work, showing our starting point and the relevant
health information content it is necessary a careful check on the data in estimating the quality of a web page. Section 4 is asso-
validity of the information before an effective recommendation. ciated with the task of automatic location of web page about
We know that check all health information on the Web to de- health, explaining its general model. Section 5 present the task
termine if is reliable it‘s very difficult, but we can evaluate a of automatic extraction data from the localized web pages pre-
number of indicators present in a web page to try to guarantee a
minimum of quality. Examples of indicators, or metrics, could 1
http://wma.comb.es/
senting problems related to the lack of standardization. Section 6 Standardization – relates to existing standards in presen-
shows a case study developed on pages related with Alzheimer‘s tation of data in web pages which may indicate ways to
disease. Finally section 7 presents the conclusions and future achieve automatic access (automatic extraction) of this da-
works. ta.
Extraction – once the web page is located, how to obtain
2. RELATED WORKS and properly manage its data to be evaluated?
An example of project related to the analysis of quality content
on health web pages is the QUATRO Project (The Quality As- Quality – is the analysis of collected data where metrics
surance and Content Description Project) and its successor, are applied to define the level of quality present in the eva-
QUATRO+ [5]. The objective of this project is to offer a com- luated web page, according to a specific user profile.
mon machine readable vocabulary to certify the quality of health As we mentioned in section 1, in this work only the phases of
information on the Web. One of the results of the QUATRO Location, Standardization and Extraction are treated. The start-
project consists on a vocabulary that presents a list of descrip- ing point in developing models was the research described in
tors and their definitions to be used as a base for creation of [10]. In [10], the main points to be collected in a Web page were
quality seals. Among the participants of the QUATRO project is analyzed, in order to determine its quality. This analysis resulted
the aforementioned Web Mèdica Acreditada – WMA. in an ontology of quality, which was reduced to the model de-
An old work, but that can still be referenced by its vision is the picted in the Figure 1 [11]. This reduction was made to simplify
so called ―Oh, yeah?‖ button, proposed by Tim Berners-Lee in the task of automatic retrieval because obtaining this main data
[6]. In this proposition, the author mention how important is the about a health web page is already possible to define a first qual-
Web say to the users something about the information being ity estimative.
presented. This ―Oh, yeah?‖ button would be responsible when As we can see in Figure 1, information about the web page is
the user is not so confident in a web page content to show a obviously necessary to define its quality. Information about
number of reasons to trust in that information. In a practical author is very important because if the content was written or
view, the button would access some meta information about the revised by a specialist in the subject, possibly the quality will be
content and show that to the user. enhanced. Is either important define if the web page is spon-
Related to information extraction, the Project AQUA (Assisting sored by some organization and if is a recognized organization,
Quality Assessment) [7] proposes to automate parts of the work for example, a governmental organization, university or indus-
manually done by organizations that offers quality labels to try.
websites, making this process easily and, consequently, increas- The E-R (Entity-Relationship) model of the Figure 1 describes
ing the number of sites with quality seals. The objective is to the main classes for determining the page quality: (i) the page
crawl the Web to locate unlabelled health web resources, sug- itself, with data related with its title, language, authorship, refer-
gest machine readable labels for them, according to predefined ences and dates of creation and update; (ii) the web page author,
labeling criteria, and monitor them. To do that, the crawling with data related with contact and expertise in the subject cov-
mechanism uses Google2 and Yahoo!3 search engines to do a ered by the web page; (iii) the organization that sponsor the Web
meta-search engine on the Web, collecting and filtering (to ig- page (if exists); (iv) the links of another pages or sites in the
nore sub-paths of URLs already in list and removes URLs hav- Web pointing to this (inLinks); and (v) which are the web links
ing already a content label) the resulting URLs. This work is from this page to other web pages (outLinks).
very similar to our work, but data related with authorship infor-
mation and inLinks and outLinks are not mentioned.
Another related work is described in [8] which aim to extract a
number of quality indicators defined by organizations like
HONCode to establish the quality content of a health website.
The work looks to detect measurable indicators in the evaluated
website, searching this information in HTML tags and meta-
tags. However, in our view, only this simple analysis could not
cover a great set of websites.
3. GENERAL MODEL
The general model presented in this work intends to cover criti-
cal points considered to achieve the final goal: evaluate the qual-
ity of a Web page about health. In a general view it is necessary
to accurately automate the following steps:
Localization – answer to the question: given an item X
stored in some dynamic set of nodes in the system, how to
find it? [9].
Figure 1. E-R model proposed to evaluate the content of
2 health web pages [11].
http://www.google.com
3
http://yahoo.com
In the development of the techniques to collect data on the mod- ing of returned results as well as for its storage in the database
el of Figure 1, was noted the difficulty in obtain information [11].
about the date of creation and update of web pages. This is very
important in determining the quality of a web page, because
expresses the content freshness. We observed that this happen in
general or by the absence of this kind of information or because
is presented in a non machine readable format, like in an image.
Thus, the set of data in which was possible to automatic extract
information in this work is related to the site language and its
inLinks and outLinks. About the authorship, it was possible to
determine the author expertise and contact (e-mail address).
The following sections describe the models and techniques de-
veloped to perform the automatic location and extraction of web
pages to obtain the set of data described on the model presented
in the Figure 1.
4. AUTOMATIC LOCATION
One of the main points in the system‘s operation was the auto-
matic location of health web pages to be evaluated and ex-
tracted. The results achieved must be returned to the evaluation
process in order of relevance, i.e. according to how they satisfy
the search criteria [11]. For more details regarding the criteria of
quality, we recommend the reading of the work developed in
[10].
Given the estimated size of the Web, find the more appropriate
page to the user‘s intention is not a trivial task. Its necessary a
sophisticate algorithm, combined with massive computational
power, to accomplish this task. For that reasons, in this work
was chose to use search engines specifically oriented for health-
pages retrieval. In this category of search engines the search is
performed on a database of previously indexed and constantly Figure 2. Model to locate data in the Web
updated pages. Some of the specialized search engines are Sear- (modified from [11])
chMedica4 and MedStory5, which are specific to the medical
field. In the next sections is present the model and prototype
developed to locate health web pages, but not focusing on spe-
5. DATA EXTRACTION
After recovering the URLs from the search engines, obtained at
cific search engines – as this choice is application-dependent.
the location step, the system starts the next task, which extracts
The specific configuration of this model to a practical applica-
the content that will be used to evaluate the quality of the web
tion is presented in section 6, with a Case Study about Alzhei-
page. As this task was idealized to occur in an automatic way a
mer‘s disease.
search is performed for specific attributes that are relevant to
determine the page quality, those attributes can be found in the
4.1 Location Model Figure 1. The initial supposition was that it should be possible to
The model presented in the Figure 2 is based on the interaction find a standardization way that allows encompassing the maxi-
among the application, a database and the Internet. It is applica- mum portion of data existent in web pages in order to make the
tion dependent because it needs the existing pattern for present- data extraction easier and faster. However, as presented in a
ing the results of the search engine used in the application, i.e., more detailed way in the 5.1 subsection, it was not possible to
the internal standard to display the results to the user. use this approach because until today there are no standardiza-
Initially, the data about the search engine (pattern related) is tion models for naming the tags or to place data in pages. There-
required by the system. Then, the criterion to be used by the fore, specific strategies were proposed, trying to attend the ne-
search engine is recovered from the database and for each crite- cessary requirements to the search of each data existent in the
ria is stored the first URLs returned as answer by this search ER model (Figure 1), as: site language; authorship data; inLinks
engine. For more details about this model see [11]. The database and outLinks.
stores the identification of the search engine used, as well as The next sections present each one of those strategies, starting
their model. This is necessary because different engines offer with troubles and adopted strategies for standardization of web
specifics return forms. Thus, it is possible to extend the system, pages content and structure.
adding new patterns of search engines or removing them from
the database. A prototype was developed for retrieval and index-
5.1 Standardization
The considered standardization has the objective to define the
4 ways in which the content of a web page could be structured.
http://www.searchmedica.com
5 That research is applied later as an information source to the
http://www.medstory.com
extraction phase. The structure and content of search engines, as a higher score, and consequently, appear in a better position in
well as web pages, is extremely dynamic. Consequently, to over- the ranking. In this work, inLinks are recovered as follows [11]:
come this constantly changes the standardization model adopted
here apply terms defined in the WorldNet [12] trying to found An request is submitted to the Browser Google Web:
synonyms for the keywords found in web pages. Examples of http://www.google.com.br/search?q=link%3A+;
the WorldNet, are: "writer", "generator", "source". After recover the results page, for each one of the listed
There are others interesting approaches that can be applied as a pages, the XPath expression: //cite is executed;
standardization strategy. An example is the application of ontol- All lists of URLs obtained in each page are stored.
ogies [13], where they can be used to support decisions about
which fields must to be analyzed; these ontologies could, either, For the recovery of outLinks the steps are [11]:
be enriched by the WorldNet terms. The use of terms from
Recover the page which intends to get the outLinks;
WorldNet is interesting to find synonyms of the fields that could
appear in the pages and be used in the extraction. Unfortunately, Execution of the expression XPath: //a[not(contains(@href
there are no guarantees that the website developer has used some = ‗DOMAIN‘))]/@href, where DOMAIN represents the
synonymous term to define fields (as author or writer, for the domain of the desired page;
author´s name field). So, is important to think in alternative
strategies. Through the expression cited in the previous step, the dup-
licated links that are hosted in the same domain of the
In this work, the recognition of HTML tags in web pages had page are removed;
generated special attention. Techniques for that are applied in
areas as Information Storage and Retrieval to identify the rele- When counting the number of outLinks, the duplicates are
vant topic of a page, since that definition is very important to the removed and the links that point to the same domain are
information filtering phase [14]. Some of the HTML tags by it counted as only one entry.
self can provide good information for the definition/search of
patterns, for example, the tags: , , , 5.4 Extraction of Author’s Information
could indicate field patterns, as: title, author´s name and contact. Data about the website author are important for determine the
quality of the presented content. If the author is considered a
5.2 Extracting the Site Language specialist on the subject, the web page will have more credibility
These step objectives find out the language in which the text of then the others, written by people not recognized as such. How-
the web page was written. In this project the system informs the ever, this kind of information is not easily determined in a ma-
URL of the health web page that must to be analyzed. The first nual way and in case of automatic treatment of data it became
step refers to search for meta tags that indicate the language of even more complex. There are some techniques that could be
the web page as defined in the Dublin Core pattern: meta tag applied to help in analyzing data about the author, as is the case
language (1)[11]. of h-index [15], which applies the number of author‘s publica-
tions and citations. In case of a complex task, where an individ-
(1) ual technique does not solve the problem, in this work was ap-
If the system didn't found the meta tag, is necessary to make a plied a solution described in [16] where techniques for data
deeper analysis on the content of the web page to verify that. In extraction about pages authorship are employed. The authorship
this case, some programs that analyses the text in order to detect model is defined by the combination of vocabularies defined in
the language can be used. Examples of analyzers used in the the Dublin Core pattern [17] and in the FOAF (Friend Of A
application are: Fuzzums6, Applied Language7 e Google Lan- Friend) ontology [18], for descriptions related to the authors
guage Detection8. After obtaining the page content subsets of expertise. Another tool for automatic extraction obtains the au-
text, as well as the keyword used in this search, are extracted thor‘s organization; his or her electronic mail address; the web-
from the content. In the most part of the cases tested the analysis site address; the author‘s number of publications and h-index
made by two of those programs was enough to detect the web [16].
page language. When the analysis made by two programs disa-
greed, one extra program was used to reach a conclusion [11]. 6. CASE STUDY
The models described in the previous sections were developed
5.3 Extraction of InLinks and OutLinks as prototypes to evaluate the proposed approach. During the
Supposing the website X, inLinks are all the links existing in the tests the topic used for search was "Alzheimer Disease". The
Web that point to X, outLinks are the links of the website X that first step was focused in the automatic location of data, the con-
point to sites that are hosted in other domains. The inLinks and sidered tasks were:
outLinks of each website should be stored for posterior analysis. Study of the existent search engines, as well as analysis of
InLinks are used for some search engines in order to build a each one of them in order to identify which ones were ap-
ranking of the results. Websites that have many inLinks receive propriate to search data about health;
6
Definition of the criteria that will be used for the searches
http://www.fuzzums.nl/~joost/talenknobbel/index.php related to the topic "Alzheimer Disease".
7
http://www.appliedlanguage.com
The section 6.1 below presents details of the search engines.
8
http://www.google.com/uds/samples/language/detect.html About the criteria, without a real population of users to develop
this case study, was chosen to use common criteria from the ―Drugs‖ - intend to provide information about drug inte-
topic. A more detailed description about the applied criteria is ractions, allergic reaction and efficiency. Keywords: alz-
presented in the subsection 6.2. heimer drugs interactions and alzheimer drug treatment.
―Case study‖ - has focus on expert users, which may be
6.1 Health and General Search Engines searching for information to base their researches about a
The queries have two kinds of results: texts that are appropriated disease, for example. Keyword: alzheimer case study.
to general public; technical texts with more detailed information
of the topic. During the search engine analysis, three of them ―Tips‖ – has the objective to answer questions as: ways to
have been choose. That was necessary for automating the loca- deal with a patient with Alzheimer; how is possible do
tion of web pages, because the developed tool should be pre- help; what are the risk factors of the disease. Keyword:
pared for a specific pattern of the engines. Among the analyzed alzheimer‘s practical tips.
ones, the following were adopted [11]: These search criteria allows the automatic location of websites
Medstory: Search in the Web and sorts the results based by the execution of searches over the used search engines (sub-
on several keywords, indicating which ones occur more section 6.1), as well as store the 10 first URLs, associated with
frequently. These keywords, about healthcare, belong to a each search engine. After the location and storing the URLs, the
predefined static list, grouped by categories, as: drugs, system can go further - data extraction.
symptoms, procedures and so on. This list predefined by
the website allows refining the results presented to the us- 6.3 Extraction
er. After the search, several complementary categories of
This step intends to extract the following attributes: page lan-
data related to the topic are presented to the user. These
guage, authorship data, inLinks and outLinks. Each one of those
categories consist of pre-processed information, stored in a
attributes was described during the previous sections. In order to
specific database.
develop the proposed solution as a prototype some tools were
SearchMedica: Allows setting up the search scope in used to support retrieve data in the websites, they are: Web
which the search engine will search: the whole Web or Harvest9 e XPather [11]:
sites defined by the user. Also, the search can be made
The Web-Harvest provide an API (Application Program
over a determined health area (e.g. cardiology, geriatrics,
Interface) that allows: consult Web servers; recover the
dermatology, etc.). The results page has an area in which
HTML page; convert it in XHTML(Extensible Hypertext
several related information are presented.
Markup Language) file and apply some manipulation
Google: Allows search in the whole Web, so, can find techniques of XML (Extensible Markup Language) docu-
since information for laymen until news or scientific work. ments, as XPath, XQuery and XSLT (Extensible Style-
However, there are no guarantees about the reliability of sheet Language Transformations) to extract the desired in-
the presented data in the resulting websites, unlike what formation;
happens in the Medstory, which search in a recognized
The XPather allows search for patterns in websites, being
source internally indexed. At the same time, searches that
used in a semi-automatic manner.
encompass a bigger number of pages can find more rele-
vant results for the user. In these cases, algorithms as Pa-
geRank™ [20], used by this search engine, try to rank the 7. CONCLUSIONS AND FUTURE WORK
results based on reliability. This article described the development of models, techniques
and prototypes aiming the automatic retrieval of content pre-
sented in web pages with health subject. After the automatic
6.2 Search Criteria location, standardization and extraction of page‘s data the objec-
As described in the introduction of this case study, the search
tive is to deliver normalized data to perform evaluations on the
criteria applied were chosen according to the topic "Alzheimer
quality of the health associated web page. The process of devel-
Disease", so, they are application dependent. Thus, an analysis
oping the model for automatic retrieval of data anticipated a
should be done in order to choose good criteria for more cohe-
series of challenges and steps to overcome and the process was
rent results. This issue refers to business analysis [21] and not to
directed to work with these problems. For location were applied
a development issue, i.e., an analysis preferably held by an ex-
well known search engines as they use quite efficient algorithms
pert in the area, in order to provide the closer criteria that could
for Web search. The search for patterns for presenting data in
be used by laymen and experts in its researches. Also, keywords
web pages has led to the conclusion that despite the already
were separated into categories; they are listed below [11]:
existence of standards such as Dublin Core, in practice, these
―General information‖ - introductory information. Key- patterns are not strictly applied by the vast majority of pages, or
words: alzheimer and alzheimer introduction. is partially applied. Thus, the need of some kind of standardiza-
tion for the development of these pages is clearly needed. As a
―Treatment methods‖ - has two goals: for practitioners, in- future work, it would be necessary to end the process of evaluat-
terested in new methods for treatment; and for laymen, in- ing the quality of a health web page making the task of quality
terested in know about its treatment. Keyword: alzhei- evaluation itself automatic.
mer‘s treatment.
―Diagnosis‖ - looks to provide information about the first
symptoms of the disease. Keywords: alzheimer‘s diagnos-
tic and alzheimer's symptoms. 9
http://web-harvest.sourceforge.net/
8. ACKNOWLEDGMENTS [11] Pernas, A. M., Palazzo, J. M. de O., Santos, A.H., Donas-
This work was partially supported by Conselho Nacional de solo, B.M., Bezerra, C.B., Manica, E., Kalil, F., Soares,
Desenvolvimento Científico e Tecnológico - CNPq, Brazil, L.S., Svoboda, L.H., Mesquita, M.P., Torres, P.R., Petry,
Edital Universal - MCT/CNPQ - 14/2010. We would also like to R.L., Santos, R.L., Leithardt, V.R.Q. 2009. Relato sobre o
thank the group of students of the discipline CMP112, Desenvolvimento de Modelos para Obtenção Automática
PPGC/UFRGS, that developed the initial work on this subject. do Conteúdo de Sites sobre Saúde. Cadernos de Informáti-
ca (UFRGS), v. 4, p. 47-91.
9. REFERENCES [12] Wordnet: a lexical database for the English. Princeton Uni-
versity. Retrieved March 20, 2011, from:
.
[1] Web Medica Acreditada – WMA. Retrieved March 21,
[13] Tiun, S., Abdullah, R. and Kong, T.E. 2001. Automatic
2011, from: .
Topic Identification Using Ontology Hierarchy. In Pro-
[2] Internet Content Rating Association - ICRA. Retrieved ceedings of the 2nd International Conference on Computa-
March 21, 2011, from: . tional Linguistics and Intelligent Text Processing - CICL-
[3] Internet Quality Agency – IQUA. Retrieved March 21, ing '01. Springer-Verlag, London, UK.
2011, from: . [14] Liu, B., Chin, C. and Ng, H. 2003. Mining Topic-Specific
[4] Alpert, J. and Hajaj, N. We knew the Web was big... The Concepts and Definitions on the Web. In The 12th Interna-
Oficial Google Blog. Retrieved March 20, 2011, from: tional World Wide Web Conference - WWW 2003, Budap-
. [15] Hirsch, J. E. 2005. An index to quantify an individual's
[5] The Quality Assurance and Content Description Project – scientific research output. PNAS 102 (46), 16569–16572.
QUATRO+. Retrieved March 28, 2011, from: [16] Lichtnow, D., Pernas, A. M., Manica, E., Kalil, F., Olivei-
. ra, J. P. M. de, Leithardt, V. R. Q. 2010. Automatic Collec-
[6] Berners-Lee, T. 1997. Cleaning up the User Interface, Sec- tion of Authorship Information for Web Publications. In:
tion—The ―Oh, yeah?‖-Button, Retrieved March 28, 2011, Proceedings of 6th International Conference on Web In-
from: http://www.w3.org/DesignIssues/UI.html. formation Systems and Technologies – WEBIST. v. 1. p.
339-344. Lisboa: INSTICC. Valencia.
[7] Stamatakis, K., Chandrinos, K., Karkaletsis, V., Mayer,
M.A., Gonzales. D.V., Labsky, M., Amigo, E. and Pöllä, [17] Dublin Core Metadata Initiative. Retrieved March 22,
M. 2007. AQUA, a system assisting labeling experts assess 2011, from: .
health Web resources. In Proceeding of Symposium on [18] Brickley, D. and Miller, L. FOAF Vocabulary Specification
Health Information Management Research – ISHIMR 0.98. Namespace Document 9 August 2010. Marco Polo
2007. Edition. Retrieved March 20, 2011, from:
[8] Wang, Y., Liu Z. 2007. Automatic detecting indicators for .
quality of health information on the Web, International [19] K. Shvachko, H. Kuang, S. Radia, R. Chansler, The Ha-
Journal of Medical Informatics, 76(8), 575-582. doop Distributed File System, IEEE 26th Symposium on
[9] Balakrishnan, H., Kaashoek, M. F., Karger, D., Morris, R., Mass Storage Systems and Technologies, 2010, ISBN: 978-
Stoica, I. 2003. Looking up data in P2P systems. In Com- 1-4244-7152-2, p.1-10.
munications of the ACM. [20] Brin, S., Page. L. The Anatomy of a Large-Scale Hypertex-
DOI=http://doi.acm.org/10.1145/606272.606299. tual Web Search Engine. In Computer Networks and ISDN
[10] Lichtnow, D., Jouris, A., Bordignon, A., Pernas, A. M., Systems. Elsevier Science Publishers, Amsterdam, The
Levin, F. H., Nascimento, G. S., Silva, I. C. S., Gasparini, Netherlands. 1998.
I., Teixeira, J. M., Rossi, L. H. L., Oliveira, O. E. D., [21] Witten, I.H. and Frank, E. Data Mining: Practical Machine
Schreiner, P., Gomes, S. R. V., Oliveira, J. P. M. de. 2009. Learning Tools and Techniques. 2a ed. Morgan Kaufmann.
Relato e Considerações sobre o Desenvolvimento de uma 629p. 2005. ISBN-13:978-0-12-088407-0.
Ontologia para Avaliação de Sites da Área de Saúde. Ca-
dernos de Informática (UFRGS), v. 4, p. 7-46.