=Paper= {{Paper |id=Vol-1179/CLEF2013wn-CLEFIP-EiseltEt2013 |storemode=property |title=Innovandio S.A. at CLEF-IP 2013 |pdfUrl=https://ceur-ws.org/Vol-1179/CLEF2013wn-CLEFIP-EiseltEt2013.pdf |volume=Vol-1179 |dblpUrl=https://dblp.org/rec/conf/clef/EiseltO13 }} ==Innovandio S.A. at CLEF-IP 2013== https://ceur-ws.org/Vol-1179/CLEF2013wn-CLEFIP-EiseltEt2013.pdf
     The simpler the better - Retrieval Model
    comparison for Prior-Art Search in Patents
                @ CLEF-IP 2013

                     Andreas Eiselt and Gabriel Oberreuter

                Innovandio S.A., Miguel Claro 195, Santiago, Chile



      Abstract. Patentability and novelty search is an essential part of any
      patent application. It ensures that the idea that should be patented has
      not already been registered anywhere else in the world. However, this
      task is complicated by the large number of documents and the fact that
      they are written in many different languages. In this paper we survey four
      approaches that will help to automate the task and share the insights we
      have gained through our participation in the CLEF-IP Workshop 2013.


1   Introduction
The level of innovation is one of the principal measures that determines if an
idea can be patented or not. In order to estimate it, an exhaustive search in over
80 million patents from more than 100 patent authorities has to be performed.
To date, this is still usually carried out with the help of keyword searches. That
this strategy does not work is proven by the fact that 54% of the 2.5 million
annual patent applications are rejected. The main problem is that the language
used in patents is often subject specific as well as inaccurate and misleading.
The reason for this is that the applicants are usually not interested in making
their ideas public and therefore try to disguise them as good as possible. Another
problem is that patent documents may be written in many different languages,
which makes it even harder to find similar ideas.
    For our participation in the patentability and novelty search task of the
CLEF-IP workshop, which was to automatically find all documents and their
respective passages that describe concepts strongly related to those explained in
the source document, we explored the applicability of four approaches that will
be explained in the following chapter.


2   Task, Corpus and Evaluation
For the CLEF-IP patentability or novelty search task all participants were pro-
vided with a corpus divided into two sets of patent documents: the first set Dpat
contains 3.118.088 patent documents (2.680.604 from the EPO and 437.484 from
the WIPO) and the second set Dapp 210 patent documents (belonging to 69
patents). Furthermore a set of 149 topics T was given. Each ti ∈ T was defined
2

as a subset from the claims of one patent document di ∈ Dapp . The basic task
was, given a topic ti , find those passages pn in dpat ∈ Dpat that are semantically
related.
    The results were then evaluated on document as well as on passage level. On
document level the Patent Retrieval Evaluation Score (PRES) [1] was used with
a cutoff at 100, while the evaluation on passage level was based on the mean
average generalized precision (MAgP) [2].


3   Experiments & Results

In order to obtain the most relevant documents and passages from Dpat for ti ,
the retrieval process was divided into two stages: candidate-retrieval and detailed
comparison. During the candidate-retrieval the most relevant documents from
Dpat were selected and subsequently a detailed comparison was performed to
determine the most relevant passages.
    In order to reduce the space of possible candidates and improved retrieval
quality, we only considered patents from Dpat , which shared at least one Interna-
tional Patent Classification (IPC) with the patent containing ti . We furthermore
used abstracts, claims and descriptions from all patent family members, since
the text contained in ti was too short.
    As the amount of patents in the workshop task, as well as in a real-world
scenario is limited, in three of the four approaches the candidate-retrieval was
performed by calculating the text similarity between ti and all possible candidate
documents from Dpat . The detailed comparison at passage level was executed
subsequently between ti and the top 100 candidates from Dpat using the same
similarity measure. As text similarity measure, we evaluated 3 approaches which
were all based on the Vector-Space Model (VSM) and the cosine similarity:
(i) Word uni-grams (ii) Character tri-grams, (iii) Cross-Language Explicit Se-
mantic Analysis (CL-ESA) . The first approach was based on the idea of simply
comparing the used vocabulary, ignoring the fact that some documents are writ-
ten in other languages. This approach was considered the best approximation
for a keyword-search, a strategy commonly used by humans to generate a patent
search-report. The second approach was based on the findings that two text doc-
uments written in different european languages have a strong character N -gram
overlap [3]. The third approach is known as Cross-Language Explicit Seman-
tic Analysis (CL-ESA) [4]. It represents a document as a vector of similarities
to the documents of a multi-lingual reference corpus. This allows to compare
documents on a semantic level, independent of the language in which they are
written.
    In comparison to the first three approaches, the fourth was based on a heuris-
tic candidate retrieval. Therefore, we generated a set of queries for each source
document and executed them on a search engine, which had all documents from
Dpat indexed. For the top candidate documents we then executed a detailed
comparison based again on word uni-grams.
    The results of each submitted run, are presented in Table 1. They show, that
                                                                                        3

Table 1. Results of the submitted runs; eiselt-cos: uses word uni-grams; eiselt-solr
uses heuristic candidate retrieval; eiselt-c3g uses character tri-grams; eiselt-clesa uses
Cross-Language Explicit Semantic Analysis

Run                      Document Level                         Passage Level
                     PRES@100 R@100 MAP                      MAP(D) Precision(D)
eiselt.cos           0.30          0.38      0.11            0.12        0.08
eiselt.solr          0.26          0.37      0.08            0.11        0.10
eiselt.c3g           0.23          0.31      0.10            0.10        0.07
eiselt.clesa         0.21          0.29      0.05            0.10        0.08



the simplest approach (cosine similarity between vectors of word uni-grams) out-
performs any other approach. This can be explained by the fact that it guarantees
a higher ranking for documents with similar vocabulary. The same documents
will get a higher score in case of an intelligent keyword-search as it is typically ex-
ecuted by humans. This also explains the good results of the heuristic candidate
retrieval, as it aims to imitate humans behaviour too. That the approach based
on character tri-grams did not bring the expected advantage is due to the fact
that documents of the same language still share a lot more character n-grams
than semantically related documents in different languages. The interpretation of
the results obtained using CL-ESA may require further investigation. They show
that this approach is, out of the four, the worse approximation for human search
behaviour. Nevertheless, they do not reflect necessarily a bad performance. Pre-
liminary investigations of the results have shown that CL-ESA assigned a higher
rank to documents which seem to be highly related and which did not appear
in the result-set of simple keyword-searches and neither in the patent search
report. Hence, in order to obtain a better idea of the result-quality, it would be
necessary to manually judge the relatedness of the top-ranked results.

References
1. Magdy, W., Jones, G.J.: PRES: a score metric for evaluating recall-oriented infor-
   mation retrieval applications. In: Proceeding of the 33rd international ACM SIGIR
   conference on Research and development in information retrieval - SIGIR ’10, New
   York, New York, USA, ACM Press (2010) 611
2. Kamps, J., Pehcevski, J., Kazai, G., Lalmas, M., Robertson, S.: Inex 2007 evaluation
   measures. In Fuhr, N., Kamps, J., Lalmas, M., Trotman, A., eds.: Focused Access
   to XML Documents. Volume 4862 of Lecture Notes in Computer Science. Springer
   Berlin Heidelberg (2008) 24–33
3. McNamee, P., Mayfield, J.: Character n-gram tokenization for european language
   text retrieval. Information Retrieval 7(1-2) (2004) 73–97
4. Potthast, M., Stein, B., Anderka, M.: A Wikipedia-Based Multilingual Retrieval
   Model. In Macdonald, C., Ounis, I., Plachouras, V., Ruthven, I., White, R., eds.: Ad-
   vances in Information Retrieval. 30th European Conference on IR Research (ECIR
   08). Volume 4956 of Lecture Notes in Computer Science., Berlin Heidelberg New
   York, Springer (2008) 522–530