About this Journal Submit a Manuscript Table of Contents
The Scientific World Journal
Volume 2014 (2014), Article ID 132158, 10 pages
http://dx.doi.org/10.1155/2014/132158
Research Article

Study of Query Expansion Techniques and Their Application in the Biomedical Information Retrieval

Department of Computer Science, Higher Technical School of Computer Engineering, University of Vigo, 32004 Ourense, Spain

Received 31 August 2013; Accepted 15 December 2013; Published 2 March 2014

Academic Editors: G. Lu and J. Tang

Copyright © 2014 A. R. Rivas et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Abstract

Information Retrieval focuses on finding documents whose content matches with a user query from a large document collection. As formulating well-designed queries is difficult for most users, it is necessary to use query expansion to retrieve relevant information. Query expansion techniques are widely applied for improving the efficiency of the textual information retrieval systems. These techniques help to overcome vocabulary mismatch issues by expanding the original query with additional relevant terms and reweighting the terms in the expanded query. In this paper, different text preprocessing and query expansion approaches are combined to improve the documents initially retrieved by a query in a scientific documental database. A corpus belonging to MEDLINE, called Cystic Fibrosis, is used as a knowledge source. Experimental results show that the proposed combinations of techniques greatly enhance the efficiency obtained by traditional queries.

1. Introduction

Biomedical knowledge is growing at a high pace, and large collections of publications offer an excellent opportunity for discovering hidden biomedical knowledge by applying information retrieval (IR) and related technologies. IR is related to the representation, storage, organization of, and access to the information items [1]. Information items must be represented in order to allow users to have easy access to the information of interest, and the user requirements must be presented in a good format to be translated into a query which can be processed by the search engine (or IR system). The translation is presented like a set of keywords (or index terms) which represents the query and summarizes the information in which the user is interested.

Information Retrieval using only keywords is not usually very efficient. In general, information about a particular issue can be represented with different keywords which may not coincide exactly with the terms entered in the query by the user. The user query can include keywords that are not present in documents, but documents could be relevant because they have other words with the same meaning. Using query expansion (QE) techniques, a query is reformulated to improve retrieval performance and obtain additional relevant documents by expanding the original query with additional relevant terms and reweighting the terms in the expanded query. Query expansion techniques are widely used for improving the efficiency of textual information retrieval systems, helping to overcome vocabulary mismatch issues including words in queries with the same or related meaning.

This paper analyzes different techniques of document preprocessing and query expansion in order to know which ones offer better results when applied to query biomedical databases. This research aims to improve the efficiency of the queries based on classic models (where documents are retrieved even if only a small part of them is related to the query), when they are performed in public scientific databases, such as Pubmed.

The remainder of the paper is organized as follows: Section 2 presents an overview of the general Information Retrieval process. Section 3 describes the preprocessing and Query Expansion methods employed in this research for retrieving relevant documents and the experimental results. Finally, the conclusions are included in Section 4.

2. Information Retrieval Process

In Information Retrieval, the query process is composed of two main phases, indexing and matching (see Figure 1). Additionally, it is possible to expand the queries to improve the efficiency of the retrieval.

132158.fig.001
Figure 1: The information retrieval process.

The indexing step preprocesses documents and queries in order to obtain keywords (relevant words, also named terms) to be used in the query. At this point, it is important to consider the use of stemming and stopword lists in order to reduce related words to their stem, base or root form. This is achieved by launching affix removal to adapt different derivational or inflectional variants of the same word to a single indexing form and remove words that do not contain information relevant to the document [1, 2].

Matching is the process of computing the similarity between documents and queries by weighting terms, the most frequently applied algorithms being the TF-IDF and BM25 algorithms. Most retrieval systems return a ranked document list in response to a query, where the documents more similar to the query considered by the system are first on the list.

Once the first answer set is obtained, different query expansion techniques can be applied. For example, the most relevant keywords of the top documents previously retrieved can be added to the query in order to rerank the documents. This process is known as relevance feedback. The retrieval can be further enhanced by modifying the words of the queries using other keywords more representative of the document content (e.g., including MeSH Headings).

This study uses the Lemur language modeling toolkit (Lemur Project http://www.lemurproject.org/) for the indexing, mapping, and feedback processes. Lemur is a software tool designed to facilitate research in language modeling and IR, using weighting algorithms to provide methods for parsing queries, indexing documents, and retrieving documents related to queries.

In order to evaluate the results of the retrieval process, a program inside the TREC conference, treceval (trec_eval http://trec.nist.gov), is used. Trec_eval makes it possible to obtain measures such as the Total number of documents over all queries (Retrieved, Relevant and Rel_ret (relevant, and retrieved)) or MAP, R-prec, and Interpolated Recall-Precision Averages.

The following subsections introduce more details of the concepts related to document corpora, stemming, stopwords, weighting algorithms, query expansion, and measures.

2.1. Document Corpora

As seen in Figure 1, three document corpora are needed to analyze the efficiency of a query system: the original document corpus, the textual descriptions of the users queries (topics), and the relevant judgments given by the experts [3]. A document corpus represents a sampling of articles published. The format of the data uses a labeled bracketing, the topics are a description in natural language of the information that the user needs, typically one sentence, and, finally, the relevance judgments are done by potential users, called experts or judges.

Most of the public biomedical document corpora belong to MEDLINE, which is used in our study: Cystic Fibrosis (Cystic Fibrosis Collection http://grupoweb.upf.es/WRG/mir2ed/ref.php) (CF). It consists of 1239 documents published between 1974 and 1979 discussing various aspects of Cystic Fibrosis. Cystic Fibrosis documents are composed of the Abstract (AB), Title (TI), and a set of manually assigned MeSH (MeSH Home page http://www.nlm.nih.gov/mesh/) (Medical Subject Headings) of a MeSH thesaurus.

MeSH thesaurus is a controlled vocabulary used for indexing, cataloging, and searching for biomedical and health-related information and documents. It imposes uniformity to the indexing of the scientific literature. MeSH thesaurus contains approximately thousand terms and is updated annually to reflect changes in medicine and medical terminology.

MeSH has a hierarchical structure with sets of terms, naming, descriptors, that allow searching at various levels of specificity. Expert annotators assign MeSH Headings terms to the documents in order to allow the user to retrieve the information that explains the same concept with different terminology. On average, to subject headings are assigned by document, of which to of them are major subjects (MJ) and the others are minor subjects (MN). Major MeSH terms describe the main topics of the document, and minor MeSH terms provide more details about it [48].

In Table 1, an example of MEDLINE document is showed. It contains the title (TI), the major subjects (MJ), the minor subjects (MN), and the document abstract (AB).

tab1
Table 1: A sample of a MEDLINE document.

Each MeSH Heading is related to several Entry terms. Entry terms are synonyms, alternate forms, and other closely related terms with a given MeSH record. They are generally used interchangeably with the MeSH Heading for the purpose of indexing and retrieval, thus increasing the access points to MeSH indexed data.

The Cystic Fibrosis collection also contains 100 queries and the documents relevant to each query [1] (see Table 2). Further, four scores are provided for each relevant document. Relevance scores can be 0 (which indicates nonrelevance), 1 (which indicates marginal relevance), and 2 (which indicates high relevance).

tab2
Table 2: A sample of query with its relevant documents and relevance scores.
2.2. Stemming

Stemming is the process of reducing related words to their stem, base or root form through affix removal. Its aim is to adapt different derivational or inflectional variants of the same word to a single indexing form [1, 2].

There are two major stemmers in use for English IR: the Porter stemmer and the Krovetz stemmer. Porter Stemmer is a process for removing suffixes from words, such as gerunds and plurals, and replacing inflectional endings [9]. It is composed of rules, each of which deals with a specific suffix and has certain conditions to satisfy. The suffixes of words are checked against each rule in a sequential manner until it matches one; the conditions in the rule are then tested, which may result in a suffix removal or modification.

Alternatively, Krovetz Stemmer removes inflectional suffixes in three steps: the conversion of a plural to its single form, the conversion of past to present tense, and the removal of -ing. The process firstly removes the suffix and then, through a process of checking in a dictionary, returns the stem to a word [10].

2.3. Stopwords

In Information Retrieval, a document is indexed by the frequency of its words. Statistical analysis of this process shows that some words have low frequency, while others have high frequency [11]. For example, and, of, and the appear frequently in the documents without significant information. This set of words is referred to as stopwords. Elimination of stopwords can significantly reduce the size of the indexing structure, speed up the calculation and increase accuracy. Up to now, a lot of stopword lists have been developed for the English language, for example, the US National Library of Medicines official list of stopwords (NLM stopword list http://www.netautopsy.org/umlsstop.htm), and the stopword list built by Gerard Salton and Chris Buckley for the experimental SMART Information Retrieval system (SMART stopword list http://www.lextek.com/manuals/onix/stopwords2.html).

2.4. Okapi BM25 Weighting Algorithm

Okapi BM25, or BM25, is a weighting function used to rank documents according to their relevance to a given query [12]. Many researchers apply the BM25 function in different corpus to retrieve relevant documents.

BM25 is a probabilistic model, where the weight of a search term is assigned based on its frequency within the document and the frequency of the query term. The corresponding weighting function is as follows: (i), , and are parameters which depend on the queries and the database;(ii) is the occurrence frequency of the term in the document ;(iii) is the frequency of the term in the topic from which the query is derived;(iv) and are, respectively, the document length and the average document length in the corpus.

is the Robertson Sparck Jones weight [13], calculated as where is the number of documents relevant to a specific topic, is the number of relevant documents containing the term , is the total number of documents in the collection, and is the number of documents containing the term.

2.5. TF-IDF Weighting Algorithm

The TF-IDF weighting algorithm (termed frequency-inverse document frequency) is often used in Information Retrieval and text mining [1]. This weight is a statistical measure used to evaluate the importance of a word to a document in a collection or corpus. The importance increases proportionally to the number of times a word appears in the document but is offset by the frequency of the word in the corpus.

Variations of the TF-IDF weighting scheme are often used by search engines as a central tool in scoring and ranking the document relevance given a user query [14, 15]. In our experiments, the formulas for the TF-IDF weighting algorithm applied areRaw TF formulas where represents the frequency of the most frequent term in the document ,Log TF formulas Okapi TF formulas where is the query length.

Table 3 contains the correspondence between the parameters of the BM25 weighting algorithm and the TF-IDF with Okapi TF formulas, used in our experiments.

tab3
Table 3: Correspondence between parameters of the BM25 weighting and Okapi TF.

The IDF (inverse document frequency) function is as follows [1, 2]:

Thus, the weight of a term is calculated as

2.6. Query Expansion

Query expansion techniques have been studied for nearly three decades. The various methods proposed in the literature can be classified into the following three groups [16]: query specific, corpus specific, and language specific.(a)Query-specific terms can be identified by locating new terms in a subset of the documents retrieved by a specific query. This is the approach taken by relevance feedback systems, where related terms come from the contents of user-identified relevant documents. This has been shown to be quite effective, but it requires that users indicate which documents are relevant. More recently, search improvements are being achieved [17, 18] without the user’s relevance judgments.(b)Corpus-specific terms are found by analyzing the contents of a particular full-text database to identify terms used in similar ways. It may be hand-built, a time-consuming and ad hoc process, or created automatically. Traditional automatic thesaurus construction techniques group words together based on their occurrence patterns at a document level [19, 20], that is, words which often occur together in documents are assumed to be similar. These thesauri can then be used for automatic or manual query expansion.(c)Language-specific terms may be found from generally available online thesauri that are not tailored for any particular text collection. Liddy and Myaeng [21] use the Longman’s Dictionary of Contemporary English, a semantically coded dictionary. Voorhees (1994) used WordNet [22], a manually constructed network of lexical relationships. Borrajo et al. [23] study the use of dictionaries in the classification of biomedical texts with three different dictionaries (BioCreative [24], NLPBA [25] and an ad hoc subset of the UniProt database named Protein [26]).

This research adopts an automatic query-specific terms approach for locating related terms. We are particularly interested in these techniques because they are commonly used to add useful words to a query. Unfortunately, casual users seldom provide a system with the relevance judgments needed in relevance feedback. In such situations, ad hoc or blind feedback is commonly used to expand the user query. This method takes the form of pseudorelevance feedback, where the actual input from the user is not required. In this method, a small set of documents is retrieved using the original user query; these documents are all assumed to be relevant without any intervention by the user [27]. The content of the assessed documents is used to adjust the weights of terms in the original query and/or to add keywords to the query. The new query is reformulated towards relevant documents and away from the nonrelevant ones [10, 28, 29].

The Lemur toolkit used in our experiments implements the Rocchio formulation for pseudo relevance feedback. It first applies the standard retrieval model [1, 2] to retrieve documents for a given query . Given the retrieved documents and the original query, the expanded query is computed as where is the number of retrieved documents for a given query and is the parameter used to weight the importance of the retrieved documents.

2.7. Measures

In order to evaluate results, trec_eval (http://trec.nist.gov/) is used. It makes it possible to obtain several measures related to information retrieval [30]. The most commonly used are the following.

2.7.1. Average Precision

For systems that return a ranked sequence of documents, it is preferable to consider the order in which the returned documents are presented. This measure averages the precision values from the rank positions where a relevant document is retrieved: (i) is the rank;(ii) is the number of documents retrieved;(iii) is a binary function on the relevance of a given rank;(iv) is the precision (proportion of a retrieved set that is relevant) at a given cut-off rank.

2.7.2. Mean Average Precision (MAP)

It summarizes rankings from multiple queries by averaging average precision:

2.7.3. R-precision

R-precision (R-prec) is the precision after R documents have been retrieved, where R is the number of relevant documents for the topic.

3. Methods and Results

This section presents an overview of the tests performed, with respect to the processes of indexing, matching, and query expansion presented in Section 2.

3.1. Indexing Processes Testing

The first tests are based on a study of the benefits produced by the use of stemming and stopwords in the indexing of documents and queries.

We analyze the impact of stemming algorithms (Porter and Krovetz) and stopword lists (NLM and SMART) in the retrieval of documents from the corpus Cystic Fibrosis. The Okapi BM25 weighting algorithm is used with default parameters (, , and ) and is applied to the Abstract field.

In Table 4 we can see that stemming is an effective technique to improve MAP. The performances are usually different between weak (Krovetz) and strong (Porter) stemming methods [12, 31, 32], but in our case the results are similar. In terms of MAP, strong stemming is a bit better, but in terms of R-prec, weak stemming is a bit better.

tab4
Table 4: Accuracy of query search using different stemming functions and stopword lists in Cystic Fibrosis. Evaluation measures used are MAP (mean average precision), R-prec (R precision), and (number of relevant documents retrieved).

If we compare different stopword removal methods, we can see that removing stopwords improves the performance. From our experiments, using the largest stopword list (SMART) results is better than using the list with fewer stopwords. The Porter stemmer with SMART stopword list provides the best results. The last four combinations (highlighted with bold style) do not have significant differences, so we conduct the remaining tests with these four combinations.

3.2. Matching Processes Testing
3.2.1. Parameterization of the Weighting Algorithms

The weighting algorithms used for ranking the retrieved documents in our tests are Okapi BM25 and TF-IDF, explained in Section 2.

In Okapi BM25, the values of the internal parameters , , and should be adjusted based on the document collection and the type of queries where it is applied [2, 3335]. A significant number of experiments have been done, and suggest general values of and between and (usually , although is set between and in the case of long queries) and (although small values can sometimes report improvements).

Finding the set of optimal parameters is costly to compute, since they have local maxima that are singularity values [35]. Hence, we are using a simplistic optimization approach. The best values obtained in our tests with the Cystic Fibrosis corpora are , , and (see Table 5).

tab5
Table 5: Testing values of the parameters for BM25. MAP is used as the evaluation measure.

For the TF-IDF weighting algorithm with Okapi TF formula, the parameters obtained for Okapi BM25 are the best approach, using the correspondence shown in Table 3. Moreover, the values obtained with the Log TF and Raw TF formulas without parameters were studied, verifying that they are worse than the approximation BM25 (see Table 6).

tab6
Table 6: MAP values for the TF-IDF BM25, Raw TF, and formulas.

Many researchers use the BM25 algorithm in articles, steering their studies to retrieve information in several fields, not only in the Abstract [31, 36, 37]. By this assumption, we test how the MAP measure increases if we look for documents related to the queries in the Abstract, Title, and Mesh fields using the BM25, TF-IDF BM25, TF-IDF Log TF and TF-IDF Raw TF formulas (see Table 7).

tab7
Table 7: MAP values to retrieving in Abstract, Title, and MeSH fields using different weighting algorithms.

The results obtained with Okapi BM25 are consistent with those presented by Trotman in [36, 37], which show a value of in the MAP measure with the same collection.

In Tables 6 and 7 we can see that the best MAP results are obtained using TF-IDF BM25 TF formula, so we continue our study with this approximation.

3.3. Query Expansion Processes Testing
3.3.1. Pseudorelevance Feedback

To improve the results of the previous processes, we proceed to make pseudo relevance feedback using the Rocchio algorithm implemented in Lemur, explained in Section 2.

For the first time, it is necessary to parameterize the algorithm. The parameters for Rocchio are number of documents in feedback , number of terms selected in feedback , and coefficient adjustment (0,4].

Table 8 shows the best values of the parameters obtained for the Cystic Fibrosis collection.

tab8
Table 8: Finding the best values of the parameters for TF-IDF BM25 feedback.

After the parameterization, we test how the MAP increases if we look for documents related to the queries in the Abstract, Title, and Mesh fields, using the TF-IDF BM25 weighting algorithm (see Table 9). These results are comparable with those obtained by Shin and Han in their expansion system presented in [8], which achieved a maximum value of for R-prec, with our R-prec being greater than .

tab9
Table 9: Evaluation measures using the pseudorelevance feedback in Abstract, Title, and MeSH fields.
3.3.2. Use of MeSH to Expand Queries

As many authors have already worked with MeSH fields to retrieve information, we focus this part of the research on testing its efficiency in query expansion. Our method is based on the work of Kwangcheol Shin and Han [8], which proves the advantage of using MeSH Headings to expand the queries instead of working with terms not related to the MeSH fields.

The new strategy consists of expanding the queries with the MeSH Headings related to its terms. For the first time, we analyze the query and extract its keywords. Each keyword is automatically introduced at the PubMed online tool to match it with the Entry terms stored in MEDLINE. If the keyword is an Entry Term, we search the associated MeSH Headings (major and minor). Finally, the query is reformulated adding the descriptors contained in the MeSH fields extracted in the previous step.

In Table 10 the results of the experiment with the proposed strategy are shown. The query expansion is applied in the Abstract, Title, and MeSH fields of the documents and the weighting algorithm is TF-IDF BM25. As shown, the results are similar to those obtained with pseudo relevance feedback (Table 9), and are comparable with those obtained by Shin and Han in their expansion system [8].

tab10
Table 10: Evaluation measures using query expansion with descriptors, applied in MeSH, Title, and Abstract fields.

Finally, a Recall-Precision graph (see Figure 2) is included showing the improvement obtained with the query expansion methods using the MeSH, Abstract, and Title fields (Tables 9 and 10) with respect to the document retrieval using only the Abstract field with TF-IDF BM25 (see Table 7). The figure shows that curves of the query expansion algorithms for the combination Porter stemmer-NLM stopwords are closest to the upper right hand corner of the graph (where recall and precision are maximized), which indicates the best performance.

132158.fig.002
Figure 2: Recall-Precision curve obtained with the query expansion methods using the MeSH, Abstract, and Title fields.

4. Conclusions

We have developed and evaluated preprocessing and query expansion techniques for retrieving documents in several fields of biomedical articles belonging to the corpus Cystic Fibrosis, a corpus of MEDLINE documents. We test the benefit of using stemming and stopwords in the preprocessing of documents and queries, following the investigations of other authors.

Studies were carried out to compare the weighting algorithms Okapi BM25 and TF-IDF available in the Lemur tool, concluding that TF-IDF with formula given by BM25 approximation is superior in its results.

Document retrieval based on Abstract, MeSH, and Title fields seems more effective than looking at each of these fields individually. In addition, the use of relevance feedback, a technique widely used by researchers in this field, produces a great improvement in the retrieval of scientific documents. The Rocchio algorithm allows obtaining good results, improving MAP and other measures.

Finally, we perform a study to improve searching expanding queries with MeSH terms. For this, we have enhanced queries locating Entry terms in them and obtaining MeSH Headings in PubMed in order to expand the original query and to map it with the documents. The results are good, making use of the Title, Abstract and MeSH fields to improve the list of documents retrieved, compared to baseline methods.

In this paper, authors have used a very simplistic approach to determine the BM25 parameters values. Tuning the BM25 free parameters (, , and ) is a difficult and computationally expensive problem that requires advanced multidimensional optimization techniques. Retrieval accuracy can be improved using more advanced parameterization methods.

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

Acknowledgment

This work has been funded from the European Union Seventh Framework Programme (FP7/REGPOT-2012-2013.1) under Grant agreement no. 316265, BIOCAPS.

References

  1. R. A. Baeza-Yates and B. Ribeiro-Neto, Modern Information Retrieval, Addison-Wesley Longman, 1999.
  2. C. D. Maning, P. Raghavan, and H. Schtze, An Introduction to Information Retrieval, Cambridge University Press, Cambridge, UK, 2008.
  3. D. Hiemstra and D. van Leeuwen, Creating A Dutch information Retrieval Test Corpus, 2002.
  4. A. A. Chang, K. M. Heskett, and T. M. Davidson, “Searching the literature using medical subject headings versus text word with PubMed,” Laryngoscope, vol. 116, no. 2, pp. 336–340, 2006. View at Publisher · View at Google Scholar · View at Scopus
  5. M. H. Coletti and H. L. Bleich, “Medical subject headings used to search the biomedical literature,” Journal of the American Medical Informatics Association, vol. 8, no. 4, pp. 317–323, 2001. View at Scopus
  6. J. O. Ebbert, D. M. Dupras, and P. J. Erwin, “Searching the medical literature using PubMed: a tutorial,” Mayo Clinic Proceedings, vol. 78, no. 1, pp. 87–91, 2003. View at Scopus
  7. L. V. Gault, M. Shultz, and K. J. Davies, “Variations in Medical Subject Headings (MeSH) mapping: from the natural language of patron terms to the controlled vocabulary of mapped lists,” Journal of the Medical Library Association, vol. 90, no. 2, pp. 173–180, 2002. View at Scopus
  8. K. Shin and S. Y. Han, “Improving information retrieval in medline by modulating mesh term weights,” in Natural Language Processing and Information Systems, F. Meziane and E. Mtais, Eds., vol. 3136, Lecture Notes in Computer Science, pp. 137–166, Springer, Berlin, Germany, 2004.
  9. M. F. Porter, “An algorithm for suffx stripping,” Program, vol. 14, no. 3, pp. 130–137, 1980.
  10. J. Leveling and G. J. F. Jones, “Sub-word indexing and blind relevance feedback for English, Bengali, Hindi, and Marathi IR,” ACM Transactions on Asian Language Information Processing, vol. 9, no. 3, article 12, 2010. View at Publisher · View at Google Scholar · View at Scopus
  11. G. K. Zipf, Human Behavior and the Principle of Least Effort, Addison-Wesley, Reading, Mass, USA, 1949.
  12. S. E. Robertson, S. Walker, S. Jones, M. M. Hancock-Beaulieu, and M. Gatford, Okapi at Trec-3, 1996.
  13. S. E. Robertson and K. S. Jones, “Relevance weighting of search terms,” Journal of the American Society for Information Science, vol. 27, no. 3, pp. 129–146, 1976. View at Scopus
  14. J. Ramos, Using Tf-Idf to Determine Word Relevance in Document Queries, 2003.
  15. S. Robertson, “Understanding inverse document frequency: on theoretical arguments for IDF,” Journal of Documentation, vol. 60, no. 5, pp. 503–520, 2004. View at Publisher · View at Google Scholar · View at Scopus
  16. S. Gauch, J. Wang, and S. M. Rachakonda, “A corpus analysis approach for automatic query expansion and its extension to multiple databases,” ACM Transactions on Information Systems, vol. 17, no. 3, pp. 250–269, 1999. View at Scopus
  17. J. Guo, G. Xu, X. Cheng, and H. Li, “Named entity recognition in query,” in Proceedings of the 32nd Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR '09), J. Allan, J. A. Aslam, M. Sanderson, C. Zhai, and J. Zobel, Eds., pp. 267–274, ACM, 2009.
  18. S. Karimi, J. Zobel, and F. Scholer, “Quantifying the impact of concept recognition on biomedical information retrieval,” Information Processing and Management, vol. 48, no. 1, pp. 94–106, 2012. View at Publisher · View at Google Scholar · View at Scopus
  19. C. J. Crouch and B. Yang, “Experiments in automatic statistical thesaurus construction,” in Proceedings of the 15th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR '92), pp. 77–88, New York, NY, USA, June 1992. View at Scopus
  20. Y. Qiu and H. P. Frei, “Concept based query expansion,” in Proceedings of the 16th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 160–169, July 1993. View at Scopus
  21. E. D. Liddy and S.-H. Myaeng, “Tipster panel—dr-link's linguistic-conceptual approach to document detection,” in TREC, pp. 113–130, 1992.
  22. G. A. Miller, R. Beckwith, C. Fellbaum, D. Gross, and K. J. Miller, “Introduction to wordnet: an on-line lexical database,” International Journal of Lexicography, vol. 3, no. 4, pp. 235–244, 1990. View at Publisher · View at Google Scholar · View at Scopus
  23. L. Borrajo, R. Romero, E. L. Iglesias, and C. M. Redondo Marey, “Improving imbalanced scientific text classification using sampling strategies and dictionaries,” Journal of Integrative Bioinformatics, vol. 8, no. 3, p. 176, 2011. View at Scopus
  24. L. Hirschman, A. Yeh, C. Blaschke, and A. Valencia, “Overview of BioCreAtIvE: critical assessment of information extraction for biology,” BMC Bioinformatics, vol. 6, supplement 1, article S1, 2005. View at Publisher · View at Google Scholar · View at Scopus
  25. G. D. Zhou, “Recognizing names in biomedical texts using hidden markov model and SVM plus sigmoid,” in Proceedings of the COLING International Joint Workshop on Natural Language Processing in Biomedicine and its Applications (NLPBA/BioNLP '04), N. Collier, P. Ruch, and A. Nazarenko, Eds., pp. 1–7, Geneva, Switzerland.
  26. R. Apweiler, A. Bairoch, C. H. Wu et al., “UniProt: the universal protein knowledgebase,” Nucleic Acids Research, vol. 32, pp. D115–D119, 2004. View at Scopus
  27. M. Mitra, A. Singhal, and C. Buckley, “Improving automatic query expansion,” in Proceedings of the 21st Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 206–214, ACM Press, 1998.
  28. S. Abdou, “Evaluation of stemming, query expansion and manual indexing approaches for the genomic task,” in Proceedings of the 15th Text REtrieval Conference Proceedings (TREC '05), pp. 863–871, 2005.
  29. Z. Yang, H. Lin, Y. Li, B. Liu, and Y. Lu, “Trec genomics track experiments at dutai,” in Text REtrieval Conference, 2005.
  30. T. Brants and Google Inc, “Natural language processing in information retrieval,” in Proceedings of the 14th Meeting of Computational Linguistics in the Netherlands, pp. 1–13, 2004.
  31. Y. Fan, X. Huang, and York University at TREC, “Enterprise email discussion search,” in Proceedings of the 15th Text Retrieval Conference (TREC '06), M. Ellen Voorhees and P. Lori Buckland, Eds., vol. 500-272, National Institute of Standards and Technology (NIST), 2006.
  32. S. E. Robertson and S. Walker, “Okapi/keenbow at trec-8,” 1999.
  33. E. Greengrass, “Information retrieval: a survey,” Tech. Rep. tr-r52-008-001, united States Department of Defense, 2001.
  34. S. E. Robertson and S. Walker, Microsoft Cambridge at Trec-9: Filtering Track, 2001.
  35. S. Robertson and H. Zaragoza, “The probabilistic relevance framework: BM25 and beyond,” Foundations and Trends in Information Retrieval, vol. 3, no. 4, pp. 333–389, 2009. View at Publisher · View at Google Scholar · View at Scopus
  36. A. Trotman, An Artificial Intelligence Approach to Information Retrieval, 2004.
  37. A. Trotman, Learning to Rank. Information Retrieval, vol. 8, 2005.