Information Analysis of High-Dimensional Data and ApplicationsView this Special Issue
A Novel Approach for Protein-Named Entity Recognition and Protein-Protein Interaction Extraction
Many researchers focus on developing protein-named entity recognition (Protein-NER) or PPI extraction systems. However, the studies about these two topics cannot be merged well; then existing PPI extraction systems’ Protein-NER still needs to improve. In this paper, we developed the protein-protein interaction extraction system named PPIMiner based on Support Vector Machine (SVM) and parsing tree. PPIMiner consists of three main models: natural language processing (NLP) model, Protein-NER model, and PPI discovery model. The Protein-NER model, which is named ProNER, identifies the protein names based on two methods: dictionary-based method and machine learning-based method. ProNER is capable of identifying more proteins than dictionary-based Protein-NER model in other existing systems. The final discovered PPIs extracted via PPI discovery model are represented in detail because we showed the protein interaction types and the occurrence frequency through two different methods. In the experiments, the result shows that the performances achieved by our ProNER and PPI discovery model are better than other existing tools. PPIMiner applied this protein-named entity recognition approach and parsing tree based PPI extraction method to improve the performance of PPI extraction. We also provide an easy-to-use interface to access PPIs database and an online system for PPIs extraction and Protein-NER.
Every year, a large number of biological experiments have been conducted, and many papers about proteins or genes have been published. With the amount and exponential increase of biology literature, it is almost impossible for biologists to keep up with the all the updated information in their research. Text mining technology that automatically extracts key information has been utilized in many biology fields. Until now, many systems based on special functions using text mining are published [1–3]. Text mining involves analyzing a large collection of documents in a manner that reveals specific information, such as the relationships and patterns buried in the collection, which is normally imperceptible to readers . Two of the most important applications of text mining are bionamed entity recognition (Bio-NER) and mining correlations or associations such as protein-protein interactions (PPIs) from the literature .
Bionamed entity recognition is a very important task because it is directly related to extracting the PPIs. If we cannot identify bionamed entities correctly, then we cannot extract the correct PPIs. Our goals are the number of recognized bionamed entities and accuracy of recognition. Thus, the main issues in this topic are as follows: how to increase the recognition accuracy and how to identify as many proteins and genes as possible. However, until now, it is difficult to satisfy the two questions together.
Three basic approaches are usually applied: the dictionary-based approach, rule-based approach, and machine learning-based approach. The dictionary-based approach refers to the use of word lists or databases containing protein or gene names, for comparison and identification. The rule-based approach is used, for example, on extracting events based on the cooccurrence of strings, prefixes, or syntactic tags. Machine learning-based approach is a method which applies machine learning algorithm to classify bionamed entities from sentences. Comparing with above two approaches, machine learning-based approach is the most popular topic because biology named entities are usually not represented by recommended names in the raw text data. Many bionamed entity recognition systems and methods have been proposed. The applied machine learning algorithms are HMM [6, 7], SVM [8, 9], MEMM , CRFs [11–13], and so forth. To increase the accuracy of recognition, several researches applied two machine learning algorithms together [9, 14].
The development of PPI extraction system is more popular than Protein-NER systems. Many PPI extraction systems have been published too, which have specific advantages, for example, PPI Finder , CBioC , and iHOP , among others. PPI Finder focuses on extracting human PPIs and gives an interface for searching the extracted PPIs. It also constructs a database to store the PPIs and related information. CBioC extracts binary relationships between biological entities automatically from the biomedical literature and provides a platform that allows community collaboration in the annotation of the extracted relationships. iHOP is an online service that provides gene-guided network as a natural way of accessing PubMed abstracts and brings all the advantages of the internet to scientific literature research. However, these systems have minor consideration in improving Protein-NER methodologies.
From 2004, the BioCreAtIvE challenge began which is held by BioCreAtIvE (A critical assessment of text mining methods in molecular biology). Three main tasks were posed at the first BioCreAtIvE challenge: the entity extraction task, the gene name normalization task, and the functional annotation of gene products task. Until now, five BioCreative Challenge Evaluations and Workshops (BioCreative I, II, II.5, III, and IV) were held [28, 29]. The last workshop contain three tasks: (I-Triage) a collaborative biocuration-text mining development task for document prioritization for curation; (II-Workflow) a biocuration workflow survey and analysis task; and (III-Interactive TM) an interactive text mining and user evaluation task.
Many machine learning-based Protein-NER methods are proposed. However, the most published tools for PPI extraction did not apply new machine learning-based Protein-NER approach (e.g., iHOP , PPI Finder , and PPInterFinder ). These studies still prefer to use dictionary-based Protein-NER method because machine learning-based Protein-NER is more time-consuming and the performance is not enough high [5, 25]. But actually, if you only use dictionary-based Protein-NER method to identify the proteins in the text, we will miss so many protein names which not are included in the “dictionary.” Furthermore, most published studies on PPI extraction [18, 21, 23, 24] do not consider the Protein-NER step and just developed methods for PPIs extraction. In other words, these two related topics are researched independently. So it is difficult to merge and be applied by users.
In the study on the extraction of PPI patterns, the step of Protein-NER is important . If the most of protein-named entities we found are not correct or many protein-named entities are not identified by Protein_NER method, the PPI extraction task cannot be continued. It is the first step of all biomedical information extraction tasks including PPI extraction so that it directly affects the quality of results . For example, if we just find out half of the protein-named entities, then our result will just contain less than half of the information of PPIs. So in our study, we constructed a high quality Protein-NER model and merged into PPI process. Furthermore, it used two different methods, so that it not only can identify the protein names that are included in protein dictionary, it can also identify the proteins which are not included in protein dictionary. Combining this Protein-NER method to our PPI extraction process based on passing tree can efficiently improve the extraction of PPI patterns, because this approach will save time to find protein names again in the step of PPI extraction.
Many statistic methods have been used to find PPIs [31, 32]. They often find two proteins’ occurrence frequency in a sentence or some mathematic formula to calculate the correlations. For example, if two proteins appear together in a sentence or in a paragraph frequently, the two proteins interact with each other. However it cannot achieve the more accurate interactions between proteins, because two proteins can have many other relationships except interaction. Text feature (interaction word) based PPI extraction methods are frequently applied to discover the protein interaction pattern because this method can discover clearer interactions between the proteins. But if we need to check all the sentences, it will be time-consuming. Therefore, we propose our approach combining these two methods in discovering weighted PPIs by occurrence frequency. Comparing with text feature based PPI exaction methods, this approach can improve the efficiency. Comparing with statistic methods, it also improves the performance. Currently, most of these researches seldom show the information of PPI types (e.g., bind, link, and inhibit); the type of PPI is very important information for biology researchers. In our approach, we also improve it by including information about protein interaction type referred in the literature.
The proposed PPI extraction system is constructed via three main models: the natural language processing (NLP) model, protein-named entity recognition model, and protein interaction discovery model. The workflow of PPIMiner is as shown in Figure 1.
2.1. Data Preprocessing: Natural Language Processing
As known, NLP is the first step of text mining. In this study, the NLP model mainly contains two parts: sentence parsing and part-of-speech (POS) tagging. Here, Penn Treebank English POS tag set  is used for part-of-speech (POS) tag because this POS tag set is a standard tag set. The PPIMiner contains a Protein-NER model to identify the protein name, so we do not need special biological POS tag set for biology text. We applied open source of Stanford Log-Linear Part-Of-Speech Tagger which is based on Penn Treebank English POS tag set for our NLP model. The GENIA corpus 3.2  is used as training dataset of protein name recognition model, and the format of dataset is an XML file. So, it is easy to classify the protein names from the text.
2.2. ProNER: Protein-Named Entity Recognition
Studies on PPI exaction systems mainly used the dictionary-based Protein-NER method to search the proteins. In order to increase the number of proteins extracted, we used two methods, dictionary-based Protein-NER method and machine learning-based Protein-NER method, to construct the protein-named recognition model, which has different advantages, respectively. User interface is as Figures 2(a) and 2(b).
2.2.1. Dictionary-Based Protein-NER
The result of dictionary-based Protein-NER method is the most believable result, comparing with other methods’. And it is easier to implement than others. Almost published systems used this method [5, 25]. In our study, we also used this method. UniProtKB/Swiss-Prot Database  is used as PPIMiner’s protein dictionary. There are 200,839 protein entries in the database. We select the 19,304 human proteins to use. HGNC (HUGO Gene Nomenclature Committee)  is used as a gene dictionary to discover the PPI represented by gene names. 5,038 extracted gene entities are restored in our database. To search as many protein-named entities as possible, we did not use only recommendation name of protein or gene; we also used other names in protein or gene entity database.
2.2.2. Machine Learning-Based Protein-NER
This Protein-NER method contains three steps: candidate generation (boundary detection), for feature extraction and classification. The reason we use this method is that it can find novel protein and gene names, which are not included in the dictionaries or the proteins and genes that are not represented by standard names. So, we used the machine learning-based method to search the more bionamed entities.
Candidate Generation Using the BFSM Algorithm. Boundary detection is a difficult task for machine learning-based methods. If we analyze more than 100,000 papers, the text dataset is also too large to identify using data mining algorithms. In this paper, to solve this problem, we used Bayesian probability based Finite State Machine (BFSM) to detect the boundary of bionamed entities and generate the candidates for classification. BFSM algorithm is proposed by us in 2011 . And it is proved that this algorithm is suitable to apply for Protein-NER. Here, we use this algorithm to extract the rules of POS set of protein names from text data. And the corpora which satisfy the rules of POS set will be generated as protein-named entity candidates.
The process is as follows.(i)Mine the frequent POS patterns of bionamed entities using Apriori algorithm.(ii)Calculate the confidence and search the association rules in the -frequent itemsets () that satisfy the minimum confidence.(iii)Construct the BFSM model according to the frequent patterns and confidence values between the items.(iv)Using the BFSM model, generate the candidates of protein-named entities.
To understand BFSM easily, an example of finite state network to generate POS rules and generated POS rules is given as in Figure 3.
Feature Selection and Extraction. Until now, many features for Protein-NER are generated . However, the target bionamed entities are different; the suitable features are different too. So, according to the external rules of protein and gene molecule name, we set up several distinguishing features for protein and gene named entity recognition. We extract these features from the candidates to converted candidates list into a tabular dataset.
In this study, our extracted features as follows.(1)Sent_Uppercase: the first letter of a word is an uppercase letter, which is not located in the middle of sentence.(2)Word_Uppercase: uppercase letter is located in the middle of a word.(3)Word_Symbol: a word contains comma (,), hyphen (-), and slash (/).(4)Word_Num: numbers are located in the middle of a word.(5)Word_Alphabet: special letters are located in the middle of a word.(6)Biology_Word: the protein-named entity candidate contains special words: protein, gene, receptor, and factor.(7)Length: the length of the words.(8)Suffix_Word: suffix word of protein-named entity candidate.(9)Prefix_Word: prefix word of protein-named entity candidate.(10)Suffix_Letter: suffix letter of protein-named entity candidate.(11)Prefix_Letter: prefix letter of protein-named entity candidate.(12)POS: POS set of protein-named entity candidate.Classification. We use the SVM algorithm  to classify the protein name based on these features from the papers. Compared with other machine learning methods, SVM algorithm is suitable for our text dataset and the accuracy is higher than others. When we begin to test, this kind of method is more time-consuming than other classification algorithms. But the accuracy is more important than running time for us because our purpose is to find more correct protein-named entities using this system. From the result of experiment, we know that the performance of classification algorithm is better, when the training data is enough. The SVM algorithm we used is libSVM in the experiments.
2.3. Protein-Protein Interaction Discovery
Based on these identified protein/gene names, we used two kinds of methods, the interaction pattern-based method and frequent pattern-based method, to discover the PPI patterns. The interaction pattern-based method is that we discover the interaction patterns (interaction words) between two proteins such as “active” and “bind.” We used the parsing tree based interaction pattern scanning approach to discover the PPI patterns. To search the patterns easily, the parsing tree is used to search the algorithm structure in this research, and we then search the surround protein names to discover the correct PPIs. The frequent pattern-based method searches for the pattern that frequently occurs in the whole dataset. The frequency is the weight of the PPI patterns exacted from the interaction pattern-based method. The final discovered PPI patterns include the interaction weight and represent the interactions between proteins more accurately than the result of just using one method, interaction patterns based or frequent patterns based method (Figures 2(c) and 2(d)).
We use a parsing tree to search the interaction patterns, and this method is much suitable for the PPIMiner that includes Protein-NER model because when we search the protein-named entity candidates, we change the POS set to protein or gene mark “TAG_PROT.” It can decrease the deep of parsing tree and then decrease the complex.
In Figure 4, a simple example shows how to change the POS set to protein or gene mark “TAG_PROT.”
2.3.1. PPI Candidate Extraction
The process of PPI extraction is as follows.(i)Search the interaction patterns in a sentence.(ii)Discover the interaction patterns in the sentence that includes the cooccurring genes or proteins. Form 1: protein(s) + interaction word (verb/noun) + protein(s). Form 2: interaction word (noun) + protein + protein. Form 3: protein(s) + (null, “/,” “:,” “and”) + protein(s) + interaction word.(iii)Map the recognized protein-named entity name onto gene and recommend gene name in dictionary.(iv)Postpruning based on prepositions to filter out incorrect PPIs based.
Here, the 113 interaction patterns are getting from  and 175 interaction words are getting from HPDR50 . The total interaction words are 1,224 words that contained derivatives of these words and deleted duplicated works. To extract more patterns, we derived the interaction patterns to other kinds of POS interaction patterns.
2.3.2. Semantic PPI Information Extraction
For discovery of more semantic PPIs, substitution words are discovered to find out more PPIs which are explained by two or more sentences. The substitution word list includes “this protein,” “that protein,” “the protein(s),” “these proteins,” and “those proteins.” And we will check the subject in the previous sentence. If the subject is not protein name, the PPI will be skipped.
Protein-named entities are found out to discover PPIs firstly when researchers use statistical methods. And it is not important which one is extracted firstly between the interaction word and protein-named entities when feature based methods or pattern-based methods are used to discover PPIs. In our research, we search PPI interaction words firstly to increase search efficiency. If we search protein-named entities firstly and decide to analyze the sentence deeply, it is very time-consuming because too much substation words exist in a paper.
2.3.3. Pruning of PPIs
To improve the accuracy of prediction, a pruning method is proposed. Two kinds of conditions need to be satisfied.
Pruning 1. If the words “by,” “in,” “of,” and “as” are in front of the protein name, the extracted PPIs are pruned.
It can filter out incorrect PPIs which are only considered the combinations of protein and interaction words.
Pruning 2. If a negative word exists in the sentence, the extracted PPIs are pruned. Form 1: “not” + interaction word. Form 2: “no” + protein(s) + interaction words. Form 3: “neither” protein “nor” protein + interaction word.
In Form 3, if other negative word exists, the extracted PPIs are not pruned (e.g., neither Protein 1 nor Protein 2 interacts with Protein 3).
The purpose of pruning method is to solve the low value of recall problem in many cases.
2.3.4. Feature Extraction and Classification
To extract more correct PPIs, several features are generated which is suitable to classify the PPI. Three classification methods (SVM, C4.5, and neural network) are used to compare the result. The classification algorithms are provided by the Weka.
3. Results and Discussion
3.1.1. Dataset for Protein-NER Evaluation
In this research, we used three datasets extracted from the GENIA. GENIA corpus is the largest corpus of its type currently available, comprising 2000 abstracts with 18,545 sentences containing 39,373 named entities, and it is usually used to train and test the Protein-NER model. After sentence parsing and POS tagging, we converted the GENIA dataset into table dataset which includes 305,546 records, and 83,403 records are protein-named entities out of the all dataset (shown in Tables 1 and 2).
3.1.2. Dataset for PPI Extraction Evaluation
In this research, we used five corpora as experiment dataset which are provided by : AIMed, BioInfer, HpDR50, IEPA, and LLL (http://mars.cs.utu.fi/PPICorpora/GraphKernel.html). These corpora have unified format. The most of researches on PPI extraction were used these dataset for evaluation (shown in Table 3).
We collect 1,571,293 paper abstracts that are related to gene and protein from PubMed database. We save the paper abstracts to our database to extract PPIs. Finally, we compare the PPI patterns with HPRD PPIs to analyze the result. The HPRD contains 39,194 PPI extracted from 20,071 papers in PubMed. This file contains human protein-protein interactions in a tab delimited format.
3.2. Evaluation Criterion
Precision determines the fraction of records that actually turn out to be positive in the group the classifier has declared as a positive class:
The recall measures the fraction of positive examples correctly predicted by the classifier. It represents the ability to find protein-named Entities (Protein-NEs) in the dataset:
The -measure represents a harmonic-mean between precision and recall and takes both measures into account. A high value of -measure ensures that both of precision and recall are reasonably high:
3.3. Performance of Protein-NER
GENIA corpus data is classified by 36 class labels. In this paper, we just select data which is included in “protein-molecule” class label as protein-named entities because GENIA corpus data is used as training data for constructing the protein-named entity recognition model. Although there are 7 class labels related to protein and 5 class labels related to DNA, they did not include the external features of protein or gene names. For example, the data which are included in class “protein-family-or-group” and “DNA-family-or-group” cannot describe the protein or gene entity’s names; the data are just the protein or gene categories such as “receptor” or “gene.”
All the experiments are carried out on a unified computer platform with a 2.93 GHz CPU and a 4 GB RAM.
3.3.1. Effect of Prepruning by BFSM
In our Protein-NER model, prepruning directly affects the accuracy of protein identification. In order to know how to influence the identification of protein-named entities, we did two experiments. In the first experiment, at first we converted GENIA text dataset into table dataset which contain 12 features for training the classifier. The two class labels of converted dataset are “protein-named entity” and “Not_Protein-Named Entity.” The number of instance in dataset is 305,546. We use the Support Vector Machine (SVM) algorithm in Weka 3.6 as classifier. The performance evaluation result is very high, although we just select 5% dataset. Table 4 shows a direct proportion between sample size and -measure. But when we use the text dataset, the accuracy is difficult to be higher than 95% because those boundary detection methods are very difficult to find out the boundary of protein-named entity. In previous research, they use sliding window technique to generate the entity candidates . 305,547 candidates were generated using our method. The original text dataset includes 402,745 words. So, comparing with the number of candidates generated by sliding window, the number of candidates generated by BFSM is much small.
3.3.2. Effect of Features on Classification
We search and detect features to classify the protein or gene named entities, and we analyze the features’ effect on classification. Table 5 shows the distributions of 6 binary features in two classes, respectively. From Table 5, we know that the percentage of two kinds of values in two different class labels is much different. The percentage of value “True” of Features “Word_Symbol” and “Word_Alphabet” in class “protein-named entity” is much higher than in “Not_Protein-named entity.” To get more clear values to evaluate the features, we calculate the Gain Ratio of the features by Weka 3.6 as shown in Table 5. The Gain Ratio is just the ratio between the information gain and the intrinsic value. Here, the information gain is equal to the total entropy for an attribute if a unique classification can be made for the result attribute of each of the attribute values. The information gain for an attribute is defined as follows:where Attr is a set of all attributes and a set of all training examples, value with fines the value of a specific example for attribute , specifies the entropy. From this experiment, we found that the most informative features are “Sent_Uppercase,” “Word_Num,” and “Word_Symbol.” Furthermore, the performance evaluation values are close to 100%. It means that the features are very suitable for protein-named entity recognition. We did other feature analysis experiments about DNA or others. But the feature ranking is different according to the class target.
3.3.3. Comparison among Submodels in PPIMiner
To know the performance of submodels, we did the experiment to recognize the protein-named entities using two submodels, respectively, as in Table 6. From the result, we can get that the precision of dictionary-based model is much higher than others, but recall -measure and accuracy is too lower than others. And when we use the two submodels together, the -measure is the best. Although many machine learning-based Protein-NER methods are proposed, the PPI extraction systems prefer to use dictionary-based approach than machine learning-based Bio-NER approach to extract the correct PPI patterns. However too many protein-named entities are missed, if we just use dictionary-based approach, and too many PPI patterns are missed.
3.3.4. Comparison with Other Existing Methods
To test the performance of our Protein-NER model, we did the comparison experiment with three published systems and two resent published researches. We used the 3-fold cross-validation method to evaluate. We also used the same training dataset to construct the identification models and used same test dataset to evaluate. And the performance results of the two resent published researches about Bio-NER referred to published papers. From the result in Table 7, we can get that our Protein-NER model has the highest -measure. Here, ABNER and Penn BioTagger are applied conditional random fields to recognize the protein-named entities. And the result is better than others. Conditional random fields algorithm is based on conditional probability to calculate the possibility of protein-named entities. In our approach, BFSM we used is also based on conditional probability to generate the rules of POS for protein-named entities. Finally, we applied vector based methods to candidates so that more properties of protein-named entities are discovered.
3.4. Performance of PPI Extraction
3.4.1. Result of PPI Extraction
We use this system to extract 40,466 PPIs from 477,008 abstracts in PubMed. We extract 6,072 patterns from 20,071 papers which are included in HPRD. We did not extract the PPIs as many as manual extraction in HPRD. But the number of PPI patterns will continue to increase. And we extracted more than 7,919 rules from other abstracts. For example, we extracted 47 PPI patterns related to protein “IGF1.” In our PPIs database, several extracted PPIs maybe have incorrect PPIs. But user can check by himself from the original text provided by PPIMiner. The system is freely accessible at http://188.8.131.52:8080/PPIMiner/. We purpose extracting more possible PPIs. And our system also shows the type of relation between interacting proteins. Actually, it is more informative for biologists that the proteins are how to interact.
3.4.2. Compare with Other PPI Extraction Methods
To compare with existing methods, the standard dataset, AIMed corpus, was used. So we extract the result from the papers. As in Table 8, our proposed method has higher performance.
From the result, we can get that feature or pattern-based methods’ performance is better than subsequence kernels or all-path graph kernel. However, it is important to find out right patterns or features of PPI information. If merged all the discovered patterns and features, the performance will be improved.
Many human PPI databases (HPRD, BioGRID, and BIND) published extract PPIs manually. But it is impossible to extract all the PPIs manually from the articles because the number of biomedical articles is growing at a very fast rate. So we need to automatically extract PPIs and upgrade the PPI database regularly.
In this paper, we developed an online system of the protein-protein interaction extraction based on SVM and parsing tree, PPIMiner. The PPI extraction system is constructed via the natural language processing model, protein-named entity recognition model, and protein-protein interaction discovery model. We used the two methods, dictionary-based method and machine learning-based method, to construct the protein name recognition models which have a different advantage, respectively. We extract the PPI patterns based on these identified protein-named entities using parsing tree. Until now, our database stored 40,466 PPIs.
The advantages of our developed system are as follows. First, in our PPI extraction system, we merged protein-named entity recognition model we proposed to PPI extraction model. So, comparing with the systems which just use the dictionary-based protein-named entity recognition method, we can extract more PPIs. The accuracy of protein-named entity recognition model is higher than other existing models and published methods. Second, PPIMiner provides three main functions: PPIs search in database, protein-named entity recognition, and PPI extraction from text data. We made an interface for protein-named entity recognition model. From PPIMiner interface, users are easy to access our protein-named entity recognition model and users can apply this model for their research.
In addition, our method used 12 features for construction of protein-named entity recognition model. And from the accuracy of classification result, we can know that these features are much suitable for protein-named entity recognition.
For feature work, we plan to supplement the protein and gene dictionary to increase the accuracy of identification. We also plan to develop the PPI extraction algorithm to extract more PPIs. The extracted PPIs in our database will continue to be updated.
Conflict of Interests
The authors declare that there is no conflict of interests.
This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Science, ICT and Future Planning (no. 2013R1A2A2A01068923).
S. Anthony, V. Sintchenko, and E. Coiera, “Text mining for discovery of hostpathogen interactions,” Infectious Disease Informatics, vol. 1, pp. 149–165, 2010.View at: Google Scholar
A. Abi-Haidar and L. M. Rocha, “Collective classification of biomedical articles using T-cell cross-regulation,” in Proceedings of the Artificial Life XII: 12th International Conference on the Simulation and Synthesis of Living Systems, H. Fellermann, M. Dörr, M. Hanczyc et al., Eds., pp. 706–713, 2010.View at: Google Scholar
G. Zhou and J. Su, “Exploring deep knowledge resources in biomedical name recognition,” in Proceedings of the International Joint Workshop on Natural Language Processing in Biomedicine and Its Applications, pp. 96–99, Association for Computational Linguistics, Geneva, Switzerland, June 1996.View at: Google Scholar
J. Finkel, S. Dingare, H. Nguyen, M. Nissim, C. Manning, and G. Sinclair, “Exploiting context for biomedical entity recognition: from syntax to the web,” in Proceedings of the International Joint Workshop on Natural Language Processing in Biomedicine and Its Applications (JNLPBA '04), 2004.View at: Google Scholar
B. Settles, “Biomedical named entity recognition using conditional random fields and rich feature sets,” in Proceedings of the International Joint Workshop on Natural Language Processing in Biomedicine and Its Applications, pp. 104–107, Association for Computational Linguistics, 2004.View at: Google Scholar
R. Leaman and G. Gonzalez, “BANNER: an executable survey of advances in biomedical named entity recognition,” in Proceedings of the Pacific Symposium on Biocomputing, vol. 13, pp. 652–663, 2008.View at: Google Scholar
R. Stre, K. Yoshida, M. Miwa, T. Matsuzaki, Y. Kano, and J. Tsujii, “Extracting protein-interactions from text with the unified AkaneRE event extraction system,” IEEE/ACM Transactions on Computational Biology and Bioinformatics, vol. 7, no. 3, pp. 442–453, 2010.View at: Publisher Site | Google Scholar
Y. Niu, D. Otasek, and I. Jurisica, “Evaluation of linguistic features useful in extraction of interactions from PubMed; application to annotating known, high-throughput and predicted interactions in I2D,” Bioinformatics (Oxford, England), vol. 26, no. 1, pp. 111–119, 2010.View at: Publisher Site | Google Scholar
C. Baral, G. Gonzalez, A. Gitter, C. Teegarden, A. Zeigler, and G. Joshi-Topé, “Cbioc: beyond a prototype for collaborative annotation of molecular interactions from the literature,” in Proceedings of the Computational Systems Bioinformatics/Life Sciences Society Computational Systems Bioinformatics Conference, 2007.View at: Google Scholar
T. Munkhdalai, M. Li, E. Namsrai, O.-E. Namsrai, and K. H. Ryu, “BFSM: finite state machine learned as name boundary definer for bio named entity recognition,” in Proceedings of the 3rd International Conference on Awareness Science and Technology (iCAST '11), pp. 344–349, IEEE, Dalian, China, September 2011.View at: Publisher Site | Google Scholar
D. Tikk, P. Palaga, and U. Leser, “A fast and effective dependency graph kernel for PPI relation extraction,” BMC Bioinformatics, vol. 11, article 8, 2010.View at: Google Scholar
D. Otasek, K. Brown, and I. Jurisica, “Confirming protein-protein interactions by text mining,” in Proceedings of the 6th SIAM Conference on Text Mining, IEEE, Bethesda, Md, USA, April 2006.View at: Google Scholar
K. Toutanova, D. Klein, C. Manning, and Y. Singer, “Featurerich part-of-speech tagging with a cyclic dependency network,” in Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics on Human Language Technology, pp. 173–180, Association for Computational Linguistics, Baltimore, Md, USA, 2003.View at: Google Scholar
Uniprotkb/swiss-prot database, http://www.uniprot.org/.
Q. Bui, B. Nualláin, C. A. Boucher, and P. M. A. Sloot.
J. Eom and B. Zhang, “Pubminer: machine learning-based text mining for biomedical information analysis,” Genomics & Informatics, vol. 2, pp. 99–106, 2004.View at: Google Scholar