BCC-NER: bidirectional, contextual clues named entity tagger for gene/protein mention recognition
© The Author(s). 2017
Received: 24 October 2016
Accepted: 21 April 2017
Published: 5 May 2017
Tagging biomedical entities such as gene, protein, cell, and cell-line is the first step and an important pre-requisite in biomedical literature mining. In this paper, we describe our hybrid named entity tagging approach namely BCC-NER (bidirectional, contextual clues named entity tagger for gene/protein mention recognition). BCC-NER is deployed with three modules. The first module is for text processing which includes basic NLP pre-processing, feature extraction, and feature selection. The second module is for training and model building with bidirectional conditional random fields (CRF) to parse the text in both directions (forward and backward) and integrate the backward and forward trained models using margin-infused relaxed algorithm (MIRA). The third and final module is for post-processing to achieve a better performance, which includes surrounding text features, parenthesis mismatching, and two-tier abbreviation algorithm. The evaluation results on BioCreative II GM test corpus of BCC-NER achieve a precision of 89.95, recall of 84.15 and overall F-score of 86.95, which is higher than the other currently available open source taggers.
KeywordsBiomedical text mining Named entity recognition Conditional random fields Hybrid NER approaches Margin-infused relaxed algorithm Bidirectional parsing
Scientific literature is the major source of biomedical knowledge, and the interest in developing automated text mining solutions to extract useful information from biomedical text is increasing every year. Bio-named entity recognition (NER) is the key step for such information extraction from biomedical literature [1–5].
Biomedical-named entities include genes, proteins, RNA, cell, and cell-line. NER in the biomedical domain is generally considered to be more difficult than other domains such as newswire as there is no standard nomenclature naming biomedical entities like genes and protein names resulting in ambiguity, and further, there are millions of biomedical entity names in use and more entities are added regularly [2, 3]. Moreover biomedical entities such as gene and protein names have similar morphology and context .
The commonly used techniques for NER task are rule-based approaches , dictionary-based approaches , machine learning approaches , and recent hybrid systems which use a combination of two or more approaches. Presently, hybrid approaches give best results in NER task [4, 5]. To understand the current state-of-the-art, we briefly introduce some of the recent hybrid approaches explored for biomedical NER task. Raja et al.  used a hybrid named entity rule-based tagger with 14 hand-crafted rules and a set of post-processing methods and an abbreviation algorithm to tag the human gene/proteins from biomedical articles.
Leaman et al.  proposed a machine learning-based open source biomedical named entity system which was a combination of conditional random fields (CRF) and some post-processing methods to tag gene/proteins. Campos et al.  designed a biomedical hybrid tagger with machine learning algorithm and lexicon-based approaches. Zhu et al.  used both support vector machines (SVM) and CRF for better performance. SVM, a binary classifier, was used to separate the biological terms from non-biological terms, and CRF was used to determine the types of biological terms. Thus, the results of SVM as well as CRF were fused and a useful algorithm was developed after applying two rules.
2 Materials and methods
Text processing module which includes NLP preprocessing, feature extraction, and selection
CRF training module, which uses bidirectional CRF for learning and labeling in both directions and model integration using MIRA
Post-processing module, which includes contextual clues and abbreviation identification algorithm.
The details of each module are described in the following sections.
2.1 Text processing
2.1.1 Text preprocessing
In order to prepare the corpus for feature extraction and NER, the following preprocessing steps were applied initially: (i) sentence splitting for splitting the articles or abstracts to sentences, (ii) tokenization for splitting the sentences into individual tokens, (iii) lemmatization to convert the tokens to the basic form of the word, (iv) POS tagging, and (v) chunking. OpenNLP  was used for sentence splitting, tokenization, POS tagging, and chunking. BioLemmatizer  was employed for lemmatization.
2.1.2 Feature extraction
Examples of orthographic, morphologic, and prefix-suffix features
In addition to the above basic features, we have also used the general NER features such as N-gram  and corpus frequency words  to improve the performance of the tagger. For the N-gram feature, we used character N-grams using a sliding window of size 4. The sliding window starts from the beginning to four characters of each token .
In our initial tagging, we found some of the tokens tagged as a bio-entity in one sentence but not tagged in some other sentence due to the variation in sentence structure. For example, the gene COR was tagged as gene in one occurrence but not in the second in the same abstract (BC2GM000136143). In order to overcome this issue and uniformly tag such tokens, we used corpus frequency words feature only for noun tokens.
For corpus frequency words, we calculated the total number of times each word or sequences of words occur in the corpus. We took the words or sequences of words with a minimum threshold frequency in the range of (1–10) into consideration.
where cf(w) denotes the corpus frequency of the word w, i.e., the total number of times w occurs in the corpus. In our experiments, we set θfreq = 10.
2.1.3 Feature selection
Due to the inclusion of rich set of features including word N-grams, the number of features associated with each token is very large and many of them may not be related to the tokens. Further, redundancy may also occur during the training phase of the samples which can cause performance degradation of the tagger. In order to use only the most informative features for classification task and to discard unrelated features, we employed a principle component analysis (PCA)-based feature selection method . PCA is the most commonly used dimension reduction method. The vital scheme of PCA is to shrink the dimensionality of a feature set albeit trying to retain the variance present in the original predictor features to the greatest degree possible .
PCA converts the data to a new dimensional space in such a way that the features with the highest eigenvalue component comes to the first coordinate, next, the eigenvalue component on the second coordinate, and so on. The dimensionality of the data is then shrunk by ignoring the lower eigenvalue components. Thus, PCA provides the most essential directions that can efficiently represent the data which is shortly explained below .
The full principal component decomposition of data matrix X can be given as T = XW where W is a 2D matrix whose columns are the eigenvectors of X T X. The transformation T = XW maps a data vector x(i) from an original space of p variables to a new space of p variables which are uncorrelated over the dataset. However, not all the principal components are kept during the transformation. Only the first L principal components produced by using the first L-loading vectors that are kept gives the truncated transformation TL = XWL where the matrix TL now has n rows but only L columns. In other words, PCA learns a linear transformation where the columns of p × L matrix W form an orthogonal basis for the L features that are de-correlated . Among all the data matrices thus transformed to only L columns, this score matrix maximizes the variance in the original data that has been preserved, while minimizing the total squared reconstruction error.
2.2 Bidirectional CRF model and integration
where λk is a parameter to be estimated from training data and indicates the informativeness of the respective feature, Z(B) is a normalization factor and FK(A, B) = n i = 1 fk(Ai−1, Ai, B, i), where each fk(Ai−1, Ai, B, i) is either a state function s(Ai−1, Ai, B, i) or a transition function t(Ai−1, Ai, B, i).
2.2.2 Bidirectional CRF
Further, we have used the second-order CRF for both forward and backward learning. In a higher order model of “n,” each label depends on a specific number of “n” previous labels. Thus, the probability will consider not only the previous observation and its features but also n-previous observations and features. Though the higher order models provide improved results, the training complexity of higher order models increases exponentially. So we used the second-order CRF which is generally used in NER task .
2.2.3 Model integration
The common ways to combine the results of bidirectional parsing includes simple set operations such as intersection and union. Usually, intersection will improve the precision and reduce the recall, while using union will improve the recall and reduce the precision . In general, union and intersection methods failed to improve the performance because they lead to a trade-off between recall and precision. For better model integration, we used MIRA algorithm proposed by Crammer and Singer . MIRA solves the above mentioned trade-off problem, since it combines the forward and backward models by adding the feature weights of both models . Further, MIRA successfully reduces the training time by exploiting the no update procedure if the instance is classified as correct and also reduces the memory space by following no fixed step size for the update procedure. Hence, we have used MIRA for model combination and compared its performance with union in terms of both results and processing time.
The score of the current correct training y must be greater than the score of any other possible y′ by at least the loss (number of errors) of that y′ in comparison to y.
Surrounding text features
Abbreviation resolution algorithm
2.3.1 Surrounding text features
Relation words: In biomedical text, existence of some relation keywords (binding, activate, etc.) implies that some protein names might occur . We compiled around 400 interaction keywords from biomedical texts. If any relation keyword was present in the sentence, then its previous and next words were checked for protein/gene names occurring three or more times in the training set. They were then tagged as gene names if occurring so.
Connective words: Similar to the relation word features, here, we checked for the linguistic cue connective words such as “and” and “or” in the sentence. If these words were present in the sentence, then the previous and next words were checked for protein/gene names that occur three or more times in the training set and tagged as gene names.
2.3.2 Parenthesis mismatching
One problem with CRF modeling is that it wrongly identifies the parenthesis, and it leads to parenthesis mismatching problem. For example, in the case of an opening curly brace being tagged and the closing curly brace not tagged, we need to remove the mismatched curly brace. We used left parenthesis and right parenthesis extension method to remove the mismatched parenthesis tagging . This is shown in the following example 1.
Sentence: The hepatocyte nuclear factor-3 (HNF-3)/forkhead (fkh) proteins consist of an extensive family of tissue-specific and developmental gene regulators.
Before post-processing: (|B HNF|I -| I 3|I) |O
After post-processing: (|O HNF|B -| I 3|I) |O
2.3.3 Two-tier abbreviation algorithm
Identifying the LF (long form) and SF (short form): CRF tagger most of the time tags only long form or short form and misses either one. To tackle this and identify the missed long-form and short-form abbreviations, we used a modified version of the simple abbreviation algorithm which is used in BioC  and named as “extract abbreviation method”. This is shown in the following Example 2.
Sentence: Brown adipose tissue (BAT) and brown-like cells in white adipose tissue (WAT) can dissipate energy.
Before Post-processing: Brown |O adipose |O tissue |O (|B BAT|I) |I
After Post-processing: Brown |B adipose |I tissue |I (|B BAT|I) |I
Abbreviation disambiguation: The second technique in two-tier abbreviation algorithm is abbreviation disambiguation. Sometimes two proteins or two genes have the same abbreviation, for example, “angiotensin converting enzyme (ACE)”and “acetylcholinesterase (ACE).”.In the above example, “ACE” denotes both angiotensin converting enzyme (ACE) and acetylchlinesterase (ACE). To overcome this problem, we used the abbreviation disambiguation method word sense disambiguation (WSD) . In WSD, along with other features, we used domain specific features such as CUI (concept unique identifiers)  and MeSH terms . WSD identifies all related words in the text which could be mapped to CUI or MeSH terms and disambiguates them to their correct sense of the long form or short form.
3 Results and discussions
BCC-NER training and testing is based on BioCreative II GM corpus which contains 15,000 training sentences and 5000 testing sentences . While training and testing, we employed our feature set with bidirectional CRF models in both forward and backward directions. Finally, we applied MIRA algorithm to integrate both models to construct the combined model.
CRF + Forward parsing + post-processing,
CRF + Backward parsing + post-processing,
CRF+ Union (Forward + Backward) + post-processing and
CRF + Combined model MIRA + post-processing.
Where, TP refers to the number of proportion of biological entities correctly identified by our hybrid approach, FN refers to the number of proportion of biological entities that the approach failed to identify and FP refers to the number of proportion of biological entities that were incorrectly identified by this approach
System performance on various models
CRF + forward parsing + post-processing
CRF + backward parsing + post-processing
CRF+ union (Forward + backward) + post-processing
CRF + combined model MIRA + post-processing
Comparison with other systems:
Comparison of our system with other open source systems
In this paper, we describe our hybrid named entity recognition system named BCC-NER for tagging biomedical entities. BCC-NER integrated all major happenings in current NER task and includes three modules. For example, we have used a rich set of features combining the major 32 basic ones, word N-grams, and corpus frequency words. The state-of-the-art feature selection and extraction algorithm PCA was applied to reduce the high number of features associated with each tokens.
The latest results on biomedical NER clearly indicate that better performance can be achieved by combining several systems. In these lines, BCC-NER employs bidirectional CRF model combined with MIRA. We are the first one to explore such a combination using MIRA in biomedical NER which gives improvised outcomes than the traditional union and intersection methods. Another important feature that contributes to our hybrid approach is the consideration of contextual clues. In contextual clues, our tagger finds the interaction words and checks if the previous and next words are present in the post-keyword list (pkl). If present, it is tagged as a gene.
Early complement components, <gene>C1q</gene> and C4, and IgA secretory piece were absent.
In the above example, our contextual clues find the connective word “and,” and then checks for the previous and next words for their presence in pkl. The previous word C1q has been already tagged as gene. So the next word “c4” is then checked for its presence in pkl and tagged as gene after this step.
Early complement components, <gene>C1q</gene> and <gene>C4</gene>, and IgA secretory piece were absent.
Following this step, we re-implemented the same step again but with relation words. If any relation keywords are found in the sentence, then both its previous and next words were checked for protein/gene names occurring three or more times in the training set.
Finally, we applied the parenthesis post-processing and two-tier abbreviation algorithm as explained above in the Sections 2.3.2 and 2.3.3. Thus, the issues of parenthesis mismatching and abbreviation disambiguation were subdued.
We propose a hybrid named entity tagging approach BCC-NER which on evaluation indicates that the inclusion of rich set of features and utilization of bidirectional CRF combined with MIRA gives best results. Additional performance improvement was achieved by post-processing steps including surrounding text features, parenthesis mismatching and two-tier abbreviation algorithm.
Although in BCC-NER we tried to integrate various state-of-the-art methods on existing tools, some aspects can be further explored. We are currently investigating other approaches including domain knowledge information through the use of dictionaries or machine learning-based solutions. In addition, we plan to include methods like co-training, superior abbreviation algorithms and different features to generate improved results.
Availability of data and materials
The open source online tool will be available at http://www.biominingbu.org:8080/BCC-NER
Dr NJ conceived the idea and supervised the overall research work. GM, SA, and BB carried out the whole work including the programming, training, and testing and preparing the manuscript. All authors read and approved the final manuscript.
GM is a research scholar doing his Ph.D at the Data mining and Text Mining Laboratory, Bharathiar University, Coimbatore, India. His area of research focuses on data mining, text mining and machine learning methods for biomedical literature. SA is doing her Ph.D in Data mining and Text mining lab, Department of Bioinformatics, Bharathiar University, Coimbatore. Her research work focuses on machine learning, data mining, text mining, text classification, and biostatistics. BB is a research scholar doing his Ph.D at DRDO-BU Center for Life Sciences, Bharathiar University campus, Coimbatore, India. His research interests are in the areas of literature mining, machine learning, and biomedical data mining. Dr. NJ is currently working as a Professor and Head at Department of Bioinformatics, Bharathiar University, Coimbatore, India. He heads the Data mining and Text Mining Laboratory which focuses on data mining, text mining and machine learning methods for microarray data analysis and interpretation and other high-throughput data from biology. He completed his PhD in Bioinformatics from the University of Ulster, Belfast, United Kingdom and spent one year as a visiting pre-doctoral fellow at Northwestern Medical School, Northwestern University, Chicago, USA.
The authors declare that they have no competing interests.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Open AccessThis article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.
- B Settles, ABNER: an open source tool for automatically tagging genes, proteins and other entity names in text. Bioinformatics 21(14), 3191–3192 (2005)View ArticleGoogle Scholar
- R Leaman, G Gonzalez, Pacific symposium on biocomputing, in BANNER: an executable survey of advances in biomedical named entity recognition (Pacific symposium, Hawaii, 2008), pp. 652–663Google Scholar
- D Campos, S Matos, JL Oliveira, Gimli: open source and high-performance biomedical name recognition. BMC Bioinformatics 14, 54 (2013)View ArticleGoogle Scholar
- K Raja, S Subramani, J Natarajan, A hybrid named entity tagger for tagging human proteins/genes. Int. J. Data Min. Bioinform. 10, 315 (2014)View ArticleGoogle Scholar
- Zhu F, Shen B, Combined SVM-CRFs for Biological Named Entity Recognition with Maximal Bidirectional Squeezing. PLoS ONE. 7(6) (2012)Google Scholar
- JT Chang, H Schutze, RB Altman, GAPSCORE: finding gene and protein names one word at a time. Bioinformatics 20, 216–225 (2004)View ArticleGoogle Scholar
- Tanabe L, Xie N, Thom LH, Matten W, Wilbur WJ, GENETAG: a tagged corpus for gene/protein named entity recognition. BMC Bioinf. 6(1) (2005).Google Scholar
- J Atserias, B Casas, E Comelles, M González, L Padró, M Padró, Proceedings of LREC, in FreeLing 1.3: Syntactic and semantic services in an open-source NLP library (LREC, Italy, 2006), pp. 48–55Google Scholar
- H Liu, T Christiansen, WA Baumgartner, K Verspoor, BioLemmatizer: a lemmatization tool for morphological processing of biomedical text. J Biomed Semant 3, 3 (2012)View ArticleGoogle Scholar
- S Jiampojamarn, N Cercone, V Kešelj, Conf. of the Pacific Assoc. for Computational Linguistics, in Biological named entity recognition using n-grams and classification methods (PACLING, Tokyo, 2005)Google Scholar
- RE Moe, International Conference on Text, Speech, and Dialogue, in Clustering in a News Corpus (Springer International Publishing, Brno, 2014), pp. 301–307Google Scholar
- Y Lu, I Cohen, XS Zhou, Q Tian, Proceedings of the 15th ACM International Conference on Multimedia, in Feature selection using principal feature analysis (ACM, Augsburg, 2007), pp. 301–304Google Scholar
- LH Bu, GZ Li, XQ Zeng, Reducing error of tumor classification by using dimension reduction with feature selection. Lecture Notes Operations Res 7(232-241), 124 (2007)Google Scholar
- SK Saha, P Mitra, S Sarkar, A comparative study on feature reduction approaches in Hindi and Bengali named entity recognition. Knowl.-Based Syst. 27, 322–332 (2012)View ArticleGoogle Scholar
- Y Bengio, A Courville, P Vincent, Representation learning: a review and new perspectives. IEEE Trans. Pattern Anal. Mach. Intell. 35, 1798–1828 (2013)View ArticleGoogle Scholar
- J Lafferty, A McCallum, F Pereira, Proceedings of the Eighteenth International Conference on Machine Learning, in Conditional random fields: Probabilistic models for segmenting and labeling sequence data (ICML, Williamstown, 2001), pp. 282–289Google Scholar
- R Klinger, CM Friedrich, Proceedings of RANLP, in Feature Subset Selection in Conditional Random Fields for Named Entity Recognition (RANLP, Bulgeria, 2009), pp. 185–191Google Scholar
- C-N Hsu, Y-M Chang, C-J Kuo, Y-S Lin, H-S Huang, I-F Chung, Integrating high dimensional bi-directional parsing models for gene mention tagging. Bioinformatics 24, i286–i294 (2008)View ArticleGoogle Scholar
- Crammer K, Singer Y, Ultraconservative online algorithms for multiclass problems. J. Mach. Learn. Res. 3, 951–991 (2003)Google Scholar
- R McDonald, K Crammer, F Pereira, Proceedings of the 43rd Annual Meeting on Association for Computational Linguistics, in Online large-margin training of dependency parsers (ACL, Michigan, 2005), pp. 91–98Google Scholar
- Z Yang, H Lin, Y Li, Exploiting the contextual cues for bio-entity name recognition in biomedical literature. J. Biomed. Inform. 41, 580–587 (2008)View ArticleGoogle Scholar
- Comeau DC, Dogan RI, Ciccarese P, Cohen KB, Krallinger M, Leitner F, Lu Z, Peng Y, Rinaldi F, Torii M, Valencia A, Verspoor K, Wiegers TC, Wu CH, Wilbur WJ, BioC: a minimalist approach to interoperability for biomedical text processing. Database. (2013) bat064Google Scholar
- M Stevenson, Y Guo, A Al Amri, R Gaizauskas, Proceedings of the Workshop on Current Trends in Biomedical Natural Language Processing, in Disambiguation of biomedical abbreviations (ACL, Colorado, 2009), pp. 71–79Google Scholar
- BT McInnes, T Pedersen, J Carlis, AMIA Annual Symposium Proceedings, in Using UMLS Concept Unique Identifiers (CUIs) for word sense disambiguation in the biomedical domain (AMIA, Chicago, 2007), p. 533Google Scholar
- Nelson, S. J., Powell, T., Srinivasan, S., Humphreys, B. L, Unified Medical Language System (UMLS) Project. In Encyclopedia of library and information sciences pp. 5320–5327(2010).Google Scholar
- Wang, X., Yang, C., Guan, R., A comparative study for biomedical named entity recognition. Int. J. Mach. Learn. Cybern., 1–10 (2015) doi:10.1007/s13042-015-0426-6