Next Article in Journal
Multi-Response Optimization of Ultrasonic Assisted Enzymatic Extraction Followed by Macroporous Resin Purification for Maximal Recovery of Flavonoids and Ginkgolides from Waste Ginkgo biloba Fallen Leaves
Next Article in Special Issue
Small Universal Bacteria and Plasmid Computing Systems
Previous Article in Journal
Profiling of Heterobranchia Sea Slugs from Portuguese Coastal Waters as Producers of Anti-Cancer and Anti-Inflammatory Agents
Previous Article in Special Issue
To Decipher the Mycoplasma hominis Proteins Targeting into the Endoplasmic Reticulum and Their Implications in Prostate Cancer Etiology Using Next-Generation Sequencing Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

ParaBTM: A Parallel Processing Framework for Biomedical Text Mining on Supercomputers

1
School of Computer Science, National University of Defense Technology, Changsha, Hunan 410073, China
2
School of Computer Science and Network Security, Dongguan University of Technology, Dongguan, Guangdong 523808, China
*
Author to whom correspondence should be addressed.
Submission received: 9 April 2018 / Revised: 22 April 2018 / Accepted: 25 April 2018 / Published: 27 April 2018
(This article belongs to the Special Issue Molecular Computing and Bioinformatics)

Abstract

:
A prevailing way of extracting valuable information from biomedical literature is to apply text mining methods on unstructured texts. However, the massive amount of literature that needs to be analyzed poses a big data challenge to the processing efficiency of text mining. In this paper, we address this challenge by introducing parallel processing on a supercomputer. We developed paraBTM, a runnable framework that enables parallel text mining on the Tianhe-2 supercomputer. It employs a low-cost yet effective load balancing strategy to maximize the efficiency of parallel processing. We evaluated the performance of paraBTM on several datasets, utilizing three types of named entity recognition tasks as demonstration. Results show that, in most cases, the processing efficiency can be greatly improved with parallel processing, and the proposed load balancing strategy is simple and effective. In addition, our framework can be readily applied to other tasks of biomedical text mining besides NER.

Graphical Abstract

1. Introduction

With the rapid development of biotechnology, the amount of biomedical literature is growing exponentially. For instance, PubMed (https://0-www-ncbi-nlm-nih-gov.brum.beds.ac.uk/pubmed/), the most recognized biomedical literature database, indexes over 28 million entries for biomedical literature. Most of that information is presented in the form of unstructured texts. It is almost impossible for any domain expert to digest such a massive amount of information within a short period of time. Therefore, automated tools are essential for a systematic understanding of literature. To deal with the literature big data challenge, text mining methods are commonly applied to extract relevant knowledge from vast amounts of literature, and this has become a prominent trend in recent years [1].
Typical tasks of biomedical text mining include named entity recognition and relation extraction. One of the most fundamental tasks of biomedical text mining is named entity recognition (NER). Its task is to recognize target entities that represents key concepts from unstructured biomedical texts, such as proteins, genes, mutations, diseases, etc. There are some existing start-of-art biomedical tools that use text mining methods to identify some specific types of entities, such as mutations [2,3], genes [4,5], and diseases [6,7]. Most of these tools can achieve satisfactory recognition performance (F score over 80%) on standard datasets.
Relation extraction (RE) is a process that typically follows NER and aims to discover semantic connections between entities. Nowadays, there are a number of RE tools using different methods to identify biomedical entity interactions [8,9], such as drug–gene relationships [10,11,12], gene–disease relationships [9,13,14] and protein–protein interaction [15]. Some of them can achieve high F scores (over 80%) on several annotated datasets.
NER and RE are the preliminary steps in mining information from literature. With the uncovered facts, it is possible to construct a complex knowledge graph, which can assist new knowledge discovery and hypotheses generation. In order to achieve this goal, it is necessary to process as many articles as possible. However, text mining procedures are time consuming. BioContext, for instance, an integrated text mining system for large-scale extraction and contextualization of biomolecular events, took nearly 3 months to complete a full run of the system, which analyzed 20 million MEDLINE abstracts and several hundred thousand PMC open access full-texts using 100 concurrent processes [16]. In addition, some text mining tools, like GNormPlus [17] require a substantial amount of memory (≥5 GB), due to the necessity of loading a large gene dictionary and complementary data structures. Consequently, commodity servers cannot fulfil the computation and storage demands of large-scale text mining. Cloud-based solutions in Map-Reduce mode can partially fulfil computational resource demands. However, practically speaking, many text mining components were written in different languages, and they are dependent on a complex collection of third-party libraries, which prevents them from being readily transplanted into a high-level framework, like Hadoop and Spark. In addition, we dived into the details of load balancing, which cannot be readily supported by Map-Reduce.
An alternative solution to address this computational challenge is to harness the power of high performance computers. High performance computers (HPC) like Tianhe-2 [18] represent high-end computing infrastructures that have traditionally been used to solve compute-intensive scientific and engineering problems. The system configuration of Tianhe-2 is listed in Table 1.
Although the software stack on Tianhe-2 is designed and optimized for compute-intensive tasks, its high-performance architecture does provide the capability and capacity of big data processing. Nonetheless, to employ Tianhe-2 for big data processing is not a trivial task, which requires expert knowledge of the system architecture and parallel programming. The programming model is MPI-based (message passing interface) [18], which adds an extra dimension of complexity to normal programming languages like C/C++, Python, Java, etc. Most existing text mining tools are implemented without parallel processing. Therefore, it is necessary to develop an enabling framework that can support parallel text mining without the need to rewrite the original code. In this paper, we develop a parallel processing framework for text mining on the Tianhe-2 supercomputer. The framework integrates text mining tools as plugins. It unifies the input–output stream, implements the parallel processing across multiple compute nodes using the MPI model, and it applies a carefully crafted load balancing strategy to improve the parallelization efficiency. Without a loss of generality, we demonstrate the effectiveness of our framework using multiple NER tools as the demonstration plugins, which can recognize genes, mutations and diseases appearing in biomedical literature. More sophisticated tools of biomedical text mining can be readily integrated into the framework. In the remaining of this paper, we will introduce how paraBTM works and evaluate its performance on Tianhe-2.

2. Results and Discussion

To verify the effectiveness of the parallel framework, we constructed a corpus named 60K, which consists of 60,000 randomly selected articles from PubMed. For NER plugins, we chose three state-of-the-art tools (GNormPlus [17], tmVar [2], DNorm [7]), developed by NCBI (National Centre for Biotechnology Information). We measured the performance in terms of the total processing time and the average processing time (across all processes), and the total time includes the time of initialization and the actual processing time of different plugins.
The 60 K corpus is presented in the NXML format, which is a standard format provided by NCBI. Titles, abstracts, and full-texts from NXML files are extracted and re-written in the PubTator format. All input and output files processed by paraBTM should follow the PubTator format and the PubTator format starts with:
  • <PMID>|t|<Title of the paper>
  • <PMID>|a|<The rest of the paper>
The output file will be appended with annotated information like named entities followed in a tab-separated way.
A basic fact is that the time overhead of text mining is not proportional to the number of input articles. We verified this via a single process run over several groups of randomly selected articles. The result is depicted in Figure 1. Here, different colors represent different processing plugins. Related numbers are also listed in Table 2.
As the number of input files increases, the time cost also increases but not linearly. For example, when the number of input articles is equal to 10, it takes about 36 min for tagging entities, and when the number of articles increases to 100, the spent time is about 3 h (180 min). This can be attributed to another important observation, that is, the total processing time is approximately proportional to the total input size (sum of file lengths as measured by number of characters), which is illustrated in Figure 2 (size unit is MB, mega-bytes) and Table 3. The workload of each plugin can be better estimated by the total length of input files, which is the basis for our load balancing strategy in the following part.
Figure 3 shows the time spent on paraBTM processing with different numbers of parallel processes on an input dataset of 16 MBs (including 175 articles) which is composed of articles randomly selected from the 60 K corpus. Parallel processing greatly reduces the processing time and different load-balancing strategies do affect the parallel efficiency. paraBTM costs about 500 s (under the Short-Board balancing strategy) when 64 processes are employed, which is around 1/16 the processing time of 2 processes. To note, each process needs to carry out initialization for every plugin, which means you cannot reduce the total processing time any further if the initialization time cost becomes the majority part.
To profile different load balancing strategies, we summarize their effects under different parallel scales, as listed in Table 4. In all 6 test cases, the Short-Board strategy is the best in 4 cases and 2nd best in 2 remaining cases. We employ the load balancing efficiency (LBE) to quantify the effects of different strategies. Here, LBE is defined as:
LBE = AET/MET
Here, AET is the average execution time and MET is the maximum execution time. According to the above definition, the maxima of LBE is 1 (achieved if AET is equal to MET) and a greater LBE represents a better load balancing efficiency.
Figure 4 shows that the Short-Board strategy exhibits the best LBE in almost all test cases. However, LBE values drop significantly when the number of parallel processes is greater than 16 in the 16M test set. The reason is that this test set contains only 175 articles, which means each process will only process two articles on average. If the input data set is big enough, the LBE will be maintained at a satisfactory level.
We also conducted an experiment on the whole 60 K corpus (61,078 articles). Table 5 shows that it took over 12 h to process 61,078 papers through three NER plugins (128 nodes under the Short-Board strategy, each node runs 5 processes). According to the results, we can see that parallelization greatly enhances the processing efficiency. To note, the speed-ups of different plugins differ as each plugin has its own characteristic computation and memory access patterns. To carry out a full-scale processing on the whole PMC-OA full-text dataset (over 1 million), it will take about 200 h if we only use 128 nodes. Fortunately, the computation capacity of Tianhe-2 is enormous, and we can reduce the total time down to several hours by harnessing the power of a few thousand nodes (over 16,000 available on Tianhe-2). We plan to carry out a full analysis on the whole PubMed dataset (the real large-scale biomedical texts) in the future. However, the cost of such a full run is currently beyond our funding support. We are currently in the application process of a bigger grant for this large-scale analysis. In our previous study, we have demonstrated that using text mining on a larger dataset does provide more comprehensive and insightful results compared with using a small dataset (say, can be handled by a few people) using thyroid cancer as a case study [19].

3. Materials and Methods

3.1. Data Sources and Storage

The biomedical literature has typical characteristics of large quantity, professional content, public resources, easy-accessibility, etc. Because of these characteristics, biomedical literature data has become one of the most noticeable data in biomedical field. For example, PubMed Central (PMC) is a free digital repository that archives publicly accessible articles. Until now, PMC has contained over 4.1 million references to full-text journal papers, covering a wide range of biomedical fields, and the literature data is stored in NXML format, from which we can extract some parts according to our interest.
However, most of the state-of-art NER tools do not support parallel processing, and it would take an enormous amount of time if we want to process the massive set of biomedical literature. One feasible solution is to harness the computing power provided by HPC systems by implementing a parallel NER processing framework. With this framework, text mining tools can be easily integrated into the framework and developers will not need to consider the details of parallel processing.
There are different levels of parallelism in text mining tasks. First, each input article is relatively independent; secondly, multiple sentences in each of the articles can be approximately regarded as independent. However, in practice, we usually use a single file as a processing unit, the reason is that many text mining tools spend a substantial amount of time to initialize on each processing pass. In addition, the memory size also limits the number of processes that can run in parallel on each computing node. For instance, on Tianhe-2 each node is equipped with 24 cores and 64 GB of memory, and the stable memory that users can control is about 50 GB (the operating system and other necessary tools need to use about 10 GB). The memory costs of a typical TmVar and gnormplus run for NER can be up to 5 GB and 10 GB. Therefore, at most 5 GNormPlus processes and 10 TmVar processes can run on one node. Figure 5 shows the implementation and deployment of a text mining system (paraBTM) in large-scale parallel environment.

3.2. Parallel Processing

3.2.1. MPI-Based Multi-Node Computation

The message passing interface (MPI) is a standard model for parallel programming on HPCs. It is well established over 20 years, and has been implemented in different sorts of programming languages including C/C++ and Python. Our method can run on any supercomputer or cluster configured with MPI support. To note, different supercomputers might have different node configurations. When running on other platforms, the configuration (RAM, number of concurrent processes) might have to change accordingly.
In this work, we use MPI4PY (http://mpi4py.scipy.org/docs/) to implement the parallel processing. MPI4PY is a well-regarded, clean, and efficient implementation of MPI for Python. Our framework can simultaneously submit many jobs to cores distributed across computing nodes in Tianhe-2.

3.2.2. Load Balancing Strategy

A typical challenge in parallel computation is load unbalance, that is, workload is unevenly distributed among nodes, making some nodes very busy for a long time and others idle [20]. In this paper, we address this problem by designing an effective load balancing strategy.
Given a set F of files to be processed F 0 , F 1 , F 2 F N 1 , F n , we initialize processes P 0 , P 1 , P 2 , , P n 1 , P n , and the number of processes is N , the problem is to allocate each file F x to an appropriate process P r a n k .
A naive solution is to randomly distribute target files into nodes. We can simply distribute files to by modulo operation rank = mod ( P , size ) , P and size represent the position of the target file in the file list and number of processes respectively, and file F p finally should be sent to P r a n k . According to the formula, each file F   to be processed is distributed to process r a n k = P   %   s i z e in turn. As the files are arranged in a random order, this process is actually a simulation of random distribution. This is a naïve strategy and easy to implement. However, this strategy does not take into consideration the length of each file, and will very likely cause an unbalanced load distribution, which would detriment the overall parallel efficiency. For instance, if the total length of files assigned to one specific node is far larger than others, then the overall running time will be prolonged until this slowest node finishes. Figure 6 shows an example of the naïve random load balancing strategy.
A slightly more complex load balancing method is the round-robin (RR) method. Round-robin algorithm is a term that originally comes from the field of operating systems. Here, the general idea inspires us to mix small files with large files together into one process. After sorting files by size (see Figure 7), the system will assign files into processes in a snakelike way, making the size of files loaded in every process remains relatively balanced. Figure 8 shows an example of RR algorithm.
The round-robin method also allocates the same number of files, and its serpentine way of load assignment ensures that the total size of the files in each process remains relatively balanced, since files were sorted by size in advance. However, in some circumstances, the lengths of input articles can be very biased, say, some files are extremely long while many others are short. In such cases, the RR method fails.
Instead of assignments based on the number of files, we proposed our “Short-Board” method. Firstly, the files that need to be processed are sorted in descending order according to the length of each file, and then files that need to be processed in the file list are sequentially fetched out and dispatched to the process whose current load is the smallest. Figure 9a–d shows an example of Short-Board algorithm. The pseudo code of Short-Board is shown in Figure 10.

4. Conclusions

In this paper, we present paraBTM, a parallel framework for biomedical text mining developed on the Tianhe-2 supercomputer. It supports different types of components as plugins and its usage is straightforward. The parallel efficiency is guaranteed by a carefully devised load balancing strategy. We evaluated the performance of paraBTM on both small- and large-scale datasets. Experimental results validate that paraBTM effectively improve the processing speed of biomedical named entity recognition. On large scale of datasets, ParaBTM managed to process 60178 PubMed full-text articles in about 12 h. paraBTM is open-source and available at https://github.com/biotm/paraBTM.

Author Contributions

Y.X. and C.W. developed the algorithms and drafted the manuscript. They developed the codes of paraBTM together with X.Y. and W.W.; E.Z. and J.Y. proposed the idea of the work and revised the whole manuscript. All the authors have read and approve the manuscript.

Funding

This research was funded by the National Key R&D Program of China grant number 2018YFB1003203 and National Natural Science Foundation of China grant number 31501073, 61672528 and 61773392.

Acknowledgments

Special thanks to Zhuo Song and Gen Li (from GeneTalks Ltd., Beijing) for providing knowledge support on genomic variations.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Rebholz-Schuhmann, D.; Oellrich, A.; Hoehndorf, R. Text-mining solutions for biomedical research: Enabling integrative biology. Nat. Rev. Genet. 2012, 13, 1–11. [Google Scholar] [CrossRef] [PubMed]
  2. Wei, C.-H.; Harris, B.R.; Kao, H.-Y.; Lu, Z. tmVar: A text mining approach for extracting sequence variants in biomedical literature. Bioinformatics 2013, 29, 1433–1439. [Google Scholar] [CrossRef] [PubMed]
  3. Thomas, P.; Rocktäschel, T.; Hakenberg, J.; Lichtblau, Y.; Leser, U. SETH detects and normalizes genetic variants in text. Bioinformatics 2016, 32, 2883–2885. [Google Scholar] [CrossRef] [PubMed]
  4. Wei, C.-H.; Kao, H.-Y. Cross-species gene normalization by species inference. BMC Bioinform. 2011, 12 (Suppl. 8), S5. [Google Scholar] [CrossRef] [PubMed]
  5. Pan, X.; Shen, H.B. OUGENE: A disease associated over-expressed and under-expressed gene database. Sci. Bull. 2016, 61, 752–754. [Google Scholar] [CrossRef]
  6. Leaman, R.; Lu, Z. TaggerOne: Joint named entity recognition and normalization with semi-Markov Models. Bioinformatics 2016, 32, 2839–2846. [Google Scholar] [CrossRef] [PubMed]
  7. Leaman, R.; Islamaj Dogan, R.; Lu, Z. DNorm: Disease name normalization with pairwise learning to rank. Bioinformatics 2013, 29, 2909–2917. [Google Scholar] [CrossRef] [PubMed]
  8. Quan, C.; Wang, M.; Ren, F. An unsupervised text mining method for relation extraction from biomedical literature. PLoS ONE 2014, 9, e102039. [Google Scholar] [CrossRef] [PubMed]
  9. Xu, D.; Zhang, M.; Xie, Y.; Wang, F.; Chen, M.; Zhu, K.Q.; Wei, J. DTMiner: Identification of potential disease targets through biomedical literature mining. Bioinformatics 2016, 32, 3619–3626. [Google Scholar] [CrossRef] [PubMed]
  10. Xu, R.; Wang, Q. A knowledge-driven conditional approach to extract pharmacogenomics specific drug-gene relationships from free text. J. Biomed. Inform. 2012, 45, 827–834. [Google Scholar] [CrossRef] [PubMed]
  11. Percha, B.; Garten, Y.; Altman, R.B. Discovery and explanation of drug-drug interactions via text mining. Pac. Symp. Biocomput. 2012, 410–421. [Google Scholar] [CrossRef]
  12. Segura-Bedmar, I.; Martínez, P.; de Pablo-Sánchez, C. Using a shallow linguistic kernel for drug-drug interaction extraction. J. Biomed. Inform. 2011, 44, 789–804. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Bravo, À.; Piñero, J.; Queralt-Rosinach, N.; Rautschka, M.; Furlong, L.I. Extraction of relations between genes and diseases from text and large-scale data analysis: Implications for translational research. BMC Bioinform. 2015, 16, 55. [Google Scholar] [CrossRef] [PubMed]
  14. Pletscher-Frankild, S.; Pallejà, A.; Tsafou, K.; Binder, J.X.; Jensen, L.J. DISEASES: Text mining and data integration of disease–gene associations. Methods 2015, 74, 83–89. [Google Scholar] [CrossRef] [PubMed]
  15. Franceschini, A.; Szklarczyk, D.; Frankild, S.; Kuhn, M.; Simonovic, M.; Roth, A.; Lin, J.; Minguez, P.; Bork, P.; von Mering, C.; et al. STRING v9.1: Protein-protein interaction networks, with increased coverage and integration. Nucleic Acids Res. 2013, 41, 808–815. [Google Scholar] [CrossRef] [PubMed]
  16. Gerner, M.; Sarafraz, F.; Nenadic, G.; Bergman, C.M. BioContext: An integrated text mining system for large-scale extraction and contextualisation of biomolecular events. Bioinformatics 2012, 28, 2154–2161. [Google Scholar] [CrossRef] [PubMed]
  17. Wei, C.-H.; Kao, H.-Y.; Lu, Z. GNormPlus: An Integrative Approach for Tagging Genes, Gene Families, and Protein Domains. BioMed Res. Int. 2015, 2015, 918710–918717. [Google Scholar] [CrossRef] [PubMed]
  18. Liao, X.; Xiao, L.; Yang, C.; Lu, Y. MilkyWay-2 supercomputer: System and application. Front. Comput. Sci. 2014, 8, 345–356. [Google Scholar] [CrossRef]
  19. Wu, C.; Schwartz, J.M.; Brabant, G.; Nenadic, G. Molecular profiling of thyroid cancer subtypes using large-scale text mining. BMC Med. Genom. 2014, 7 (Suppl. 3), S3. [Google Scholar] [CrossRef] [PubMed]
  20. Kaur, S.; Kaur, G. A Review of Load Balancing Strategies for Distributed Systems. IJCA 2015, 121, 45–47. [Google Scholar] [CrossRef]
Sample Availability: Samples of the compounds are not available from the authors.
Figure 1. The time cost of processing different numbers of input articles in serial.
Figure 1. The time cost of processing different numbers of input articles in serial.
Molecules 23 01028 g001
Figure 2. The time cost of processing different input sizes in serial.
Figure 2. The time cost of processing different input sizes in serial.
Molecules 23 01028 g002
Figure 3. Effects of different load balancing strategies.
Figure 3. Effects of different load balancing strategies.
Molecules 23 01028 g003
Figure 4. Load balancing efficiencies.
Figure 4. Load balancing efficiencies.
Molecules 23 01028 g004
Figure 5. The implementation and deployment of paraBTM in large-scale parallel environment.
Figure 5. The implementation and deployment of paraBTM in large-scale parallel environment.
Molecules 23 01028 g005
Figure 6. Random load balancing strategy.
Figure 6. Random load balancing strategy.
Molecules 23 01028 g006
Figure 7. A sorted file list.
Figure 7. A sorted file list.
Molecules 23 01028 g007
Figure 8. Round-robin algorithm.
Figure 8. Round-robin algorithm.
Molecules 23 01028 g008
Figure 9. A demonstration of the Short-Board load balancing algorithm.
Figure 9. A demonstration of the Short-Board load balancing algorithm.
Molecules 23 01028 g009
Figure 10. The pseudo code of Short-Board.
Figure 10. The pseudo code of Short-Board.
Molecules 23 01028 g010
Table 1. The system configuration of Tianhe-2.
Table 1. The system configuration of Tianhe-2.
ItemsContent
ManufacturerNUDT
Cores3,120,000
Memory1,024,000 GB
CPUIntel Xeon E5-2692v2 12 C 2.2 GHz
InterconnectTH Express-2
Linpack Performance(Rmax)33,862.7 TFlop/s
Theoretical Peak(Rpeak)54,902.4 TFlop/s
HPCG [TFlop/s]580.109
Operating SystemKylin Linux
MPIMPICH2 with a customized GLEX channel
Table 2. Time distribution of processing different numbers of input articles in serial.
Table 2. Time distribution of processing different numbers of input articles in serial.
Number of PapersTime(s)
DistributionGNormPlustmVarDnormTotal
102.111742.16416.7158.292219.27
204.22764.13444.8665.143278.33
407.764865.14488.0883.195444.17
8017.177864.37744.86177.458803.85
Table 3. Time cost distribution of processing different input sizes in serial.
Table 3. Time cost distribution of processing different input sizes in serial.
Sizes of Papers (M)Time (s)
DistributionGNormPlustmVarDnormTotal
11.931511.35239.7459.511812.53
24.802696.64319.3162.943083.69
416.244944.35633.5786.245680.40
821.778604.43985.63130.339742.16
Table 4. Profiling of different load balancing strategies.
Table 4. Profiling of different load balancing strategies.
(a) Maximum times on different numbers of parallel processes with different strategies.
Number of ProcessesMaximum Time among All Processes (s)
RandomRound-RobinShort-Board
28676.608651.428466.89
44466.815211.964596.09
82792.292706.412386.45
161737.941457.011475.07
32932.85942.10930.25
64609.06579.8553.42
(b) Average times on different numbers of parallel processes with different strategies.
Number of ProcessesAverage Time of All Processes (s)
RandomRound-RobinShort-Board
28513.108374.088392.46
44001.384535.424393.07
82287.192354.142242.42
161265.461174.121270.33
32624.73666.91640.63
64379.08376.06366.84
(c) Load balancing efficiencies on different numbers of parallel processes with different strategies.
Number of ProcessesEfficiency (Average/Maximum)
RandomRound-RobinShort-Board
20.980.970.99
40.900.870.96
80.820.870.94
160.730.810.86
320.670.710.69
640.620.650.66
Table 5. The processing time of 61,078 papers running on 128 processes.
Table 5. The processing time of 61,078 papers running on 128 processes.
Number of ProcessesTime (s)
DistributionGNormPlustmVar2.0DnormTotal
118,934.185,874,482.04654,145.3882,455.36,630,016.9
128364323,73316,21423343,823
Speed-up (x)5.20247.5240.34353.89151.29

Share and Cite

MDPI and ACS Style

Xing, Y.; Wu, C.; Yang, X.; Wang, W.; Zhu, E.; Yin, J. ParaBTM: A Parallel Processing Framework for Biomedical Text Mining on Supercomputers. Molecules 2018, 23, 1028. https://0-doi-org.brum.beds.ac.uk/10.3390/molecules23051028

AMA Style

Xing Y, Wu C, Yang X, Wang W, Zhu E, Yin J. ParaBTM: A Parallel Processing Framework for Biomedical Text Mining on Supercomputers. Molecules. 2018; 23(5):1028. https://0-doi-org.brum.beds.ac.uk/10.3390/molecules23051028

Chicago/Turabian Style

Xing, Yuting, Chengkun Wu, Xi Yang, Wei Wang, En Zhu, and Jianping Yin. 2018. "ParaBTM: A Parallel Processing Framework for Biomedical Text Mining on Supercomputers" Molecules 23, no. 5: 1028. https://0-doi-org.brum.beds.ac.uk/10.3390/molecules23051028

Article Metrics

Back to TopTop