Next Article in Journal
An Intelligent Warning Method for Diagnosing Underwater Structural Damage
Previous Article in Journal
A FEAST Algorithm for the Linear Response Eigenvalue Problem
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Hybrid Genetic-Whale Optimization Model for Ontology Learning from Arabic Text

by
Rania M. Ghoniem
1,2,*,
Nawal Alhelwa
3 and
Khaled Shaalan
4
1
Department of Computer, Mansoura University, Mansoura 35516, Egypt
2
Department of Information Technology, College of Computer and Information Sciences, Princess Nourah Bint Abdulrahman University, Riyadh 84428, Saudi Arabia
3
Department of Arabic, College of Arts, Princess Nourah Bint Abdulrahman University, Riyadh 84428, Saudi Arabia
4
Faculty of Engineering & IT, The British University in Dubai, Dubai 345015, UAE
*
Author to whom correspondence should be addressed.
Submission received: 18 July 2019 / Revised: 8 August 2019 / Accepted: 11 August 2019 / Published: 29 August 2019

Abstract

:
Ontologies are used to model knowledge in several domains of interest, such as the biomedical domain. Conceptualization is the basic task for ontology building. Concepts are identified, and then they are linked through their semantic relationships. Recently, ontologies have constituted a crucial part of modern semantic webs because they can convert a web of documents into a web of things. Although ontology learning generally occupies a large space in computer science, Arabic ontology learning, in particular, is underdeveloped due to the Arabic language’s nature as well as the profundity required in this domain. The previously published research on Arabic ontology learning from text falls into three categories: developing manually hand-crafted rules, using ordinary supervised/unsupervised machine learning algorithms, or a hybrid of these two approaches. The model proposed in this work contributes to Arabic ontology learning in two ways. First, a text mining algorithm is proposed for extracting concepts and their semantic relations from text documents. The algorithm calculates the concept frequency weights using the term frequency weights. Then, it calculates the weights of concept similarity using the information of the ontology structure, involving (1) the concept’s path distance, (2) the concept’s distribution layer, and (3) the mutual parent concept’s distribution layer. Then, feature mapping is performed by assigning the concepts’ similarities to the concept features. Second, a hybrid genetic-whale optimization algorithm was proposed to optimize ontology learning from Arabic text. The operator of the G-WOA is a hybrid operator integrating GA’s mutation, crossover, and selection processes with the WOA’s processes (encircling prey, attacking of bubble-net, and searching for prey) to fulfill the balance between both exploitation and exploration, and to find the solutions that exhibit the highest fitness. For evaluating the performance of the ontology learning approach, extensive comparisons are conducted using different Arabic corpora and bio-inspired optimization algorithms. Furthermore, two publicly available non-Arabic corpora are used to compare the efficiency of the proposed approach with those of other languages. The results reveal that the proposed genetic-whale optimization algorithm outperforms the other compared algorithms across all the Arabic corpora in terms of precision, recall, and F-score measures. Moreover, the proposed approach outperforms the state-of-the-art methods of ontology learning from Arabic and non-Arabic texts in terms of these three measures.

1. Introduction

In recent times, the internet has become people’s principle source of information. A huge quantity of web pages and databases is accessed every day. The instant growth in the quantity of information accessed via the Internet has caused difficulty and frustration for those trying to find a particular piece of information. Likewise, the various kinds of information resources that exist on the Internet constitute an enormous quantity of information in the form of web pages, e-libraries, blogs, e-mails, e-documents, and news articles, all containing huge amounts of data [1]. Such information is unstructured or semi-structured, which means that the knowledge discovery process is challenging. To deal with this challenge, the semantic web was invented as an extension of the ordinary web [2].
Ontology is a method for extending web syntactic interoperability to semantic interoperability. Ontologies are exploited to represent huge data in such a way that allows machines to interpret its meaning, allowing it to be reused and shared [3]. They are formal and explicit specifications of concepts and relations [4] and play a crucial role in improving natural language processing (NLP) task performance, such as information extraction and information retrieval. Ontologies are usually restricted to a particular domain of interest. The preliminary identification of ontology is expressed as Characterization of Conceptualization. The ontology learning from texts is “The acquisition of a domain model from textual corpus” [5].
Building ontologies can be accomplished manually, automatically, or in a semi-automatic way. However, the manual building of ontologies has the drawbacks of being time-consuming, expensive, and error-prone [6]. Furthermore, it demands the cooperation of ontology engineers and domain experts. In order to avoid these shortcomings, ontology learning has evolved to automate or semi-automate the construction of ontologies. Ontology learning includes knowledge extraction through two principle tasks: concepts extraction (which constitute the ontology) and extracting the semantic relations that link them [7,8,9].
Despite the Arabic language’s importance as the sixth most spoken language in the world [2] and the tremendous growth of Arabic content via the web in recent years, it has been given little attention in the ontology learning field [10,11,12]. Several contributions are available on domain ontologies in English [13,14,15] and other languages. However, Arabic is not commonly considered by specialists in this field. Furthermore, the automatic extraction of semantic relationships from Arabic corpora has not been extensively investigated in comparison to other languages such as English. The majority of attempts to construct Arabic ontology is still implemented manually [2,16]. Manually developing conceptual ontologies is not only a time-consuming but also a labor-intensive job. Furthermore, extra challenges are encountered when extracting knowledge from Arabic texts due to the nature of the Arabic language, the words’ semantic vagueness, and the lack of tools and resources which support Arabic. Consequently, the Arabic language suffers from a lack of ontologies and applications in the semantic web [17,18].
In summary, only a few studies have considered automatic ontology learning from Arabic text [4,9,12,19,20,21,22,23]. These works fall into one of the following three categories: handcrafted rule-based methods [12,20,21], machine learning methods [9,19,20,22], and hybrid rule-based/machine learning methods [4,23]. The studies that have introduced rule-based approaches for ontology learning are based on extracting the semantic relationships between Arabic concepts or Arabic named entities, and utilize the same technique, which can identify linguistic patterns from a given corpus. These patterns are then converted to rules and transducers. The drawbacks of the rule-based methods include being time-consuming and having the requirement to fully cover all rules which may represent any kind of relationship. The works that have proposed machine-learning approaches for ontology learning are based on conventional classification algorithms that categorize Arabic relations into corresponding types, but do not provide any solutions to overcome the drawbacks of these classification algorithms, such as their low performance when analyzing large textual datasets and high-dimension data. Some works have attempted to overcome the shortcomings of the two previous methods by integrating inference rules and machine learning algorithms into hybrid approaches. Although this hybridization has somewhat optimized the overall performance, more advanced hybrid approaches to optimize Arabic ontology learning are still required.
By comparison to the other languages, several studies have been conducted for learning ontology from English text [24,25,26], which has achieved the largest number of contributions among the other languages. Some of these studies presented rule-based approaches [24], and the others proposed machine-learning-based approaches [25,26]. In [24], the authors presented a rule-based approach for learning the English ontology in which the inductive logic programming was used to obtain ontology mapping. This method described the ontology in OWL format and then interpreted it into first-order logic. Thereafter, it generated generalized logical rules depending on background knowledge, just as mappings do. In [25], an exemplar-based algorithm was introduced to link the text to semantically similar classes in an ontology built for the domain of chronic pain medicine. In [26], a machine learning approach based upon a neural network was presented to learn ontology through the encoder–decoder configuration. Accordingly, the natural language definitions were translated into Description Logics formulae through syntactic transformation. These methods of building ontologies are domain-specific. Therefore, they are not applicable with the Arabic language and do not support the Arabic texts.
Recently, the hybrid approaches of different bio-inspired optimization algorithms [27,28,29] demonstrated competitive performances in different applications of computer science, where two or more algorithms of the following are used as hybrid to optimize the problem in the domain of interest: the genetic algorithm (GA) [30,31,32], social spider optimization [33,34], ant colony optimization (ACO) [35,36], and whale optimization algorithm (WOA) [37,38]. These methods have several merits, including having a small parameter set, simple frameworks, and capability to avoid the shortcoming of being trapped in the local optima. Thus, they are suitable for several real applications and have the robustness to solve many problems of global optimization without the need to change the original algorithm structure.
In between these algorithms, the WOA was introduced in [39] for solving the global optimization problem through emulating the humpback whales behavior. These humpback whales are well known of a hunting method, namely, bubble-net feeding [39]. This behavior operates in three phases, including coral loop, lobtail, and capture loop [39]. The extra information on this behavior can be found in [40]. In comparison to the other bio-inspired optimization algorithms, such as Particle Swarm Optimization (PSO), the WOA algorithm has a good exploration capability of the search space [37]. However, it suffers from poor exploitation and the probability to be trapped into local optima.
In addition, GA is another heuristic algorithm for combinatorial optimization [31]. In comparison to the other similar algorithms like Tabu Search (TS) [41,42] and simulated annealing (SA) [43], we can find that all of them are applied for several combinatorial optimization problems. Furthermore, they also have different properties. First, a great computational cost is required by GA to find the optimal solution. Secondly, the best solution quality provided by the GA is superior to the SA and is comparable to the TS. Moreover, the domain-specific knowledge can be incorporated by the GA in all combinatorial or optimization phases to dictate the strategy of search, in contrary to TS and SA, which lack this feature. Therefore, based on the proven superiority of the GA and WOA in many applications [30,31,32,37,38] and to overcome the drawbacks of the ordinary WOA, this work further demonstrates the robustness of the proposed hybrid genetic-whale optimization algorithm (G-WOA) to optimize ontology learning from Arabic texts, in which the GA algorithm is used to optimize the exploitation capability of the ordinary WOA algorithm and solve its premature convergence issue by combining the genetic operations of GA into the WOA.
This paper contributes to the state-of-the-art of Arabic ontology learning through the following:
  • Firstly, a text mining algorithm is proposed particularly for extracting the concepts and their semantic relations from the Arabic documents. The extracted set of concepts with the semantic relations constitutes the structure of the ontology. In this regard, the algorithm operates on the Arabic documents by calculating the concept frequency weights depending on the term frequency weights. Thereafter, it calculates the weights of concept similarity using the information-driven from the ontology structure involving the concept’s path distance, the concept’s distribution layer, and the mutual parent concept’s distribution layer. Eventually, it performs the mapping of features by assigning the concept similarity to the concept features. Unlike the ordinary text mining algorithms [9,10], this property is crucial because merging the concept frequency weights with the concept similarity weights supports the detection of Arabic semantic information and optimizes the ontology learning.
  • Secondly, this is the first study to propose bio-inspired algorithms for optimization of Arabic ontology learning, in which a hybrid G-WOA algorithm is proposed in this context, to optimize the Arabic ontology learning from the raw text, by optimizing the exploration-exploitation trade-off. It can benefit from a priori knowledge (initial concept set obtained using the text mining algorithm) to create innovative solutions for the best concept/relation set that can constitute the ontology.
  • Thirdly, investigating the comparable performance between the proposed G-WOA and five other bio-inspired optimization algorithms [32,39,44,45,46], when learning ontology from Arabic text, where its solutions are also compared to those obtained by the other algorithms, across different Arabic corpora. To the best of our knowledge, the proposed and compared bio-inspired algorithms have not been investigated in Arabic or non-Arabic ontology learning yet.
  • Fourthly, the proposed ontology learning approach is applicable with the other languages, where it can be applied to extract the optimal ontology structure from the non-Arabic texts.

2. Literature Review

Due to the rapid surge of textual data in recent years, several studies have concentrated on how to create taxonomy from labeled data [47,48,49,50]. In this context, there were many attempts to deal with multi-label learning/classification problems. In [47], the authors concentrated on how to learn classifiers the balanced label through label representation, using a proposed algorithm, namely, Parabel. This algorithm could learn the balanced and deep trees. The trees learned using this algorithm were prone to prediction performance degradation because of forceful aggregation for labels of head and tail into longer decision paths and generic partitions. In [48], the authors introduced a shallow tree algorithm, namely Bonsai, which can deal with the label space diversity and scales to a large number of labels. The Bonsai algorithm was able to treat with diversity in the process of partitioning by allowing a larger fan-out at every node.
In [49,50], the authors used the hierarchical and flat classification strategies with the large-scale taxonomies, relying on error generalization bounds for the multiclass hierarchical classifiers. The main goal of some of these works was the large-scale classification of data into a large number of classes, while the others concentrated on how to learn the classifier the given trees. In contrary to these works, the main goal of this paper was to introduce an approach for extracting the optimal structure that constitutes the ontology from the raw textual data by employing the text mining and bio-inspired optimization techniques.

2.1. Literature Review on Arabic Text Mining

Although several works have been devoted to text mining from English and Latin languages [51,52], little attention has been paid to mining the Arabic texts. This is mainly because of the Arabic structural complexity and the presence of several Arabic dialects. Table 1 presents state-of-the-art information on Arabic text mining [53,54,55,56,57,58,59]. The majority of works in this context have concentrated on using the Vector Space Model [57], Latent Semantic Indexing [56], and Term Frequency (TF)/Inverse Document Frequency (TF/IDF) [54,55]. However, these algorithms still suffer from two shortcomings: the dimension curse and the semantic information lack. Therefore, in this study, we proposed a specific text mining algorithm that begins with the conceptualization stage to extract the initial concept set constituting the ontology and captures their semantic information.

2.2. Literature Review on Arabic Ontology Learning

Ontology learning from text is a very important area in computer science. Published works on ontology learning from Arabic texts are still rare. As previously mentioned, the contributions of the state-of-the-art Arabic ontology learning from texts can be distinguished into one of the following categories. The works under these categories were examined in the following section and using Table 2.
The rule-based approaches [12,20,21,60,61,62] rely on patterns comprising all the possibly-correlated linguistic sequences commonly executed in a form of finite-state transducers or even regular expressions. Despite those methods being beneficial for a limited domain, besides their better analysis quality, they cannot act in a good way, in particular, the creation of the manually hand-crafted patterns is so laborious with regard to effort and time. Hence, through the applications of such approaches, it is difficult to manipulate enormous amounts of data.
For automating the relations extraction, some studies [9,19,22,63,64] have used machine learning algorithms involving (1) unsupervised, (2) semi-supervised, and (3) supervised learning. For the unsupervised methods, the popular approach takes clusters from the patterns of the same relationship and then generalizes them. However, the semantic representations of relational patterns, in addition to the scalability to big data, make these methods face a challenge in reference to the reliability of the obtained patterns [55]. Although these algorithms can manipulate large quantities of data, the conversion of the output relations to ontologies represents a labor-intensive task.
To encounter the drawbacks of the unsupervised approaches, the studies investigated the semi-supervised methods or bootstrapping techniques that need seeding-points sets rather than training sets. The seeds are linguistic patterns or even relation-instances which are applied in an iterative way for acquisition of more basic elements until all objective relations are found. The shortcoming of the bootstrapping approaches deeply relies on the chosen initial seeds, which might reflect precisely the information of the corpus. On the other side, the extraction caliber is low. The supervised techniques [63] are the last category under the machine learning-based approaches, which depends on a completely labeled-corpus. Thus, extracting the relations is regarded as a matter of classification, according to the supervised techniques. Amongst them, we mention conditional random fields, support vector machine (SVM) [64], decision tree [19], in addition to Maximum Entropy (MaxEnt). These algorithms give a low performance in case of the high-dimensional corpora.
On the other side, the researchers have successfully addressed some of the previously discussed challenges such as the long sentences of Arabic and the non-fixed location of semantic relations in sentences. Therefore, they have integrated the rule-based method with machine learning to get hybrid approaches [4,23,65]. These hybrid methods have demonstrated enhanced performance in comparison to the single rule-based or the machine learning-based approaches. Generally, recent literature demonstrates a huge interest in the hybrid artificial intelligence-based models to solve problems in several domains. In [27], a hybrid algorithm integrates the merits of GA, including the great global converging ratio together with ACO to introduce solutions for the supplier selection problems. In [28], a genetic-ant colony optimization model was proposed to overcome the word sense disambiguation that represents a serious natural language processing problem. Therefore, it is important to propose hybrid intelligent approaches to introduce numerous choices for unorthodox handling of Arabic ontology learning problem, which comprise vagueness, uncertainty, and high dimensionality of data.
In this context, these hybrid bio-inspired optimization algorithms can present innovative solutions to support the Arabic language. They can overcome the key shortcoming of existing methods for Arabic ontology learning as they can deal with the high-dimensional or sparse data that makes it hard to capture the relevant information, which helps to learn ontology via dimensionality reduction, depending on selecting only the optimal concepts and semantic relations that contribute to the ontology structure and ignoring the non-related ones. Therefore, this paper contributes to the state-of-the-art on Arabic ontology learning with a hybrid model based on GA and WOA. This model was experimented to ontology learning using a number of the publicly available Arabic and non-Arabic corpora.

3. Preliminaries

3.1. Genetic Algorithm

The GAs [30,31,32] are random-search algorithms that are inspired by natural genetic mechanism and biological natural selection, which belong to the computational intelligence algorithms. The GA emulates the reproduction, crossover, and mutation in the process of genetic mechanism and natural selection. In the GAs, the individual is the optimized solution of the problem, namely the chromosome or genetic string. The GA can be expressed as an eight tuple: G A = { C ,   F i t n e s s ,   P ,   P o p S i z e ,   L ,   α ,   β ,   S } , where C is the encoding method for the individuals within population, F i t n e s s is a fitness function for evaluating individuals, P is the initial solution, P o p S i z e is the population size, L , α and β indicate the operators of selection, crossover and mutation, respectively, and S defines the GA termination condition. A GA begins with the initial population of chromosomes or strings and then produces successive populations of chromosomes. The basic GA comprises the following three operations:
  • Reproduction. The reproduction means keeping chromosomes without changes and transferring them to the next generation. Inputs and outputs of this procedure are the same chromosomes.
  • Crossover. This process concatenates two chromosomes to produce a new two ones through switching genes. On this basis, the input for this step is two chromosomes, whereas the output is two different ones.
  • Mutation. This process reverses randomly one gene value of a chromosome. Thus, the input chromosome is completely different from the output one.
When determining not to conduct crossover, the chromosomes of parents are duplicated to the off-spring without change. Evolution speed of genetic search is altered by varying the probability of crossover. Practically, the crossover value is close to 1. Contrarily, the mutation ratio is usually fairly small.

3.2. Whale Optimization Algorithm

The WOA was proposed in [39]. It is inspired by the humpback whales’ behavior. In comparison to the other bio-inspired algorithms, the WOA improves the candidate solutions in each step of optimization. In this context, the emulation of bubble-nets was implemented using a spiral movement. This procedure imitates the helix-shaped movement of the actual humpback whales.

3.2.1. Encircling Prey

Assume that a whale c ( i ) has a position which is updated through moving it simultaneously in a spiral around its prey c b e s t . Mathematically, this procedure is expressed as follows:
c ( i   +   1 )   =     S .   e h r .   c o s ( 2   π   r )     c b e s t ( i )
where S   =   |   c b e s t     c ( i )   | refers to the distance between c ( i ) and c b e s t at iteration i , r         [ 1 ,   1 ] represents a random number, and h is a constant variable defining a logarithmic spiral shape. The positions of the wales are updated by the encircling behavior based upon c b e s t ( i ) as follows:
S   =   | K .   c b e s t     c ( i )   |  
c ( i   +   1 )   =     c b e s t ( i )           A .   S
K and A represent coefficient vectors and are defined using
K     =     2 m
A     =     2 o   m         o
where m denotes a random vector and e is decreased linearly from 2 till 0 along iterations i , then the value of o is computed using
o     =     o       i   o o m a x

3.2.2. Bubble-Net Attacking Method

For the bubble-net attacking, the whales are able to swim simultaneously around the prey over a spiral-shaped path and throughout a shrinking circle. Equation (7) defines this behavior:
c ( i     +     1 )   =   { c b e s t         A .   S i f     m     <     0 . 5 S .   e p r .   c o s ( 2   π   r       +     c b e s t ( i )   i f     m     >     0 . 5  
where m         [ 0 ,   1 ] refers to the probability of choosing the mechanism of swimming on all the prey’s sides (weather spiral model-based swimming or shrinking encircling-based swimming). Nevertheless, humpback whales search for prey in a random manner.

3.2.3. Searching for Prey

In reality, humpback whales swim randomly so that they search for prey. The positions of the whales are updated using a randomly chosen whale c r a n d ( i ) as given below:
S     =     |   K .   c r a n d ( i )         c ( i )   |
c ( i   +   1 )   =   c r a n d ( i )         A .   S .
Eventually, based upon the value of e (decreases from 2 till 0), K , A and the probability m , the position of every i t h whale is updated. If m   >     0 . 5 , then go to Equation (1). Otherwise, go to either Equations (2) and (3) or Equations (8) and (9) depending on the value of | K | . This procedure is repeated until the stopping condition.

3.3. Arabic Ontology Learning

Ontology learning is one of the most important issues in Arabic language processing. In the literature, to construct the ontology of any conceptual domain, this is based on three dominant linguistic theories:

3.3.1. The Semantic Field Linguistic Theory

The semantic field linguistic theory [17], in which the word meaning is deemed within a specific perspective of the world, was presented by Jost Trier [5]. Accordingly, it is determined by its relationship to the words within the field/domain (conceptual area). It presumes that each word is constructed inside semantic fields based upon a primitive feature set. Moreover, the position of the word within the field determines its meaning, and the relations it creates with the remaining words in this field. Utilizing componential analysis, what is meant by a word is established in reference to some specified atomic components or decompositions representing the features that distinguish a considered word. Such features form the base for structuring a particular semantic domain. The individual word meaning can be identified as an integration of the representative features. Such formulae are indicated as componential definitions for the semantic units and denoting formalized dictionary definitions.

3.3.2. The SEMANTIC analysis Linguistic Theory

This is a strategy to extract and represent the meaning of word contextual usage by applying statistical methods to the textual corpus [66]. The main idea is to aggregate words into contexts within which a specified word is or does not belong. This depends on a set of constraints that decides the similarities of word meanings and sets words to each other.

3.3.3. The Semantic Relations Theory

Underlying semantic relations for Arabic text show a great deal of variety [67]. The three semantic relationships considered in the current work can be explained with the following examples of biomedical concepts from our corpus:
  • Synonymy. This relationship type aims concepts that hold nearly similar meanings. For instance, the concepts شهيق inspiration and استنشاق inhalation are synonyms.
  • Antonyms. This relationship aims concepts that demonstrate opposite meanings, i.e., antonyms, like خبيث malignant, and حميد benign.
  • Inclusion. This type of relation means that one entity-type comprises sub entity-types. For example, the concept صمام رئوي pulmonary valve with the concept قلب heart, can indicate a part-to-whole or Is-a relationship. Figure 1 presents an example of some biomedical knowledge concepts available in our corpus which are linked with an Is-a relationship.

4. Proposed Model for Arabic Ontology Learning

This section introduces the proposed model for ontology learning from Arabic text. The proposed model integrates: (1) a proposed text mining algorithm for extracting the concepts and the semantic relations which they are linked with, from the text documents, and (2) a proposed hybrid genetic-whale optimization algorithm to select the optimal concept/relationship set that constitute the Arabic ontology.

4.1. Pre-Processing

Pre-processing of Arabic texts in the three datasets investigated in this study is performed in two steps:
  • Eliminating stop-words. Words like pronouns and conjunctions are extremely common and if we remove these words from text we can focus on important concepts. Examples of stop words are: ‘في’ → ‘in’, ‘هذا’ → this, ‘بين’ → ‘between’, ‘مع’ → ‘with’, ‘إلى’ → ‘to’, ‘أو’ → ‘or’, ‘و’ → ‘and’, etc.
  • Stemming. This task leaves out the primitive form of a word. Thus, words or terms that share identical root but differ in their surface-forms due to their affixes can be determined. Such a procedure encompasses eliminating two things: a prefix, like ‘الـ’, at the start of words and as suffix such as ‘ية’at the end of words. An instance of eliminating a prefix and a suffix is the input word ‘’السرطانية ‘cancerous’ which is stemmed to ‘سرطان’ ‘cancer’.

4.2. Proposed Text Mining Algorithm

The algorithm extracts concepts and their semantic relations that constitute the ontology from each document of Arabic text, in three steps: Term weighting, concept similarity weights, and feature mapping.

4.2.1. Term Weighting

The weight in text mining is a well-known statistical measure for evaluating how important a term (word) is for a textual document in a corpus. Thus, we assigned a weight to each term of a document. This procedure is called term weighting. Thereby, every document is expressed in a vector form relying on the terms encompassed inside. Formally speaking, the vector that characterizes the document will be in the following format:
d o c n   =     { T W 1 ,   T W 2 ,   . . . . . .   T W a ,   . . . . .   T W | C | }
where T W a refers to the weighting of the term that has the number m in the d o c document of index n , C represents the term set, and | C | denotes the cardinality of C .
To obtain a vector involves the terms of C , the TF-IDF is utilized as weighting. Assume that the term frequency T F a expresses the occurrences number of T a within the document, and the document frequency D F a is the document number in which the given term T a can be seen at least once. Thus, we can compute the inverse document frequency I D F a , as illustrated in Equation (11) using D F a [68]:
I D F a   =   log   ( | D O C | D F a )
where | D O C | denotes the number of documents assigned as a training set, and T W a is computed by Equation (12):
T W a   =   T F a     I D F a .
Subsequently, the irrelevant and redundant features are eliminated from the text document, thus, we can represent the document set as a “document-term” matrix as follows:
[ T 1 T 2 . . T a ]   =   [ T W ( 1 , 1 ) T W ( 1 , 2 ) .   .   .   .   .   .   . T W ( 1 , a ) T W ( 2 , 1 ) T W ( 2 , 2 ) .   .   .   .   .   .   . T W ( 2 , a ) . . .   .   .   .   .   .   . . . . .   .   .   .   .   .   . . T W ( a , 1 ) T W ( a , 2 ) .   .   .   .   .   .   . T W ( a , A ) ] .
Depending on the resulting weights for feature frequency, the algorithm maps the document’s terms to corresponding concepts. As illustrated in Algorithm 1, T W and C W are two matrices to the same document, and S T and C T indicate the sets of terms and concepts, respectively. The algorithm reveals that through mapping, the document’s terms to correlative concepts, the document’s vector of terms will be converted into a vector of concepts. Thus, the algorithm will replace the document set of Equation (13) by the document-concept matrix in Equation (14):
[ T 1 T 2 . . T a ]   =   [ C W ( 1 , 1 ) C W ( 1 , 2 ) .   .   .   .   .   .   . C W ( 1 , a ) C W ( 2 , 1 ) C W ( 2 , 2 ) .   .   .   .   .   .   . C W ( 2 , a ) . . .   .   .   .   .   .   . . . . .   .   .   .   .   .   . . C W ( a , 1 ) C W ( a , 2 ) .   .   .   .   .   .   . C W ( a , A ) ]
where C W ( l ,   m ) denotes the frequency weight for “concept 1” in document m , a represents the documents number, and A is the concepts number.
Algorithm 1: The Proposed Arabic Text Mining Algorithm
Input: A term weighting matrix T W of training set corresponding to term set S T   =   { T 1 ,   T 2 ,   .   .   .   . ,   T a }
Output: Matrix of mapped features W S obtained by assigning concept similarity weights to the concepts
  • //Mapping terms to concepts
  • //The matrix C W is initially another copy of T W
  • Update a matrix C W with the resulting concept weighting set corresponding to concept set C T   =   { C 1 ,   C 2 ,   ,     C a } , as follows:
  • While ( S T     Φ )
  •   While ( C T     Φ )
  •     For A   =   1 to c o u n t   ( S T )
  •       For B   =   1 to c o u n t   ( C T )
  •        IF M a t c h i n g       ( T A ,   C B )   =   1   //The two elements are equal
  •          C B     m a p p i n g       ( T A )
  •          C W ( A , B )   =   C W ( A ,   B )   +   T W ( A ,   B )
  •          R e m o v e     T A   f r o m     S T
  •        ELSE
  •          R e m o v e     T A   f r o m   S T
  •        END IF
  •       END FOR
  •     END FOR
  •   END
  •  END
  • //Calculation of semantic similarities between n concepts of C W
  • //Matrix of resulting weights of concept similarity
  • [ ] S
  • For A   =   1 to c o u n t   ( S T )
  •     For B   =   1 to c o u n t   ( C T )
  •       //Computation of semantic similarities between each two concepts in C W
  • W P D   =   λ M m i n ( l a y e r   ( C W ( A ,   A ) ) ,   l a y e r   ( C W ( A ,   B ) ) // M   is   top   layer   number
  •  
  • S i m i l a r i t y ( C W ( A ,   A ) ,   C W ( A ,   B ) )   =   l a y e r ( N e a r e s t ( C W ( A ,   A ) ,   C W ( A ,   B ) ) ) W P D ( C W ( A ,   A ) ,   C W ( A ,   B ) ) M
  •       Append the similarity between C W ( A ,   A ) and C W ( A ,   B ) to S as:
  •        W S ( A ,   B )   =   S i m i l a r i t y ( C W ( A ,   A ) ,   C W ( A ,   B ) )
  •   End For
  • End For
  • Assign resulting concept similarity weights to the concepts according to Equation (20).

4.2.2. Concept Similarity Weights

In this study, experts in the domain of Arabic language implemented the conceptual characterization of the Arabic ontology. The concepts and semantic relations of the ontology hierarchy were then built using the Protégé tool [69]. Considering the concept hierarchy structure of the biomedical information depicted in Figure 1, the concept similarities can be computed based on the distances between nodes. In this regard, computing distances between nodes has been introduced in several studies through different methods depending on the domain of application [70]. In the current study, computing similarities among concepts that constitute the ontology structure encompasses three elements: (1) the path distance between concepts, (2) the concept’s distribution layer, and (3) the mutual parent concept’s distribution layer. For each concept node within the ontology, we can trace and obtain all its paths to the root concept node, then generate the routing table of the ontology.
Therefore, the concepts weighted path distance (WPD) is calculated by considering the following factors:
If the path distance ( P D ) that the concepts have is long, they will have less similarity, as in the following example, where C is the concept node of index i in the ontology structure.
IF     P D ( C 4 ,   C 16 ) < P D ( C 1 ,   C 16 )     THEN                           S i m i l a r i t y ( C 4 ,   C 16 )   >   S i m i l a r i t y ( C 1 ,   C 16 ) END   IF
Neglecting the path distance factor, the deeper the neighboring concepts localize at the distribution layer-level, the higher the similarity they have, as
S i m i l a r i t y ( C 16 ,   C 4 )   >   S i m i l a r i t y ( C 4 ,   C 1 ) >   S i m i l a r i t y ( C 1 ,   C 5 ) .
For concepts that have a mutual parent, the deeper they localize at the distribution layer, the higher the similarity they have, as an instance:
S i m i l a r i t y ( C 16 ,   C 15 )   >   S i m i l a r i t y ( C 4 ,   C 2 ) .
Assuming two adjacent concepts q A and q B , we can compute the WPD of the concepts using Equation (15) of Algorithm 1, where l a y e r   ( q A ) , and l a y e r   ( q B ) denote the distribution layer number for concepts q A , and q B , respectively. M represents the number of upper layer in the entire ontology hierarchy besides λ which is a scalar that is set through experimentation. For our work, it was assigned a value of 1.
Eventually, we estimated the similarity between the given concepts q A and q B using Equation (16) (see Algorithm 1) where l a y e r ( N e a r e s t ( q A ,   q B ) ) indicates the distribution layer number of the closest common concept of concepts q A and q B . After computing the concept similarities for all the concepts in the document’s ontology hierarchy, we can construct a matrix of “concept-concept” as show below:
W S   =   [ S i m i l a r i t y ( q 1 ,   q 1 ) S i m i l a r i t y ( q 1 ,   q 2 ) .   .   .   .   .   .   . S i m i l a r i t y ( q 1 ,   q n ) S i m i l a r i t y ( q 2 ,   q 1 ) S i m i l a r i t y ( q 2 ,   q 2 ) .   .   .   .   .   .   . S i m i l a r i t y ( q 2 ,   q n ) . . .   .   .   .   .   .   . . . . .   .   .   .   .   .   . . S i m i l a r i t y ( q n ,   q 1 ) S i m i l a r i t y ( q n ,   q 2 ) .   .   .   .   .   .   . S i m i l a r i t y ( q n ,   q N ) ]
  =   [ 1 S i m i l a r i t y ( q 1 ,   q 2 ) .   .   .   .   .   .   . S i m i l a r i t y ( q 1 ,   q n ) S i m i l a r i t y ( q 2 ,   q 1 ) . .   .   .   .   .   .   . S i m i l a r i t y ( q 2 ,   q n ) . 1 .   .   .   .   .   .   . . . . .   .   .   .   .   .   . . S i m i l a r i t y ( q n ,   q 1 ) S i m i l a r i t y ( q n ,   q 2 ) .   .   .   .   .   .   . 1 ]

4.2.3. Feature Mapping (Assigning Similarity Weights to the Mapped Concepts)

As for the “concept-concept” matrix, the values within W S ought to be either above or equivalent to “0”, where “0” denotes to non-similar concepts whereas “1” denotes to similar concepts, and W S is delineated as asymmetric-positive semi-definite. Hence, we can express W S as in Equation (19):
W S   =   E * D G   * E 1   =   E * D G   *   D G   *   E 1 .
where
  • D G a diagonal matrix whose elements denote the non-negative eigenvalues of W S ,
  • E an orthogonal matrix whose columns point to the corresponding eigenvectors,
  • D G a diagonal matrix whose diagonal items are the square root for D G diagonal elements.
Eventually, the document set that is expressed as in Equation (14) will be rewritten as
[ T ^ 1 T ^ 2 . . T ^ n ]   =   [ T 1   *   E D G T 2   *   E D G . . T n   *   E D G ]   =   [ T 1 T 2 . . T n ] E D G   =   [ C W ( 1 , 1 ) C W ( 1 , 2 ) . . . . . . . . C W ( 1 , a ) C W ( 2 , 1 ) C W ( 2 , 2 ) . . . . . . . . C W ( 2 , a ) . . . . . . . . . . . . . . . . . . . . . . C W ( a , 1 ) C W ( a , 2 ) . . . . . . . . C W ( a , A ) ] E D G
where
  • T ^ n   =   T n   *   E D G   ,
  • C W ( 1 ,   1 ) the frequency weight of “concept l ” in document m ,
  • a the documents number,
  • A the concepts number.

4.3. The Proposed Hybrid Genetic-Whale Optimization Algorithm for Arabic Ontology Learning

In the ordinary WOA, the exploitation phase relies on computing the distance between the whale (search agent) and the best one known in this iteration. To optimize the exploitation capability of WOA and solve the premature convergence issue of the WOA, in this study, the genetic operations of GA were combined into WOA. The core of the proposed algorithm, G-WOA, is the hybridization of the WOA’s operators along with GA’s operators [71] to optimize the ontology learning from Arabic text by optimizing the WOA’s exploration-exploitation trade-off. The operator of G-WOA is mainly a hybrid operator (as shown in lines 7 to 29 of Algorithm 2), which integrates GA’s mutation, crossover, selection, and the WOA’s components, called, encircling prey, bubble-net attacking, and searching for prey.
Algorithm 2: The proposed hybrid G-WOA Algorithm for ontology learning from Arabic text
Input: A vector R assigns the document’s mapped features.
//The G-WOA algorithm parameters:
P o p S i z e population size, C r   crossover rate, M R   mutation rate, E     The stopping criterion, h   constant defines the logarithmic spiral shape, r   random variable, where r         [ 1 ,   1 ] , K coefficient vector of WOA, and e is linearly decreased from 2 to 0 along iterations ( t ).
//Fitness function parameters
w f     weight of false alarm rate, w d     weight of detection rate, and w c     selected features weight.
Output: R * the solution with the optimal concept/semantic relation set contributing to the ontology.
  • Represent each document d 1 ,   d 2 ,   ,   d O by a single whale c to obtain a pool O of whales. C   =   { c 1 ,   c 2 ,   ,   c O } .
  • Evaluate the fitness for each whale c i     C using Equation (21).
  • Get the best individual c b e s t and set it as c G 0 .
  • Initialize the counter of iteration, t     = 1
  • While (stopping criterion E is not met)
  • { C ˜     φ
  •   For each p     1 to P o p S i z e
  •    Choose a random integer u r a n d from { 1 ,   2 ,     U } .
  •    Randomly select two whales c r a n d 1 ,   j t , c r a n d 2 ,   j t     C   ( c r a n d 1     c r a n d 2 ) :
  •    Update K , A , e , h   , and r .
  •    For each gene j in the solution c p ,   j
  •    IF r a n d   ( 0 ,   1 )         C r or j   =   =   u r a n d Then
  •      o f f s p r i n g p , j t   =   c r a n d 1 ,   j t   +   M p ,   j ( c r a n d 2 ,   j t   c r a n d 1 ,   j t   ) .
  •    ELSE
  •     IF m   <     0 . 5 THEN
  •      IF | K |       1 THEN
  •        Choose a random individual c r a n d t       c p t .
  •         S t p ,   j     =     |   K p t .   c r a n d ,   j t     c p ,   j t   | .
  •         o f f s p r i n g p ,   j t   =   c G ,   j t     A p t .   S p ,   j t .
  •      ELSE
  •         S t p ,   j     =     |   K p t .   c G ,   j t     c p ,   j t   |
  •         o f f s p r i n g p ,   j t   =   c G ,   j t     A p t .   S p ,   j t
  •      End IF
  •     ELSE
  •       o f f s p r i n g p ,   j t   =   | c G ,   j t     c p ,   j t   | .   e x p   ( h r ) .   c o s ( 2   π   r )   +   c G , j t
  •     End IF
  •    End IF
  •   End For
  • End For
  • Evaluate the fitness of the offspring o f f s p r i n g p t .
  • Return to population.
  • For each p     1 to P o p S i z e
  • IF o f f s p r i n g p t         M R THEN
  •    Replace c p t with o f f s p r i n g p t .
  • End IF
  • End For
  • Choose the best individual c b e s t t among the updated population.
  • IF c b e s t t         M R THEN
  •    Replace c G t with c b e s t t .
  • End IF
  • t     = t     +     1
  • End while

4.3.1. Initial Population

The GA is embedded into the WOA algorithm in order to develop a number of whales (search agents) in the form of chromosomes. Every chromosome is a hypothesis for the best solution (preys). Therefore, every search agent contains genes, each of which represents a concept/semantic relation of the ontology. A set of random agents c p , j t is generated initially. After generating the random solutions, the hybrid G-WOA starts to search for the best solution through a number of iterations (t).

4.3.2. Fitness Evaluation

An internal classifier was used to evaluate the fitness value of each agent (whale). In this work, it was proven that the SVM showed the best performance among the other classifiers. We used fitness function for measuring each agent’s false alarm rate, detection rate, and the number of concepts selected in each iteration until reaching the best solution. The optimal solution will be the one that decreases the False Alarm Rate (FAR), increases the Detection Rate (DR), and decreases the number of selected concepts. A standalone weighted fitness function was used to deal with this Multi-Criteria Decision Making. Three weights w f , w d , and w c were used to define FAR, DR, and the number of selected features, respectively.
F i t n e s s   =   w f [ F a l s e     A l a r m     R a t e ]   +   w d [ D e t e c t i o n     R a t e ]   +   w c [ F ]
where
F a l s e     A l a r m     R a t e     ( F A R )     =     F a l s e     P o s i t i v e F a l s e     P o s i t i v e       +     T r u e     N e g a t i v e
D e t e c t i o n     R a t e     ( D R )   =     T r u e     P o s i t i v e T r u e     P o s i t i v e       +       F a l s e     N e g a t i v e
F   =   [ 1     k   =   1 M G k M ]
G k { 0 i f     t h e     c o n c e p t     i s     s e l e c t e d     t h r o u g h     s e l e c t i n g     i t s     r e p r e s e n t a t i v e     g e n e     o f     t h e     w h a l e 1 i f     t h e     c o n c e p t     i s     n e g l e c t e d     t h r o u g h     n e g l e c t i n g     i t s     r e p r e s e n t a t i v e     g e n e     o f     t h e     w h a l e
M Number of concepts.

4.3.3. Mutation

The mutation operator, which is the core of the G-WOA algorithm, was used to produce a mutant vector. In this regard, a mutation rate M R is defined as a prerequisite. If the gene of the picked solution is lower than the M R value, then the algorithm will mutate each gene within the parent solution using Equation (26). Where o f f s p r i n g p , j t is the new generated solution c r a n d 1 ,   j t and c r a n d 2 ,   j t are two randomly selected parents,   M p ,   j is a random value in the range [ 0 ,   1 ] , t denotes the current iteration number, and p represents the whale number.
o f f s p r i n g p , j t   =   c r a n d 1 ,   j t   +   M p ,   j   ( c r a n d 2 ,   j t   c r a n d 1 ,   j t   )

4.3.4. Crossover

In the encircle prey phase, the uniform crossover operator is performed between the mutant vector, namely, o f f s p r i n g p ,   j t , and a randomly selected solution c r a n d ,   j t . The ordinary WOA algorithm uses a random variable to compute the distance between the best whale and the search agent without considering the fitness value for neither the current solution nor the functioned one. On the contrary, the G-WOA implements the crossover operator of GA in the encircle prey phase so that it selects a neighbor solution around the optimal solution. The crossover rate C r is defined as a parameter for the G-WOA algorithm. The parent solution is integrated with the neighbor solution to generate the child based on the C r value, using the following equation:
o f f s p r i n g p ,   j t   =   [ o f f s p r i n g p ,   j t r a n d   ( 0 ,   1 )         C r     o r     j   =   =   u r a n d c G ,   j t     A p t .   S p ,   j t c G ,   j t     A p t .   S p ,   j t | c G ,   j t     c p ,   j t   | .   e x p   ( h r ) .   c o s ( 2   π   r )   +   c G , j t r a n d   ( 0 ,   1 )     >     C r     o r     j   =   =   u r a n d   &   m < 0.5   &   | K |     1 r a n d   ( 0 ,   1 )     >     C r     o r     j   =   =   u r a n d   &   m < 0.5   &   | K |   <   1 r a n d   ( 0 ,   1 )     >     C r     o r     j   =   =   u r a n d   &   m 0.5  

4.3.5. Selection

The selection operator was implemented in G-WOA to determine if the target or offspring survived to the following iteration. The selection operator in G-WOA is expressed as in Equations (28) and (29). If every gene value of the generated solution is higher than the mutation value, then the G-WOA will replace the parent solution with the generated one. This comparison will be performed for each solution in the population. Then, the best solution is selected from the updated population based on the fitness value computed using Equation (21). The new best generated solution c b e s t t will be replaced with the old one c G t if the each gene value of the best solution is lower than the mutation value:
c p t   =   [ o f f s p r i n g   p t i f     o f f s p r i n g   p t         M R c p t o t h e r w i s e
c G t =   [ c b e s t t i f     c b e s t t         M R c G t o t h e r w i s e .  

4.3.6. Termination Phase

In the G-WOA algorithm, the new position of i   t h individual in the following generation is the fittest one between parent c p t and child o f f s p r i n g p t . In this context, solutions should regard boundary constraints. In case the constraints are violated, Equation (30) can be used to apply the following repairing rule:
c p ,   j     =     { u j     +     r a n d ( 0 ,   1 )     ×     ( l j     u j ) i f       c p ,   j   <     u u   l j     +     r a n d ( 0 ,   1 )     ×     ( l j     u j ) i f       c p ,   j     >     l u
where u j and l j represents upper and lower bounds of the solution’s j   t h dimension, respectively. c p ,   j refers to the j   t h dimension of the p   t h solution. r a n d ( 0 ,   1 ) represents a random number (between 0 and 1). Furthermore, the G-WOA algorithm checks the current iteration index. If the current iteration index reached the limit of the predefined criterion ( E ) , then the new solutions generated are chosen, which are the solutions with the highest fitness. Then, the database is updated with the new solutions for Arabic ontology structure. Otherwise, the G-WOA algorithm will proceed the iteration process.

5. Experimental Results

This section discusses the validation results of the proposed approach for Arabic ontology learning based on text mining and G-WOA algorithms. Extensive experiments have been conducted using different bio-inspired optimization algorithms and over different Arabic corpora. Furthermore, to discuss and evaluate the how the proposed approach works for the non-Arabic setting, we applied it to two publicly available non-Arabic corpora and compared the results to the state-of-the-art works that use the same corpora. The details of the experiments are illustrated in the following section.

5.1. Corpora

The Arabic corpora tested in this work are automatic content extraction (ACE) [72,73] corpora, ANERcorp [74,75] dataset, and a private corpus of Arabic biomedical texts. In the previously published computational linguistic work, the ACE and ANERcorp were frequently utilized for the purposes of evaluation and comparison with the existing systems. Three ACE corpora were investigated in this study: ACE 2003 (Broadcast News (BN), and Newswire (NW)), as well as ACE 2004 (NW). They are publicly available and were all tested by the proposed algorithm. For each dataset, the types of concepts (named entities) and their representation are demonstrated in Table 3. With the goal of identifying certain types of Arabic biomedical named entities in this work, we created a private corpus for evaluating the proposed approach of Arabic ontology learning. This task was accomplished by collecting a number of the Arabic open source texts in the biomedical domain, which were assessed by expert physicians. The private corpora information was illustrated in Table 3, where we represent each class in each Arabic domain by a number of documents that contained the number of unique words the concept mining and ontology learning algorithms will operate on.
Furthermore, the non-Arabic corpora tested in this work include two publicly available ones that belong to the biomedical domain and are related to the protein–protein interactions. These corpora are Learning Language in Logic (LLL) [76] and the Interaction Extraction Performance Assessment (IEPA) [77]. The LLL corpus presents the task of gene interaction from a group of sentences related to Bacillus subtilis transcription. The IEPA dataset comprises 303 abstracts obtained from the repository of PubMed, each one including a particular pair of co-occurring chemicals.

5.2. Performance Measures

The performance validation measures used in this paper are precision ( P R E ), recall ( R E C ), and F-score ( F ), [58]. The F-score is used in information retrieval to represent the harmonic incorporation of the values computed from precision ( P R E ), and recall ( R E C ) measures. These metrics were calculated for each k -fold using Equations (31)–(33), then we finally estimated the overall average of their values:
P R E = T r u e     P o s i t i v e T r u e     P o s i t i v e       +     F a l s e     P o s i t i v e
R E C = T r u e     P o s i t i v e T r u e     P o s i t i v e       +     F a l s e     N e g a t i v e
F = 2     T r u e     P o s i t i v e 2     T r u e     P o s i t i v e     +     F a l s e     P o s i t i v e     +     F a l s e     N e g a t i v e .

5.3. Cross Validation

In this work, we used k-fold cross-validation to evaluate the quality of the solution obtained using the G-WOA algorithm, in which k is equivalent to ten. Each corpus was randomly separated into ten sub-samples which were equally sized. From each corpus, a single sub-sample was set as a validation set so that it was used in performance testing, then the k     1 sub-samples were employed as a training set. This procedure was repeated 10 times. In each fold, each k sub-sample was employed exactly once as the validation set. The k -outcomes of the folds were then averaged so that they provided a single rating.

5.4. Discussion

5.4.1. Comparison to the State-of-the-Art

The comparison to the state-of-the-art was composed of three experiments: (1) comparisons with the other bio-inspired optimization algorithms existing in the literature regarding Arabic ontology learning, (2) comparisons with the previously published approaches on Arabic ontology learning from the text, and (3) comparisons with the state-of-the-art on learning ontology from non-Arabic settings. Firstly, to validate the performance of the proposed G-WOA algorithm in learning ontology from Arabic text, we compared the solution results returned by it to those returned by the ordinary GA and WOA. Moreover, extensive comparisons were conducted by comparing the performance of the G-WOA algorithm to three other bio-inspired algorithms: PSO [44], moth flame optimization (MFO) [45], and the hybrid differential evolution-whale optimization (DE-WOA) [46]. To compare these bio-inspired algorithms, the parameter setting had to be determined for each. Table 4 presents the parameter list used in this work, which was taken from [32,37,44,78]. In each experiment, the tested algorithm was first implemented into one of the previously mentioned corpora. Then, the measures of P R E , R E C , and F were computed using Equations (31)–(33). This process was repeated for each dataset. Then, the average values of the three measures across all datasets were computed.
For each algorithm of the G-WOA, GA, WOA, PSO, MFO, and DE-WOA, we demonstrated the detailed validation results obtained across all the investigated Arabic corpora. The results are demonstrated in Table A1, Table A2, Table A3, Table A4, Table A5 and Table A6 of Appendix A, respectively. To sum up, Table 5 presents the total average measures of each algorithm across all the corpora. From Table 5 and Table A1, Table A2, Table A3, Table A4, Table A5 and Table A6, it is apparent that the proposed G-WOA algorithm outperformed the other algorithms in all folds and across all the datasets. The P R E , R E C , and F results provided by the hybrid G-WOA algorithm were higher when compared to those from the ordinary GA, WOA, PSO, MFO, and the hybrid DE-WOA algorithm. Taking the ACE 2003 (BN) as an example, the results of P R E , R E C , and F were 98.14%, 99.03%, and 98.59%, respectively. The F-score ( F ) values obtained using the G-WOA were also 98.79%, 98.44%, 98.57%, and 98.63% for ACE 2003 (NW), ACE 2004 (NW), ANERcorp, and the private corpus.
Compared to the GA algorithm, the obtained F-score ( F ) values were 93.75%, 93.63%, 93.73%, 93.65%, and 93.41% for the ACE 2003 (BN), ACE 2003 (NW), ACE 2004 (NW), ANERcorp, and the private corpus. On the other hand, the WOA achieved F-score ( F ) values of 96.66%, 96.87%, 96.79%, 96.94%, and 96.81%, respectively for the aforementioned corpora. From these results, the improvement in F-score ( F ) of the ontology learning in comparison to the basic GA algorithm was 4.84%, 5.16%, 4.71%, 4.92%, and 5.22%, respectively, for the same corpora. In comparison to the ordinary WOA, the improvement reached 1.93%, 1.92%, 1.65%, 1.63%, 1.82%, respectively, for the five corpora. These results indicate that the G-WOA was able to accelerate the process of global searching when learning ontology, with its ability to balance effectively both exploration and exploitation.
Furthermore, the results also show that the MFO algorithm was more optimal than WOA in ontology learning. This is due to the good ability of MFO to switch between both exploration and exploitation, contrary to the WOA, which was trapped early in local optima throughout the optimization process [37]. Therefore, the MFO algorithm occupies the second rank after the G-WOA in terms of Arabic ontology learning. On the other side, the DE-WOA has a lower performance than G-WOA and MFO across all the datasets. Although the DE algorithm had robust global searchability, it was weak in the exploitation, and converged slowly. Thus, the DE algorithm needs to be optimized for it to be hybridized with other algorithms, as reported in [46]. Thus, the DE-WOA has the third rank in terms of the Arabic ontology learning.
In contrast, the convergence speed is a crucial criterion for evaluating the performance of any optimization method. Therefore, the convergence time for the proposed hybrid G-WOA algorithm was computed and compared to the time obtained by all algorithms versus the false alarms rate. The false alarms rate was computed in this paper using Equation (22). As depicted in Figure 2 and Figure A1 (of Appendix A), when following the WOA algorithm across all the Arabic datasets, we see that it took a lower convergence speed in comparison to the proposed hybrid G-WOA algorithm. This can be interpreted by the poor exploitation ability for the ordinary WOA algorithm, which requires a long time to search for the offspring and parents. On the contrary, the hybrid G-WOA algorithm overcame this drawback by combining the genetic operations into the WOA algorithm.
Secondly, to investigate the efficacy of the proposed approach that integrates the text mining and G-WOA algorithms to learn ontology from the Arabic text, we performed a comparison between it and the more recent works that use the same Arabic corpora, in terms of precision ( P R E ), recall ( R E C ), and F-score ( F ). Table 6 shows the comparison. Compared to the other methods presented in the literature, as Table 6 shows, the proposed approach yielded superior results in terms of P R E , R E C , and F measures. These results demonstrate the robustness of integrating text mining and G-WOA algorithms.
Thirdly, to test the efficiency of the proposed approach to learn ontology from the non-Arabic text, we applied it to the two aforementioned publicly available corpora. Furthermore, we compared its performance to the other approaches presented in the literature to learn ontology from the non-Arabic text, in terms of P R E , R E C , and F measures. The comparison is shown in Table 7, while the application results are demonstrated in Table A7 of Appendix A. The results demonstrate that the proposed G-WOA achieved superior results when applied to the non-Arabic corpora.

5.4.2. Contributions to the Literature

From the previous results, the proposed approach outperformed the state-of-the-art approaches to Arabic ontology learning. Likewise, it was also noted that very little research has used the evolutionary approaches, whereas no hybrid bio-inspired algorithms were investigated by the previously published works. The majority of studies have depended on the hybridization of rule-based and machine learning approaches [4,23], which have shortcomings, as previously discussed in the introduction section. Some works used deep learning algorithms [78,79] like the long-short-term-memory and convolutional neural network, but the results are still below expectation. In [78], a deep neural network-based method was proposed. The application results to the ANERcorp were 95.76%, 82.52%, and 88.64%, in terms of P R E , R E C , and F measures. These results are also lower than those obtained using the approach proposed in this work. In [79], the F measure results obtained using the presented deep learning approach were 91.2%, 94.12%, 91.47%, and 88.77%, respectively, for the ACE 2003 (NW), ACE 2003 (BN), ACE 2004 (NW), and ANERcorp. These results are also lower than those obtained by applying the approach proposed in this work to the same corpora, which reveals the efficiency of our approach. The results presented in [78,79] reveal the need for enhancing the performance of the deep learning methods and to overcome their shortcomings such as being stuck in the local optima, when applied to the natural language processing, for instance, through using the bio-inspired optimization algorithms.
The proposed ontology learning approach is also applicable with non-Arabic texts. Furthermore, the comparisons to the state-of-the-art approaches on learning ontology using the same non-Arabic corpora demonstrate higher results in favor of the proposed approach. These results confirm that the proposed approach outperforms the state-of-the-art methods on learning ontology from the non-Arabic texts.

5.4.3. Implications for Practice

As for ontology learning using the G-WOA algorithm, the contributions of GA and WOA enabled the GA-WOA to jump out easily of the local minima. Accordingly, it found a promising search direction toward global optimization. Specifically, the G-WOA algorithm has a robust capability to attain equilibrium between both global and local exploitation. Therefore, the proposed hybrid G-WOA algorithm outperformed the other compared algorithms in terms of speed.
The implications for practice show that the synergy of text mining and G-WOA algorithms can operate on either the Arabic or non-Arabic document by extracting the concepts and their semantic relations and then providing the solutions with the best set of concepts between the initial one. The obtained solutions can optimize the ontology construction from the Arabic or the non-Arabic text by returning only the important concepts that contribute to the ontology structure while ignoring the redundant or less important ones.

6. Conclusions

The majority of the state-of-the-art works on Arabic ontology learning from texts have depended on the hybridization of the handcrafted rules and machine learning algorithms. Contrary to the literature, this study presented a novel approach for Arabic ontology learning from texts which advances the state-of-the-art in two ways. First, a text mining algorithm was proposed for extracting the initial concept set from the text documents together with their semantic relations. Secondly, a hybrid G-WOA was proposed to optimize the ontology learning from Arabic text. The G-WOA integrates the genetic search operators like mutation, crossover, and selection into the WOA algorithm to achieve the equilibrium between both exploration and exploitation, in order to find the best solutions that exhibit the highest fitness. The experimental results revealed the following conclusions.
Firstly, as for learning Arabic ontology from texts, the proposed GA-WOA outperformed the ordinary GA, and WOA across all the Arabic datasets in terms of P R E , R E C , and F measures. When comparing the solution results obtained using the G-WOA to those obtained using the ordinary GA, we found an improvement in F-score ( F ) by up to 4.84%, 5.16%, 4.71%, 4.92%, and 5.22%, respectively, for ACE 2003 (NW), ACE 2004 (NW), ANERcorp, and the private corpus. Furthermore, the improvement reached 1.93%, 1.92%, 1.65%, 1.63%, 1.82%, respectively for the same corpora when using the ordinary WOA algorithm. Secondly, the G-WOA also outperformed the PSO, DE-WOA, and MFO across all the Arabic corpora, in terms of the three measures. The MFO occupies the second rank after the G-WOA, in terms of ontology learning from Arabic text. This was interpreted by the good ability of MFO to switch between both exploration and exploitation. Thirdly, the G-WOA outperformed the other algorithms in convergence speed. Taking the WOA as an example, it is found to have low convergence due to its poor exploitation. Thus, the G-WOA algorithm is superior when compared to the other bio-inspired algorithms in terms of convergence speed.
Furthermore, the G-WOA exhibited low rates of false alarms across all the Arabic datasets, in comparison to the other algorithms. Fourthly, the proposed Arabic ontology learning approach, which is based on the synergy of text mining and G-WOA algorithms, outperformed the state-of-the-art in terms of precision ( P R E ), recall ( R E C ), and F-score ( F ). This was due to its high capability to extract the concepts along with the semantic relations from the Arabic documents, then creating a population of search agents (solutions) that include genes represent the initial concepts. Moreover, the G-WOA starts to search for the best solution through a set of iterations, including embedding the genetic operators into the WOA architecture. Eventually, the algorithm returns the solution which recommends the best set of concepts/relations that can contribute to the ontology. Eventually, the proposed ontology learning approach is applicable to the non-Arabic texts. It achieved higher performance that outperformed the state-of-the-art approaches on learning ontology from the non-Arabic text.

Limitations and Future Research Directions

The proposed approach for Arabic ontology learning cannot deal with learning the hierarchical feature representation from the text. One advantage of the deep learning algorithms is that they are able to generate high-level feature representation from raw texts directly. Therefore, we tend to present a deep neural network model using latent features to improve learning ontology from Arabic texts. The proposed model will work on embedding the words and positions as latent features, therefore, it will not rely on feature engineering. To overcome the limitations of the deep network model, such as being stuck in the local optima, different bio-inspired optimization algorithms will be tested and compared in this regard.

Author Contributions

Conceptualization, R.M.G.; data curation, R.M.G., N.A., and K.S.; formal analysis, R.M.G., N.A., and K.S.; funding acquisition, N.A.; investigation, R.M.G.; methodology, R.M.G.; project administration, N.A. and K.S.; resources, R.M.G.; software, R.M.G.; supervision, R.M.G. and K.S.; validation, R.M.G.; visualization, R.M.G.; writing —original draft, R.M.G.; writing—review & editing, R.M.G. and K.S.

Funding

This research project was funded by the Deanship of Scientific Research at Princess Nourah bint Abdulrahman University, through the Research Funding Program.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A: Tables of Results

This section presents the tables and figures that summarize the application results of the proposed hybrid G-WOA and the other bio-inspired algorithms to the Arabic and non-Arabic corpora.
Table A1. Performance evaluation of ontology learning using the proposed G-WOA over the five corpora.
Table A1. Performance evaluation of ontology learning using the proposed G-WOA over the five corpora.
Hybrid G-WOA
FoldACE 2003 (BN)ACE 2003 (NW)ACE 2004 (NW)ANERcorpPrivate Corpus
P R E (%) R E C (%) F (%) P R E (%) R E C (%) F (%) P R E (%) R E C (%) F (%) P R E (%) R E C (%) F (%) P R E (%) R E C (%) F (%)
19999.1299.0699.4298.8499.1397.7898.5598.1698.4998.9598.7298.8399.0198.92
298.399.0498.6799.5398.6699.0998.8699.7899.3298.5797.0597.898.0698.7298.39
397.9399.7998.8599.3198.9899.1497.0999.3998.2398.8297.8998.3597.0399.6598.32
497.6199.1198.3599.2897.6898.4798.3999.1998.7998.7698.9698.8698.5298.9798.74
598.3898.2198.2999.498.4598.9298.5498.3698.4599.1898.2498.7197.9799.5598.75
698.0998.1198.198.597.3197.997.9498.2198.0799.2898.2498.7697.3299.4398.36
797.7699.2898.5199.949999.4797.0599.4798.2599.8297.5298.6698.5799.1198.84
897.898.9498.3799.7498.1198.9297.1499.5498.3398.9998.7698.8797.8398.7698.29
998.6799.5299.0998.5297.9598.2398.6698.298.4398.1398.4398.2898.4898.4498.46
1097.999.2298.5699.0498.2298.6397.5199.298.3599.8997.5698.7198.6299.8699.24
Average98.1499.0398.5999.2798.3298.7997.998.9998.4498.9998.1698.5798.1299.1598.63
Table A2. Performance evaluation of ontology learning using the GA over the five corpora.
Table A2. Performance evaluation of ontology learning using the GA over the five corpora.
GA
FoldACE 2003 (BN)ACE 2003 (NW)ACE 2004 (NW)ANERcorpPrivate Corpus
P R E (%) R E C (%) F (%) P R E (%) R E C (%) F (%) P R E (%) R E C (%) F (%) P R E (%) R E C (%) F (%) P R E (%) R E C (%) F (%)
193.0893.8793.4893.293.5293.3693.8394.9694.493.7192.4493.0893.8293.2993.56
293.2393.7593.4994.0592.1693.193.1493.4193.2893.8993.8393.8693.293.7793.49
392.1993.5292.8694.9293.089493.3195.4794.3893.2292.3292.7793.9692.6593.31
493.3193.5693.4493.6193.2793.4492.3894.3893.3795.3693.0394.1993.8192.2293.01
593.1495.7194.4193.0193.9593.4892.6794.6593.6595.4593.6294.5394.7492.0993.4
692.5593.8893.2295.4892.3893.9193.5194.2393.8794.792.7393.7193.493.7293.56
793.1795.9794.5594.292.6893.4493.4594.2793.8693.5193.2293.3794.1292.6693.39
893.5193.6193.5695.892.2994.0293.2994.4893.8993.1893.693.3993.3893.493.39
992.2995.9694.0994.4892.9793.7292.3294.1493.2394.8392.2493.5293.5292.0592.78
1093.2595.4894.3694.4193.1893.892.7893.9993.3994.5593.594.0395.3792.9894.16
Average92.9794.5393.7594.3292.9593.6393.0794.4093.7394.2493.0593.6593.9392.8893.41
Table A3. Performance evaluation of ontology learning using the WOA over the five corpora.
Table A3. Performance evaluation of ontology learning using the WOA over the five corpora.
WOA
FoldACE 2003 (BN)ACE 2003 (NW)ACE 2004 (NW)ANERcorpPrivate Corpus
P R E (%) R E C (%) F (%) P R E (%) R E C (%) F (%) P R E (%) R E C (%) F (%) P R E (%) R E C (%) F (%) P R E (%) R E C (%) F (%)
195.2597.1696.297.4495.8796.6595.4398.0996.749697.696.7996.0997.0296.55
296.2997.0396.6697.8695.7496.7996.2298.0397.1295.0297.2596.1295.7797.296.48
396.7697.2897.0297.2495.0296.1295.4997.796.5896.0997.5396.895.8997.0996.49
496.1697.7896.9698.6495.7497.1795.3897.8496.5996.3297.1996.7595.1397.3196.21
595.3297.8296.5598.5895.6597.0996.1197.7396.9195.6497.1296.3795.9698.1997.06
695.5597.5796.5597.0796.2696.6695.0897.3796.2196.3897.0796.7295.6398.5597.07
796.449796.7297.0695.3996.2296.8297.5397.1796.7698.7997.7696.0197.8296.91
895.3497.7796.5498.9995.6597.2995.7597.4696.695.2798.0196.6296.9798.4597.7
995.5397.4396.479996.197.5395.4698.9797.1896.4398.9797.6895.3298.897.03
1096.5797.2896.9297.9396.4197.1695.7197.8396.7696.5798.9997.7795.6797.6296.64
Average95.9297.4196.6697.9895.7896.8795.7597.8696.7996.0597.8596.9495.8497.8196.81
Table A4. Performance evaluation of ontology learning using the PSO over the five corpora.
Table A4. Performance evaluation of ontology learning using the PSO over the five corpora.
PSO
FoldACE 2003 (BN)ACE 2003 (NW)ACE 2004 (NW)ANERcorpPrivate Corpus
P R E (%) R E C (%) F (%) P R E (%) R E C (%) F (%) P R E (%) R E C (%) F (%) P R E (%) R E C (%) F (%) P R E (%) R E C (%) F (%)
195.7995.795.7594.3794.0994.2394.6394.7794.795.8894.1595.0194.0195.4194.71
294.4695.594.9895.7194.1794.9494.4195.6695.0495.4294.4194.9294.1595.494.78
395.3594.0694.7194.0395.9394.9895.7695.7795.7794.194.6494.3795.3995.195.25
494.1694.4794.3294.8695.4695.1695.8595.1895.5295.4595.5395.4995.9895.295.59
595.2595.0395.1495.9194.5995.2595.5194.2494.8895.4594.3394.8995.3395.495.37
694.5495.5995.0795.9594.6495.395.694.7595.1894.4794.3194.3994.9795.9195.44
79694.8795.4495.7294.7295.2294.2794.9294.694.4795.2594.8695.795.195.4
894.1994.194.1594.1394.1794.1594.894.994.8594.1295.2194.6795.5794.1594.86
994.2995.4694.8895.7595.8995.8294.3694.5394.4594.7995.3695.0894.3494.194.22
1095.4795.7895.6395.559595.2895.494.2994.8595.8895.8595.8794.2594.2294.24
Average94.9595.069595.294.8795.0395.0694.994.989594.994.9694.979594.99
Table A5. Performance evaluation of ontology learning using the MFO over the five corpora.
Table A5. Performance evaluation of ontology learning using the MFO over the five corpora.
DE-WOA
FoldACE 2003 (BN)ACE 2003 (NW)ACE 2004 (NW)ANERcorpPrivate Corpus
P R E (%) R E C (%) F (%) P R E (%) R E C (%) F (%) P R E (%) R E C (%) F (%) P R E (%) R E C (%) F (%) P R E (%) R E C (%) F (%)
197.3696.3596.8697.296.7196.9697.3396.7497.0498.2197.2497.7398.996.4397.65
297.4296.0196.7197.5897.6997.6498.5897.0497.8198.4597.6398.0497.596.0396.76
397.0297.3397.1897.9696.1697.0697.0697.0197.0498.9197.8798.3998.5997.998.25
497.2597.2197.2397.0797.397.1997.6297.7597.6997.3796.6196.9997.897.6497.72
597.8497.1497.4997.2697.4597.3697.4896.1596.8297.4797.2797.3797.7397.1597.44
697.4296.8797.1597.9996.9597.4798.4397.4597.9497.7297.8697.7998.2597.1997.72
797.4996.1496.8297.0197.2497.1398.996.7497.8198.0896.3597.2197.296.1696.68
897.3296.7397.0397.5696.8997.2397.5396.2396.8898.8296.5897.6997.897.6797.74
997.2796.796.9997.0796.0996.5898.3297.0397.6897.3197.997.6198.3696.4997.42
1097.2796.4396.8597.8896.196.9997.1396.6196.8797.0197.997.4698.4897.6198.05
Average97.3796.6997.0397.4696.8697.1697.8496.8897.3697.9497.3297.6398.06197.0397.54
Table A6. Performance evaluation of ontology learning using the DE-WOA over the five corpora.
Table A6. Performance evaluation of ontology learning using the DE-WOA over the five corpora.
MFO
FoldACE 2003 (BN)ACE 2003 (NW)ACE 2004 (NW)ANERcorpPrivate Corpus
P R E (%) R E C (%) F (%) P R E (%) R E C (%) F (%) P R E (%) R E C (%) F (%) P R E (%) R E C (%) F (%) P R E (%) R E C (%) F (%)
197.1295.2896.296.5795.4696.0297.4395.1696.2996.3395.1595.7498.1396.3697.24
297.2795.1696.2197.2195.6996.4596.2795.0695.6797.0996.3996.7497.2396.997.07
397.2995.596.3997.7495.9596.8497.7196.5197.1196.396.5396.4297.596.0396.76
497.8995.596.6996.3596.2596.396.0895.1195.696.6495.4396.0498.995.5197.18
597.195.3996.2497.5795.1596.3597.7196.2596.9897.3795.7296.5497.7596.3797.06
697.0596.5396.7996.4896.7496.6197.4496.5296.9897.496.3196.8698.2496.1797.2
797.4795.1696.3196.7495.8596.396.5896.2296.497.5495.4896.598.3495.296.75
897.0795.896.4497.1695.6696.4197.6796.329796.9596.1296.5497.195.2296.16
997.196.4496.7796.6196.2496.4397.9196.4397.1797.8196.1196.9697.2995.8496.56
1097.7196.9697.3496.2296.8196.5296.3196.4596.3896.4496.996.6797.595.1396.31
Average97.3195.7796.5496.8795.9896.4297.119696.5696.9996.0196.597. 895.8796.83
Figure A1. The convergence time versus the FAR rate for all algorithms using (a) ACE 2003 (NW), (b) ACE 2004 (NW), (c) ANERcorp, and (d) the private corpus. Cont.
Figure A1. The convergence time versus the FAR rate for all algorithms using (a) ACE 2003 (NW), (b) ACE 2004 (NW), (c) ANERcorp, and (d) the private corpus. Cont.
Algorithms 12 00182 g0a1aAlgorithms 12 00182 g0a1b
Table A7. Performance evaluation of ontology learning using the proposed G-WOA and the non-Arabic corpora.
Table A7. Performance evaluation of ontology learning using the proposed G-WOA and the non-Arabic corpora.
FoldNon-Arabic Corpus
LLLIEPA
P R E   ( % ) R E C   ( % ) F   ( % ) P R E   ( % ) R E C   ( % ) F   ( % )
197.2897.4497.3697.7797.897.78
298.4597.7898.1197.6397.9297.77
398.3398.7798.5598.0897.5997.83
498.1998.5998.3998.3897.8698.12
598.3198.9598.6398.8597.4698.15
697.8296.9397.3798.3497.1497.74
798.2898.2498.2698.3397.8998.11
898.8496.8797.8598.697.0697.82
998.5898.5198.5498.7497.7498.24
1097.9297.9197.9197.8697.9497.9
Average98.29898.198.2697.6497.95

References

  1. Hawalah, A. A Framework for Building an Arabic Multi-disciplinary Ontology from Multiple Resources. Cogn. Comput. 2017, 10, 156–164. [Google Scholar] [CrossRef]
  2. Al-Zoghby, A.M.; Elshiwi, A.; Atwan, A. Semantic Relations Extraction and Ontology Learning from Arabic Texts—A Survey. In Intelligent Natural Language Processing: Trends and Applications Studies in Computational Intelligence; Springer: Cham, Switzerland, 2017; pp. 199–225. [Google Scholar]
  3. Mezghanni, I.B.; Gargouri, F. CrimAr: A Criminal Arabic Ontology for a Benchmark Based Evaluation. Procedia Comput. Sci. 2017, 112, 653–662. [Google Scholar] [CrossRef]
  4. Mezghanni, I.B.; Gargouri, F. Deriving ontological semantic relations between Arabic compound nouns concepts. J. King Saud Univ.-Comput. Inf. Sci. 2017, 29, 212–228. [Google Scholar] [CrossRef]
  5. Gruber, T.R. A translation approach to portable ontology specifications. Knowl. Acquis. 1993, 5, 199–220. [Google Scholar] [CrossRef]
  6. Hazman, M.; El-Beltagy, S.R.; Rafea, A. A Survey of Ontology Learning Approaches. Int. J. Comput. Appl. 2011, 22, 36–43. [Google Scholar] [CrossRef]
  7. Benaissa, B.-E.; Bouchiha, D.; Zouaoui, A.; Doumi, N. Building Ontology from Texts. Procedia Comput. Sci. 2015, 73, 7–15. [Google Scholar] [CrossRef] [Green Version]
  8. Zamil, M.G.A.; Al-Radaideh, Q. Automatic extraction of ontological relations from Arabic text. J. King Saud Univ.-Comput. Inf. Sci. 2014, 26, 462–472. [Google Scholar] [CrossRef] [Green Version]
  9. Benabdallah, A.; Abderrahim, M.A.; Abderrahim, M.E.-A. Extraction of terms and semantic relationships from Arabic texts for automatic construction of an ontology. Int. J. Speech Technol. 2017, 20, 289–296. [Google Scholar] [CrossRef]
  10. Al-Zoghby, A.M.; Shaalan, K. Ontological Optimization for Latent Semantic Indexing of Arabic Corpus. Procedia Comput. Sci. 2018, 142, 206–213. [Google Scholar] [CrossRef]
  11. Albukhitan, S.; Helmy, T.; Alnazer, A. Arabic ontology learning using deep learning. In Proceedings of the International Conference on Web Intelligence-WI 17, Leipzig, Germany, 23–26 August 2017. [Google Scholar]
  12. Sadek, J.; Meziane, F. Extracting Arabic Causal Relations Using Linguistic Patterns. ACM Trans. Asian Low-Resour. Lang. Inf. Process. 2016, 15, 1–20. [Google Scholar] [CrossRef]
  13. Kaushik, N.; Chatterjee, N. Automatic relationship extraction from agricultural text for ontology construction. Inf. Process. Agric. 2018, 5, 60–73. [Google Scholar] [CrossRef]
  14. Alami, N.; Meknassi, M.; En-Nahnahi, N. Enhancing unsupervised neural networks based text summarization with word embedding and ensemble learning. Expert Syst. Appl. 2019, 123, 195–211. [Google Scholar] [CrossRef]
  15. Chi, N.-W.; Jin, Y.-H.; Hsieh, S.-H. Developing base domain ontology from a reference collection to aid information retrieval. Autom. Constr. 2019, 100, 180–189. [Google Scholar] [CrossRef]
  16. Al-Arfaj, A.; Al-Salman, A. Towards Ontology Construction from Arabic Texts-A Proposed Framework. In Proceedings of the 2014 IEEE International Conference on Computer and Information Technology, Xi’an, China, 11–13 September 2014. [Google Scholar]
  17. Al-Rajebah, N.I.; Al-Khalifa, H.S. Extracting Ontologies from Arabic Wikipedia: A Linguistic Approach. Arab. J. Sci. Eng. 2013, 39, 2749–2771. [Google Scholar] [CrossRef]
  18. Albukhitan, S.; Alnazer, A.; Helmy, T. Semantic Web Annotation using Deep Learning with Arabic Morphology. Procedia Comput. Sci. 2019, 151, 385–392. [Google Scholar] [CrossRef]
  19. Boujelben, I.; Jamoussi, S.; Hamadou, A.B. Enhancing Machine Learning Results for Semantic Relation Extraction. In Natural Language Processing and Information Systems Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2013; pp. 337–342. [Google Scholar]
  20. Albarghothi, A.; Saber, W.; Shaalan, K. Automatic Construction of E-Government Services Ontology from Arabic Webpages. Procedia Comput. Sci. 2018, 142, 104–113. [Google Scholar] [CrossRef]
  21. Bentrcia, R.; Zidat, S.; Marir, F. Extracting semantic relations from the Quranic Arabic based on Arabic conjunctive patterns. J. King Saud Univ.-Comput. Inf. Sci. 2018, 30, 382–390. [Google Scholar] [CrossRef]
  22. Kramdi, S.E.; Haemmerl, O.; Hernandez, N. Approche générique pour l’xtraction de relations partir de texts. Journées Francoph. D’ingénierie Des Connaiss. 2009, 97–108. Available online: https://hal.archives-ouvertes.fr/hal-00384415/document (accessed on 26 June 2019).
  23. Boujelben, I.; Jamoussi, S.; Hamadou, A.B. A hybrid method for extracting relations between Arabic named entities. J. King Saud Univ.-Comput. Inf. Sci. 2014, 26, 425–440. [Google Scholar] [CrossRef] [Green Version]
  24. Karimi, H.; Kamandi, A. A learning-based ontology alignment approach using inductive logic programming. Expert Syst. Appl. 2019, 125, 412–424. [Google Scholar] [CrossRef]
  25. Juckett, D.A.; Kasten, E.P.; Davis, F.N.; Gostine, M. Concept detection using text exemplars aligned with a specialized ontology. Data Knowl. Eng. 2019, 119, 22–35. [Google Scholar] [CrossRef]
  26. Petruccia, G.; Rospocher, M.; Ghidini, C. Expressive Ontology Learning as Neural Machine Translation. SSRN Electron. J. 2018, 52–53, 66–82. [Google Scholar] [CrossRef]
  27. Luan, J.; Yao, Z.; Zhao, F.; Song, X. A novel method to solve supplier selection problem: Hybrid algorithm of genetic algorithm and ant colony optimization. Math. Comput. Simul. 2019, 156, 294–309. [Google Scholar] [CrossRef]
  28. Alsaeedan, W.; Menai, M.E.B.; Al-Ahmadi, S. A hybrid genetic-ant colony optimization algorithm for the word sense disambiguation problem. Inf. Sci. 2017, 417, 20–38. [Google Scholar] [CrossRef]
  29. Gaidhane, P.J.; Nigam, M.J. A hybrid grey wolf optimizer and artificial bee colony algorithm for enhancing the performance of complex systems. J. Comput. Sci. 2018, 27, 284–302. [Google Scholar] [CrossRef]
  30. Elrehim, M.Z.A.; Eid, M.A.; Sayed, M.G. Structural optimization of concrete arch bridges using Genetic Algorithms. Ain Shams Eng. J. 2019. [Google Scholar] [CrossRef]
  31. Liu, P.; Basha, M.D.E.; Li, Y.; Xiao, Y.; Sanelli, P.C.; Fang, R. Deep Evolutionary Networks with Expedited Genetic Algorithms for Medical Image Denoising. Med. Image Anal. 2019, 54, 306–315. [Google Scholar] [CrossRef] [PubMed]
  32. Ghoniem, R.M. Deep Genetic Algorithm-Based Voice Pathology Diagnostic System. In Natural Language Processing and Information Systems Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2019; pp. 220–233. [Google Scholar]
  33. Gupta, R.; Nanda, S.J.; Shukla, U.P. Cloud detection in satellite images using multi-objective social spider optimization. Appl. Soft Comput. 2019, 79, 203–226. [Google Scholar] [CrossRef]
  34. Nguyen, T.T. A high performance social spider optimization algorithm for optimal power flow solution with single objective optimization. Energy 2019, 171, 218–240. [Google Scholar] [CrossRef]
  35. Jayaprakash, A.; Keziselvavijila, C. Feature selection using Ant Colony Optimization (ACO) and Road Sign Detection and Recognition (RSDR) system. Cogn. Syst. Res. 2019, 58, 123–133. [Google Scholar] [CrossRef]
  36. Chen, L.; Xiao, C.; Li, X.; Wang, Z.; Huo, S. A seismic fault recognition method based on ant colony optimization. J. Appl. Geophys. 2018, 152, 1–8. [Google Scholar] [CrossRef]
  37. Aziz, M.A.E.; Ewees, A.A.; Hassanien, A.E. Whale Optimization Algorithm and Moth-Flame Optimization for multilevel thresholding image segmentation. Expert Syst. Appl. 2017, 83, 242–256. [Google Scholar] [CrossRef]
  38. Elaziz, M.A.; Mirjalili, S. A hyper-heuristic for improving the initial population of whale optimization algorithm. Knowl.-Based Syst. 2019, 172, 42–63. [Google Scholar] [CrossRef]
  39. Mirjalili, S.; Lewis, A. The Whale Optimization Algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  40. Goldbogen, J.A.; Friedlaender, A.S.; Calambokidis, J.; Mckenna, M.F.; Simon, M.; Nowacek, D.P. Integrative Approaches to the Study of Baleen Whale Diving Behavior, Feeding Performance, and Foraging Ecology. BioScience 2013, 63, 90–100. [Google Scholar] [CrossRef] [Green Version]
  41. Habib, Y.; Sadiq, M.S.; Hakim, A. Applications and Science of Neural Networks, Fuzzy Systems, and Evolutionary Computation; Society of Photo Optical: Bellingham, WA, USA, 1998. [Google Scholar] [CrossRef]
  42. Xue, X.; Chen, J. Using Compact Evolutionary Tabu Search algorithm for matching sensor ontologies. Using Compact Evolutionary Tabu Search algorithm for matching sensor ontologies. Swarm Evol. Comput. 2019, 48, 25–30. [Google Scholar] [CrossRef]
  43. Afia, A.E.; Lalaoui, M.; Chiheb, R. A Self Controlled Simulated Annealing Algorithm using Hidden Markov Model State Classification. Procedia Comput. Sci. 2019, 148, 512–521. [Google Scholar] [CrossRef]
  44. Ghoniem, R.M.; Shaalan, K. FCSR-Fuzzy Continuous Speech Recognition Approach for Identifying Laryngeal Pathologies Using New Weighted Spectrum Features. In Proceedings of the International Conference on Advanced Intelligent Systems and Informatics 2017 Advances in Intelligent Systems and Computing, Cairo, Egypt, 9–11 September 2017; 2017; pp. 384–395. [Google Scholar]
  45. Das, A.; Mandal, D.; Ghoshal, S.; Kar, R. Concentric circular antenna array synthesis for side lobe suppression using moth flame optimization. AEU-Int. J. Electron. Commun. 2018, 86, 177–184. [Google Scholar] [CrossRef]
  46. Pourmousa, N.; Ebrahimi, S.M.; Malekzadeh, M.; Alizadeh, M. Parameter estimation of photovoltaic cells using improved Lozi map based chaotic optimization Algorithm. Sol. Energy 2019, 180, 180–191. [Google Scholar] [CrossRef]
  47. Prabhu, Y.; Kag, A.; Harsola, S.; Agrawal, R.; Varma, M. Parabel: Partitioned label trees for extreme classification with application to dynamic search advertising. In Proceedings of the 2018 World Wide Web Conference on World Wide Web-WWW, Lyon, France, 23–27 April 2018. [Google Scholar]
  48. Khandagale, S.; Xiao, H.; Babbar, R. Bonsai-Diverse and Shallow Trees for Extreme Multi-label Classification. Available online: https://arxiv.org/abs/1904.08249 2109 (accessed on 6 August 2019).
  49. Babbar, R.; Partalas, I.; Gaussier, E.; Amini, M.-R. On Flat versus Hierarchical Classification in Large-Scale Taxonomies. In Proceedings of the 27th Annual Conference on Neural Information Processing Systems (NIPS 26), Lake Tao, NV, USA, 5–10 December 2013; pp. 1824–1832. [Google Scholar]
  50. Babbar, R.; Partalas, I.; Gaussier, E.; Amini, M.-R.; Amblard, C. Learning taxonomy adaptation in large-scale classification. J. Mach. Learn. Res. 2016, 17, 1–37. [Google Scholar]
  51. Moradi, M.; Ghadiri, N. Different approaches for identifying important concepts in probabilistic biomedical text summarization. Artif. Intell. Med. 2018, 84, 101–116. [Google Scholar] [CrossRef]
  52. Mosa, M.A.; Anwar, A.S.; Hamouda, A. A survey of multiple types of text summarization with their satellite contents based on swarm intelligence optimization algorithms. Knowl.-Based Syst. 2019, 163, 518–532. [Google Scholar] [CrossRef]
  53. Ababneh, J.; Almomani, O.; Hadi, W.; El-Omari, N.K.T.; Al-Ibrahim, A. Vector Space Models to Classify Arabic Text. Int. J. Comput. Trends Technol. 2014, 7, 219–223. [Google Scholar] [CrossRef]
  54. Fodil, L.; Sayoud, H.; Ouamour, S. Theme classification of Arabic text: A statistical approach. In Proceedings of the Terminology and Knowledge Engineering, Berlin, Germany, 19–21 June 2014; pp. 77–86. Available online: https://hal.archives-ouvertes.fr/hal-01005873/document (accessed on 25 May 2019).
  55. Al-Tahrawi, M.M.; Al-Khatib, S.N. Arabic text classification using Polynomial Networks. J. King Saud Univ. -Comput. Inf. Sci. 2015, 27, 437–449. [Google Scholar] [CrossRef] [Green Version]
  56. Al-Anzi, F.S.; Abuzeina, D. Toward an enhanced Arabic text classification using cosine similarity and Latent Semantic Indexing. J. King Saud Univ.-Comput. Inf. Sci. 2017, 29, 189–195. [Google Scholar] [CrossRef] [Green Version]
  57. Abuzeina, D.; Al-Anzi, F.S. Employing fisher discriminant analysis for Arabic text classification. Comput. Electr. Eng. 2018, 66, 474–486. [Google Scholar] [CrossRef]
  58. Al-Anzi, F.S.; Abuzeina, D. Beyond vector space model for hierarchical Arabic text classification: A Markov chain approach. Inf. Process. Manag. 2018, 54, 105–115. [Google Scholar] [CrossRef]
  59. Alkhatib, M.; Barachi, M.E.; Shaalan, K. An Arabic social media based framework for incidents and events monitoring in smart cities. J. Clean. Prod. 2019, 220, 771–785. [Google Scholar] [CrossRef]
  60. Ben Hamadou, A.; Piton, O.; Fehri, H. Multilingual extraction of functional relations between Arabic named entities using Nooj platform. In Proceedings of the Nooj 2010 International Conference and Workshop, Komotini, Greece, 27–28 May 2010; Gavriilidou, Z., Chadjipapa, E., Papadopoulou, L., Silberztein, M., Eds.; 2010; pp. 192–202. [Google Scholar]
  61. Boujelben, I.; Jamoussi, S.; Ben Hamadou, A. Rules based approach for semantic relations extraction between Arabic named entities. In Proceedings of the International NooJ 2012 Conference, Paris, France, 14–16 June 2012; pp. 123–133. [Google Scholar]
  62. Mesmia, F.B.; Zid, F.; Haddar, K.; Maurel, D. ASRextractor: A Tool extracting Semantic Relations between Arabic Named Entities. Procedia Comput. Sci. 2017, 117, 55–62. [Google Scholar] [CrossRef]
  63. Celli, F. Searching for Semantic Relations between Named Entities in I-CAB 2009. Available online: http://clic.cimec.unitn.it/fabio (accessed on 28 April 2019).
  64. Shahine, M.; Sakre, M. Hybrid Feature Selection Approach for Arabic Named Entity Recognition. In Computational Linguistics and Intelligent Text Processing Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2018; pp. 452–464. [Google Scholar]
  65. Kadir, R.A.; Bokharaeian, B. Overview of Biomedical Relations Extraction using Hybrid Rule-based Approaches. J. Ind. Intell. Inf. 2013, 1, 169–173. [Google Scholar] [CrossRef]
  66. Landauer, T.K.; Dumais, S.T. A solution to Platos problem: The latent semantic analysis theory of acquisition, induction, and representation of knowledge. Psychol. Rev. 1997, 104, 211–240. [Google Scholar] [CrossRef]
  67. Alkhatib, M.; Monem, A.A.; Shaalan, K. A Rich Arabic WordNet Resource for Al-Hadith Al-Shareef. Procedia Comput. Sci. 2017, 117, 101–110. [Google Scholar] [CrossRef]
  68. Salton, G.; Buckley, C. Term-weighting approaches in automatic text retrieval. Inf. Process. Manag. 1988, 24, 513–523. [Google Scholar] [CrossRef] [Green Version]
  69. Wei, Y.-Y.; Wang, R.-J.; Hu, Y.-M.; Wang, X. From Web Resources to Agricultural Ontology: A Method for Semi-Automatic Construction. J. Integr. Agric. 2012, 11, 775–783. [Google Scholar] [CrossRef]
  70. Zhang, X.; Chan, F.T.; Yang, H.; Deng, Y. An adaptive amoeba algorithm for shortest path tree computation in dynamic graphs. Inf. Sci. 2017, 405, 123–140. [Google Scholar] [CrossRef] [Green Version]
  71. Shojaedini, E.; Majd, M.; Safabakhsh, R. Novel adaptive genetic algorithm sample consensus. Appl. Soft Comput. 2019, 77, 635–642. [Google Scholar] [CrossRef] [Green Version]
  72. AbdelRahman, S.; Elarnaoty, M.; Magdy, M.; Fahmy, A. Integrated Machine Learning Techniques for Arabic Named Entity Recognition. IJCSI Int. J. Comput. Sci. Issues 2010, 7, 27–36. [Google Scholar]
  73. Oudah, M.; Shaalan, K. Person Name Recognition Using the Hybrid Approach. In Natural Language Processing and Information Systems Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2013; pp. 237–248. [Google Scholar]
  74. Benajiba, Y.; Rosso, P.; Benedíruiz, J.M. ANERsys: An Arabic Named Entity Recognition System Based on Maximum Entropy. In Computational Linguistics and Intelligent Text Processing Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2007; pp. 143–153. [Google Scholar]
  75. Abdul-Hamid, A.; Darwish, K. Simplified Feature Set for Arabic Named Entity Recognition. Available online: https://www.aclweb.org/anthology/W10-2417 (accessed on 24 April 2019).
  76. Nédellec, C.; Nazarenko, A. Ontologies and Information Extraction. 2005. Available online: https://hal.archives-ouvertes.fr/hal-00098068/document (accessed on 18 April 2019).
  77. Ding, J.; Berleant, D.; Nettleton, D.; Wurtele, E. Mining Medline: Abstracts, Sentences, Or Phrases? Biocomputing 2002 2001, 7, 326–337. [Google Scholar] [CrossRef]
  78. Gridach, M. Deep Learning Approach for Arabic Named Entity Recognition. In Computational Linguistics and Intelligent Text Processing Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2018; pp. 439–451. [Google Scholar]
  79. Khalifa, M.; Shaalan, K. Character convolutions for Arabic Named Entity Recognition with Long Short-Term Memory Networks. Comput. Speech Lang. 2019, 58, 335–346. [Google Scholar] [CrossRef]
  80. Taghizadeh, N.; Faili, H.; Maleki, J. Cross-Language Learning for Arabic Relation Extraction. Procedia Comput. Sci. 2018, 142, 190–197. [Google Scholar] [CrossRef]
  81. Oudah, M.; Shaalan, K. Studying the impact of language-independent and language-specific features on hybrid Arabic Person name recognition. Lang. Resour. Eval. 2016, 51, 351–378. [Google Scholar] [CrossRef]
  82. Shaalan, K.; Oudah, M. A hybrid approach to Arabic named entity recognition. J. Inf. Sci. 2013, 40, 67–87. [Google Scholar] [CrossRef] [Green Version]
  83. Lima, R.; Espinasse, B.; Freitas, F. A logic-based relational learning approach to relation extraction: The OntoILPER system. Eng. Appl. Artif. Intell. 2019, 78, 142–157. [Google Scholar] [CrossRef]
  84. Ma, C.; Zhang, Y.; Zhang, M. Tree Kernel-based Protein-Protein Interaction Extraction Considering both Modal Verb Phrases and Appositive Dependency Features. In Proceedings of the 24th International Conference on World Wide Web-WWW 15 Companion, Florence, Italy, 18–22 May 2015. [Google Scholar]
  85. Li, L.; Guo, R.; Jiang, Z.; Huang, D. An approach to improve kernel-based Protein–Protein Interaction extraction by learning from large-scale network data. Methods 2015, 83, 44–50. [Google Scholar] [CrossRef] [PubMed]
  86. Qian, L.; Zhou, G. Tree kernel-based protein–protein interaction extraction from biomedical literature. J. Biomed. Inform. 2012, 45, 535–543. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Representation of some biomedical concepts in our corpus which have an Is-a semantic relationship.
Figure 1. Representation of some biomedical concepts in our corpus which have an Is-a semantic relationship.
Algorithms 12 00182 g001
Figure 2. The convergence time versus the FAR rate for all algorithms using the ACE 2003 (BN) corpus.
Figure 2. The convergence time versus the FAR rate for all algorithms using the ACE 2003 (BN) corpus.
Algorithms 12 00182 g002
Table 1. A state-of-the-art on Arabic text mining.
Table 1. A state-of-the-art on Arabic text mining.
ReferenceYear of PublicationArabic Text Mining AlgorithmCorpusAccuracy
[53]2014Cosine Coefficient, Jacaard Coefficient, and Dice Coefficient Saudi Newspapers (SNP)Cosine coefficient outperformed Jaccard and Dice coefficients with 0.917, 0.979, and 0.947 for Precision, Recall, and F-measure, respectively.
[54]2014Term frequency (TF), and Term Frequency/Inverse Document Frequency (TF/IDF) for feature extraction, Semi-Automatic Categorization (SAC), and Automatic Categorization (AC) for feature selection.News books: Arabic Dataset for Theme Classification (subsets 1 & 2) Global recognition score is used to measure the ratio of correctly-classified documents: employing TF/IDF (95%), and TF (88%)
[55] 2015TF/IDF, Chi Square for selecting feature, besides a local class-based policy for feature reductionAl-Jazeera News Recall of 0.967%, and F-measure of 0.959
[56] 2017Latent Semantic IndexingAlqabas newspaper in Kuwait82.50%
[57] 2018Vector Space Model (VSM)Set by Alqabas newspaper, in Kuwait 84.4%
[58]2018VSMAlqabas newspaper90.29%
[59]2019Removing the stop words existed in the collected tweets, extracting the keywords and sorting them into one of the corresponding categories: classified words or unclassified words. Then, applying named entity recognition as well as data analysis rules on the classified words to generate final report. The lexical features along with Twitter-specific features were employed in classification. A private database of collected Arabic tweets96.49%
Table 2. A state-of-the-art on Arabic ontology learning.
Table 2. A state-of-the-art on Arabic ontology learning.
MethodologyWorksYear of PublicationContribution
Rule-based approaches[60,61] 2010 & 2012Extracting a set of linguistic patterns from text then rewriting it into finite state transducers
[12]2016The authors developed a model of pattern recognizer that targets to signal the existence of cause–effect information in sentences from non-specific domain texts. The model incorporated 700 linguistic patterns to distinguish the sentence parts representing the cause, besides to these representing the effect. To construct patterns, various sets of the syntactic features were considered through analyzing the untagged corpus.
[62]2017The authors introduced a rule-based system namely, ASRextractor, to extract and annotate semantic relations relating Arabic named entities. The semantic relation extraction was based upon an annotated corpus of Arabic Wikipedia. The corpus incorporated 18 types of semantic relations like synonymy and origin.
[20]2018A statistical parsing method was adopted to estimate the key-phrase/keyword from the Arabic dataset. The extracted dataset was converted to an OWL ontology format. Then, the mapping rules were used to link the components of ontology to corresponding keywords.
[21]2018A set of rules/conjunctive patterns were defined for extracting the semantic relations of the Quranic Arabic according to a deep study for Arabic grammar, POS tagging, as well as the morphology features appears in the corpus of Quranic Arabic.
Machine learning-based approaches[63]2009With the objective of semantic relation extraction, the authors amalgamated two supervised methods, to be specific, the basic Decision Tree as well as Decision Lists-based Algorithm. They estimated Three semantic relations (i.e., location, social and role) among named entities (NEs).
[22]2009On the basis of the dependency graph producing by syntactic analysis, the authors adopted a learning pattern algorithm, denoted ( LP ) 2 for generating annotation rules.
[19]2013A rule mining approach has been proposed to be applied on an Arabic corpus using lexical, numerical, and semantic features. After the learning features were extracted from the annotated instances, a set of rules were generated automatically by three learning algorithms, namely, Apriori, decision tree algorithm C4.5, and Tertius.
[9]2017A statistical algorithm was used to extract the simple and complex terms, namely, “the repeated segments algorithm”. For selecting segments that have sufficient weight, the authors used the Weighting Term Frequency–Inverse Document Frequency algorithm (WTF-IDF). Further, a learning approach was proposed based upon the analysis of examples for learning extraction markers to detect new pairs of relations.
[64]2018Genetic algorithm (GA) was proposed to minimize the computation time needed to search out the informative and appropriate Arabic text features needed for classification. The SVM was used as machine learning algorithm that evaluates the accuracy of the Arabic named entities recognition.
Hybrid approaches[65]2013Three methodologies were encompassed: kernel method, co-occurrence, and later rule-based. These methods were utilized for extracting simple and complicated relations regard the biomedical domain. For mapping the data into a feature space of high-dimensionality, Kernel-based algorithms have been used.
[23]2014The authors proposed a hybrid rule-based/machine learning approach and a manual technique for extracting semantic relations between pairs on named entities.
[4]2017A rules patterns set was defined from compound concepts for inducing of general relations. It utilized a gamification mechanism to specify relations based on prepositions semantics. The Formal Concept Analysis/Relational Concept Analysis approaches were employed for modeling the hierarchical as well as transversal relations of concepts.
Table 3. Information of the Arabic corpora tested in this work.
Table 3. Information of the Arabic corpora tested in this work.
CorpusTotal
ACE 2003 (BN)ACE 2003 (NW)ACE 2004 (NW)ANERcorpPrivate Corpus
Person51771118653602 6695
Date2058357- 435
Time11528- 44
Price317105- 125
Measurement142851- 93
Percent33554 92
Location107312924934425 7283
Organization18149313132025 4012
Healthcare Provider----80978097
Health Disorder----13,50213,502
Cancers----90729072
Surgeries----70557055
Table 4. The parameter list used in this work.
Table 4. The parameter list used in this work.
GAWOAPSOMFOGA-WOADE-WOA
Population size: 100Population size: 100Particles number P: 10Population size: 100Population size: 100Population size: 100
Maximum generations: 500The random variable r   :   [ 1 ,   1 ] Iterations number t: 10The constant defines the logarithmic spiral shape h: 1 Maximum generations: 500Scaling factor for DE: Random between 0.2 and 0.8
Crossover probability C r : 0.9Logarithmic spiral shape h: 1Acceleration c 1 : 2The random variable r   :   [ 1 ,   1 ] Crossover probability C r : 0.9DE mutation scheme: DE/best/1/bin
Mutation rate M R : 0.05e: Decreased from 2 to 0Acceleration c 2 : 2 Mutation rate M R : 0.05The random variable r   :   [ 1 ,   1 ]
Reproduction ratio: 0.18 Maximal weight of inertia: 0.7 Maximum iterations number: 10Logarithmic spiral shape h: 1
Selection: weighted Roulette Wheel Minimal weight of inertia: 0.1 The random variable r   :   [ 1 ,   1 ] e: Decreased from 2 to 0
Logarithmic spiral shape h: 1
e: Decreased from 2 to 0
Table 5. Average measures for each algorithm across all datasets (detailed results of algorithms can be followed in Table A1, Table A2, Table A3, Table A4, Table A5 and Table A6 of the Appendix A).
Table 5. Average measures for each algorithm across all datasets (detailed results of algorithms can be followed in Table A1, Table A2, Table A3, Table A4, Table A5 and Table A6 of the Appendix A).
AlgorithmAverage Measures
P R E   ( % ) R E C   ( % ) F   ( % )
Proposed GA-WOA98.4898.7398.6
GA93.7193.5693.63
WOA96.3197.3496.81
PSO95.0494.9594.99
MFO97.7396.9697.34
DE-WOA97.0795.9396.57
Table 6. Comparison to the state-of-the-art on Arabic ontology learning.
Table 6. Comparison to the state-of-the-art on Arabic ontology learning.
ReferenceYearCorpusApproach Accuracy
[79]2019ACE 2003 (NW), ACE 2003 (BN), ACE 2004 (NW), and ANERcorp. The Long-Short-Term-Memory neural tagging model was augmented with the Convolutional Neural Network to extract the character-level features. F = 91.2%, 94.12%, 91.47%, and 88.77%, respectively, for the four corpora.
[78]2018ANERcorp A deep neural network-based method. P R E = 95.76%,
R E C = 82.52%, and
F = 88.64%.
[80]2018ANERcorp Integration of some tree and polynomial kernels for feature representation. The universal dependency parsing was used for the relation extraction. F = 63.5%.
[64]2016ANERcorpA hybrid approach of the GA and SVM. F = 82%.
[81]2016ACE 2003 (NW), ACE 2003 (BN), ACE 2004 (NW), and ANERcorp.Hybridization of the rule-based and machine learning approaches. The feature space comprised the language-specific and language independent features. The decision tree classifier was used as a machine learning algorithm. P R E = 92.7%, R E C = 88.1%, and F = 90.3% for the ACE 2003 (BN), while they are 92.9%, 93.4%, 93.2%, for the ACE 2003 (NW), respectively. P R E = 82.8%, R E C = 82%, and 82.4%, for the ACE 2004 (NW), while they are 94.9%, 94.2%, and 94.5%, respectively, for the ANERcorp.
[82]2013ACE 2003 (NW), ACE 2003 (BN), ACE 2004 (NW), and ANERcorp.Hybrid rule-based/machine learning approach. The features comprised: Rule-based features estimated from the rule-based decisions, morphological features derived from morphological analysis of decisions, POS features, contextual features, Gazetteer features, and word-level features. The J48, Libsvm, and Logistic classifiers were used.The highest results were achieved when applying the proposed method to the ANERcorp: P R E = 87%, R E C = 60%, and F = 94%.
Proposed approach2019 A text mining algorithm to extract the initial concept set from the Arabic documents. A proposed G-WOA algorithm to get the best solutions that optimize the ontology learning through selecting only the optimal concept set with their semantic relations, which contribute to the ontology structure. P R E = 98.14%,
R E C = 99.03%, and
F = 98.59%, for the ACE 2003 (BN) while their values are 99.27%, 98.32%, and 98.79%, respectively, for the ACE 2003 (NW).
P R E = 97.9%,
R E C = 98.99%, and
F = 98.44%, for the ACE 2004 (NW), while their values are 98.99%, 98.16%, and 98.57% for ANERcorp, and 98.12%, 99.15%, 98.63%, for the private corpus.
Table 7. Comparison to the state-of-the-art in non-Arabic settings.
Table 7. Comparison to the state-of-the-art in non-Arabic settings.
ReferenceYearCorpusApproach Accuracy
[83]2019LLL and IEPAA logic-based relational learning method for Relation Extraction utilizing the Inductive Logic Programming, namely OntoILPER, to generate symbolic extraction rules. F = 79.9% and 76.1%, respectively, for the two corpora.
[84]2015LLL and IEPAAn optimized tree kernel-based Protein–protein extraction approach. The modal verbs together with the appositive dependency features were used for defining some relevant rules that expand and optimize the shortest dependency path in between two proteins. F = 82.3% and 68.7%, respectively, for the two corpora.
[85]2015LLL and IEPAThree word representational techniques including vector clustering, distributed representation, and Brown clusters. The SVM was used for unsupervised learning. F = 87.3% and 76.5%, respectively, for the two corpora.
[86]2012LLL and IEPATree kernel-based extraction approach, in which the tree representation produced from constituent syntactic parser, was refined utilizing the shortest dependency route in between two proteins estimated from the dependency parser. F = 84.6% and 69.8%, respectively, for the two corpora.
Proposed approach2019LLL and IEPAIntegrating text mining and G-WOA algorithms.98.1% and 97.95%, respectively, for the two corpora.

Share and Cite

MDPI and ACS Style

Ghoniem, R.M.; Alhelwa, N.; Shaalan, K. A Novel Hybrid Genetic-Whale Optimization Model for Ontology Learning from Arabic Text. Algorithms 2019, 12, 182. https://0-doi-org.brum.beds.ac.uk/10.3390/a12090182

AMA Style

Ghoniem RM, Alhelwa N, Shaalan K. A Novel Hybrid Genetic-Whale Optimization Model for Ontology Learning from Arabic Text. Algorithms. 2019; 12(9):182. https://0-doi-org.brum.beds.ac.uk/10.3390/a12090182

Chicago/Turabian Style

Ghoniem, Rania M., Nawal Alhelwa, and Khaled Shaalan. 2019. "A Novel Hybrid Genetic-Whale Optimization Model for Ontology Learning from Arabic Text" Algorithms 12, no. 9: 182. https://0-doi-org.brum.beds.ac.uk/10.3390/a12090182

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop