Next Article in Journal
Blockchain Behavioral Traffic Model as a Tool to Influence Service IT Security
Previous Article in Journal
PACER: Platform for Android Malware Classification, Performance Evaluation and Threat Reporting
Previous Article in Special Issue
Feature Fusion Text Classification Model Combining CNN and BiGRU with Multi-Attention Mechanism
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Semantic Search Enhanced with Rating Scores

National Research Council (CNR), Istituto di Analisi dei Sistemi ed Informatica “Antonio Ruberti”, Via dei Taurini 19, I-00185 Rome, Italy
*
Author to whom correspondence should be addressed.
Future Internet 2020, 12(4), 67; https://0-doi-org.brum.beds.ac.uk/10.3390/fi12040067
Submission received: 11 March 2020 / Revised: 10 April 2020 / Accepted: 13 April 2020 / Published: 15 April 2020
(This article belongs to the Special Issue New Perspectives on Semantic Web Technologies and Applications)

Abstract

:
This paper presents SemSime, a method based on semantic similarity for searching over a set of digital resources previously annotated by means of concepts from a weighted reference ontology. SemSime is an enhancement of SemSim and, with respect to the latter, it uses a frequency approach for weighting the ontology, and refines both the user request and the digital resources with the addition of rating scores. Such scores are High, Medium, and Low, and in the user request indicate the preferences assigned by the user to each of the concepts representing the searching criteria, whereas in the annotation of the digital resources they represent the levels of quality associated with each concept in describing the resources. The SemSime has been evaluated and the results of the experiment show that it performs better than SemSim and an evolution of it, referred to as S e m S i m R V .

1. Introduction

The most significant improvement within the Semantic Web research area pertains to reasoning and searching abilities. In this perspective, semantic similarity reasoning, which relies on the knowledge coded in a reference ontology [1], is a different technique with respect to the well-known deductive reasoning used in expert systems. In [2], we proposed SemSim, a semantic search method based on a Weighted Reference Ontology (WRO). In SemSim, both the resources in the search space and the requests of users are represented by means of an Ontology Feature Vector ( O F V ), which is a set of concepts from the WRO. We distinguish the user request, also denoted as Request Vector ( R V ), from the description of a resource, also referred to as Annotation Vector, indicated by A V . In the search process, SemSim contrasts the R V against each A V , and the result is a ranking of the resources that exhibit the highest similarity degree with respect to the request defined by the user.
In [2], we analyzed two different approaches in order to weigh the reference ontology, namely the frequency-based and the uniform probabilistic approaches. In the experiment described in that paper, we show that SemSim by the frequency-based approach outperforms the SemSim by the uniform probabilistic approach, as well as the most representative similarity methods from the literature.
In this work, we present a new method, referred to as SemSime. It relies on the frequency-based approach and revises SemSim along two directions. According to the first direction, in contrasting the R V with the A V , SemSime takes into consideration the cardinality of the set of the concepts (features) in the user request rather than the maximal cardinality of the compared O F V . This choice allows us to give more relevance to the features which are requested by the user rather than the extra features contained in the annotation vectors available in the search space. Along the second direction, SemSim has been enhanced with the rating scores H i g h (H), M e d i u m (M), and L o w (L) in the O F V , with regard to both the request and the search space resources. Within the request, rating scores denote the preferences given by the user to the concepts of the WRO used to specify the query whereas, within the annotation vectors, rating scores represent the levels of quality associated with the concepts when they describe the resources. Consider an example rooted in the tourism domain, where the user is searching for a vacation package by specifying the following features: I n t e r n a t i o n a l H o t e l (H), L o c a l T r a n s p o r t a t i o n (M), C u l t u r a l A c t i v i t y (H), and E n t e r t a i n m e n t (L). On the basis of the given rating scores, he/she gives a high preference to resorts which are international hotels offering cultural activities, and less priority to the remaining features, in particular to the entertainments. Analogously, a holiday package annotated with H o r s e R i d i n g (H), M u s e u m (M), and T h a i M e a l (L) is characterized by a high quality level with regard to the horse riding service, rather than the facilities in visiting museums or having Thai meals at lunch or dinner. Note that, in [3], a proposal concerning rating scores was given, where the concepts of the WRO are weighted according to the uniform probabilistic approach [4], rather than the frequency-based one. Furthermore, in our approach we assumed that, given a facility (for instance, H o r s e R i d i n g ) included in a tourist package, the higher the user’s priority about that facility, the higher the expectancy about the quality of the same facility and, therefore, the greater the availability of the user for considering more expensive solutions.
In this paper, we have experimented SemSime in the domain of tourism and we have compared it to the SemSim method defined in [2] and a further evolution of SemSim, referred to as SemSim R V . Essentially, SemSim R V is the original SemSim method where, in line with the first direction adopted in SemSime illustrated above, more priority has been given to the features indicated by the user in his/her request. The results of the experiment show that SemSime outperforms both these methods.
The organization of the rest of the paper is the following: Section 2 is about the related work; Section 3 shows the SemSime method and recalls SemSim; Section 4 illustrates the experimental results; Section 5 presents the conclusion and future activities.

2. Related Work

In the literature, several proposals regarding semantic similarity reasoning have been defined, as for instance [5,6,7,8,9]. In general, the matchmaking between vectors of concepts is computed on the basis of their intersection as performed in Dice and Jaccard methods [10], without considering neither the information content of the concepts in the ontology nor the hierarchical relationship among them. In [11], the Weighted Sum method has been proposed where the similarity of hierarchically related concepts is evaluated, although by using a fixed value (i.e., 0.5). In [12], the similarity between sets of ontology concepts is computed on the basis of the shortest path in a graph by applying an extended version of the Dijkstra algorithm [13]. Finally, other proposals, such as [14], consider the inverse document frequency (IDF) method, and combine it with the term frequency (TF) approach.
With respect to the mentioned papers, in this work the semantic matchmaking method is performed in accordance with the information content approach defined by Lin in [15], which is an evolution of [16]. The Lin’s method results in a better correlation with human judgment when compared with other approaches, such as the edge-counting [17,18,19]. Furthermore, concerning the evaluation of the similarity between vectors of concepts, we borrowed the Hungarian algorithm for solving the maximum weighted matching problem in bipartite graphs [20].
In [21], the representation of users’ preferences relies on the definition of users’ profiles, which are built according to an ontology-based model. The ontology has been exploited in order to establish relationships among user profiles and features. Then, the level of interest of the user is modeled by taking into account both features and historical data. Reference [22] copes with the evaluation of the similarity between users’ profiles. In particular, it proposes a collaborative filtering method in order to identify similar users rating in a given set of items, and to analyze the ratings associated with these users. In [23], the traditional feature-based measures for computing similarity are used, and the authors present an ontology-based approach for evaluating similarity between classes where attributes replace features. The underpinning of this approach is that the more attributes two classes have in common, the higher similarity they have.
Finally, it turns out that SemSime is a novel approach because, as opposed to the above mentioned proposals, it introduces rating scores for enriching the semantic annotations of both user requests and resources.

3. The Semantic Similarity Method

The SemSime method is an extension of SemSim proposed in [2,24], which is based on the information content approach applied to ontologies [15]. According to this approach, concepts of the ontology are associated with weights such that, along the hierarchy, as weights of concepts decrease their information content increase. A Weighted Reference Ontology (WRO) is a pair:
WRO = <Ont, w>
and, in particular:
  • Ont = <C, H>, where C is a set of c o n c e p t s , also referred to as f e a t u r e s , and H is the set of pairs of concepts of C that are related according to the I S A hierarchy (specialization hierarchy);
  • w is a function, referred to as weight, such that, given a concept c in the ontology, w(c) is a number in [0,…,1].
In Figure 1, a W R O related to the tourism domain is shown, whose weights are defined by using the frequency-based approach.
Given a WRO and a concept c C , the information content ic(c) is defined as [25]:
i c ( c ) = - log w ( c )
The Resource Space represents all the searchable and available digital resources. To define the semantic content of a given resource, a structure gathering a set of concepts from the W R O is associated with it. This structure will be referred to as Ontology Feature Vector (OFV) and is represented as follows:
o f v = ( c 1 , , c n ) , where c i C , i = 1 , , n
Similarly, an O F V describes a user request. As mentioned in the Introduction, in the following, A V (Annotation Vector), and R V (Request Vector) will denote the semantics of a resource and a user request, respectively.
In order to search for the resources in the Resource Space, the SemSim method has been defined. This method evaluates the semantic similarity between an AV and a RV by using the semsim function which relies on the consim function. The latter allows the evaluation of the similarity between concepts, say c i , c j , as defined below:
c o n s i m ( c i , c j ) = 2 × i c ( l c a ( c i , c j ) ) i c ( c i ) + i c ( c j )
where lca (lowest common ancestor) is the least abstract concept in the hierarchy subsuming c i and c j . Given an instance of A V , say a v , and an instance of R V , say r v , semsim computes consim for each pair of concepts belonging to the Cartesian product of a v and r v . Indeed, we aim at identifying the set of pairs of concepts from a v and r v that maximizes the sum of consim according to the Hungarian algorithm for the maximum weighted matching problem in bipartite graphs [20]. In particular, given:
r v = { c 1 r , , c m r }
a v = { c 1 a , , c n a }
where {c 1 r ,…, c m r } ∪ {c 1 a ,…, c n a } ⊆C, let S be defined as the following Cartesian product:
S = r v × a v
and let P ( r v , a v ) be defined as:
P ( r v , a v ) = { P S : ( c i r , c j a ) , ( c h r , c k a ) P , c i r c h r , c j a c k a , | P | = m i n { n , m } } .
Thus, semsim(rv, av) is defined as follows:
s e m s i m ( r v , a v ) = m a x P P ( r v , a v ) { ( c i r , c j a ) P c o n s i m ( c i r , c j a ) } m a x { n , m }

SemSime

The SemSime method, which is based on SemSim, has been conceived in order to manage a more refined representation of an OFV, where concepts are associated with scores. In SemSime, an OFV is indicated as OFVe and, similar to the previous section, RVe and AVe as well.
In our approach, the presence of the scores in the AVe and RVe represents a refined annotation and a refined request vector, respectively. In the case of the AVe, each score, denoted by s a , stands for the quality level of the feature characterizing the resource, while in the case of RVe, denoted by s r , it indicates the priority degree assigned by the user to the requested feature. As mentioned above, these scores are: H ( H i g h ), M ( M e d i u m ), or L ( L o w ). Therefore, a given OFVe, say ofve, is represented as follows:
o f v e = { ( c 1 , s 1 ) , , ( c n , s n ) }
where s i is s i a or s i r , c i is a c i a or c i r , i = 1 , , n .
The semsime function evaluates the similarity between an AVe and a RVe, represented as ave and a rve, respectively, by coupling their elements. The value calculated by semsime is achieved by multiplying the consim value by the score matching value, namely sm, illustrated in Table 1, on the basis of both s a and s r scores. Formally, given:
r v e = { ( c 1 r , s 1 r ) , , ( c m r , s m r ) }
a v e = { ( c 1 a , s 1 a ) , , ( c n a , s n a ) }
s e m s i m e ( r v e , a v e ) is defined as follows:
s e m s i m e ( r v e , a v e ) = ( c i r , c j a ) P M c o n s i m ( c i r , c j a ) s m ( s i r , s j a ) m
where ( c i r , s i r ) ∈ r v e , ( c j a , s j a ) ∈ a v e , and P M P ( r v , a v ) is utilized to evaluate the semsim value between rv and av, without considering the rating scores of both r v e and a v e .
As mentioned above, in the experimentation presented in the next section, we also address an evolution of S e m S i m , referred to as S e m S i m R V , where in contrasting the R V with the A V , analogously to SemSime, the number of features defined in the user request are considered rather than the maximal cardinality of the compared O F V . Formally, the s e m s i m R V ( r v , a v ) function is defined as the semsim(rv, av) (see formula (2)), where max{n,m} is replaced by the cardinality of the request vector m, i.e.,
s e m s i m R V ( r v , a v ) = m a x P P ( r v , a v ) { ( c i r , c j a ) P c o n s i m ( c i r , c j a ) } m
Note that m a x { n , m } in Equation (2) has been replaced in both Equations (3) and (4) by m, which is the cardinality of the request vector, in order to give relevance to the user request and, in particular, to the features searched by the user.

4. SemSime Evaluation

In the experimentation, we asked 15 colleagues at work to identify their privileged tourist packages (request vector, r v ) by choosing 4 or (at most) 5 concepts of the ontology illustrated in Figure 1. In particular, assume the user desires to reside in an international hotel, in a location connected by flight and local transportation services, and the opportunity of enjoying cultural activities and entertainments. The related request vector is given below:
rv = (InternationalHotel, LocalTransportation, CulturalActivity, Entertainment, Flight)
By associating a score among H, M, and L with each concept in the r v , the user can indicate the level of his/her priority for that concept. Accordingly, the user can express his/her priority in the above r v by the following:
rve = (InternationalHotel (H), LocalTransportation (M), CulturalActivity (H), Entertainment (L), Flight (H))
where InternationalHotel, CulturalActivity and Flight received a higher preference with respect to LocalTransportation and Entertainment, the latter with the lowest score.
Then, we requested our colleagues to examine the 10 packages reported in Table 2, where each concept has been associated with a score among H, M, and L, specifying, similar to T r i p A d v i s o r , the quality level of the offered services. Our colleagues were also asked to choose 5 packages, out of 10, more similar to their r v , and to assign a score to each chosen one representing the degree of similarity between the package and their r v (Human Judgment—HJ). The request vectors defined by the colleagues are shown in Table 3.
For the purpose of illustration, let us consider the request vector defined by the user U 3 . This user desires to get to his/her destination by flight, reside in an international hotel, commute by public transportation, enjoy international meals and local attractions. As we observe, in the request vector of U 3 all the features have been refined by the score H, except F l i g h t which has been refined by M.
Table 4 shows the experimental results related to the user U 3 . The first column contains the packages chosen by the user, that are P 1 , P 3 , P 4 , P 5 , P 8 , and, respectively, the related similarity values 0.95, 0.8, 0.5, 0.4, 0.2. The other columns illustrate the 10 packages ranked, respectively, according to S e m S i m , S e m S i m R V and S e m S i m e , with the associated similarity values with U 3 .
In Table 5 S e m S i m , S e m S i m R V , and S e m S i m e correlations with H J are given. We observe that S e m S i m e has higher values in 73% of cases. Table 6 shows the precision and recall for all the users, where “-” stands for undefined because the set of retrieved resources is empty. Note that with respect to S e m S i m the precision of S e m S i m e increases in 5 cases and decreases in 1 case, while the recall increases in most of the cases. Regarding to S e m S i m R V the precision of S e m S i m e is the same while the recall increases in 4 cases and is the same for the remaining cases.
Note that, due to the presence of the rating scores, the similarity values computed according to S e m S i m e never increase with respect to S e m S i m and S e m S i m R V . For this reason, the evaluation of the precision and the recall has been performed on the basis of two different thresholds: 0.5 with regard to H J , S e m S i m , and S e m S i m R V , as indicated in Table 4, and 0.4 concerning S e m S i m e . This distinction allows us to balance the decrease of the similarity values obtained by S e m S i m e due to the presence of rating scores. In particular, considering U 3 in Table 4, the precision remains invariant (i.e., 1), while the recall increases (0.33 vs. 0.66 vs. 1.00). Both the precision and the recall are equal to 1 because S e m S i m e retrieves all and only the resources relevant to the user whereas, for instance, the recall of S e m S i m R V is equal to 0.66 because, among P 1 , P 3 , and P 4 , P 1 is not retrieved.
Overall, in selecting resources more similar to users’ requests, the experimental results show that S e m S i m e improves the performances of both S e m S i m and S e m S i m R V . This improvement is achieved because, with respect to S e m S i m and S e m S i m R V , S e m S i m e relies on the additional knowledge carried by the preference and quality rating scores.

5. Conclusions and Future Work

In this work a semantic similarity method, referred to as SemSime, has been introduced. It relies on a previous proposal of the authors, namely SemSim. In particular, SemSime has been endowed with the rating scores H i g h (H), M e d i u m (M), L o w (L) in the O F V for representing both the preferences of the user and the quality of the resources in the search space. An experimentation of SemSime has been performed in the tourism domain by comparing it with the original method SemSim, and a variant of it, referred to as S e m S i m R V . The experiment reveals that SemSime outperforms both SemSim and S e m S i m R V , with respect to precision, recall and correlation.
Currently, we are planning to apply the proposed approach on a large dataset with regard to the annotated resources, and also by considering a large participation in the human judgment activity. In particular, we are setting up an experiment by focusing on the Digital Library of the Association for Computing Machinery (ACM). In this context, the reference ontology is represented by the ACM Computing Classification System (ACM-CCS), and the semantic annotations of the resources (more than one thousand papers) are the sets of keywords selected by the authors according to the ACM-CCS.
As a future investigation, SemSime will be applied to the automotive domain, in the framework of a collaboration with an Italian SME working in this sector, in order to improve the decision making process from both the dealer and the customer sides. In this domain, the idea of giving more relevance to the offer rather than the request will be investigated in order to enable the customer to be aware about the new features from car companies.

Author Contributions

A.F., E.P. and F.T. contributed equally to this work. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Gruber, T.R. A translation approach to portable ontologies. Knowl. Acquis. 1993, 5, 199–220. [Google Scholar] [CrossRef]
  2. Formica, A.; Missikoff, M.; Pourabbas, E.; Taglino, F. Semantic search for matching user requests with profiled enterprises. Comput. Ind. 2013, 64, 191–202. [Google Scholar] [CrossRef]
  3. Missikoff, M.; Formica, A.; Pourabbas, E.; Taglino, F. Enriching semantic search with preference and quality scores. In Encyclopedia with Semantic Computing and Robotic Intelligence; World Scientific: Singapore, 2017. [Google Scholar]
  4. Formica, A.; Missikoff, M.; Pourabbas, E.; Taglino, F. Weighted Ontology for Semantic Search; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2008; Volume 5335, pp. 1289–1303. [Google Scholar]
  5. Alani, H.; Brewster, C. Ontology ranking based on the Analysis of Concept Structures. In Proceedings of the K-CAP’05, Banff, AB, Canada, 2–5 October 2005. [Google Scholar]
  6. Euzenat, J.; Shvaiko, P. Ontology Matching; Springer: Heidelberg, Germany, 2007. [Google Scholar]
  7. Fang, W.-D.; Zhang, L.; Wang, Y.-X.; Dong, S.-B. Towards a Semantic Search Engine Based on Ontologies. In Proceedings of the 4th International Conference on Machine Learning, Guangzhou, China, 18–21 August 2005. [Google Scholar]
  8. Madhavan, J.; Halevy, A.Y. Composing Mappings among Data Sources. In Proceedings of the VLDB’03—International Conference on Very Large Databases, Berlin, Germany, 9–12 September 2003; pp. 572–583. [Google Scholar]
  9. Maguitman, A.G.; Menczer, F.; Roinestad, H.; Vespignani, A. Algorithmic Detection of Semantic Similarity. In Proceedings of the WWW’05—World Wide Web Conference, Chiba, Japan, 10–14 May 2005. [Google Scholar]
  10. Maarek, Y.S.; Berry, D.M.; Kaiser, G.E. An Information Retrieval Approach for Automatically Constructing Software Libraries. IEEE Trans. Softw. Eng. 1991, 17, 800–813. [Google Scholar] [CrossRef]
  11. Castano, S.; De Antonellis, V.; Fugini, M.G.; Pernici, B. Conceptual Schema Analysis: Techniques and Applications. ACM Trans. Databases Syst. 1998, 23, 286–333. [Google Scholar] [CrossRef]
  12. Cordi, V.; Lombardi, P.; Martelli, M.; Mascardi, V. An Ontology-Based Similarity between Sets of Concepts. In Proceedings of the WOA’05, Camerino, Italy, 14–16 November 2005; pp. 16–21. [Google Scholar]
  13. Dijkstra, E.W. A note on two problems in connexion with graphs. Numerische Mathematik 1959, 1, 269–271. [Google Scholar] [CrossRef] [Green Version]
  14. Manning, C.D.; Raghavan, P.; Schtze, H. Introduction to Information Retrieval; Cambridge University Press: New York, NY, USA, 2008. [Google Scholar]
  15. Lin, D. An Information-Theoretic Definition of Similarity. In Proceedings of the 15th International Conference on Machine Learning, Madison, WI, USA, 24–27 July 1998; Shavlik, J.W., Ed.; Morgan Kaufmann: San Francisco, CA, USA, 1998; pp. 296–304. [Google Scholar]
  16. Resnik, P. Using information content to evaluate semantic similarity in a taxonomy. In Proceedings of the IJCAI’95—The 14th International Joint Conference on Artificial Intelligence, Montreal, QC, Canada, 20–25 August 1995; pp. 448–453. [Google Scholar]
  17. Pourabbas, E.; Taglino, F. A semantic platform for enterprise knowledge interoperability. In Shaping Enterprise Interoperability in the Future Internet 5, Proceedings of the I-ESA’12—Enterprise Interoperability V, Valencia, Spain, 20–23 March 2012; Poler, R., Doumeingts, G., Katzy, B., Chalmeta, R., Eds.; Springer: London, UK, 2012. [Google Scholar]
  18. Rada, L.; Mili, V.; Bicknell, E.; Bletter, M. Development and application of a metric on semantic nets. IEEE Trans. Syst. Man Cybern. 1989, 19, 17–30. [Google Scholar] [CrossRef] [Green Version]
  19. Wu, Z.; Palmer, M. Verb semantics and lexicon selection. In Proceedings of the 32nd Annual Meeting of the Association for Computational Linguistics, Las Cruces, NM, USA, 27–30 June 1994; pp. 133–138. [Google Scholar]
  20. Dulmage, A.L.; Mendelsohn, N.S. Coverings of bipartite graphs. Can. J. Math. 1958, 10, 517–534. [Google Scholar] [CrossRef]
  21. Ali, G.; ElKorany, A. Semantic-based Collaborative Filtering for Enhancing Recommendation. In Proceedings of the KEOD’14—International Conference on Knowledge Engineering and Ontology Development, Rome, Italy, 21–24 October 2014; pp. 176–185. [Google Scholar]
  22. Eckhardt, A. Similarity of users’ (content-based) preference models for Collaborative filtering in few ratings scenario. Expert Syst. Appl. 2012, 39, 11511–11516. [Google Scholar] [CrossRef]
  23. Akmala, S.; Shih, L.H.; Batres, R. Ontology-based similarity for product information retrieval. Comput. Ind. 2014, 65, 91–107. [Google Scholar] [CrossRef] [Green Version]
  24. Formica, A.; Missikoff, M.; Pourabbas, E.; Taglino, F. Semantic Search for Enterprises Competencies Management. In Proceedings of the KEOD’10—International Conference on Knowledge Engineering and Ontology Development, Valencia, Spain, 25–28 October 2010; pp. 183–192. [Google Scholar]
  25. Shannon, C.E. A Mathematical Theory of Communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef] [Green Version]
Figure 1. A Weighted Reference Ontology ( W R O ) in the tourism domain.
Figure 1. A Weighted Reference Ontology ( W R O ) in the tourism domain.
Futureinternet 12 00067 g001
Table 1. The score matching function sm.
Table 1. The score matching function sm.
s r
HML
H1.00.80.5
s a M0.81.00.8
L0.50.81.0
Table 2. Tourist packages.
Table 2. Tourist packages.
Packages AV e
P1(Campsite (H), LocalTransportation (H), Concert (M),Flight(M), ArcheologicalSite (H), Food (L))
P2(FarmHouse (H), Trekking (M), HorseRiding (H), MediterraneanMeal (H))
P3(InternationalHotel (M), LocalTransportation (M), ArcheologicalSite (H), Flight (H), RegularMeal (M))
P4(InternationalHotel (H), CarRental (H), Museum (H), Cinema (L), Flight (H), InternationalMeal (L), ArtGallery (H), ShoppingCenter (L))
P5(Bed&Breakfast (M), LocalTransportation (H), Train (H), RockConcert(H), Cinema (M), BikeRiding (L))
P6(Bed&Breakfast (H), Bazaar (H), Museum (M), Theater (M), OpenAirActivity (L))
P7(SeasideCottage (M), Riding (H), Boating (H), Train (M), VegetarianMeal (H))
P8(Pension (M), CarRental (H), Shopping (L), CulturalActivity (H), LightMeal L)
P9(Campsite (H), LocalTransportation (M), OpenAirActivity (H), Train (M), VegetarianMeal (H))
P10(Pension (M), ArtGallery (H), EthnicMeal (H), ClassicalConcert (H), LocalTransportation (L), Entertainment (L))
Table 3. Request vectors.
Table 3. Request vectors.
Users RV e
U1(InternationalHotel (M), Train (M), ArcheologicalSite (H), MediterraneanMeal (H))
U2(Campsite (H), BikeRiding (M), Trekking (H), MediterraneanMeal (M))
U3(InternationalHotel (H), Flight (M), LocalTransportation (H), InternationalMeal (H), Attraction (H))
U4(Bed&Breakfast (M), Flight (H), Museum (M), EthnicMeal (L))
U5(Bed&Breakfast (M), Riding (H), Boating (M), VegetarianMeal (H), Shopping (L))
U6(InternationalHotel (M), Flight (M), Trekking (H), MediterraneanMeal (H), ShoppingCenter (L))
U7(InternationalHotel (H), LocalTransportation (M), CulturalActivity (H), Entertainment (L), Flight (H))
U8(Farmhouse (H), Train (H), Boating (H), MediterraneanMeal (M))
U9(Bed&Breakfast (H), Train (H), LocalTransportation (H), OpenAirActivity (M), EthnicMeal (M), RockConcert (M))
U10(Bed&Breakfast (H), Museum (M), MediterraneanMeal (H), Bazaar (H))
U11(Bed&Breakfast (M), Car Rental (H), ArcheologicalSite (H), MediterraneanMeal (H), ShoppingCenter (M))
U12(InternationalHotel (L), Flight (M), ArcheologicalSite (H), EthnicMeal (M))
U13(Campsite(H), Train (L), Trekking (M), RockConcert (M), Food (L))
U14(InternationalHotel(M), Flight(H), LocalTransportation(H), EthnicMeal(M), OpenAirActivity(H))
U15(Bed&Breakfast(H), OpenAirActivity(M), EthnicMeal(M), Bazaar(M), Entertainment(M))
Table 4. User U3.
Table 4. User U3.
HJ SemSim SemSim RV SemSim e
P10.95P30.76P30.76P30.61
P30.8P40.40P40.64P40.50
P40.5P10.36P10.44P10.42
P50.4P90.26P90.26P50.23
P80.2P50.20P50.24P90.22
P100.18P100.22P100.12
P70.07P70.07P70.06
P80.07P80.07P80.05
P20.02P20.02P20.02
P60.00P60.00P60.00
Table 5. Correlation.
Table 5. Correlation.
User Correlation
SemSim SemSim RV SemSim e
U10.620.710.71
U20.740.800.91
U30.730.740.82
U40.220.190.53
U50.560.450.81
U60.800.580.69
U70.930.980.98
U80.940.900.89
U90.750.750.83
U100.430.410.54
U110.750.720.71
U120.500.510.57
U130.610.790.58
U140.250.530.57
U150.720.800.78
Table 6. Precision and Recall.
Table 6. Precision and Recall.
User Precision Recall
SemSim SemSim RV SemSim e SemSim SemSim RV SemSim e
U1-1100.330.33
U21110.330.330.66
U31110.330.661
U4-0.50.500.330.33
U51110.330.330.33
U6-1100.330.33
U70.50.750.750.3311
U81110.330.661
U91110.660.660.66
U101110.330.330.33
U111110.330.330.66
U1210.660.660.330.660.66
U131110.330.660.66
U14-0.330.3300.330.33
U151110.330.330.33

Share and Cite

MDPI and ACS Style

Formica, A.; Pourabbas, E.; Taglino, F. Semantic Search Enhanced with Rating Scores. Future Internet 2020, 12, 67. https://0-doi-org.brum.beds.ac.uk/10.3390/fi12040067

AMA Style

Formica A, Pourabbas E, Taglino F. Semantic Search Enhanced with Rating Scores. Future Internet. 2020; 12(4):67. https://0-doi-org.brum.beds.ac.uk/10.3390/fi12040067

Chicago/Turabian Style

Formica, Anna, Elaheh Pourabbas, and Francesco Taglino. 2020. "Semantic Search Enhanced with Rating Scores" Future Internet 12, no. 4: 67. https://0-doi-org.brum.beds.ac.uk/10.3390/fi12040067

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop