Show simple item record

dc.contributor.authorGoikoetxea Salutregi, Josu
dc.contributor.authorLastra Díaz, Juan José
dc.contributor.authorAgirre Bengoa, Eneko
dc.contributor.authorTaieb, Mohamed Ali Hadj
dc.contributor.authorGarcía Serrano, Ana
dc.contributor.authorBen Aouicha, Mohamed
dc.contributor.authorSánchez, David
dc.date.accessioned2024-12-03T17:08:13Z
dc.date.available2024-12-03T17:08:13Z
dc.date.issued2020-09-30
dc.identifier.citationInformation Systems 96 : (2021) // Article ID 101636es_ES
dc.identifier.issn0306-4379
dc.identifier.issn1873-6076
dc.identifier.urihttp://hdl.handle.net/10810/70754
dc.description.abstractThis work is a companion reproducibility paper of the experiments and results reported in Lastra-Diaz et al. (2019a), which is based on the evaluation of a companion reproducibility dataset with the HESML V1R4 library and the long-term reproducibility tool called Reprozip. Human similarity and relatedness judgements between concepts underlie most of cognitive capabilities, such as categorization, memory, decision-making and reasoning. For this reason, the research on methods for the estimation of the degree of similarity and relatedness between words and concepts has received a lot of attention in the fields of artificial intelligence and cognitive sciences. However, despite the huge research effort done, there is a lack of a self-contained, reproducible and extensible collection of benchmarks which being amenable to become a de facto standard for large scale experimentation in this line of research. In order to bridge this reproducibility gap, this work introduces a set of reproducible experiments on word similarity and relatedness by providing a detailed reproducibility protocol together with a set of software tools and a self-contained reproducibility dataset, which allow that all experiments and results in our aforementioned work to be reproduced exactly. Our aforementioned primary work introduces the largest, most detailed and reproducible experimental survey on word similarity and relatedness reported in the literature, which is based on the implementation of all evaluated methods into the same software platform. Our reproducible experiments evaluate most of methods in the families of ontology-based semantic similarity measures and word embedding models. We also detail how to extend our experiments to evaluate other unconsidered experimental setups. Finally, we provide a corrigendum for a mismatch in the MC28 similarity scores used in our original experimentses_ES
dc.description.sponsorshipThis work has been partially supported by the Spanish project VEMODALEN (TIN2015-71785-R), the Basque Government (type A IT1343-19), BBVA BigKnowledge bigknowledge project, and the Spanish Research Agency LIHLITH project (PCIN-2017-118/AEI) in the framework of EU ERA-Net CHIST-ERA.es_ES
dc.language.isoenges_ES
dc.publisherElsevieres_ES
dc.rightsinfo:eu-repo/semantics/openAccesses_ES
dc.rights.urihttp://creativecommons.org/licenses/by-nc-nd/4.0/
dc.titleA large reproducible benchmark of ontology-based methods and word embeddings for word similarityes_ES
dc.typeinfo:eu-repo/semantics/articlees_ES
dc.rights.holder© 2020 Elsevier under CC BY-NC-ND licensees_ES
dc.relation.publisherversionhttps://doi.org/10.1016/j.is.2020.101636es_ES
dc.identifier.doi10.1016/j.is.2020.101636
dc.departamentoesLenguajes y sistemas informáticoses_ES
dc.departamentoeuHizkuntza eta sistema informatikoakes_ES


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record

© 2020 Elsevier under CC BY-NC-ND license
Except where otherwise noted, this item's license is described as © 2020 Elsevier under CC BY-NC-ND license