Search results
1 – 10 of 18Emad Behboudi, Amrollah Shamsi and Gema Bueno de la Fuente
In 2016, Bohannon published an article analyzing the download rate of the top ten countries using the illegal Sci-Hub website. Four years later, this study approaches the search…
Abstract
Purpose
In 2016, Bohannon published an article analyzing the download rate of the top ten countries using the illegal Sci-Hub website. Four years later, this study approaches the search behavior of these ten countries as they query about Sci-Hub in Google's search engine, the world's most widely used search engine. The authors also tracked the possible consequences of using Sci-Hub, such as plagiarism.
Design/methodology/approach
The search terms “Sci-Hub”, “Plagiarism” and “Plagiarism Checker” were explored with Google Trends. The queries were performed globally and individually for the ten target countries, all categories and web searches. The time range was limited between 1/1/2016 (after the date of publication of Bohannon's work) and 29/03/2020. Data were extracted from Google Trends and the findings were mapped.
Findings
Searching for the word Sci-Hub on Google has increased nearly eightfold worldwide in the last four years, with China, Ethiopia and Tunisia having the most searches. Sci-Hub's search trends increased for most of the T10C, with Brazil and Iran having the highest and lowest average searches, respectively.
Originality/value
Access to the research literature is required to the progress of research, but it should not be obtained illegally. Given the increasing incidence of these problems in countries at any level of development, it is important to pay attention to ethics education in research and establish ethics committees. A comprehensive review of the research process is required to reduce the urge to circumvent copyright laws and includes training and educating research stakeholders in copyright literacy. To address these goals, national and international seriousness and enthusiasm are essential.
Details
Keywords
Gema Bueno de la Fuente, Carmen Agustín-Lacruz, Mariângela Spotti Lopes Fujita and Ana Lúcia Terra
The purpose of this study is to analyse the recommendations on knowledge organisation from guidelines, policies and procedure manuals of a sample of institutional repositories and…
Abstract
Purpose
The purpose of this study is to analyse the recommendations on knowledge organisation from guidelines, policies and procedure manuals of a sample of institutional repositories and networks within the Latin American area and observe the level of follow-up of international guidelines.
Design/methodology/approach
Presented is an exploratory and descriptive study of repositories’ professional documents. This study comprised four steps: definition of convenience sample; development of data codebook; coding of data; and analysis of data and conclusions drawing. The convenience sample includes representative sources at three levels: local institutional repositories, national aggregators and international network and aggregators. The codebook gathers information from the repositories’ sample, such as institutional rules and procedure manuals openly available, or recommendations on the use of controlled vocabularies.
Findings
The results indicate that at the local repository level, the use of controlled vocabularies is not regulated, leaving the choice of terms to the authors’ discretion. It results in a set of unstructured keywords, not standardised terms, mixing subject terms with other authorities on persons, institutions or places. National aggregators do not regulate these issues either and limit to pointing to international guidelines and policies, which simply recommend the use of controlled vocabularies, using URIs to facilitate interoperability.
Originality/value
The originality of this study lies in identifying how the principles of knowledge organisation are effectively applied by institutional repositories, at local, national and international levels.
Details
Keywords
Gonzalo Mochón, Eva M. Méndez and Gema Bueno de la Fuente
The purpose of this paper is to propose a methodology for assessing thesauri and other controlled vocabularies management tools that can represent content using the Simple…
Abstract
Purpose
The purpose of this paper is to propose a methodology for assessing thesauri and other controlled vocabularies management tools that can represent content using the Simple Knowledge Organization System (SKOS) data model, and their use in a Linked Open Data (LOD) paradigm. It effectively analyses selected set of tools in order to prove the validity of the method.
Design/methodology/approach
A set of 27 criteria grouped in five evaluation indicators is proposed and applied to ten vocabulary management applications which are compliant with the SKOS data model. Previous studies of controlled vocabulary management software are gathered and analyzed, to compare the evaluation parameters used and the results obtained for each tool.
Findings
The results indicate that the tool that obtains the highest score in every indicator is Poolparty. The second and third tools are, respectively, TemaTres and Intelligent Theme Manager, but scoring lower in most of the evaluation items. The use of a broad set of criteria to evaluate vocabularies management tools gives satisfactory results. The set of five indicators and 27 criteria proposed here represents a useful evaluation system in the selection of current and future tools to manage vocabularies.
Research limitations/implications
The paper only assesses the ten most important/well know software tools applied for thesaurus and vocabulary management until October 2016. However, the evaluation criteria could be applied to new software that could appear in the future to create/manage SKOS vocabularies in compliance with LOD standards.
Originality/value
The originality of this paper relies on the proposed indicators and criteria to evaluate vocabulary management tools. Those criteria and indicators can be valuable also for future software that might appear. The indicators are also applied to the most exhaustive and qualified list of this kind of tools. The paper will help designers, information architects, metadata librarians, and other staff involved in the design of digital information systems, to choose the right tool to manage their vocabularies in a LOD/vocabulary scenario.
Details
Keywords
Jane Greenberg, Eva Mendez Rodriguez and Gema Bueno de la Fuente
This editorial underscores the importance of linked data and linked open data (LD/LOD) in contemporary librarianship and information science. It aims to present the motivation for…
Abstract
Purpose
This editorial underscores the importance of linked data and linked open data (LD/LOD) in contemporary librarianship and information science. It aims to present the motivation for this special issue of Library High Tech (LHT), specifically the theme of linking and opening vocabularies (LOV) as a component of the LOD landscape. The editorial also seeks to describe the content of the papers and their contributions in the areas of current practice, trends, and R&D (research and development).
Design/methodology/approach
The editorial is organized into four main sections: Introductory aspects about linking open data; the context and inspiration for this special issue of LHT; a description of the accepted papers highlighting their contribution for LOV; and a conclusion.
Findings
The editorial presents several reflective insights for advancing LOD and LOV practices that may lead toward a more integrated semantic web.
Originality/value
The editorial presents a synthesis of the three editors' views on approaching the subject of linking open vocabularies. The framework for the editorial is reflected in the original call for papers that supported this special issue.
Details
Keywords
Daniel Vila-Suero and Asunción Gómez-Pérez
Linked data is gaining great interest in the cultural heritage domain as a new way for publishing, sharing and consuming data. The paper aims to provide a detailed method and…
Abstract
Purpose
Linked data is gaining great interest in the cultural heritage domain as a new way for publishing, sharing and consuming data. The paper aims to provide a detailed method and MARiMbA a tool for publishing linked data out of library catalogues in the MARC 21 format, along with their application to the catalogue of the National Library of Spain in the datos.bne.es project.
Design/methodology/approach
First, the background of the case study is introduced. Second, the method and process of its application are described. Third, each of the activities and tasks are defined and a discussion of their application to the case study is provided.
Findings
The paper shows that the FRBR model can be applied to MARC 21 records following linked data best practices, librarians can successfully participate in the process of linked data generation following a systematic method, and data sources quality can be improved as a result of the process.
Originality/value
The paper proposes a detailed method for publishing and linking linked data from MARC 21 records, provides practical examples, and discusses the main issues found in the application to a real case. Also, it proposes the integration of a data curation activity and the participation of librarians in the linked data generation process.
Details
Keywords
Jorge Luis Morato, Sonia Sanchez-Cuadrado, Christos Dimou, Divakar Yadav and Vicente Palacios
– This paper seeks to analyze and evaluate different types of semantic web retrieval systems, with respect to their ability to manage and retrieve semantic documents.
Abstract
Purpose
This paper seeks to analyze and evaluate different types of semantic web retrieval systems, with respect to their ability to manage and retrieve semantic documents.
Design/methodology/approach
The authors provide a brief overview of knowledge modeling and semantic retrieval systems in order to identify their major problems. They classify a set of characteristics to evaluate the management of semantic documents. For doing the same the authors select 12 retrieval systems classified according to these features. The evaluation methodology followed in this work is the one that has been used in the Desmet project for the evaluation of qualitative characteristics.
Findings
A review of the literature has shown deficiencies in the current state of the semantic web to cope with known problems. Additionally, the way semantic retrieval systems are implemented shows discrepancies in their implementation. The authors analyze the presence of a set of functionalities in different types of semantic retrieval systems and find a low degree of implementation of important specifications and in the criteria to evaluate them. The results of this evaluation indicate that, at the moment, the semantic web is characterized by a lack of usability that is derived by the problems related to the management of semantic documents.
Originality/value
This proposal shows a simple way to compare requirements of semantic retrieval systems based in DESMET methodology qualitatively. The functionalities chosen to test the methodology are based on the problems as well as relevant criteria discussed in the literature. This work provides functionalities to design semantic retrieval systems in different scenarios.
Details
Keywords
Arash Joorabchi and Abdulhussain E. Mahdi
This paper aims to report on the design and development of a new approach for automatic classification and subject indexing of research documents in scientific digital libraries…
Abstract
Purpose
This paper aims to report on the design and development of a new approach for automatic classification and subject indexing of research documents in scientific digital libraries and repositories (DLR) according to library controlled vocabularies such as DDC and FAST.
Design/methodology/approach
The proposed concept matching-based approach (CMA) detects key Wikipedia concepts occurring in a document and searches the OPACs of conventional libraries via querying the WorldCat database to retrieve a set of MARC records which share one or more of the detected key concepts. Then the semantic similarity of each retrieved MARC record to the document is measured and, using an inference algorithm, the DDC classes and FAST subjects of those MARC records which have the highest similarity to the document are assigned to it.
Findings
The performance of the proposed method in terms of the accuracy of the DDC classes and FAST subjects automatically assigned to a set of research documents is evaluated using standard information retrieval measures of precision, recall, and F1. The authors demonstrate the superiority of the proposed approach in terms of accuracy performance in comparison to a similar system currently deployed in a large scale scientific search engine.
Originality/value
The proposed approach enables the development of a new type of subject classification system for DLR, and addresses some of the problems similar systems suffer from, such as the problem of imbalanced training data encountered by machine learning-based systems, and the problem of word-sense ambiguity encountered by string matching-based systems.
Details
Keywords
Francisco Delgado Azuara, José Ramón Hilera González and Raul Ruggia
– This report aims to present the state of play of semantic interoperability problems in social security data exchanges.
Abstract
Purpose
This report aims to present the state of play of semantic interoperability problems in social security data exchanges.
Design/methodology/approach
The subject is presented as an open issue and taking into account the circumstances of the exchanges. Electronic exchanges in European Union social security are used as a case study.
Findings
Semantic level of these exchanges seems to be the weakness and the use of controlled vocabularies is proposed as possible solution. The creation and maintenance of metadata sets are finally considered as a compromise solution.
Originality/value
The article proposes a solution for the current semantic problems in electronic exchanges of social security information. The solution could be useful for social security institutions all over the world.
Details
Keywords
– The paper aims to investigate issues surrounding publishing thesauri as linked data and explores a systematic way for publishing linked open thesauri
Abstract
Purpose
The paper aims to investigate issues surrounding publishing thesauri as linked data and explores a systematic way for publishing linked open thesauri
Design/methodology/approach
Based on the four principles of linked data, a novel approach is proposed to incorporate terminology services with thesauri and provide more descriptive information about thesauri and relating services. A proof-of-concept system was built for demonstrating the approach.
Findings
Carefully designed URIs and terminology services are necessary for publishing linked open thesauri. However, there is still a lack of standards for describing linked open thesauri and terminology services. Since it is possible to publish thesauri at the hands of creators and maintainers, distributed thesauri may be a good complement to the traditional registry approach.
Research limitations/implications
Only a simple knowledge organisation system (SKOS) core model is considered in the paper. There may be some adjustment for applying the findings to thesauri using the SKOS extended model and other models.
Originality/value
The paper proposes a more systematic way for publishing linked open thesauri. It is important to consider a linked open thesaurus as a dynamic system rather than a static data dump. In addition, functionality for using thesauri is equally important for publishing thesauri.
Details
Keywords
Thomas Baker, Pierre-Yves Vandenbussche and Bernard Vatant
The paper seeks to analyze the health of the vocabulary ecosystem in terms of requirements, addressing its various stakeholders such as maintainers of linked open vocabularies…
Abstract
Purpose
The paper seeks to analyze the health of the vocabulary ecosystem in terms of requirements, addressing its various stakeholders such as maintainers of linked open vocabularies, linked data providers who use those vocabularies in their data and memory institutions which, it is hoped, will eventually provide for the long-term preservation of vocabularies.
Design/methodology/approach
This paper builds on requirements formulated more tersely in the DCMI generic namespace policy for RDF vocabularies. The examination of requirements for linked open vocabularies focuses primarily on property-and-class vocabularies in RDFS or OWL (sometimes called metadata element sets), with some consideration of SKOS concept schemes and Dublin Core application profiles. It also discusses lessons learned through two years of development of the linked open vocabularies (LOV), of which main features and key findings are described.
Findings
Key findings about the current practices of vocabulary managers regarding metadata, policy and versioning are presented, as well as how such practices can be improved to ensure better discoverability and usability of RDF vocabularies. The paper presents new ways to assess the links and dependencies between vocabularies. It also stresses the necessity and importance of a global governance of the ecosystem in which vocabulary managers, standard bodies, and memory institutions should engage.
Research limitations/implications
The current paper is focused on requirements related to a single type of vocabulary but could and should be extended to other types such as thesauri, classifications, and other semantic assets.
Practical implications
Practical technical guidelines and social good practices are proposed for promotion in the Vocabulary Ecosystem (for example, by vocabulary managers).
Originality/value
This paper brings together the research and action of several important actors in the vocabulary management and governance field, and is intended to be the basis of a roadmap for action presented at the Dublin Core conference of September 2013 in Lisbon (DC 2013).
Details