Most pictures shared online are accompanied by temporal metadata (i.e., the day and time they were taken), which makes it possible to associate an image content with real-world events. Maliciously ...manipulating this metadata can convey a distorted version of reality. In this work, we present the emerging problem of detecting timestamp manipulation. We propose an end-to-end approach to verify whether the purported time of capture of an outdoor image is consistent with its content and geographic location. We consider manipulations done in the hour and/or month of capture of a photograph. The central idea is the use of supervised consistency verification, in which we predict the probability that the image content, capture time, and geographical location are consistent. We also include a pair of auxiliary tasks, which can be used to explain the network decision. Our approach improves upon previous work on a large benchmark dataset, increasing the classification accuracy from 59.0% to 81.1%. We perform an ablation study that highlights the importance of various components of the method, showing what types of tampering are detectable using our approach. Finally, we demonstrate how the proposed method can be employed to estimate a possible time-of-capture in scenarios in which the timestamp is missing from the metadata.
The European Genome-phenome Archive in 2021 Freeberg, Mallory Ann; Fromont, Lauren A; D’Altri, Teresa ...
Nucleic acids research,
01/2022, Letnik:
50, Številka:
D1
Journal Article
Recenzirano
Odprti dostop
Abstract
The European Genome-phenome Archive (EGA - https://ega-archive.org/) is a resource for long term secure archiving of all types of potentially identifiable genetic, phenotypic, and clinical ...data resulting from biomedical research projects. Its mission is to foster hosted data reuse, enable reproducibility, and accelerate biomedical and translational research in line with the FAIR principles. Launched in 2008, the EGA has grown quickly, currently archiving over 4,500 studies from nearly one thousand institutions. The EGA operates a distributed data access model in which requests are made to the data controller, not to the EGA, therefore, the submitter keeps control on who has access to the data and under which conditions. Given the size and value of data hosted, the EGA is constantly improving its value chain, that is, how the EGA can contribute to enhancing the value of human health data by facilitating its submission, discovery, access, and distribution, as well as leading the design and implementation of standards and methods necessary to deliver the value chain. The EGA has become a key GA4GH Driver Project, leading multiple development efforts and implementing new standards and tools, and has been appointed as an ELIXIR Core Data Resource.
Graphical Abstract
Graphical Abstract
The European Genome-phenome Archive serves human genomics and health communities by offering data submission, validation, discovery, and access services to researchers worldwide. Data reuse value is maximised by implementing community standards and ensuring interoperability of data and tools to accelerate biomedical and translational research.
This paper is a survey of standards being used in the domain of digital cultural heritage with focus on the Metadata Encoding and Transmission Standard (METS) created by the Library of Congress in ...the United States of America. The process of digitization of cultural heritage requires silo breaking in a number of areas—one area is that of academic disciplines to enable the performance of rich interdisciplinary work. This lays the foundation for the emancipation of the second form of silo which are the silos of knowledge, both traditional and born digital, held in individual institutions, such as galleries, libraries, archives and museums. Disciplinary silo breaking is the key to unlocking these institutional knowledge silos. Interdisciplinary teams, such as developers and librarians, work together to make the data accessible as open data on the “semantic web”. Description logic is the area of mathematics which underpins many ontology building applications today. Creating these ontologies requires a human–machine symbiosis. Currently in the cultural heritage domain, the institutions’ role is that of provider of this open data to the national aggregator which in turn can make the data available to the trans-European aggregator known as Europeana. Current ingests to the aggregators are in the form of machine readable cataloguing metadata which is limited in the richness it provides to disparate object descriptions. METS can provide this richness.
Minimally, a research data repository exists to make a collection of data assets available to potential users. If a dataset cannot be discovered and found, it cannot be reused (Garnett et al. 2017). ...Harvestable metadata catalogues are a key strategy for achieving greater global findability of data assets, as they create a surveyable access point to discover data products within large data collections. Such catalogues can be especially effective if they are tailored for interoperability with feature-rich infrastructures (e.g. meta-catalogues, see Kapiszewski & Karcher 2020; CRFCB 2014) that are highly visible and widely used, and also themselves integrated within the larger ecosystem of research infrastructures. This study offers insight into a set of World Data System (WDS) research data repositories ongoing and successful implementations of harvestable metadata services, which apply established and emerging research data standards and practices to fit global, local and domain-specific interoperability contexts. Establishing a harvestable metadata service involves making choices in a space where standards and technologies are continuously evolving. The repositories in this study leverage the resources they have, within the policy and funding constraints of their institution, to serve the changing needs of heterogeneous user groups. This document encapsulates and completes the work that was carried out by the WDS International Technology Office (ITO) Harvestable Metadata Services Working Group (HMetS-WG).
Assessing metadata is of paramount importance for several critical reasons. Metadata plays a pivotal role in various aspects, including data retrieval and search, data organization, interoperability, ...data preservation, and the overall user experience. The purpose of this scoping review is to identify the most commonly measured dimensions of metadata quality in existing studies on metadata quality assessment. The study also investigates the types of data sources and countries contributing most to the literature on metadata quality assessment and the types of documents used to communicate their findings. The methodology involves the application of PRISMA model for qualitatively evaluating 55 studies on metadata quality assessment. The co-occurrence analysis is made on the title and abstract of selected articles using VOSviewer 1.6.18 version, visualization software. The review found that completeness, accuracy, consistency, accessibility, conformance, provenance, and timeliness are commonly used dimensions in metadata quality assessment. However, there is no consensus on their exact definition and measurement, indicating a need for further investigation into less commonly assessed quality dimensions. Digital repositories and open government data are the most commonly studied data sources, with the United States being the leading contributor and journal articles being the most commonly used document type. The cluster analysis based on co-occurrence of terms in title and abstract found three research areas, “Metadata Quality Assessment,” “Metadata Quality Dimensions,” and “Metadata Quality Applications, Frameworks, and Approaches” as prominent areas of research. The originality of the study lies in its methodology that involves rigorous screening of articles on metadata quality. It is a first attempt to qualitatively synthesize literature on metadata quality. The article emphasizes the importance of metadata quality research and the need to improve the flexibility of metadata quality assessment tools to facilitate better metadata quality assurance measures.
Purpose
The purpose of this study was to investigate metadata creation practices in a functional academic institution repository in Malawi, with a specific focus on the Lilongwe University of ...Agriculture and Natural Resources (LUANAR) library.
Design/methodology/approach
The study used a qualitative approach with a case study design. The study adopted a case study strategy that focuses on the in-depth, holistic and in-context examination of one or more cases. The researcher used non-probability purposive sampling to include all three LUANAR Digital Repository (LDR) staff at LUANAR library because they were thought to be knowledgeable about the LDR metadata work. The three library staff members directly involved in repository metadata were investigated for the study. Data collection techniques used in a case study approach included semi-structuring face-to-face interviews and documentary analysis. Data from interviews and documentary reviews were manually analyzed and presented in thematic categories based on the study’s objectives.
Findings
Qualified Dublin Core (DC) was chosen by all participants as the only metadata structure scheme that they will use to create and implement metadata in the repository. DC application profile was the only scheme used to enforce uniform naming and capitalization conventions in the application of Qualified DC element definitions. The scheme, however, was discovered to be the Qualified DC default format in the DSpace system. All participants indicated that the Agricultural Organization of the United Nations Vocabulary is used. Participants highlighted that institutional repository system compatibility, the subject matter of the resources, resource types and staff expertise influenced the selection criteria for the metadata schemes. The repository policy had been developed but had yet to be adopted by the LUANAR management.
Research limitations/implications
The current study was limited to LUANAR library. A wider study across public and private universities in Malawi is needed to ascertain the role of metadata policy, technical knowledge and metadata specialist institutional repositories.
Practical implications
Metadata policy is to aid in the understanding of the data, ensuring that appropriate security measures are used to protect the data and for metadata harvesting purposes.
Social implications
Academic libraries should lobby for management support towards metadata policy for institutional repositories.
Originality/value
Very little is known about challenges affecting the growth of institutional repositories and standards adopted, including metadata harvesting. This paper bridges the gap in metadata standards for institutional repositories in developing countries.