The process by which original, source data is inputed by a human into a digital, structured format for later use and analysis.
Terms
38 resultsThe control of data entry and maintenance to ensure the data meets defined standards, qualities, or constraints.
Systematic combining of data from different independent and potentially heterogeneous sources, to create a more compatible, unified view of these data for research purposes
Entire period of time that data exists. This lifecycle describes the flow of data starting from planning, collecting, processing, analysing, preserving, sharing and finally reusing the data.
Statement describing how data will be managed throughout a specified life cycle including terms regarding archiving and potential preservation of the data in a data repository. The plan is considered to be a 'living' document, i.e. one which can be updated when necessary.
A person primarily responsible for database creation, data collection, validation, integration, coding, review, quality evaluation, and archiving, as well as for preparation of reports and data displays.
The process of transporting data between computers, storage devices, or formats.
The ability of information to be accessed and exchanged between systems
Set of high-level principles that establish a guiding framework for data-related issues.
Data preparation is the process of manipulating (or pre-processing) data (which may come from disparate data sources) to improve their quality or reduce bias in subsequent analysis.
A collection of one or more data items that are packaged for or by a specific application.
Detailed description of a dataset or dataset series together with additional information that will enable it to be created, supplied to and used by another party
The statistical analysis and assessment of the quality of data values within a dataset for consistency, uniqueness and logic.
An agent or agents that produces data or reference metadata
Process whereby data are subjected to an assessment process to determine whether they should be acquired by a repository; followed by a rigorous acquisition and ingest process that results in products being publicly made available and supported for the long-term by that repository.
The reliability and application efficiency of data.
Role within an organization responsible for ensuring metadata quality, integrity, and access in a manner that is consistent with policies and practices established through data governance to ensure the data is fit for reuse.
A process whereby data is changed from one format or structure to another.
Process used to determine if data are inaccurate, incomplete, or unreasonable.
An identifiable collection of data
An entity is a physical, digital, conceptual, or other kind of thing with some fixed aspects; entities may be real or imaginary.
A planned data integration process that extracts concretizations of information from a data source, prepares and transforms them, then loads the transformed concretizations into a target data store.
The FAIR principles are guidelines to improve the Findability, Accessibility, Interoperability, and Reuse of digital assets. The principles emphasise machine-actionability (i.e., the capacity of computational systems to find, access, interoperate, and reuse data with none or minimal human intervention) because humans increasingly rely on computational support to deal with data as a result of the increase in volume, complexity, and creation speed of data. The principles refer to three types of entities: data (or any digital object), metadata (information about that digital object), and infrastructure.
The thing being sampled or transformed in an act of sampling.
User-interface element in which data is entered or presented
A type of data transformation that preserves the content of a dataset while converting its representation format.
The completion of production of a new entity by an activity.
A grouping of entities that share a common set of characteristics in a particular context.
Designates an assigned code that is uniquely and meaningfully defined only in the local source system or within the scope of an internal management system.
Data that defines and describes other data
A self-contained component (unit or item) that is used in combination with other components.
A Web-based resources that preserve, manage, and provide access to many types of digital materials in a variety of formats.
A proposition that states a set of conditions that have to be fulfilled to achieve some specific objective.
The capability of different information systems to communicate information consistent with the intended meaning, which encompasses the data structure elements involved, the terminology appearing as data and the associated identifiers
A planned process with objective to check that the accuracy or the quality of a claim or prediction satisfies some criteria and which is assessed by comparing with independent results
An articulation that is by social convention understood to be potentially assignable to a variable
A software pipeline that is composed of a wide array of software algorithms to process biological raw data.
Full set of logically connected steps needed to transform input data into a result, for which one or more outputs of one data transformation step can be the input of one or more subsequent steps