1. Data quality refers to the level of quality of data.
2. The fitness for use of information; information that meets the requirements of its authors, users, and administrators. (Martin Eppler)
Data Quality and Quality Criteria
Data, datasets, descriptive data and metadata have to meet "quality criteria, such as accuracy, completeness, consistency, and actuality. [...] Additionally, Open Data credibility and transparency are crucial in order for Open Data to be trustworthy. Origin, originality, and data changes must be traceable and quality features testable." (Fraunhofer FOKUS, see also here)
Data Quality in the open energy database (oedb)
The database set-up is designed to support users in achieving good data quality:
- The interfaces (API and web interface) of the database structure the data and the metadata. In this way the database tries to ensure for instance the submission of all relevant metadata belonging to a dataset. Another example is the automatically generated primary key, which ensures that every row in a table has got an unique identifier. For further information see also DatabaseRules.
- Plausibility and integration tests are applied to identify mistakes in the data, e.g. the geometrical correctness test which makes the manual testing with SQL-functions obsolete.
- When the number of users becomes large enough, a third level of data testing and rating based on user evaluations might be implemented.
Users can contribute to data quality for instance by following the DatabaseRules. For further reading and guidelines with regard to data management and publication see for instance: Open Knowledge Foundation, Open Data Foundation and Software Carpentry (e.g. here).