This release includes:

Bugfixes

  • Some fields were missing a standardConceptFieldName for the sourceValueCompleteness check, causing those checks to fail with an error. The missing field names have now been added
  • Many unit plausibility checks were missing legitimate units from the list of plausible units - these have now been updated
  • Sample results files in the shinyApps folder which had been causing issues for some users in rendering the Shiny app have now been removed
  • The fkDomain check has been disabled for the episode.episode_object_concept_id and measurement.value_as_concept_id fields, as these fields have multiple acceptable domains. The check can only currently support a single accepted domain
  • The withinVisitDates check was previously incorrectly categorized as a Conformance check; it has been recategorized as Plausibility

New executeDqChecks Parameter

There is now a parameter, checkSeverity, which can be used to limit the execution of DQD to fatal, convention, and/or characterization checks. Fatal checks are checks that should never fail, under any circumstance, as they relate to the relational integrity of the CDM. Convention checks are checks on critical OMOP CDM conventions for which failures should be resolved whenever possible; however, some level of failure is unavoidable (i.e., standard concept mapping for source concepts with no suitable standard concept). Characterization checks provide users with an understanding of the quality of the underlying data and generally will need their thresholds modified to match expectations of the source.

Documentation

  • We added 2 more check documentation pages - all DQ checks now have documentation! Check out the newly added pages here and please reach out with feedback as we continue improving our documentation!
  • We fixed a bug in the exclude checks sample code in CodeToRun.R

This release includes:

Bugfixes

  • Checks
    • plausibleStartBeforeEnd was failing if SOURCE_RELEASE_DATE was before CDM_RELEASE_DATE in the CDM_SOURCE table. This is the opposite of the correct logic! The check is now updated to fail if the CDM_RELEASE_DATE is before the SOURCE_RELEASE_DATE
    • plausibleTemporalAfter was throwing a syntax error in BigQuery due to the format of a hardcoded date in the SQL query. This query has now been updated to be compliant with SqlRender and the issue has been resolved
  • A dependency issue was causing viewDqDashboard to error out in newer versions of R. This has now been resolved
  • SqlOnly mode was failing due to the format of the new check plausibleGenderUseDescendants, which takes multiple concepts as an input. This has now been fixed

New Results Field

  • A new field has been added to the DQD results output - executionTimeSeconds. This field stores the execution time in seconds of each check in numeric format. (The existing executionTime field stores execution time as a string, making it difficult to use in analysis.)

Check Threshold Updates

The default thresholds for 2 checks were discovered to be inconsistently populated and occasionally set to illogical levels. These have now been fixed as detailed below.

  • The default thresholds for sourceValueCompleteness have been updated as follows:
    • 10% for _source_value columns in condition_occurrence, measurement, procedure_occurrence, drug_exposure, and visit_occurrence tables
    • 100% for all other _source_value columns
  • The default thresholds for sourceConceptRecordCompleteness have been updated as follows:
    • 10% for _source_concept_id columns in condition_occurrence, drug_exposure, measurement, procedure_occurrence, device_exposure, and observation tables
    • 100% for all other _source_concept_id columns

New Documentation

We have continued (and nearly completed) our initiative to add more comprehensive user documentation at the data quality check level. A dedicated documentation page is being created for each check type. Each check’s page includes detailed information about how its result is generated and what to do if it fails. Guidance is provided for both ETL developers and data users.

Check out the newly added pages here and please reach out with feedback as we continue improving our documentation!

This release includes:

New Checks

4 new data quality check types have been added in this release:

  • plausibleStartBeforeEnd: The number and percent of records with a value in the cdmFieldName field of the cdmTableName that occurs after the date in the plausibleStartBeforeEndFieldName.
  • plausibleAfterBirth: The number and percent of records with a date value in the cdmFieldName field of the cdmTableName table that occurs prior to birth.
  • plausibleBeforeDeath: The number and percent of records with a date value in the cdmFieldName field of the cdmTableName table that occurs after death.
  • plausibleGenderUseDescendants: For descendants of CONCEPT_ID conceptId (conceptName), the number and percent of records associated with patients with an implausible gender (correct gender = plausibleGenderUseDescendants).

The 3 temporal plausibilty checks are intended to replace plausibleTemporalAfter and plausibleDuringLife, for a more comprehensive and clear approach to various temporality scenarios. plausibleGenderUseDescendants is intended to replace plausibleGender, to enhance readability of the DQD results and improve performance. The replaced checks are still available and enabled by default in DQD; however, in a future major release, these checks will be deprecated. Please plan accordingly.

For more information on the new checks, please check the Check Type Definitions documentation page. If you’d like to disable the deprecated checks, please see the suggested check exclusion workflow in our Getting Started code here.

Check Updates

  • The number of measurements checked in plausibleUnitConceptIds has been reduced, and the lists of plausible units for those measurements have been re-reviewed and updated for accuracy. This change is intended to improve performance and reliability of this check. Please file an issue if you would like to contribute additional measurements + plausible units to be checked in the future
  • Some erroneous plausibleValueLow thresholds have been corrected to prevent false positive failures from occurring

New Documentation

We have begun an initiative to add more comprehensive user documentation at the data quality check level. A dedicated documentation page is being created for each check type. Each check’s page will include detailed information about how its result is generated and what to do if it fails. Guidance is provided for both ETL developers and data users.

9 pages have been added so far, and the rest will come in a future release. Check them out here and please reach out with feedback as we continue improving our documentation!

This release includes:

New Feature

A new function writeDBResultsToJson which can be used to write DQD results previously written to a database table (by setting writeToTable = TRUE in executeDqChecks or by using the writeJsonResultsToTable function) into a JSON file in the standard DQD JSON format.

Bugfixes

  • DQD previously threw an error if the CDM_SOURCE table contained more than 1 row. It has now been updated to select a random row from CDM_SOURCE to use for its metadata and warn the user upon doing this. Whether or not CDM_SOURCE should ever contain more than 1 row is still an unresolved discussion in the community. Either way, DQD should be allowed to run if the table has been improperly populated - and perhaps check(s) should be added for its proper use once a convention is finalized
  • Fixed additional field level checks (fkDomain, fkClass, plausibleTemporalAfter) to incorporate user-specified vocabDatabaseSchema where appropriate
  • Additional minor bugfixes & refactors

This release includes:

  • Minor documentation updates
  • A patch for an issue in one of DQD’s transitive dependencies, vroom
  • Test suite upgrades to run remote DB tests against OMOP v5.4, and to add Redshift to remote DB tests

This release includes:

Threshold file updates

The following changes involve updates to the default data quality check threshold files. If you are currently using an older version of DQD and update to v2.4.0, you may see changes in your DQD results. The failure threshold changes are fixes to incorrect thresholds in the v5.4 files and thus should result in more accurate, easier to interpret results. The unit concept ID changes ensure that long-invalid concepts will no longer be accepted as plausible measurement units.

  • The incorrect failure thresholds for measurePersonCompleteness and measureValueCompleteness were fixed in the v5.4 table & field level threshold files. This issue has existed since v5.4 support was initially added in March 2022
    • Many measurePersonCompleteness checks had a threshold of 0 when it should have been 95 or 100
    • Many measureValueCompleteness checks had a threshold of 100 when it should have been 0, and many had no threshold (defaulting to 0) when it should have been 100
    • The thresholds have now been updated to match expectations for required/non-required tables/fields
  • In the v5.2, v5.3, and v5.4 table level threshold files, measurePersonCompleteness for the DEATH table has been toggled to Yes, with a threshold of 100
  • In the v5.2, v5.3, and v5.4 concept level threshold files, all references to unit concept 9117 in plausibleUnitConceptIds have been updated to 720870. Concept 9117 became non-standard and was replaced with concept 720870, on 28-Mar-2022
  • In the v5.2, v5.3, and v5.4 concept level threshold files, all references to unit concepts 9258 and 9259 in plausibleUnitConceptIds have been removed. These concepts were deprecated on 05-May-2022

Bugfix

  • Call to new function convertJsonResultsFileCase in Shiny app was appended with DataQualityDashboard::. This prevents potential issues related to package loading and function naming conflicts

Some minor refactoring of testthat files and package build configuration and some minor documentation updates were also added in this release.

This release includes:

New features

  • New SQL-only Mode: Setting sqlOnly and sqlOnlyIncrementalInsert to TRUE in executeDqChecks will return (but not run) a set of SQL queries that, when executed, will calculate the results of the DQ checks and insert them into a database table. Additionally, sqlOnlyUnionCount can be used to specify a number of SQL queries to union for each check type, allowing for parallel execution of these queries and potentially large performance gains. See the SqlOnly vignette for details
  • Results File Case Converter: The new function convertJsonResultsFileCase can be used to convert the keys in a DQD results JSON file between snakecase and camelcase. This allows reading of v2.1.0+ JSON files in older DQD versions, and other conversions which may be necessary for secondary use of the DQD results file. See function documentation for details

Bugfixes

  • In the v2.1.0 release, all DQD variables were converted from snakecase to camelcase, including those in the results JSON file. This resulted in errors for users trying to view results files generated by older DQD versions in DQD v2.1.0+. This issue has now been fixed. viewDqDashboard will now automatically convert the case of pre-v2.1.0 results files to camelcase so that older results files may be viewed in v2.3.0+

This release includes:

New features

  • cohortTableName parameter added to executeDqChecks. Allows user to specify the name of the cohort table when running DQD on a cohort. Defaults to "cohort"

Bugfixes

  • Fixed several bugs in the default threshold files:
    • Updated plausible low value for specimen quantity from 1 to 0
    • Removed foreign key domains for episode object concept ID (multitude of plausible domains make checking this field infeasible)
    • Updated date format for hard-coded dates to YYYYMMDD to conform to SqlRender standard
    • Added DEATH checks to v5.2 and v5.3
    • Fixed field level checks to incorporate user-specified vocabDatabaseSchema and cohortDatabaseSchema where appropriate
  • Removed outputFile parameter from DQD setup vignette (variable not set in script)
  • Removed hidden BOM character from several threshold csv files, and updated csv read method to account for BOM character moving forward. This character caused an error on some operating systems

And some minor documentation updates for clarity/accuracy.

  1. Fixing bug in cdmDatatype check SQL that was causing NULL values to fail the check.
  1. Updating author list in DESCRIPTION.

This release includes:

Bugfixes

  • cdmDatatype check, which checks that values in integer columns are integers, updated so that float values will now fail the check
  • Quotes removed from offset column name in v5.4 thresholds file so that this column is skipped by DQD in all cases (use of reserved word causes failures in some SQL dialects)
  • Broken images fixed in addNewCheck vignette

HADES requirements

  • All snakecase variables updated to camelcase
  • Global variable binding R Check note resolved

This release includes:

New check statuses

  • Not Applicable identifies checks with no data to support them
  • Error identifies checks that failed due to a SQL error

New Checks

  • measureConditionEraCompleteness checks to make sure that every person with a Condition_Era record have a record in Condition_Occurrence as well
  • withinVisitDates looks at clinical facts and the visits they are associated with to make sure that the visit dates occur within one week on either side of the visit
  • plausibleUnitConceptIds identifies records with invalid Unit_Concept_Ids by Measurement_Concept_Id

outputFolder input parameter

  • The outputFolder parameter for the executeDqChecks function is now REQUIRED and no longer has a default value. This may be a breaking change for users who have not specified this parameter in their script to run DQD.

Removal of measurement plausibility checks

  • Most plausibleValueLow and plausibleValueHigh measurement values were removed from the concept check threshold files, due to feedback from the community that many of these ranges included plausible values and as such were causing unexpected check failures. An initiative is planned to reinterrogate these ranges and add them back once the team has higher confidence that they will only flag legitimately implausible values

Integrated testing was also added and the package was refactored on the backend

No material changes from v1.4, this adds a correct DESCRIPTION file with the correct DQD version

This release provides support for CDM v5.4 and incorporates minor bug fixes related to incorrectly assigned checks in the control files.

This fixes a small bug and removes a duplicate record in the concept level checks that was throwing an error.

This release includes additional concept level checks to support the OHDSI Symposium 2020 study-a-thon and bug fixes to the writeJSONToTable function. This is the release that study-a-thon data partners should use.

This is a bug fix release that updates how notes are viewed in the UI and adds CDM table, field, and check name to the final table.

This release of the Data Quality Dashboard incorporates the following features: - Addition of notes fields in the threshold files - Addition of notes to the UI - Functionality to run the DQD on a cohort - Fixes the writeToTable, writeJsonToTable functions

This is the first release of the OHDSI Data Quality Dashboard tool.