Data Preparation

How to cleanse data prior to importing data files.

Data Cleansing

Before running the service, implementors SHOULD review through the import data files for any explicit anomalies. Anomalies that result in invalid rows may contains records where a referenceIdentifier was provided, but did not contain one or more of the required fields such as first firstName, last name or sex.

A few examples of records that will not validate (anonymized for privacy purposes) include:

Example Records

Error Description

351233, Mark, Smith, M, 1899-12-31

Anomalous DOB format

525389,ZzAna, ZzGomez,F, 10/10/10

Addition of Zz may indicate duplicates

423452,TRUE,Sam, Ng, F, 1/22/94

TRUE in FirstName field

These types of records SHOULD be corrected manually or completely removed from the import data set.

The adapter's processing of the JSON schema file ensures these types they do not affect the end result and will indicate any field-type errors . Processing a file with invalid fields will return the following error and will not

Invalid records that need to be corrected before processing:
{525389, ZzAna, ZzMaria, ZzGomez, F, 10/10/10}

Entity-Specific Reference Identifiers

An entity-specific referenceIdentifier is required for every record. These reference identifiers MUST be unique per record and specific to the import file's system of record.

In the case that an import file does not have a referenceIdentifier available, an autogenerated field of type uint (unsigned 64 bit integer) or type id MUST be provided. The adapter will not process import files that will not. Providing an import file without a reference identifier will return the following error:

Invalid records that need to be corrected before processing:
{ , Jane, Emily, Doe, F, 09/22/95}

Last updated

Was this helpful?