Best Practice Workflows: Data Integration
  • 28 Aug 2023
  • 2 Minutes to read
  • Dark
    Light

Best Practice Workflows: Data Integration

  • Dark
    Light

Article Summary

A Best Practice Workflow (BPW) is a practical guide to achieving some broad goal in Ursa Studio. The types of activities covered are sufficiently open-ended and heterogenous that a fully prescriptive, cookbook-style guide could never be complete; instead BPWs provide detailed guidance on best practices generally relevant to the activity, as well as the background and rationale supporting that guidance.

This documentation is intended for both new and experienced Ursa Studio users, though as teams become more fluent with Ursa Studio they will likely find superior variations and innovations beyond the practices listed here that are better suited to their local working and data environments.

Each Workflow is broken down into a number of Tasks, listed below. Generally, the ordering of the Tasks within their respective Workflow reflect the recommended chronological ordering of the work.

Accessing Source Data Workflow

  1. Gather information on the source data
  2. Determine appropriate storage locations for flat files
  3. Generate a Source ID value for each source system
  4. Determine the appropriate Namespace for integration assets
  5. Create Registered Table objects
  6. Create Import objects and perform initial loads
  7. Configure the object visibility settings
  8. Create data exploration objects
  9. Validate the source data
  10. Determine the grain size of each source data object

Semantic Mapping Workflow

  1. Determine what Semantic Mapping objects are needed
  2. Identify the Natural objects impacted by each source data object
  3. Determine the mappings for identifier fields
  4. Determine the mappings for claims or billing transaction fields
  5. Checkpoint to review the interpretation of source data
  6. Create or extend Semantic Mapping Templates
  7. Create and run Semantic Mapping objects
  8. Validate the semantic mapping results

Data Mastering Workflow

  1. Determine which concepts require mastering
  2. Determine which source tables contain information needed for mastering
  3. Determine the combinations of fields that should trigger a merge
  4. Configure and run the patient and provider mastering objects
  5. Create and run mastering objects for other concepts
  6. Validate the data mastering results
  7. Checkpoint to review the implementation of semantic mapping and data mastering

Creating Local Transform Objects Workflow

  1. Determine the sequence of Local Transform objects needed for each Natural object
  2. Create and run Local Transform objects
  3. Validate the Local Transform object results

Connecting to the Natural Object Layer

  1. Add new fields to existing Natural objects
  2. Create Natural objects for new source system concepts
  3. Connect upstream objects to Natural objects
  4. Review and add validation rules to Natural objects
  5. Run Natural objects and validate results
  6. Run downstream assets and review data diagnostic measure results
  7. Validate aggregate measure results against a gold standard
  8. Perform case review on a random sample of patients
  9. Checkpoint to review end-to-end integration results
  10. Determine which Semantic Mapping and Local Transform objects should be views
  11. Clean up integration assets

Was this article helpful?