Execution flow enables a Provider Data Scientist /Business Analyst to design the Data Pipeline in a specific sequence


Note : The other execution options shown in the below screenshots are not available for GCP environment.

File Validation

This feature enables a Provider Data Scientist /Business Analyst to validate the raw input file ingested into the Bristlecone NEO® Data Lake.


    • Empty File Check : Validates if the raw input file is empty
    • Missing Column Header : Validates the raw input file for columns without a header
    • Duplicate Column Check : Validates the raw input file for duplicated columns

Configuring File Validation

Step 1 : Click on File Validation button as shown below



Step 2 : Click on File Validation tab as shown below :



Step 3 : Configuring File Validation Details

    • Stop on Failure : This feature stops the execution of the Data Pipeline when the File validation fails
    • Empty File Check : This feature checks if the input file is empty
    • Missing Column Header Check : This feature checks for columns which do not have headers
    • Duplicate Column Check : This feature check if there are any columns duplicated in the input file

Post Data Pipeline Execution , click on Data Pipeline Execution Details to receive a pop up. Expand File Validation Step to find the dictionary validation details


Note : User can enable /disable these toggle buttons as per the requirement


Dictionary Validation

       This feature enables a Provider Data Scientist /Business Analyst to validate the raw input file and stop the data pipeline from triggering, if the file turns to be invalid based on the following conditions.


S.No

Feature Based Test

Status

Comments

1

Extra Column Check

PASS

No extra column

2

Extra Column Check

FAIL

Input file has columns which are not defined in dictionary

3

Missing Column Check

PASS

No missing column

4

Missing Column Check

FAIL

Input file has one / more missing columns

5

Null Value Check

PASS

No nullable values found in the non-nullable columns

6

Null Value Check

FAIL

Null values found in the non-nullable columns

7

Allowed Value Check

PASS

A column has only allowed values and (nulls if marked as nullable)

8

Allowed Value Check

FAIL

A column has values other than allowed values

9

Unique Value Check

PASS

No duplicates

10

Unique Value Check

FAIL

Added duplicates entries found in the input file

Note: Applicable for columns marked as Unique in the Data Dictionary

11

Data Type Check

PASS

No difference in Data Types

12

Data Type Check

FAIL

Added different data type included which has not been mentioned in the Data Dictionary

13

Scale and Precision Check

Not Implemented

Will show status message as SKIPPED

14

Value Check

Not Implemented

Will show status message as SKIPPED

15

Length Check

PASS

Length of a given attribute in the input file matches the length specified in the associated Data Dictionary

16

Length Check

FAIL

Length of a given attribute in the input file is a mismatch with the length specified in the associated Data Dictionary


Configuring Data Dictionary Validation

Step 1 : Click on Dictionary Validation button during the creation/updating the Data Pipeline as shown below :


Step 2 : Click on Dictionary Validation tab as shown below :


Step 3 : Configure Dictionary Validation Details:

Dictionary Validation Details contain the following toggle buttons that allow the user to enable or disable the following features

    • Stop on Failure : This feature stops the execution flow of the Data Pipeline when the raw input data file of the specific data pipeline fails during Data Dictionary Validation
    • Extra Column Check : This feature compares the input file with the associated Data Dictionary and checks for extra columns in the input file
    • Data Type Check : This feature checks for columns and their associated data types as per the specific data dictionary connected to the input file
    • Unique Value Check : This feature checks if the data dictionary has unique values

      Note : Unique Value Check is case sensitive

    • Allowed Value Check : This feature checks for allowed values for various columns in an input file

      Ex : A column called currency has all the values in $ , upon choosing allowed values for the currency column user can have other currency formats such as INR , dinars ,euros, pounds.

    • Length Check : Checks for the length of the value in all the columns mentioned in the associated Data Dictionary for the specific input file

      Note : Applicable to selected data types such as : String, Character, Integer

    • Missing Column Check : Compares the input file with the associated Data Dictionary and checks for the missing columns [ if any ] in the input file
    • Null Value Check : Checks if null values are allowed for a specific file as per the associated data dictionary

Steps to configure Dictionary Validation

Step 3a : Select the specific dictionary associated with the input file as shown below :


Step 3b : Select the toggle buttons to apply various validations on the associated dictionary of the specific input file as shown below:

The screen capture below shows a few of the toggle buttons activated for illustration purposes



Step 4 : Post Dictionary Configuration, click on Create button to complete Dictionary Validation flow


Serverless Custom Job

Enables a Provider Data Scientist to execute a pre-configured AWS glue script as a part of the Data Pipeline Execution Flow

Post Data Pipeline creation, click on Serverless Custom Job as shown below


Click on Serverless Custom Job tab to add Jobs as shown below