September 10, 2021

A Top 10 US Financial Services Organization

Event Date:
Hosted By:
Register Now

A Global 1000 Financial Services organization moves credit risk information from 26 different systems to AWS. Data goes through a number of normalization and transformation processes before being pushed to a AWS S3 based data lake. A reporting solution is then used to provide various regulatory reports to the related groups.

The customer wanted to track over 2200 segments of credit risk data based on the country, product type and risk code combination. The data had to be validated as soon as it was ingested into S3. They wanted to flag the following types of data errors stemming from either the source system changes or transformation errors.

• Invalid Country-Product-Risk score combination. If a new combination is created (because of source system changes or transformation issues), downstream stake holders need to be altered. In addition, customer wanted to flag any new country code, product code or risk code.

• Abnormal changes in the number of transactions in each of the 2200+ Country-Product-Risk Code based segments. Anomalous credit risk transactions within each of the 2200+ credit risk segments.

• Anomalous changes in the sum of the credit risk in each segment .

• Technical Data Validation Checks ( Null value percentage for each segment ), duplicate records etc.

Data Sentinel Data Quality’s ML algorithms autonomously identified over 2200 credit risk subsegments from the banks transaction dataset and created data fingerprints to track each risk segment individual.

Data Sentinel autonomously constructed detailed data quality fingerprints for monitoring and validating transactions. This was equivalent to over 22,000 traditional Data Quality rules. These DQ rules were self- learning and evolved with the evolution in data over time.

Within 2 weeks of deployment, Data Sentinel identified 31 errors that the SME’s had not anticipated (“unknown- unknowns”) and so the existing data quality solution did not catch. The hidden risk of bad and inconsistent data propagating downstream was minimized.

September 10, 2021

A Top 10 US Financial Services Organization

Date:
Hosted By:
Register Now

A Global 1000 Financial Services organization moves credit risk information from 26 different systems to AWS. Data goes through a number of normalization and transformation processes before being pushed to a AWS S3 based data lake. A reporting solution is then used to provide various regulatory reports to the related groups.

The customer wanted to track over 2200 segments of credit risk data based on the country, product type and risk code combination. The data had to be validated as soon as it was ingested into S3. They wanted to flag the following types of data errors stemming from either the source system changes or transformation errors.

• Invalid Country-Product-Risk score combination. If a new combination is created (because of source system changes or transformation issues), downstream stake holders need to be altered. In addition, customer wanted to flag any new country code, product code or risk code.

• Abnormal changes in the number of transactions in each of the 2200+ Country-Product-Risk Code based segments. Anomalous credit risk transactions within each of the 2200+ credit risk segments.

• Anomalous changes in the sum of the credit risk in each segment .

• Technical Data Validation Checks ( Null value percentage for each segment ), duplicate records etc.

Data Sentinel Data Quality’s ML algorithms autonomously identified over 2200 credit risk subsegments from the banks transaction dataset and created data fingerprints to track each risk segment individual.

Data Sentinel autonomously constructed detailed data quality fingerprints for monitoring and validating transactions. This was equivalent to over 22,000 traditional Data Quality rules. These DQ rules were self- learning and evolved with the evolution in data over time.

Within 2 weeks of deployment, Data Sentinel identified 31 errors that the SME’s had not anticipated (“unknown- unknowns”) and so the existing data quality solution did not catch. The hidden risk of bad and inconsistent data propagating downstream was minimized.

Let's talk

Ready To Discuss Your Data Challenges?

you may also like

Blog

What is Data Ethics?

Data may be utilized to make decisions and have a large influence. However, this valuable resource is not without its drawbacks. How can businesses acquire, keep, and use data in an ethical manner?

News

Susan Kirk joins Data Sentinel as SVP of Customer Success

Data Sentinel enhances its customer support capability with the addition of a recognized industry leader.

Webinar

Demystifying the De Identification of Data - How to Protect your Organization’s Data

Kevin Downey, Chief Technical Officer at Data Sentinel, Liana Di Giorgio, Senior Associate at Norton Rose Fulbright, and Patricia Thaine, MSc, CEO and Co-founder of Private.ai discuss the art of de-identification of data.