cancel
Showing results for 
Search instead for 
Did you mean: 

Data Services - Scoring of Validation Transforms/Rules

Former Member
0 Kudos

I'm using BODS as a data quality tool. So my data flows looks like this :

I'm looking for the easiest way to generate a score for this validation rule. I'm aware that I can use the "Count" feature in Audit to store the count of Pass/Fail records and then computing the score. Isn't there another way to generate a score for validation rules in BODS ?

I found this "Score" check box inside the function/rule in the validation transform. What does this check box mean ? and where does it store the score ?

Accepted Solutions (1)

Accepted Solutions (1)

Former Member
0 Kudos

You use the score in the following case:

Prerequiste: multiple paramters/bindings in your validation function

--> In case of a failed row, you can configure to which column this error belongs to. Enable the score field for all columns, where the error should be counted to.

See Reference guide p.581/582

Former Member
0 Kudos

I just checked the reference guide. Thank you that was helpful but will not server what I'm looking for.

I'm looking for generating a score similar to the score that SAP Information Steward generates for each validation rule.

Former Member
0 Kudos

Thats gonna be a little more complicated. I would do the following:

Create an error_log table where you store all failed rules with the corresponding failed column(s) and an error_id. In a static error table, you store the score contribution of a certain error per error_id.

After the staging process you're able to produce some kind of dq/error reporting based on those two tables.

Cheers, @rogermathis

Former Member
0 Kudos

What do you mean by " In a static error table, you store the score contribution of a certain error per error_id." ? I do it myself or this score contribution is generated by the tool ?

Thanks again.

Former Member
0 Kudos

I mean a manually maintained static error table, where you store error messages, scores etc. I'm thinking of a datamodel like that:

D_ERROR (static error table)

error_id, error_msg, error_type, error_score, etc.

1, 'data validation failed','data validation error',4

LOG_ERROR (fact data, populated during ETL load)

job_id, error_id, error_column, validation_rule, error_table, dataflow, workflow, error_pk[1-5], error_values[1-5]

1,1,'ZIP','bad zip code','ADDRESS','DF_STG_ADDRESS','WF_STG_ADDRESS',3234234,'44x'

In a report, you join both tables and aggregate the data as you wish...

Actually, I just want to give you an idea of a concept. Feel free to adapt. It's matter of your requirement.

Personally, I always implement a generic error, dq logging and metadata concept surronding all Jobs, Workflows and Dataflows in order to be able to create a central and easy to use metadata, error and dq reporting.

Former Member
0 Kudos

Thanks for the idea. Manually is very hard because I'll have +1000 validation rules

I thought BODS does something like this :

If 10 records out of 100 failed, then it outputs the score for this validation rule as 9 out of 10 automatically. (Like Information Steward). It looks like I cannot accomplish that with BODS.

Former Member
0 Kudos

Hi,

I'm sure that information is retained in the DS repository tables - let me have a look later today,

regards,

Adrian

Former Member
0 Kudos

Hi,

have a look at the tables AL_QD_VRULE - lists the rules and AL_QD_STATS - contains the results of the individual runs.

These will give you your percentage pass/fail figures,

regards,

Adrian

Former Member
0 Kudos

Thanks Adrian

This is what I was looking for.

I wonder why don't Data Services expose this kind of information in the tool itself without the need to go to the Repo and do some extra work to extract the data.

Answers (0)