Any good coach will tell you that these three factors are key to success in sport. Quick ball, accurate passes, consistent defence are the keys to victory.
Similarly, for your data quality processes these three are key.
Does your data quality solution ensure accurate data, irrespective of how or when it entered the environment!
It is no longer good enough to rely on a batch/etl process to cleanse data as it moves between systems.
For many organisations this means that source data can never be cleaned – as legal, political or system complexities inhibit the ability to cleanse at source. These companies are turning to real time data cleansing platforms that validate, correct and identify duplicate records before it is captured into the database.
If you fix the problem before it enters the source system the complexities inhibiting cleansing are largely removed.
Does your data quality solution support the reuse of business rules and data cleansing processes across your enteprise?
Can a business data steward define business rules in an easy to use interface and share those rules for deployment to your ETL processes, your web services, your ERP, your MDM applications and your legacy.
Can this be done without redevelopment and the possibility of errors or misinterpretations. The only way to ensure that data is handled consistently (and therefore to retain accuracy) is to use the same rules everywhere.
Speed of Execution
Data validations must not impact your users – they should not have to wait for data validations to occur.
If the solution selected requires exorbitant hardware support to run quickly in your environment, or cannot scale to handle your data volumes then it cannot solve your problem.
Has the solution you are looking at been specifically designed for high speed execution with a minimal hardware footprint. If not, you should consider an alternative.