Community
In today's era of quick e-commerce, ordering anything is just a click away. Imagine ordering a crate of your favorite beer on a bright Sunday afternoon, and getting delivered in just 20 minutes. Upon arrival, you find the crate is broken, but fortunately, the cans inside are intact. You probably will not complain much. What if besides broken crate, few cans had dents too. You will be agitated but once you have your favorite beer chilled to perfection, you will probably be more considerate about the broken crate and the dented cans.
Now, consider a different scenario where the crate arrives in perfect condition with intact cans, but the taste of the beer is off and unpleasant. Which scenario would you prefer? Most likely, you would be more forgiving in the first scenario, as the quality of the beer saved the day, despite the broken crate and a few dents. However, in the second scenario, no good packaging could compensate for the compromised quality of the beer.
Let's apply this analogy to data quality. Here, the crate represents data storage, typically data warehouses and Data Marts, the cans are the tables and columns, and the beer is the data itself. You can tolerate a less-than-optimal data warehouse and table structure if the data is accurate and business is not complaining. In the second scenario, you have the latest cloud data storage, and tables and columns are replaced with data products, but the data itself is questionable. How would you react?
Common sense dictates that we should focus on fixing the beer and the data, as they are the main heroes of the story. Unfortunately, we often concentrate on fixing the crates and cans or data warehouses and data structures. While addressing these aspects is important, it will not save us from embarrassment, fines, and penalties if a regulator comes knocking on our door or business complains about loss in revenue and operational inefficiencies.
This explains why many analysts estimate the cost of fixing bad data quality for an organization to be meagre USD 14 to 20 million. This cost is for fixing the crates and cans not Beer and Data. This raises a critical question: why would any organization invest efforts to solve data quality issues when 14 million on a balance sheet do not even appear or appear as the last two decimal places? If we really consider the impact of bad beer and bad data, we will have dissatisfied customers leading to fall in customer and business satisfaction, frequent returns and manual reconciliations, dip in revenue and on top of it, FDA and Fed knocking on your doors.
Why is data quality misunderstood?
Although quality can be somewhat nebulous, it is fundamentally significant. Data quality is a complex and context-dependent concept often misunderstood across business, technology, process, and data science domains, with each attributing different issues to it. Numerous studies have squarely blamed low AI adoption on poor data quality. Before seeking a solution, it is crucial to understand what everyone means by quality.
The discussion raises critical questions about whether data quality equates to data reliability, if it unfairly bears blame for broken processes or managerial conflicts, and whether tools or process fixes alone can resolve it. Ultimately, true data quality encompasses understanding the context, intent, and dimensions of data, suggesting it is the sum total of all data management domains rather than a single isolated aspect
Diverse Perspectives on Data Quality
Each executive views data quality issues through their unique lenses:
To make data usable, we need to understand the context, intent and dimensions of data
Applying data governance principles:
Effective data quality management requires integrating all domains of data governance rather than treating data quality as an isolated issue.
Identify, Understand, and Catalog Your Core Data Assets: As explained in our analogy, it is important to identify the hero, data in this case, and clearly articulate the intent, context and dimensions of quality to be monitored.
Beyond Paper Ownership: Ownership does not just mean assigning a name to a data asset. It involves a deep emotional attachment to the assets one owns and ensuring that they are always in the best condition.
Standards, Policies, Procedures: What gets measured gets monitored. Following this logic, it is important to define standards, policies, and procedures for your data. However, this does not mean copying them from various sources. They should be contextual to your business.
Guardrails and Controls: It is crucial to establish guardrails and controls to protect your data and adhere to standards and policies in the context of business and regulations.
Processes: Processes must cover the end-to-end data lifecycle to ensure quality of data at all the stages and the workflows must support the processes.
Operating Model: Ultimately, it is people who will ensure the quality of data is protected through processes and technology. This is where an operating model comes into play. It is not a bureaucratic setup but a group of people who genuinely care about data.
In a nutshell, we must love our data exactly the way we love our beer. Fixing the quality of beer is of utmost importance, and for that it is important to find out where in the supply chain management the quality of Beer went wrong. Whether there was an issue in sourcing ingredients, brewing, packaging, distribution, or retail. Effective management of this chain is crucial for breweries to meet demand, minimize costs, and maintain product. Similarly, for fixing and maintaining quality of data, it’s important to map the e2e data lifecycle to understand the changes and transformations happening on the data through lineage, traceability, auditability, or provenance.
Poor data quality is far more debilitating than a few million dollars on Balance sheet. Hence Data Quality must be embraced as a way of life to enable organization to support business strategy and regulatory compliance.
This content is provided by an external author without editing by Finextra. It expresses the views and opinions of the author.
Nauman Hassan Director at Paymentology
23 hours
Joris Lochy Product Manager at Intix | Co-founder at Capilever
08 September
Sandeep Hinduja Vice President & Head of Banking (US) at Newgen Software Inc.
05 September
John Bertrand MD at Tec 8 Limited
Welcome to Finextra. We use cookies to help us to deliver our services. You may change your preferences at our Cookie Centre.
Please read our Privacy Policy.