
Data Governance: Simple and Practical

(sondem/Shuttertock)
In 2022, the global data governance market grew to $3.2 billion, a figure that’s dwarfed by the significant cost to operate data governance programs. In spite of the spend, many programs fail to generate high business value because they lack practicality and simplicity: the art of maximizing the amount of work not done.
Where data governance programs often get off track is right before they get on, by defining the charter and roadmap too broadly. The key question is which elements of a data governance program should you chase out of the gate: ownership and stewardship? Committees and councils? Policies and standards? Data quality? Metadata or master data management? Data security and access? Issue management? Technology?
Choose the single element fundamental to fulfill the primary purpose of the program. To do this, answer two questions: what outcomes does the business need and what are the most urgent? From there, codify the primary purpose in your data governance charter and establish its pursuit as the first leg of your roadmap.
Focus on the Purpose and Postpone the Rest
Purpose-driven data governance programs narrow their focus to deliver urgent business needs and defer much of the rest, with a couple caveats. First, data governance programs are doomed to fail without senior executive buy-in and continuous engagement of key stakeholders. Without them, no purpose can be fulfilled. Second, data governance programs must identify and gain commitment from relevant (but perhaps not all) data owners and stewards, but that doesn’t necessarily mean roles and responsibilities need to be fully fleshed out right away.
Identify the primary purpose then focus on it – sounds like a simple formula, but it’s not obvious. Many data governance leaders are quick to define and pursue their three practices or five pillars or seven elements, and why shouldn’t they? They need those capabilities, but wanting it all comes at the sacrifice of getting it now. Generate business value with your primary purpose before expanding.
I’ve consulted with multiple customers whose different symptoms trace back to the same blight: a tech company that couldn’t get the trust in trusted data, an insurer that struggled to improve data quality, and a multinational financial services firm that couldn’t remain compliant.
Each approached data governance with a big bang approach instead of simplistically and practically.
Trusted Data
A tech company described their struggles to me this way: in quarterly business reviews cross-functional leaders couldn’t agree on the numbers, what metrics meant, where to source them, or how to interpret them. The primary purpose of their data governance program needed to be: to improve trust in the data. Their pathway looked like:
- Identify the critical data elements in dispute
- Negotiate and agree to data definitions; codify them in business glossaries
- Make those data elements available in source of truth datasets (if you show up to QBR with a different number, you’re wrong)
- Repipe widely used dashboards to the source of truth
Simplicity. The data governance elements missing from this pathway were everything that didn’t swiftly move the needle on trusted data. What use are policies without trust? Why improve the quality of data that no one understands?
Data Quality
An insurer explained to me their dashboards weren’t always refreshed, and when they were, wide fluctuations in values made it impossible to make informed decisions. Their primary purpose needed to be: to improve the quality of the data. Their pathway looked like:
- Identify unreliable dashboards and critical data elements
- Uncover the most adversely impacting data quality dimensions, for instance:
- Data Timeliness – Frequently failing pipelines without service-level agreements (SLAs) mean users can’t gauge data staleness
- Data Accuracy – Improper tooling to exchange data over a hybrid multi-cloud environment leads to data interpretation inconsistency
- Benchmark data quality measurements; put profiling and dashboarding in place
- Begin the search for the right hybrid multi-cloud technology vendor
- Improve data quality: define SLAs, reschedule pipelines, standardize data assets
Simplicity. Missing from the pathway were committees and councils, which might improve quality, but not as directly or quickly as attacking the root cause head on.
Regulatory Compliance
For a financial services firm, new regulations were inconsistently interpreted and wastefully implemented across the globe. Their primary purpose needed to be: to become truly compliant. Their pathway looked like:
- Establish data governance committees, councils, and working groups with clear responsibilities and accountabilities
- Document policies and standards around compliance definitions and implementation methods
- Implement and enforce
Remember, keep it simple, purposeful, and laser focused, and evolve as business needs change. That way you build your pillars while delivering business value continually.
About the author: Shayde Christian is the Chief Data & Analytics Officer for Cloudera, where he guides data-driven cultural change to generate maximum value from data. He enables Cloudera customers to get the absolute best from their Cloudera products such that they can generate high value use cases for competitive advantage. Previously a principal consultant, Shayde formulated data strategy for Fortune 500 clients and designed, constructed, or turned around failing enterprise information management organizations. Shayde enjoys laughter and is often the cause of it.
Related Items:
The Rise and Fall of Data Governance (Again)
Governance, Privacy, and Ethics at the Forefront of Data in 2021
Finding the Data Access Governance Sweet Spot
June 27, 2025
- EarthDaily Ignites a New Era in Earth Observation with Landmark Satellite Launch
- Domo Deepens Collaboration with Snowflake to Accelerate AI-Driven Analytics and Data Integration on the AI Data Cloud
- AIwire Launches Annual People to Watch Program
June 26, 2025
- Thomson Reuters: Firms with AI Strategies Twice as Likely to See AI-driven Revenue Growth
- DataBahn Raises $17M Series A to Advance AI-Native Data Pipeline Platform
- BCG Report: Companies Must Go Beyond AI Adoption to Realize Its Full Potential
- H2O.ai Breaks New World Record for Most Accurate Agentic AI for Generalized Assistants
- Foresight Raises $5.5M Seed Round to Bring Unified Data and AI to the Private Market
- Treasure Data Launches MCP Server: Let Your LLM Talk to Your Data
- Fujitsu Strengthens Global Consulting with Focus on AI, Data, and Sustainability
- HPE Expands ProLiant Gen12 with New AMD Servers
- Emergence AI Launches CRAFT for Natural Language Data Workflow Automation
- Overture Maps Launches GERS, a Global Standard for Interoperable Geospatial IDs, to Drive Data Interoperability
June 25, 2025
- Datadobi Launches StorageMAP 7.3 with Policy-Driven Workflow Automation
- Lucidworks Finds 65% of Companies Lack Foundation for Agentic AI
- OpenRouter Raises $40M to Scale Up Multi-Model Inference for Enterprise
- Gartner Predicts Over 40% of Agentic AI Projects Will Be Canceled by End of 2027
- Linux Foundation Launches the Agent2Agent Protocol Project
- Abstract Security Introduces LakeVilla for Scalable, Instant Access to Cold Data
June 24, 2025
- Inside the Chargeback System That Made Harvard’s Storage Sustainable
- What Are Reasoning Models and Why You Should Care
- Databricks Takes Top Spot in Gartner DSML Platform Report
- Snowflake Widens Analytics and AI Reach at Summit 25
- Top-Down or Bottom-Up Data Model Design: Which is Best?
- It’s Snowflake Vs. Databricks in Dueling Big Data Conferences
- Why Snowflake Bought Crunchy Data
- Change to Apache Iceberg Could Streamline Queries, Open Data
- LinkedIn Introduces Northguard, Its Replacement for Kafka
- The Evolution of Time-Series Models: AI Leading a New Forecasting Era
- More Features…
- Mathematica Helps Crack Zodiac Killer’s Code
- Solidigm Celebrates World’s Largest SSD with ‘122 Day’
- AI Agents To Drive Scientific Discovery Within a Year, Altman Predicts
- DuckLake Makes a Splash in the Lakehouse Stack – But Can It Break Through?
- ‘The Relational Model Always Wins,’ RelationalAI CEO Says
- Confluent Says ‘Au Revoir’ to Zookeeper with Launch of Confluent Platform 8.0
- The Top Five Data Labeling Firms According to Everest Group
- Supabase’s $200M Raise Signals Big Ambitions
- Toloka Expands Data Labeling Service
- Data Prep Still Dominates Data Scientists’ Time, Survey Finds
- More News In Brief…
- Astronomer Unveils New Capabilities in Astro to Streamline Enterprise Data Orchestration
- Yandex Releases World’s Largest Event Dataset for Advancing Recommender Systems
- Astronomer Introduces Astro Observe to Provide Unified Full-Stack Data Orchestration and Observability
- Seagate Unveils IronWolf Pro 24TB Hard Drive for SMBs and Enterprises
- BigID Reports Majority of Enterprises Lack AI Risk Visibility in 2025
- Databricks Unveils Databricks One: A New Way to Bring AI to Every Corner of the Business
- Snowflake Openflow Unlocks Full Data Interoperability, Accelerating Data Movement for AI Innovation
- Gartner Predicts 40% of Generative AI Solutions Will Be Multimodal By 2027
- Zscaler Unveils Business Insights with Advanced Analytics for Smarter SaaS Spend and Resource Allocation
- Databricks Donates Declarative Pipelines to Apache Spark Open Source Project
- More This Just In…