by DL Keeshin
January 26, 2026
In my last post, I talked about the kDS BETA program and what organizations can expect from a seven-phase implementation. Today, I want to step back from the practical details of deployment and explore the theoretical foundation that makes kDS fundamentally different from traditional data governance tools.
What if I told you that most enterprise data tools are missing the most important context about your data—not what it is or where it flows, but why it exists, what it means to your business, and what problems it's supposed to solve? This is the insight that led to the development of what I call the Organizational Data Flow Architecture approach—a data management methodology that transforms how organizations understand and work with their data.
While developing kDS, I kept returning to a fundamental question: Are there any established methodologies that propose a "complete data flow topology map overlaid on organizational structure"?
After extensive research, I discovered something surprising—there isn't one. At least, not in the way that addresses the real challenges enterprise organizations face with data governance.
The enterprise data landscape has several well-established approaches, each valuable but incomplete for comprehensive data management:
Data Flow Diagrams (DFD) represent the classic structured analysis method. They show processes, data stores, and flows—but they don't integrate corporate or legal hierarchy. The focus is purely on technical systems, not organizational structure.
Data Lineage tools have become highly developed in enterprise data catalogs like Collibra, Informatica, and Alation. They track data's journey from source through transformation to destination. But they don't map to parent-subsidiary-division-role hierarchies. The focus remains on technical lineage, not organizational boundaries.
RACI matrices in data governance map Responsible, Accountable, Consulted, and Informed roles to data governance activities. They link roles to activities—but they don't show data flow topology. It's a static responsibility assignment, not a dynamic representation of data movement.
IBM's Data Topology Framework introduces concepts of data "zones" and layers spanning cloud, edge, and device deployments. But again, it focuses on technical deployment topology without integrating organizational legal structure.
These are primarily governance and documentation tools. What's missing is a data management approach that helps organizations understand not just where data is, but why it exists and what business problems it's intended to solve.
What became clear is that we needed something new—a synthesis of corporate and legal hierarchy with data flow topology, discovered through structured engagement with the people who actually work with data every day.
This hybrid framework combines data lineage concepts (technical flow), organizational hierarchy (legal and operational structure), data governance (ownership and accountability), and business process mapping (who uses what data for what purpose).
The closest existing concepts might be Process Mining, which discovers actual business process flows from event logs, or Organizational Network Analysis, which maps informal communication and data flows. Enterprise Architecture frameworks like TOGAF and Zachman touch on these ideas but focus more on system architecture than data flow mapped to organizational meaning.
Here's the key insight: mapping data flows to organizational architecture doesn't just explain what data exists—it explains why it exists, what it means to the organization, and what business problems it's designed to solve.
Traditional data tools answer important questions about what data exists (data catalogs), where data flows (lineage tools), and how data transforms (ETL documentation). But they miss the crucial context of why the data flow exists (business purpose), what it means (organizational context), who depends on it (stakeholder impact), and what problems it solves (business value).
When you map data to organizational hierarchy, suddenly everything gains context:
At the Level 1 - Parent Company, data serves consolidated reporting, board governance, and investor relations. This is "official company data"—audited, regulated, public-facing. The stakes are highest here because this data represents the company to the outside world. Problem being solved: providing accurate, auditable information to shareholders, regulators, and the board for strategic decision-making and compliance.
At the Level 2 - Subsidiary/Division, data flows serve legal compliance, define regulatory boundaries, and manage liability protection. Data crossing subsidiary boundaries has profound legal implications—inter-company transactions, transfer pricing, data sovereignty requirements. A transaction moving from one subsidiary to another isn't just a technical data transfer; it's a legal event with tax and regulatory consequences. Problem being solved: maintaining legal separation between entities, managing liability exposure, and ensuring compliance with jurisdiction-specific regulations.
It's worth noting that we've found some organizations structure their subsidiaries around major products or portfolios rather than purely functional or geographic lines. For example, a company might have separate legal entities for different product lines—allowing each to operate with its own P&L, risk profile, and regulatory requirements. This product-based subsidiary structure makes understanding cross-subsidiary data flows even more critical, as data about shared customers, common suppliers, or consolidated financials crosses legal boundaries between what are essentially competing or complementary product families.
At the Level 3 - Business Unit, data represents actual business operations—P&L accountability, strategic initiatives, revenue and costs, customer relationships. This is where business strategy becomes visible in data patterns. Problem being solved: enabling operational management, measuring business unit performance, and supporting tactical decision-making for profitability and growth.
At the Level 4 - Role, data is created, validated, and understood by the people who work with it daily. This is the source of truth, where data quality is determined and business context is richest. Problem being solved: enabling day-to-day operations, supporting frontline decisions, and capturing the operational reality that feeds all higher levels.
To make this concrete, let's examine how the kDS platform works with sample data that ships with every deployment. Newco Foods is a food manufacturing company with two subsidiaries: Food Manufacturing and Food Safety. Each subsidiary has multiple business units, and each unit has roles that create and use data.
Consider production quality data. Without organizational context, the story is purely technical: "Quality metrics flow from production floor systems to quality database to executive dashboard."
With organizational architecture mapping, the same data flow reveals its business meaning and the problems being solved:
The Quality Assurance Technician (Level 4 - Role) in the Quality Control business unit captures inspection data because they're ensuring product meets safety standards. This data represents the moment a batch passes or fails quality checks—a critical decision point for food safety. Problem being solved: preventing unsafe products from reaching consumers and maintaining lot-level traceability for potential recalls.
The Quality Control business unit (Level 3 - Business Unit) aggregates daily quality metrics for trend analysis and compliance reporting. This ensures the subsidiary maintains its quality certifications and meets regulatory requirements. Problem being solved: maintaining FDA compliance, identifying quality trends before they become critical issues, and supporting continuous improvement initiatives.
When quality data moves from the Food Safety subsidiary to the Food Manufacturing subsidiary (Level 2 - Subsidiaries), it crosses a legal boundary. This isn't just data sharing—it's inter-company compliance coordination with implications for liability, insurance, and regulatory oversight. If Food Safety certifies a batch and Food Manufacturing ships it, both subsidiaries share legal responsibility. Problem being solved: coordinating quality assurance across legal entities to ensure unified food safety standards while maintaining proper legal separation and liability management.
Finally, Newco Foods Corporation (Level 1 - Parent) consolidates quality metrics for board reporting, investor communications, and corporate compliance. This becomes "official company performance data"—the basis for claiming food safety standards to customers, regulators, and investors. Problem being solved: demonstrating enterprise-wide quality performance to external stakeholders, managing corporate reputation, and ensuring consistent messaging about food safety commitments.
Same data flow. Completely different understanding of what it means, why it matters, and what business problems it solves.
The organizational architecture mapping makes data governance relevant to everyone in the enterprise. The CFO cares about which data crosses subsidiary boundaries for inter-company accounting and tax optimization. The Chief Compliance Officer focuses on food safety certifications crossing jurisdictions and sensitive formulation data protection. The CTO needs to understand system dependencies across business units and where bottlenecks occur. Food Safety and Quality teams need clarity on who owns ingredient and allergen data when it crosses boundaries, especially critical for product recalls and regulatory audits.
Traditional data lineage is syntactic: "Table A joins with Table B to create View C." Organizational-architecture-mapped lineage is semantic: "Operations data from the Division level is aggregated into management reports at the Parent level because executives need visibility into business unit performance and investors require segment reporting under SEC rules—solving the business problem of transparent performance measurement and regulatory compliance."
This semantic layer answers questions traditional tools cannot. For compliance, you can pinpoint exactly which allergen data flows to external partners and why. For M&A, you understand how an acquisition's processes map to your subsidiary structure and what integration points matter. For audits, you can demonstrate the complete chain of custody from manufacturing floor to financial statements with controls at each organizational boundary.
kDS doesn't just map data flows—it maps the organizational meaning of data flows and the business problems they solve.
This transforms data discovery from a technical documentation exercise into a strategic data management capability. We're capturing not just what data exists, but what the data means to the business, why the organization structured data flow this way, who depends on this data for what business purpose, and most critically—what problems this data is designed to solve.
Understanding what problems your data solves is essential for effective data management. It helps you prioritize which data quality issues matter most, determine what data is truly critical to operations versus nice-to-have, identify where redundant data sources create unnecessary complexity, and make informed decisions about data architecture investments.
Because this approach appears to fill a gap in existing methodologies, kDS represents the opportunity to establish a new category in enterprise data management—one that speaks the language of business, not just technology. This could be formalized as "Organizational Data Flow Architecture" or "Business-Integrated Data Management"—whatever we call it, the fundamental insight remains: understanding your data landscape requires understanding your organizational landscape and the business problems your data is designed to solve.
If this becomes a recognized methodology, it could change how enterprises approach data management entirely. Instead of starting with technical metadata scanning, organizations would begin with organizational structure. Instead of treating data lineage as a purely technical exercise, they would see it as a business intelligence capability. Instead of separating data management from organizational design, they would recognize them as inherently linked. And most importantly, instead of cataloging data without context, they would understand what business problems each data flow is designed to solve.
This is the key differentiation that could make kDS extremely valuable to enterprise buyers. We're solving a problem that Collibra, Alation, and the big consulting firms haven't cracked because they approach data from pure technology or pure governance perspectives—not from organizational architecture and business problem-solving. The BETA program isn't just about testing software—it's about proving out this methodology in real-world enterprise environments, learning how organizational architecture shapes data architecture, and demonstrating how making these relationships explicit transforms data management outcomes.
If you're working in enterprise data management, I'd welcome your thoughts on this framework. Does it resonate with challenges you're facing? Are there aspects of organizational-data mapping we haven't considered? What would make this approach more valuable in your environment?
The methodology is evolving, and the best insights will come from practitioners grappling with real data management challenges in complex organizations.
As always, thank you for stopping by.