Are you eager to delve deep into the challenges...
BI Built to Order, On-Demand: Automating Data Warehouse Delivery
This week, Dr. Barry Devlin published a provocative new paper on data warehouse automation – “BI Built to Order, On-Demand: Automating Data Warehouse Delivery.”
You can grab it here, if you’re curious. And you should be. Because in the paper Devlin does two things: first, he considers a few Inconvenient Truths about how data warehouses are built and managed – or misbuilt and mismanaged – and, second, he makes the case for data warehouse automation as a common-sense fix for today’s often mismanaged data warehouse development.
When Devlin described his vision for the business information system he called a “data warehouse” – back in early 1988 – we just didn’t have the tools to efficiently design, build, and manage warehouse systems. Everything, or almost everything, had to be done by hand: there weren’t any ETL tools, data integration suites, studios, platforms or workbenches. But even once we got primitive versions of these tools – starting in 1993 or thereabouts – things didn’t magically get better. In fact, by 2003, we were already starting to come to grips with the empirical fact that data warehouse projects took too long to build, failed to deliver on many of the promises Devlin had outlined in his paper, and, most important, were too hard to change. We know: WhereScape-the-company grew out of the integration experiences of our founders, who specialized in fixing just these problems.
But a great point that Devlin makes is that most of these problems were byproducts of what might be called an “out of phase” development process. Simply put: building data warehouse systems was and to some degree still is a disintegrated affair. In larger organizations, it is performed by separate teams or groups of developers, each working with their own set of tools, each using their own methodology, and each building at their own pace. According to Devlin, this is one of the biggest impediments to traditional analytic development.
“Modeling, database design and development of population routines required multiple, disconnected iterations involving business users, modelers, database administrators and ETL programmers at different times, each using different and unconnected tools. These gaps and tool transitions slowed the process and gave rise to design errors and inconsistencies,” Devlin writes.
The upshot is that this model compromises both the consistency of data and the timeliness of application delivery. Devlin sees data warehouse automation software – which centralizes data warehouse and analytical development in a single tool – promoting an iterative, agile development methodology, and implements a shared metadata repository – as a prescriptive Rx for this problem.
“The common environment and shared metadata repository offered by data warehouse automation overcomes this … by integrating the design and delivery of the data model, database structure, and the population process in one place – whether for a warehouse or mart,” he writes. “All the design and population metadata is stored together in a single repository, allowing development to flow smoothly and iteratively from user requirements, through database design, to creation of population routines. By integrating all the steps of the design and development process, consistent and quality data can be delivered quickly to the business for immediate review and early acceptance.”
Data warehouse automation software isn’t a turnkey fix. Devlin recognizes this. All the same, it’s a way to eliminate out-of-phase development, centralize the development process, and enforce a consistent, delivery-focused development paradigm. It gives you a solid foundation on which to build your data warehouse. Data warehouse automation software has other benefits that aren’t at all confined strictly to development. As Devlin notes, it promotes collaboration between business and IT, making it possible to produce data-driven – or business-data-driven – apps.
I’ll say more about this in a follow up post.
Amplifying WhereScape’s Power with Yellowfin: Unveiling New Analytics Opportunities for Your Business
In an age dominated by vast amounts of information, the emphasis on data-driven decision-making has never been greater. The landscape of Business Intelligence (BI) and data analytics has seen a remarkable evolution, emphasizing solutions that can seamlessly integrate...
Data Mesh and Data Fabric: Changing the Game in Data Product Development
Data Mesh vs Data Fabric Data Mesh and Data Fabric are reshaping how organizations approach data product development. In an era where data-driven decisions are central to business success, these innovative paradigms are becoming increasingly crucial. By enabling...
WhereScape Announces the Release of RED 10.0.0.0
WhereScape is pleased to announce the general availability of WhereScape RED 10.0.0.0. This release is the culmination of man-years of effort. It confirms WhereScape’s commitment to continuing to develop new technologies and tools and its commitment to delivering the...
Effective AI through Data Modeling
As we journey deeper into the digital age, the importance of data modeling within the broader landscape of artificial intelligence (AI) has become more pronounced than ever. The success of AI-driven initiatives is tightly woven with the quality and structure of the...
Is Data Vault 2.0 Still Relevant?
TL;DR Yes. Data Vault 2.0 Data Vault 2.0 is a database modeling method published in 2013. It was designed to overcome many of the shortcomings of data warehouses created using relational modeling (3NF) or star schemas (dimensional modeling). Speci fically, it...
Data Vault 2.0 Resources
Data Vault Revisited: A Six-Year Journey into the Secure Data Repository In 2017, Dr. Barry Devlin provided valuable insights about Data Vaults, a concept that sparked interest among businesses and IT professionals. Data Vaults were envisioned as secure repositories...
Understanding Data Vault 2.0
How to Avoid Pitfalls During Data Vault 2.0 Implementation Implementing a data vault as your Data Modeling approach has many advantages, such as flexibility, scalability, and efficiency. But along with that, one must be aware of the challenges that come along with...
Navigating the AI Landscape
The Pivotal Role of Data Modeling In the rapidly evolving digital age, artificial intelligence (AI) has emerged as a game-changer, deeply impacting the business landscape. Its ability to automate operations, refine decision-making processes, and significantly enhance...
Information Management Maturity
Unlocking Your Business Potential: Understanding and Enhancing Information Management Maturity In a recent report by Gartner, they emphasize the crucial role of information in the current business environment, stating, "Through 2025, organizations that are data-driven...
Data Warehousing Best Practices
In modern times, organizations are daily generating huge volumes of data. Appreciating the significance of data, companies are storing data from different departments which can be analyzed to gather insights to help the organization in better decision-making. This...
Related Content

Amplifying WhereScape’s Power with Yellowfin: Unveiling New Analytics Opportunities for Your Business
In an age dominated by vast amounts of information, the emphasis on data-driven decision-making has never been greater. The landscape of Business Intelligence (BI) and data analytics has seen a remarkable evolution, emphasizing solutions that can seamlessly integrate...

Data Mesh and Data Fabric: Changing the Game in Data Product Development
Data Mesh vs Data Fabric Data Mesh and Data Fabric are reshaping how organizations approach data product development. In an era where data-driven decisions are central to business success, these innovative paradigms are becoming increasingly crucial. By enabling...

WhereScape Announces the Release of RED 10.0.0.0
WhereScape is pleased to announce the general availability of WhereScape RED 10.0.0.0. This release is the culmination of man-years of effort. It confirms WhereScape’s commitment to continuing to develop new technologies and tools and its commitment to delivering the...

Effective AI through Data Modeling
As we journey deeper into the digital age, the importance of data modeling within the broader landscape of artificial intelligence (AI) has become more pronounced than ever. The success of AI-driven initiatives is tightly woven with the quality and structure of the...