PASS Data Community Summit 2025 | Seattle, WA...
Data Warehouse Development
Dr Barry Devlin is among the foremost authorities on business insight and one of the founders of data warehousing, having published the first architectural paper on the topic in 1988. Today he is a leading consultant and speaker on data warehouse development.
Barry has published a number of articles for WhereScape, to follow you will find a synopsis and introduction to some of these. Each has a link to the full blog so you can explore that specific subject in more detail.
Designing a Data Warehouse
Always keep in mind the basic goal of your project: to deliver a cross-functional, long-life foundation for data provision and decision support. Data warehouse development project types vary and will continue to mutate over time with requirements that you cannot predict now, and your data warehouse must continue to provide accurate data throughout this evolution.
This blog explains how to:
- Use templates to save time and money rather than building from scratch
- How to define and refine the logical structure of relational tables
- Choose which approach of data modelling is best for you – 3NF, Star Schema, Data Vault etc.
Building a Data Warehouse
This blog explains how every design is only as good as the reality of its source systems, their missing data and poorly defined data structures. The finished design is always a balance between the vision for the model and the constrains of the sources. The article covers:
- The five steps to follow when building a data warehouse
- How Data Warehouse Automation can help
- How to move from design to build
- Building a Data Vault with WhereScape Data Vault Express
Operating a Data Warehouse
This blog explains how to deliver your data warehouse successfully to the business and run it smoothly on a daily basis. We must avoid the problems of past ad hoc data warehouse development approaches that combined manual and semi-automated methods, and adopt advanced data management and automation practices. Find out how:
- Deployment needs to be treated as a long-term, monogamous relationship
- To address issues such as packaging and installation of the code
- To bundle sets of objects and transport from dev to QA and through to production
- To handle interdependencies between the data warehouse, data marts and data lake
- To automate the historical information that tracks performance over time
Maintaining a Data Warehouse
In some development projects, once a piece of software is up and running it needs only minor bug-fixing, but maintaining a data warehouse needs more attention than that. The nature of creative decision-making support is that users are continuously discovering new business requirements, changing their mind about what data they need and thus demanding new data elements and structures on a weekly or monthly basis. Indeed, in some cases, the demands may arrive daily! Read this blog to find out:
- What a data lake should and shouldn’t be used for
- Why and how a Data Vault gives more agility in the maintenance phase
- The role of metadata in data warehouse maintenance
- How to predict downstream impact of changes from automated documentation
Demystifying Microsoft Fabric Components for Business & Technical Users
Microsoft Fabric is rapidly becoming the go-to solution for enterprises aiming to consolidate their analytics processes under a single comprehensive platform. However, understanding the full scope and function of its components can initially seem daunting to both...
An Introduction to Microsoft Fabric: Unifying Analytics for Enterprises
In today's data-driven world, enterprises face an ever-growing demand to harness data efficiently. The complexity of managing diverse and expansive data sources often presents significant challenges. Microsoft Fabric has emerged as a comprehensive solution designed to...
WhereScape at TDWI Munich: Automate Data Vault on Databricks
WhereScape at TDWI Munich 2025: Automate a Full Data Vault on Databricks in Just 45 Minutes June 24–26, 2025 | MOC Munich, Germany As data complexity grows and business demands accelerate, scalable and governed data architectures are no longer optional—they're...
What Is OLAP? Online Analytical Processing for Fast, Multidimensional Analysis
Streamline your data analysis process with OLAP for better business intelligence. Explore the advantages of Online Analytical Processing (OLAP) now! Do you find it challenging to analyze large volumes of data swiftly? A Forrester study reveals that data teams spend...
Build AI-Ready Data: Visit WhereScape at AI & Big Data Expo
June 4–5, 2025 | Booth 202 | Santa Clara Convention Center As organizations scale their artificial intelligence and analytics capabilities, the demand for timely, accurate, governed, and AI-ready data has become a strategic priority. According to Gartner, through...
Automating Star Schemas in Microsoft Fabric: A Webinar Recap
From Data Discovery to Deployment—All in One Workflow According to Gartner, data professionals dedicate more than half of their time, 56%, to operational tasks, leaving only 22% for strategic work that drives innovation. This imbalance is especially apparent when...
What is a Data Model? How Structured Data Drives AI Success
What is a data model? According to the 2020 State of Data Science report by Anaconda, data scientists spend about 45% of their time on data preparation tasks, including cleaning and loading data. Without well-structured data, even the most advanced AI systems can...
ETL vs ELT: What are the Differences?
In working with hundreds of data teams through WhereScape’s automation platform, we’ve seen this debate evolve as businesses modernize their infrastructure. Each method, ETL vs ELT, offers a unique pathway for transferring raw data into a warehouse, where it can be...
Dimensional Modeling for Machine Learning
Kimball’s dimensional modeling continues to play a critical role in machine learning and data science outcomes, as outlined in the Kimball Group’s 10 Essential Rules of Dimensional Modeling, a framework still widely applied in modern data workflows. In a recent...
Automating Data Vault in Databricks | WhereScape Recap
Automating Data Vault in Databricks can reduce time-to-value by up to 70%—and that’s why we hosted a recent WhereScape webinar to show exactly how. At WhereScape, modern data teams shouldn't have to choose between agility and governance. That's why we hosted a live...