The EDW is Not Dead, Long Live the EDWHeading

By Cary Moore, Senior Director of Data Science

For the last decade (at least), so much of what I’ve seen and read in the marketplace, research, literature, and press is that the data lake and data scientists simplified end-user report environments, and the cloud will inevitably replace all that was wrong with the Enterprise Data Warehouse (EDW); like it was some evil imperial monolith from the “Dark Side” to be destroyed.

Reasons: Over-budget, over-complicated, often late, missed requirements, stakeholder engagement, and every other criticism you can think of for failed missions. Of course, many of these circumstances were true. Every EDW I’ve ever built has had its challenges – just like any other project – but, fortunately, most were successful. Invariably, there were always a few that proved to be more challenging than others, and customer satisfaction waned during the engagement. “Hindsight is 20/20” – If you only knew what you needed to know before the work started, then even those more difficult projects may have been (more) successful. There is always something hiding under a rock that you couldn’t see or didn’t know existed, and neither did the customer until you dug it up.

Those hidden rocks, the unknown, unknown unknowns, and all of the data problems, business processes, systemic issues, and so on, would have existed in any other situation. It’s the data issues that are the most difficult to resolve. Essential to the success of any solution approach is adding meaning (read: value) to data. Critical data must be documented, not just for its definition, structure, provenance, etc., but for its use; that is, how data is consumed by key decision-makers. Oftentimes, simply identifying what’s required, formulas, correct source, aggregation, filtering, is a painstaking process – and there is no baby Yoda with the “Force” – that can replace the necessity of knowing the business.

I’m not writing this article for the initiated, many of whom know this all too well. I’m trying to help those that are struggling, trying to deliver that all-important data lake. Point is, just because you have the greatest technical wizardry money can buy – cloud or otherwise – the best intentioned, well-architected, supreme solution – will not address gathering requirements, including defining the data. This is not a technical task, it’s a hunting expedition, and without a Mandalorian on commission, you are only successful if you understand the business. To understand the business, you must speak to many people, ask the right questions, get the complete answers and confirm them. I’ve recently seen articles discussing that young Padawan data scientists are not enjoying their roles because much of their time is spent simply getting data and requirements, not running models; this should not come as a surprise.

Simple things can trip up even the best analysts, architects, and developers. For example, a seemingly simple concept like “current balance” at a financial services company can easily become a multi-headed monster as different business units, application teams, and reporting areas define the term for their own use. Reconciling the varying business rules, data requirements, and consumption architecture is a challenge for even the most experienced architects (read Jedi). Legacy applications also present their own unique challenges. I’ve had systems that had the date in 4 fields (columns), one for the century, year, month and day, in integers, with names like “excy” and “efdy.”  I’m sure you will figure out what those mean just by looking at them. Success requires a persistent analyst to locate the SME and confirm all of the rules and the corresponding result and document it (data mapping). Of course, in today’s ever-changing business world, understanding the history of how the data came to be in its current incarnation is also critical. “Incorrect” data from systems changing over time required a series of filters to remove historical data, uncovered through many iterations. As we have all seen, it’s a rare team that maintains good documentation with enough detail and history assuming it has been captured before, let alone new systems and sources. Often it is only the availability of the long-standing members that keep tribal knowledge available. Hopefully, they all aren’t located on the planet Alderaan.

Previous technologies were not perfect and much of what we do with data is infinitely easier.  ETL and previous relational EDWs had their challenges. ETL is too manual, time-consuming, hard to change or version. On-premise EDWs and their schemas are difficult to design and change too, requiring expensive dedicated hardware and support staff. The newer tools help in those areas; reducing the amount of dedicated hardware and processing times, lessen the impact of changes and allowing more efficient validation and testing activities. Machine learning and open source software are significantly changing what you can do with data and democratizing what previously was only available to specialists. Technology is still woefully incapable of creating meaning. Producing flatter tables and curated data sets with all the necessary detail lend great assistance to the consumers of that content encouraging more self-service and reducing effort overall. Whether seasoned data scientists or newbie report developers, all will benefit from data with the understanding and meaning built into them.  Without a centralized process to do that beforehand, you will simply cause many different people to require the same time of the same SMEs asking the same questions.

Call it data governance, data quality, data cleansing, metadata management, data architecture, data catalogs or data dictionaries – for me, the problem continues to be the same exercise just with a different name. I call it “understanding your data” by adding meaning to it. It will be a while before that skillset goes away or becomes automated by technology. Data is still vastly a people business no matter how fast or fancy the technology and lacking Jedi mind tricks… May the force be with you.

Recent Posts

Executive Perspective: Why Securing Your Data is the Key to Winning with AI

As CXOs, we’re all focused on leveraging AI to drive efficiency, innovation, and competitive advantage. The conversation often starts with infrastructure (GPUs, LLMs, and copilots), but let’s be clear: AI doesn’t run on GPUs. It runs on data. And if our data isn’t secure, nothing else matters. Unsecured and unclassified data undermines trust, exposes us to risk, and jeopardizes the very AI strategies we’re betting our futures on. It’s not about chasing the next shiny tool; it’s about building a foundation that allows AI to scale safely, responsibly, and securely.

Quantifying the Value of Data in Financial Services

In the financial services sector, data is a critical asset that drives profitability, risk management, regulatory compliance, and competitive edge. However, measuring its value remains challenging for many CFOs across sectors of the financial services industry regardless of organizational size or country of operations. CFOs rely on accurate data for forecasting, budgeting, and strategic planning. Quality data leads to better decision-making, optimized capital allocation, and swift responses to market changes. It is also vital for risk management, regulatory compliance (e.g., BCBS 239, Basel III, AML/KYC), and avoiding fines and reputational damage. “Fit for Business Use” data also supports customer retention, personalized services, and improved revenue stability. Data-savvy CFOs leverage insights for long-term growth.

AI Starts with Data: Go From Hype to Results

AI continues to dominate the conversation in business. From executive meetings to strategic roadmaps, AI is no longer just a trend but a real driver of transformation. The challenge is that while nearly every organization is talking about AI, very few are prepared to use it in a way that delivers measurable outcomes and lasting impact. The difference between hype and outcomes almost always comes down to two things: the quality of your data and your organization’s readiness to execute.

Exciting Updates from Informatica World: Paradigm Embraces the Future of Agentic AI

The digital landscape is evolving rapidly, and staying ahead means embracing the latest innovations in data management and artificial intelligence. At this year’s Informatica World, Paradigm is thrilled to share the groundbreaking advancements unveiled by Informatica, centered around their latest agentic AI solutions on the Intelligent Data Management Cloud (IDMC) platform.

Modernizing PowerCenter: The IDMC Way – Better, Faster, Cheaper

For many organizations, Informatica PowerCenter has been the workhorse of their data integration for years, even decades, reliably driving ETL processes and populating data warehouses that feed BI reports. However, this longevity often leads to a complex environment that can hinder agility and innovation.

Boost Growth with Data-Driven Hiring for Boutique Consultancies

Consistency is key to a boutique consultancy. Delivering quality services day in and day out, even as client demand fluctuates, relies heavily on having the right talent at the right time. Perhaps one of the largest operational challenges for small and mid-sized consulting firms, though, is matching recruitment cycles with cyclical demand. Without scalable, data-driven talent practices, consultancies can suffer from misaligned capacity, lost revenue streams, and stalled growth.

Strategies for a Successful Journey in Building the Dream Team

In the whirlwind world of project management, the success of a project often hinges on the strength and consistency of the team behind it. Imagine embarking on a journey where building a high-performing project team is not just about assembling a group of skilled individuals; it’s about fostering collaboration, trust, and a shared sense of purpose. Based on my personal experiences, let me take you through this journey with some strategies I use to help you build and lead a high-performing project team.

The Ultimate Guide to AI-Enhanced APIM Analytics for Enterprise Success

Enterprises increasingly rely on Application Programming Interface (API) Management (APIM) to streamline their operations, enhance customer experiences, and drive innovation. Azure API Management is a comprehensive solution enabling organizations to manage, secure, and optimize their APIs efficiently. However, beyond API management, APIM analytics – particularly when integrated with AI and data analytics – empowers senior executives with critical insights for real-time decision-making.

Why PMOs that Leverage Power BI are More Successful

Project Management Offices (PMOs) are increasingly turning to Microsoft Power BI as a game-changing tool to enhance their project management processes. By leveraging data visualization and analytics, PMOs can make informed decisions, streamline reporting, and improve overall project outcomes.