In March I worked on this article with folks at Dremio, explaining the challenge of data copies in modern “data lakehouse” environments.
An organization’s data is copied for many reasons, namely ingesting datasets into data warehouses, creating performance-optimized copies and building BI extracts for analysis. Unfortunately, data replication, transformation and movement can result in longer time to insight, reduced efficiency, elevated costs, and increased security and compliance risk.
Fortunately, a next-gen data architecture enabled by the Dremio data lake service removes the need for replicated data, helping organizations to minimize complexity, boost efficiency and dramatically reduce costs.
You can download the completed whitepaper here