Using Debezium to Create a Data Lake with Apache Iceberg
Data lakes can be agile, low-cost ways for companies to store their data. By incorporating key elements into their design, you can build the right tools to keep it from becoming a data swamp.
Data lakes can be agile, low-cost ways for companies to store their data. By incorporating key elements into their design, you can build the right tools to keep it from becoming a data swamp.
This tutorial provides a practical deep dive into the internals of Apache Iceberg using Dremio Sonar as the engine.
The data lake and data warehousing space is facing major disruption spearheaded by innovative table formats like Apache Iceberg.
By creating new features, we can fine tune models and enhance their accuracy. Learn how to engineer features on your data lake using Dremio.
The story of the data lakehouse is a tale of evolution, responding to the growing demands for more adept data processing.
The Databricks platform is widely used for extract, transform, and load (ETL), machine learning, and data science.
Avoid unnecessary table rewrites with partition evolution.
The Apache Iceberg project achieves a milestone with its 1.0 release — with its robust features and stable APIs, it’s never been a better time to adopt Iceberg as your data lakehouse table format.