

Gnarly Data Waves by Dremio
Dremio (The Open Data Lakehouse Platform)
Gnarly Data Waves is a weekly show about the world of Data Analytics and Data Architecture. Learn about the technologies giving the company access to cutting-edge insights. If you work datasets, data warehouses, data lakes or data lakehouses, this show it for you!
Join us for our live recordings to participate in the Q&A:
dremio.com/events
Subscribe to the Dremio youtube channel on:
youtube.com/dremio
Take the Dremio Platform for a free test-drive:
https://www.dremio.com/test-drive/
Join us for our live recordings to participate in the Q&A:
dremio.com/events
Subscribe to the Dremio youtube channel on:
youtube.com/dremio
Take the Dremio Platform for a free test-drive:
https://www.dremio.com/test-drive/
Episodes
Mentioned books

Jun 10, 2024 • 20min
EP50 - Optimize Analytics Workloads with Dremio + Snowflake
Watch Alex Merced, Senior Technical Evangelist at Dremio on "Optimize Analytics Workloads with Dremio + Snowflake". This session will delve into the key cost drivers of Snowflake and demonstrate how integrating Apache Iceberg and Dremio with a Data Lakehouse architecture can significantly reduce your data warehousing expenses. Discover strategies to optimize your data operations and achieve cost efficiency with cutting-edge technologies.
Ready to Get-Started: https://www.dremio.com/get-started/?u...
See all upcoming episodes and past episodes: https://www.dremio.com/gnarly-data-wa...
Connect with us!
Community Forum: https://bit.ly/2ELXT0W
Github: https://bit.ly/3go4dcM
Blog: https://bit.ly/2DgyR9B
Questions?: https://bit.ly/30oi8tX
Website: https://bit.ly/2XmtEnN
Resource: https://www.dremio.com/resources/?utm...
Events: https://www.dremio.com/events/?utm_me...

Apr 17, 2024 • 28min
EP49 - What’s New in Dremio: New Capabilities for the Best Apache Iceberg Lakehouse
Dremio is making it easier than ever to build and manage an Apache Iceberg data lakehouse. Mark Shainman will share the new Dremio capabilities that help you achieve the fastest, most scalable, and easiest-to-manage lakehouse for analytics and AI.
In this video you’ll learn how:
- Dremio can help you accelerate Apache Iceberg adoption with seamless ingest
- Enhanced Reflections query acceleration can optimize performance and streamline management
- New capabilities continue to improve reliability, stability and scalability
- Dremio is delivering new capabilities to increase observability for ease of administration and management

Apr 1, 2024 • 28min
EP48 - Understanding the Dremio Data Lakehouse
We will embark on a journey that begins with a brief history of data analytics, tracing its development through the advent of the data lakehouse concept. This exploration sets the stage for a deeper understanding of the unique position Dremio occupies within this ecosystem, highlighting its innovative approach to bridging the gap between vast data lakes and the analysts striving to extract actionable insights.
The core of this presentation features a live demonstration, showcasing the end-to-end process of data connection and evaluation within the Dremio platform. Attendees will witness firsthand how Dremio facilitates a seamless flow of data from storage in data lakes to its transformation into a format ready for analysis, ultimately culminating in the delivery of valuable insights to analysts. This demonstration not only illustrates Dremio’s capabilities but also emphasizes its role in enabling a win-win scenario for both data engineers and analysts, by simplifying access to data and enhancing the efficiency of the analytics process.
In this video, we’ll cover:
- A short overview of the power of Dremio
- What is a semantic layer and why you need it
- Why Dremio is faster than anything else
Watch to gain a deeper understanding of the Dremio Data Lakehouse and discover how it can revolutionize your approach to data analytics, from enhancing data accessibility to streamlining the journey from raw data to actionable insights.

Mar 29, 2024 • 58min
GDW CE Workshop 1 - Getting Started with Dremio: Build a Data Lakehouse on your Laptop
Ready to revolutionize your data management approach and learn how to maximize your environment with Dremio?
Watch Alex Merced in this workshop where he’ll guide you step-by-step through building a lakehouse on your laptop with Dremio, Nessie and Minio. This is a great opportunity to try out many of the best features Dremio offers.
You'll learn how to:
- Read and write Apache Iceberg tables on your object storage, cataloged by Nessie,
- Create views in the semantic layer,
- And much more
GDW Community Edition Workshop Description:
This hands-on workshop, participants will embark on a journey to construct their very own data lakehouse platform using their laptops. The workshop is designed to introduce and guide participants through the setup and utilization of three pivotal tools in the data lakehouse architecture: Dremio, Nessie, and Apache Iceberg. Each of these tools plays a crucial role in enabling the flexibility of data lakes with the efficiency and ease of use of data warehouses aiming to simplify and economize data management.
You will start by setting up a Docker environment to run all necessary services, including a notebook server, Nessie for catalog tracking with Git-like versioning, Minio as an S3-compatible storage layer, and Dremio as the core lakehouse platform. The workshop will provide a practical, step-by-step guide to federating data sources, organizing and documenting data, and performing queries with Dremio; tracking table changes and branching with Nessie; and creating, querying, and managing Apache Iceberg tables for an ACID-compliant data lakehouse.
Prerequisites for the workshop include having Docker installed on your laptop. You will be taken through the process of creating a docker-compose file to spin up the required services, configuring Dremio to connect with Nessie and Minio, and finally, executing SQL queries to manipulate and query data within their lakehouse.
This immersive session aims to not just educate but to empower attendees with the knowledge and tools needed to experiment with and implement their data lakehouse solutions. By the end of the workshop, participants will have a functional data lakehouse environment on their laptops, enabling them to explore further and apply what they have learned to real-world scenarios. Whether you're looking to improve your data management strategies or curious about the data lakehouse architecture, this workshop will provide a solid foundation and practical experience.

28 snips
Mar 27, 2024 • 35min
EP47 - Learn how to reduce your Snowflake cost by 50%+ with a lakehouse
Explore how transitioning to a data lake house can slash Snowflake costs by 50%, addressing hidden costs and inefficiencies. Learn about managing data from various sources, benefits of Dremio's platform for cost savings and efficiency, and strategies to reduce costs with data reflections.

Mar 25, 2024 • 1h
EP46 - Getting Started with Dremio
Learn about the evolution of data lake houses and Dremio's unique technologies for query optimization. Explore efficient data merging, troubleshooting, and experimentation. Discover how Dremio enhances data workflows and manages views efficiently. Dive into the benefits of using Dremio for performance, usability, and cloud data consumption.

Feb 22, 2024 • 40min
EP45 - Next-Gen Data Pipelines are Virtual: Simplify Data Pipelines with dbt, Dremio, and Iceberg
The podcast delves into simplifying data pipelines with Dremio's next-gen DataOps, including data source branching, virtual data marts, dbt orchestration, and lakehouse maintenance techniques. Topics covered range from data catalog capabilities and modeling with Dremio to building a 'lake house' for data analytics, with an emphasis on streamlining data management and improving data quality.

Jan 31, 2024 • 22min
EP44 - How S&P Global is Building an Azure Data Lakehouse with Dremio
S&P Global is a leading global financial services company headquartered in New York. It provides credit ratings, benchmarks, analytics, and workflow solutions in the global capital, commodity, and automotive markets. As a company, data is an essential asset across all of S&P Global’s solutions offerings.
Watch Tian de Klerk, Director of Business Intelligence, as he shares how they built a data lakehouse for FinOps analysis with Dremio Cloud on Microsoft Azure.
Tian will share about:
- The hidden costs of extracting operational data into BI cubes
- Simplifying traditional data engineering processes with Dremio’s zero-ETL lakehouse
- How Dremio’s semantic layer and query acceleration make self-service analytics easy for end users

Jan 19, 2024 • 54min
EP43 - Empowering Analytics: Unleashing the Power of Dremio Cloud on Microsoft Azure
In this session, Dremio and Microsoft will delve into the exciting developments surrounding the public preview launch of Dremio Cloud on Microsoft Azure. This presentation will provide a comprehensive exploration of how businesses are strategically operationalizing their data lakes, with a particular focus on unlocking the vast potential residing within Azure Storage. Attendees will gain valuable insights into the transformative journey toward harnessing the full benefits of a data lakehouse.
The discussion will guide participants through the myriad possibilities that emerge when leveraging Dremio Cloud seamlessly on Azure, offering a holistic approach to executing analytics pipelines. This integration eliminates the need for costly data warehouses, presenting a revolutionary paradigm shift. A step-by-step walkthrough will illuminate the process of landing data within the lakehouse, followed by seamlessly progressing data through a virtual semantic layer. This strategic approach adds significant business meaning and value, enhancing the overall utility of the data before it is surfaced to end users.
The session will also shed light on the noteworthy performance improvements and cost savings achieved by reducing data extract expenses associated with Power BI workloads. By embracing Dremio Cloud on Azure, organizations can elevate their analytical capabilities while optimizing operational costs, marking a pivotal advancement in the realm of data management and analytics. Join us as we explore the forefront of innovation in data lake operationalization and witness the tangible benefits of this dynamic integration.
Watch and learn how Jonny Dixon, Sr. Product Manager at Dremio and Hanno Borns, Principal Product Manager at Microsoft Azure will look into:
- Problems companies face with existing analytical architectures
- How Dremio and Microsoft Azure work together
- What Dremio Cloud on Azure is, and the value it provides
- How the Dremio Cloud on Azure solution works, with a demo

Dec 20, 2023 • 26min
EP42 - What's new in Dremio: New GenAI capabilities, advance for 100% query success + now on Azure
Dremio delivers no compromise lakehouse analytics for all of your data - and recent launches are making Dremio faster, more reliable, and more flexible than ever. Watch Mark Shainman, Product Marketing Manager at Dremio, Colleen Quinn, Product Marketing Manager at Dremio share and learn what's new in Dremio.
- New Gen-AI capabilities for automated data descriptions and labeling
- Dremio Cloud SaaS service now available on Microsoft Azure
- Advances to ensure 100% query reliability with no memory failures
- Expanded Apache Iceberg capabilities to streamline Iceberg adoption and improve performance
Ready to Get-Started: https://www.dremio.com/get-started/?u...
See all upcoming episodes and past episodes: https://www.dremio.com/gnarly-data-wa...
Connect with us!
Community Forum: https://bit.ly/2ELXT0W
Github: https://bit.ly/3go4dcM
Blog: https://bit.ly/2DgyR9B
Questions?: https://bit.ly/30oi8tX
Website: https://bit.ly/2XmtEnN
Resource: https://www.dremio.com/resources/?utm...
Events: https://www.dremio.com/events/?utm_me...#datalakehouse #analytics #datawarehouse #datalake #dataengineers #dataarchitects #governance #dremiocloud #opendatalakehouse #apacheiceberg #selfservice #enterprisedata #multitables #tableformat #microsoftazure #saas #automateddata #query #whatsnew #colleenquinn #markshainman #getstarted #etl #pipelines #genai #generativeai #parquet #json #tableau


