Gnarly Data Waves by Dremio

Dremio (The Open Data Lakehouse Platform)
undefined
Jun 10, 2024 • 20min

EP50 - Optimize Analytics Workloads with Dremio + Snowflake

Watch Alex Merced, Senior Technical Evangelist at Dremio on "Optimize Analytics Workloads with Dremio + Snowflake". This session will delve into the key cost drivers of Snowflake and demonstrate how integrating Apache Iceberg and Dremio with a Data Lakehouse architecture can significantly reduce your data warehousing expenses. Discover strategies to optimize your data operations and achieve cost efficiency with cutting-edge technologies. Ready to Get-Started: https://www.dremio.com/get-started/?u... See all upcoming episodes and past episodes: https://www.dremio.com/gnarly-data-wa... Connect with us! Community Forum: https://bit.ly/2ELXT0W Github: https://bit.ly/3go4dcM Blog: https://bit.ly/2DgyR9B Questions?: https://bit.ly/30oi8tX Website: https://bit.ly/2XmtEnN Resource: https://www.dremio.com/resources/?utm... Events: https://www.dremio.com/events/?utm_me...
undefined
Apr 17, 2024 • 28min

EP49 - What’s New in Dremio: New Capabilities for the Best Apache Iceberg Lakehouse

Dremio is making it easier than ever to build and manage an Apache Iceberg data lakehouse. Mark Shainman will share the new Dremio capabilities that help you achieve the fastest, most scalable, and easiest-to-manage lakehouse for analytics and AI. In this video you’ll learn how: - Dremio can help you accelerate Apache Iceberg adoption with seamless ingest - Enhanced Reflections query acceleration can optimize performance and streamline management - New capabilities continue to improve reliability, stability and scalability - Dremio is delivering new capabilities to increase observability for ease of administration and management
undefined
Apr 1, 2024 • 28min

EP48 - Understanding the Dremio Data Lakehouse

We will embark on a journey that begins with a brief history of data analytics, tracing its development through the advent of the data lakehouse concept. This exploration sets the stage for a deeper understanding of the unique position Dremio occupies within this ecosystem, highlighting its innovative approach to bridging the gap between vast data lakes and the analysts striving to extract actionable insights. The core of this presentation features a live demonstration, showcasing the end-to-end process of data connection and evaluation within the Dremio platform. Attendees will witness firsthand how Dremio facilitates a seamless flow of data from storage in data lakes to its transformation into a format ready for analysis, ultimately culminating in the delivery of valuable insights to analysts. This demonstration not only illustrates Dremio’s capabilities but also emphasizes its role in enabling a win-win scenario for both data engineers and analysts, by simplifying access to data and enhancing the efficiency of the analytics process. In this video, we’ll cover: - A short overview of the power of Dremio - What is a semantic layer and why you need it - Why Dremio is faster than anything else Watch to gain a deeper understanding of the Dremio Data Lakehouse and discover how it can revolutionize your approach to data analytics, from enhancing data accessibility to streamlining the journey from raw data to actionable insights.
undefined
Mar 29, 2024 • 58min

GDW CE Workshop 1 - Getting Started with Dremio: Build a Data Lakehouse on your Laptop

Ready to revolutionize your data management approach and learn how to maximize your environment with Dremio? Watch Alex Merced in this workshop where he’ll guide you step-by-step through building a lakehouse on your laptop with Dremio, Nessie and Minio. This is a great opportunity to try out many of the best features Dremio offers. You'll learn how to: - Read and write Apache Iceberg tables on your object storage, cataloged by Nessie, - Create views in the semantic layer, - And much more GDW Community Edition Workshop Description: This hands-on workshop, participants will embark on a journey to construct their very own data lakehouse platform using their laptops. The workshop is designed to introduce and guide participants through the setup and utilization of three pivotal tools in the data lakehouse architecture: Dremio, Nessie, and Apache Iceberg. Each of these tools plays a crucial role in enabling the flexibility of data lakes with the efficiency and ease of use of data warehouses aiming to simplify and economize data management. You will start by setting up a Docker environment to run all necessary services, including a notebook server, Nessie for catalog tracking with Git-like versioning, Minio as an S3-compatible storage layer, and Dremio as the core lakehouse platform. The workshop will provide a practical, step-by-step guide to federating data sources, organizing and documenting data, and performing queries with Dremio; tracking table changes and branching with Nessie; and creating, querying, and managing Apache Iceberg tables for an ACID-compliant data lakehouse. Prerequisites for the workshop include having Docker installed on your laptop. You will be taken through the process of creating a docker-compose file to spin up the required services, configuring Dremio to connect with Nessie and Minio, and finally, executing SQL queries to manipulate and query data within their lakehouse. This immersive session aims to not just educate but to empower attendees with the knowledge and tools needed to experiment with and implement their data lakehouse solutions. By the end of the workshop, participants will have a functional data lakehouse environment on their laptops, enabling them to explore further and apply what they have learned to real-world scenarios. Whether you're looking to improve your data management strategies or curious about the data lakehouse architecture, this workshop will provide a solid foundation and practical experience.
undefined
28 snips
Mar 27, 2024 • 35min

EP47 - Learn how to reduce your Snowflake cost by 50%+ with a lakehouse

Explore how transitioning to a data lake house can slash Snowflake costs by 50%, addressing hidden costs and inefficiencies. Learn about managing data from various sources, benefits of Dremio's platform for cost savings and efficiency, and strategies to reduce costs with data reflections.
undefined
Mar 25, 2024 • 1h

EP46 - Getting Started with Dremio

Learn about the evolution of data lake houses and Dremio's unique technologies for query optimization. Explore efficient data merging, troubleshooting, and experimentation. Discover how Dremio enhances data workflows and manages views efficiently. Dive into the benefits of using Dremio for performance, usability, and cloud data consumption.
undefined
Feb 22, 2024 • 40min

EP45 - Next-Gen Data Pipelines are Virtual: Simplify Data Pipelines with dbt, Dremio, and Iceberg

The podcast delves into simplifying data pipelines with Dremio's next-gen DataOps, including data source branching, virtual data marts, dbt orchestration, and lakehouse maintenance techniques. Topics covered range from data catalog capabilities and modeling with Dremio to building a 'lake house' for data analytics, with an emphasis on streamlining data management and improving data quality.
undefined
Jan 31, 2024 • 22min

EP44 - How S&P Global is Building an Azure Data Lakehouse with Dremio

S&P Global is a leading global financial services company headquartered in New York. It provides credit ratings, benchmarks, analytics, and workflow solutions in the global capital, commodity, and automotive markets. As a company, data is an essential asset across all of S&P Global’s solutions offerings. Watch Tian de Klerk, Director of Business Intelligence, as he shares how they built a data lakehouse for FinOps analysis with Dremio Cloud on Microsoft Azure. Tian will share about: - The hidden costs of extracting operational data into BI cubes - Simplifying traditional data engineering processes with Dremio’s zero-ETL lakehouse - How Dremio’s semantic layer and query acceleration make self-service analytics easy for end users
undefined
Jan 19, 2024 • 54min

EP43 - Empowering Analytics: Unleashing the Power of Dremio Cloud on Microsoft Azure

In this session, Dremio and Microsoft will delve into the exciting developments surrounding the public preview launch of Dremio Cloud on Microsoft Azure. This presentation will provide a comprehensive exploration of how businesses are strategically operationalizing their data lakes, with a particular focus on unlocking the vast potential residing within Azure Storage. Attendees will gain valuable insights into the transformative journey toward harnessing the full benefits of a data lakehouse. The discussion will guide participants through the myriad possibilities that emerge when leveraging Dremio Cloud seamlessly on Azure, offering a holistic approach to executing analytics pipelines. This integration eliminates the need for costly data warehouses, presenting a revolutionary paradigm shift. A step-by-step walkthrough will illuminate the process of landing data within the lakehouse, followed by seamlessly progressing data through a virtual semantic layer. This strategic approach adds significant business meaning and value, enhancing the overall utility of the data before it is surfaced to end users. The session will also shed light on the noteworthy performance improvements and cost savings achieved by reducing data extract expenses associated with Power BI workloads. By embracing Dremio Cloud on Azure, organizations can elevate their analytical capabilities while optimizing operational costs, marking a pivotal advancement in the realm of data management and analytics. Join us as we explore the forefront of innovation in data lake operationalization and witness the tangible benefits of this dynamic integration. Watch and learn how Jonny Dixon, Sr. Product Manager at Dremio and Hanno Borns, Principal Product Manager at Microsoft Azure will look into: - Problems companies face with existing analytical architectures - How Dremio and Microsoft Azure work together - What Dremio Cloud on Azure is, and the value it provides - How the Dremio Cloud on Azure solution works, with a demo
undefined
Dec 20, 2023 • 26min

EP42 - What's new in Dremio: New GenAI capabilities, advance for 100% query success + now on Azure

Dremio delivers no compromise lakehouse analytics for all of your data - and recent launches are making Dremio faster, more reliable, and more flexible than ever. Watch Mark Shainman, Product Marketing Manager at Dremio, Colleen Quinn, Product Marketing Manager at Dremio share and learn what's new in Dremio. - New Gen-AI capabilities for automated data descriptions and labeling - Dremio Cloud SaaS service now available on Microsoft Azure - Advances to ensure 100% query reliability with no memory failures - Expanded Apache Iceberg capabilities to streamline Iceberg adoption and improve performance Ready to Get-Started: https://www.dremio.com/get-started/?u... See all upcoming episodes and past episodes: https://www.dremio.com/gnarly-data-wa... Connect with us! Community Forum: https://bit.ly/2ELXT0W Github: https://bit.ly/3go4dcM Blog: https://bit.ly/2DgyR9B Questions?: https://bit.ly/30oi8tX Website: https://bit.ly/2XmtEnN Resource: https://www.dremio.com/resources/?utm... Events: https://www.dremio.com/events/?utm_me...#datalakehouse #analytics #datawarehouse #datalake #dataengineers #dataarchitects #governance #dremiocloud #opendatalakehouse #apacheiceberg #selfservice #enterprisedata #multitables #tableformat #microsoftazure #saas #automateddata #query #whatsnew #colleenquinn #markshainman #getstarted #etl #pipelines #genai #generativeai #parquet #json #tableau

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app