Gnarly Data Waves by Dremio

Dremio (The Open Data Lakehouse Platform)

Gnarly Data Waves is a weekly show about the world of Data Analytics and Data Architecture. Learn about the technologies giving the company access to cutting-edge insights. If you work datasets, data warehouses, data lakes or data lakehouses, this show it for you! Join us for our live recordings to participate in the Q&A: dremio.com/events Subscribe to the Dremio youtube channel on: youtube.com/dremio Take the Dremio Platform for a free test-drive: https://www.dremio.com/test-drive/ read less
TechnologieTechnologie

Folgen

EP52 - The Best of the Subsurface Data Lakehouse Conference 2024
25-06-2024
EP52 - The Best of the Subsurface Data Lakehouse Conference 2024
Join us for a captivating recap of Subsurface 2024—the leading conference at the intersection of data engineering, open source technology, and modern data architecture. This webinar will distill: - highlights of the conference, - curated clips of inspiring keynotes, - insightful discussions on real-world data lakehouse implementations by industry leaders such as Nomura, NetApp, and Blue Cross. - and deep dives into the transformative potential of open source projects like Apache Iceberg, Apache XTable, and Ibis. Whether you missed the conference or want to revisit its most impactful moments, this webinar offers a unique opportunity to stay ahead of the curve in the rapidly evolving data landscape. Don't miss this chance to gain valuable insights from the experts and innovators who are shaping the future of data. - Article on Dremio Auto-Ingest: https://www.dremio.com/blog/introducing-auto-ingest-pipes-event-driven-ingestion-made-easy/ - Article on Dremio and Hybrid Data Lakehouses (Vast, Netapp, Minio): https://www.dremio.com/blog/3-reasons-to-create-hybrid-apache-iceberg-data-lakehouses/ --------------------------------------------------------------- Get Hands-on with the Data Lakehouse ---------------------------------------------------------------- - Apache Iceberg Lakehouse on your Laptop: https://bit.ly/am-dremio-lakehouse-laptop - SQLServer to Iceberg to Dashboard: https://bit.ly/am-sqlserver-dashboard - MongoDB to Iceberg to Dashboard: https://bit.ly/am-mongodb-dashboard - Postgres to Iceberg to Dashboard: https://bit.ly/am-postgres-to-dashboard - MySQL to Iceberg to Dashboard: https://bit.ly/am-dremio-mysql-dashboard - Elasticsearch to Iceberg to Dashboard: https://bit.ly/am-dremio-elastic - Apache Druid to Iceberg to Dashboard: https://bit.ly/am-druid-dremio - JSON/CSV/Parquet to Iceberg to Dashboard: https://bit.ly/am-json-csv-parquet-dremio - From Kafka to Iceberg to Dremio: https://bit.ly/am-kafka-connect-dremio - Lowering Snowflake Costs with Dremio: https://bit.ly/am-dremio-snowflake-spend
EP48 - Understanding the Dremio Data Lakehouse
01-04-2024
EP48 - Understanding the Dremio Data Lakehouse
We will embark on a journey that begins with a brief history of data analytics, tracing its development through the advent of the data lakehouse concept. This exploration sets the stage for a deeper understanding of the unique position Dremio occupies within this ecosystem, highlighting its innovative approach to bridging the gap between vast data lakes and the analysts striving to extract actionable insights. The core of this presentation features a live demonstration, showcasing the end-to-end process of data connection and evaluation within the Dremio platform. Attendees will witness firsthand how Dremio facilitates a seamless flow of data from storage in data lakes to its transformation into a format ready for analysis, ultimately culminating in the delivery of valuable insights to analysts. This demonstration not only illustrates Dremio’s capabilities but also emphasizes its role in enabling a win-win scenario for both data engineers and analysts, by simplifying access to data and enhancing the efficiency of the analytics process. In this video, we’ll cover: - A short overview of the power of Dremio - What is a semantic layer and why you need it - Why Dremio is faster than anything else Watch to gain a deeper understanding of the Dremio Data Lakehouse and discover how it can revolutionize your approach to data analytics, from enhancing data accessibility to streamlining the journey from raw data to actionable insights.
GDW CE Workshop 1 - Getting Started with Dremio: Build a Data Lakehouse on your Laptop
29-03-2024
GDW CE Workshop 1 - Getting Started with Dremio: Build a Data Lakehouse on your Laptop
Ready to revolutionize your data management approach and learn how to maximize your environment with Dremio? Watch Alex Merced in this workshop where he’ll guide you step-by-step through building a lakehouse on your laptop with Dremio, Nessie and Minio. This is a great opportunity to try out many of the best features Dremio offers. You'll learn how to: - Read and write Apache Iceberg tables on your object storage, cataloged by Nessie, - Create views in the semantic layer, - And much more GDW Community Edition Workshop Description: This hands-on workshop, participants will embark on a journey to construct their very own data lakehouse platform using their laptops. The workshop is designed to introduce and guide participants through the setup and utilization of three pivotal tools in the data lakehouse architecture: Dremio, Nessie, and Apache Iceberg. Each of these tools plays a crucial role in enabling the flexibility of data lakes with the efficiency and ease of use of data warehouses aiming to simplify and economize data management. You will start by setting up a Docker environment to run all necessary services, including a notebook server, Nessie for catalog tracking with Git-like versioning, Minio as an S3-compatible storage layer, and Dremio as the core lakehouse platform. The workshop will provide a practical, step-by-step guide to federating data sources, organizing and documenting data, and performing queries with Dremio; tracking table changes and branching with Nessie; and creating, querying, and managing Apache Iceberg tables for an ACID-compliant data lakehouse. Prerequisites for the workshop include having Docker installed on your laptop. You will be taken through the process of creating a docker-compose file to spin up the required services, configuring Dremio to connect with Nessie and Minio, and finally, executing SQL queries to manipulate and query data within their lakehouse. This immersive session aims to not just educate but to empower attendees with the knowledge and tools needed to experiment with and implement their data lakehouse solutions. By the end of the workshop, participants will have a functional data lakehouse environment on their laptops, enabling them to explore further and apply what they have learned to real-world scenarios. Whether you're looking to improve your data management strategies or curious about the data lakehouse architecture, this workshop will provide a solid foundation and practical experience.
EP43 - Empowering Analytics: Unleashing the Power of Dremio Cloud on Microsoft Azure
19-01-2024
EP43 - Empowering Analytics: Unleashing the Power of Dremio Cloud on Microsoft Azure
In this session, Dremio and Microsoft will delve into the exciting developments surrounding the public preview launch of Dremio Cloud on Microsoft Azure. This presentation will provide a comprehensive exploration of how businesses are strategically operationalizing their data lakes, with a particular focus on unlocking the vast potential residing within Azure Storage. Attendees will gain valuable insights into the transformative journey toward harnessing the full benefits of a data lakehouse. The discussion will guide participants through the myriad possibilities that emerge when leveraging Dremio Cloud seamlessly on Azure, offering a holistic approach to executing analytics pipelines. This integration eliminates the need for costly data warehouses, presenting a revolutionary paradigm shift. A step-by-step walkthrough will illuminate the process of landing data within the lakehouse, followed by seamlessly progressing data through a virtual semantic layer. This strategic approach adds significant business meaning and value, enhancing the overall utility of the data before it is surfaced to end users. The session will also shed light on the noteworthy performance improvements and cost savings achieved by reducing data extract expenses associated with Power BI workloads. By embracing Dremio Cloud on Azure, organizations can elevate their analytical capabilities while optimizing operational costs, marking a pivotal advancement in the realm of data management and analytics. Join us as we explore the forefront of innovation in data lake operationalization and witness the tangible benefits of this dynamic integration. Watch and learn how Jonny Dixon, Sr. Product Manager at Dremio and Hanno Borns, Principal Product Manager at Microsoft Azure will look into: - Problems companies face with existing analytical architectures - How Dremio and Microsoft Azure work together - What Dremio Cloud on Azure is, and the value it provides - How the Dremio Cloud on Azure solution works, with a demo