Building Modern Data Applications Using Databricks Lakehouse
Will Girten
Verkauft von Rarewaves USA United, OSWEGO, IL, USA
AbeBooks-Verkäufer seit 20. Juni 2025
Neu - Softcover
Zustand: Neu
Anzahl: Mehr als 20 verfügbar
In den Warenkorb legenVerkauft von Rarewaves USA United, OSWEGO, IL, USA
AbeBooks-Verkäufer seit 20. Juni 2025
Zustand: Neu
Anzahl: Mehr als 20 verfügbar
In den Warenkorb legenLearn the latest Databricks features, with up-to-date insights into the platform. This book will develop your skills to build scalable and secure data pipelines to ingest, transform, and deliver timely, accurate data to drive business decisions.
Bestandsnummer des Verkäufers LU-9781801073233
Get up to speed with the Databricks Data Intelligence Platform to build and scale modern data applications, leveraging the latest advancements in data engineering
With so many tools to choose from in today’s data engineering development stack as well as operational complexity, this often overwhelms data engineers, causing them to spend less time gleaning value from their data and more time maintaining complex data pipelines. Guided by a lead specialist solutions architect at Databricks with 10+ years of experience in data and AI, this book shows you how the Delta Live Tables framework simplifies data pipeline development by allowing you to focus on defining input data sources, transformation logic, and output table destinations.
This book gives you an overview of the Delta Lake format, the Databricks Data Intelligence Platform, and the Delta Live Tables framework. It teaches you how to apply data transformations by implementing the Databricks medallion architecture and continuously monitor the data quality of your pipelines. You’ll learn how to handle incoming data using the Databricks Auto Loader feature and automate real-time data processing using Databricks workflows. You’ll master how to recover from runtime errors automatically.
By the end of this book, you’ll be able to build a real-time data pipeline from scratch using Delta Live Tables, leverage CI/CD tools to deploy data pipeline changes automatically across deployment environments, and monitor, control, and optimize cloud costs.
This book is for data engineers looking to streamline data ingestion, transformation, and orchestration tasks. Data analysts responsible for managing and processing lakehouse data for analysis, reporting, and visualization will also find this book beneficial. Additionally, DataOps/DevOps engineers will find this book helpful for automating the testing and deployment of data pipelines, optimizing table tasks, and tracking data lineage within the lakehouse. Beginner-level knowledge of Apache Spark and Python is needed to make the most out of this book.
„Über diesen Titel“ kann sich auf eine andere Ausgabe dieses Titels beziehen.
Please note that we do not offer Priority shipping to any country.
We currently do not ship to the below countries:
Afghanistan
Bhutan
Brazil
Brunei Darussalam
Channel Islands
Chile
Israel
Lao
Mexico
Russian Federation
Saudi Arabia
South Africa
Yemen
Please do not attempt to place orders with any of these countries as a ship to address - they will be cancelled.