Anbieter: GreatBookPrices, Columbia, MD, USA
EUR 40,03
Währung umrechnenAnzahl: 4 verfügbar
In den WarenkorbZustand: New.
Anbieter: Lakeside Books, Benton Harbor, MI, USA
EUR 38,85
Währung umrechnenAnzahl: Mehr als 20 verfügbar
In den WarenkorbZustand: New. Brand New! Not Overstocks or Low Quality Book Club Editions! Direct From the Publisher! We're not a giant, faceless warehouse organization! We're a small town bookstore that loves books and loves it's customers! Buy from Lakeside Books!
Anbieter: Lucky's Textbooks, Dallas, TX, USA
EUR 42,45
Währung umrechnenAnzahl: Mehr als 20 verfügbar
In den WarenkorbZustand: New.
Anbieter: GreatBookPrices, Columbia, MD, USA
EUR 45,16
Währung umrechnenAnzahl: 4 verfügbar
In den WarenkorbZustand: As New. Unread book in perfect condition.
Anbieter: California Books, Miami, FL, USA
EUR 47,54
Währung umrechnenAnzahl: Mehr als 20 verfügbar
In den WarenkorbZustand: New.
Anbieter: Grand Eagle Retail, Bensenville, IL, USA
Erstausgabe
EUR 50,61
Währung umrechnenAnzahl: 1 verfügbar
In den WarenkorbPaperback. Zustand: new. Paperback. Design and implement a modern data lakehouse on the Azure Data Platform using Delta Lake, Apache Spark, Azure Databricks, Azure Synapse Analytics, and Snowflake. This book teaches you the intricate details of the Data Lakehouse Paradigm and how to efficiently design a cloud-based data lakehouse using highly performant and cutting-edge Apache Spark capabilities using Azure Databricks, Azure Synapse Analytics, and Snowflake. You will learn to write efficient PySpark code for batch and streaming ELT jobs on Azure. And you will follow along with practical, scenario-based examples showing how to apply the capabilities of Delta Lake and Apache Spark to optimize performance, and secure, share, and manage a high volume, high velocity, and high variety of data in your lakehouse with ease.The patterns of success that you acquire from reading this book will help you hone your skills to build high-performing and scalable ACID-compliant lakehouses using flexible and cost-efficient decoupled storage and compute capabilities. Extensive coverage of Delta Lake ensures that you are aware of and can benefit from all that this new, open source storage layer can offer. In addition to the deep examples on Databricks in the book, there is coverage of alternative platforms such as Synapse Analytics and Snowflake so that you can make the right platform choice for your needs.After reading this book, you will be able to implement Delta Lake capabilities, including Schema Evolution, Change Feed, Live Tables, Sharing, and Clones to enable better business intelligence and advanced analytics on your data within the Azure Data Platform.What You Will LearnImplement the Data Lakehouse Paradigm on Microsofts Azure cloud platformBenefit from the new Delta Lake open-source storage layer for data lakehouses Take advantage of schema evolution, change feeds, live tables, and moreWritefunctional PySpark code for data lakehouse ELT jobsOptimize Apache Spark performance through partitioning, indexing, and other tuning optionsChoose between alternatives such as Databricks, Synapse Analytics, and SnowflakeWho This Book Is ForData, analytics, and AI professionals at all levels, including data architect and data engineer practitioners. Also for data professionals seeking patterns of success by which to remain relevant as they learn to build scalable data lakehouses for their organizations and customers who are migrating into the modern Azure Data Platform. Intermediate user level Shipping may be from multiple locations in the US or from the UK, depending on stock availability.
EUR 53,29
Währung umrechnenAnzahl: 1 verfügbar
In den WarenkorbPAP. Zustand: New. New Book. Shipped from UK. Established seller since 2000.
EUR 61,33
Währung umrechnenAnzahl: 8 verfügbar
In den WarenkorbPaperback. Zustand: New. 1st ed. Design and implement a modern data lakehouse on the Azure Data Platform using Delta Lake, Apache Spark, Azure Databricks, Azure Synapse Analytics, and Snowflake. This book teaches you the intricate details of the Data Lakehouse Paradigm and how to efficiently design a cloud-based data lakehouse using highly performant and cutting-edge Apache Spark capabilities using Azure Databricks, Azure Synapse Analytics, and Snowflake. You will learn to write efficient PySpark code for batch and streaming ELT jobs on Azure. And you will follow along with practical, scenario-based examples showing how to apply the capabilities of Delta Lake and Apache Spark to optimize performance, and secure, share, and manage a high volume, high velocity, and high variety of data in your lakehouse with ease.The patterns of success that you acquire from reading this book will help you hone your skills to build high-performing and scalable ACID-compliant lakehouses using flexible and cost-efficient decoupled storage and compute capabilities. Extensive coverage of Delta Lake ensures that you are aware of and can benefit from all that this new, open source storage layer can offer. In addition to the deep examples on Databricks in the book, there is coverage of alternative platforms such as Synapse Analytics and Snowflake so that you can make the right platform choice for your needs.After reading this book, you will be able to implement Delta Lake capabilities, including Schema Evolution, Change Feed, Live Tables, Sharing, and Clones to enable better business intelligence and advanced analytics on your data within the Azure Data Platform.What You Will LearnImplement the Data Lakehouse Paradigm on Microsoft's Azure cloud platformBenefit from the new Delta Lake open-source storage layer for data lakehouses Take advantage of schema evolution, change feeds, live tables, and moreWritefunctional PySpark code for data lakehouse ELT jobsOptimize Apache Spark performance through partitioning, indexing, and other tuning optionsChoose between alternatives such as Databricks, Synapse Analytics, and SnowflakeWho This Book Is ForData, analytics, and AI professionals at all levels, including data architect and data engineer practitioners. Also for data professionals seeking patterns of success by which to remain relevant as they learn to build scalable data lakehouses for their organizations and customers who are migrating into the modern Azure Data Platform.
Anbieter: BargainBookStores, Grand Rapids, MI, USA
EUR 62,51
Währung umrechnenAnzahl: 5 verfügbar
In den WarenkorbPaperback or Softback. Zustand: New. The Azure Data Lakehouse Toolkit: Building and Scaling Data Lakehouses with Delta Lake, Apache Spark, Azure Databricks and Synapse Analytics, and Snow. Book.
EUR 64,36
Währung umrechnenAnzahl: 1 verfügbar
In den WarenkorbZustand: New. Brand New Original US Edition. Customer service! Satisfaction Guaranteed.
EUR 60,88
Währung umrechnenAnzahl: 2 verfügbar
In den WarenkorbZustand: New. pp. 465.
EUR 60,21
Währung umrechnenAnzahl: 2 verfügbar
In den WarenkorbZustand: New. pp. 465.
Anbieter: Rarewaves.com USA, London, LONDO, Vereinigtes Königreich
Erstausgabe
EUR 70,22
Währung umrechnenAnzahl: 8 verfügbar
In den WarenkorbPaperback. Zustand: New. 1st ed. Design and implement a modern data lakehouse on the Azure Data Platform using Delta Lake, Apache Spark, Azure Databricks, Azure Synapse Analytics, and Snowflake. This book teaches you the intricate details of the Data Lakehouse Paradigm and how to efficiently design a cloud-based data lakehouse using highly performant and cutting-edge Apache Spark capabilities using Azure Databricks, Azure Synapse Analytics, and Snowflake. You will learn to write efficient PySpark code for batch and streaming ELT jobs on Azure. And you will follow along with practical, scenario-based examples showing how to apply the capabilities of Delta Lake and Apache Spark to optimize performance, and secure, share, and manage a high volume, high velocity, and high variety of data in your lakehouse with ease.The patterns of success that you acquire from reading this book will help you hone your skills to build high-performing and scalable ACID-compliant lakehouses using flexible and cost-efficient decoupled storage and compute capabilities. Extensive coverage of Delta Lake ensures that you are aware of and can benefit from all that this new, open source storage layer can offer. In addition to the deep examples on Databricks in the book, there is coverage of alternative platforms such as Synapse Analytics and Snowflake so that you can make the right platform choice for your needs.After reading this book, you will be able to implement Delta Lake capabilities, including Schema Evolution, Change Feed, Live Tables, Sharing, and Clones to enable better business intelligence and advanced analytics on your data within the Azure Data Platform.What You Will LearnImplement the Data Lakehouse Paradigm on Microsoft's Azure cloud platformBenefit from the new Delta Lake open-source storage layer for data lakehouses Take advantage of schema evolution, change feeds, live tables, and moreWritefunctional PySpark code for data lakehouse ELT jobsOptimize Apache Spark performance through partitioning, indexing, and other tuning optionsChoose between alternatives such as Databricks, Synapse Analytics, and SnowflakeWho This Book Is ForData, analytics, and AI professionals at all levels, including data architect and data engineer practitioners. Also for data professionals seeking patterns of success by which to remain relevant as they learn to build scalable data lakehouses for their organizations and customers who are migrating into the modern Azure Data Platform.
Anbieter: GreatBookPricesUK, Woodford Green, Vereinigtes Königreich
EUR 53,02
Währung umrechnenAnzahl: Mehr als 20 verfügbar
In den WarenkorbZustand: As New. Unread book in perfect condition.
Anbieter: GreatBookPricesUK, Woodford Green, Vereinigtes Königreich
EUR 53,28
Währung umrechnenAnzahl: Mehr als 20 verfügbar
In den WarenkorbZustand: New.
EUR 64,50
Währung umrechnenAnzahl: 4 verfügbar
In den WarenkorbZustand: New. pp. 465.
Anbieter: Ria Christie Collections, Uxbridge, Vereinigtes Königreich
EUR 64,07
Währung umrechnenAnzahl: Mehr als 20 verfügbar
In den WarenkorbZustand: New. In.
Anbieter: Chiron Media, Wallingford, Vereinigtes Königreich
EUR 63,08
Währung umrechnenAnzahl: 10 verfügbar
In den WarenkorbPF. Zustand: New.
Anbieter: Revaluation Books, Exeter, Vereinigtes Königreich
EUR 67,34
Währung umrechnenAnzahl: 2 verfügbar
In den WarenkorbPaperback. Zustand: Brand New. 487 pages. 10.00x7.01x0.98 inches. In Stock.
Anbieter: CitiRetail, Stevenage, Vereinigtes Königreich
Erstausgabe
EUR 55,62
Währung umrechnenAnzahl: 1 verfügbar
In den WarenkorbPaperback. Zustand: new. Paperback. Design and implement a modern data lakehouse on the Azure Data Platform using Delta Lake, Apache Spark, Azure Databricks, Azure Synapse Analytics, and Snowflake. This book teaches you the intricate details of the Data Lakehouse Paradigm and how to efficiently design a cloud-based data lakehouse using highly performant and cutting-edge Apache Spark capabilities using Azure Databricks, Azure Synapse Analytics, and Snowflake. You will learn to write efficient PySpark code for batch and streaming ELT jobs on Azure. And you will follow along with practical, scenario-based examples showing how to apply the capabilities of Delta Lake and Apache Spark to optimize performance, and secure, share, and manage a high volume, high velocity, and high variety of data in your lakehouse with ease.The patterns of success that you acquire from reading this book will help you hone your skills to build high-performing and scalable ACID-compliant lakehouses using flexible and cost-efficient decoupled storage and compute capabilities. Extensive coverage of Delta Lake ensures that you are aware of and can benefit from all that this new, open source storage layer can offer. In addition to the deep examples on Databricks in the book, there is coverage of alternative platforms such as Synapse Analytics and Snowflake so that you can make the right platform choice for your needs.After reading this book, you will be able to implement Delta Lake capabilities, including Schema Evolution, Change Feed, Live Tables, Sharing, and Clones to enable better business intelligence and advanced analytics on your data within the Azure Data Platform.What You Will LearnImplement the Data Lakehouse Paradigm on Microsofts Azure cloud platformBenefit from the new Delta Lake open-source storage layer for data lakehouses Take advantage of schema evolution, change feeds, live tables, and moreWritefunctional PySpark code for data lakehouse ELT jobsOptimize Apache Spark performance through partitioning, indexing, and other tuning optionsChoose between alternatives such as Databricks, Synapse Analytics, and SnowflakeWho This Book Is ForData, analytics, and AI professionals at all levels, including data architect and data engineer practitioners. Also for data professionals seeking patterns of success by which to remain relevant as they learn to build scalable data lakehouses for their organizations and customers who are migrating into the modern Azure Data Platform. Intermediate user level Shipping may be from our UK warehouse or from our Australian or US warehouses, depending on stock availability.
EUR 63,79
Währung umrechnenAnzahl: 8 verfügbar
In den WarenkorbPaperback. Zustand: New. 1st ed. Design and implement a modern data lakehouse on the Azure Data Platform using Delta Lake, Apache Spark, Azure Databricks, Azure Synapse Analytics, and Snowflake. This book teaches you the intricate details of the Data Lakehouse Paradigm and how to efficiently design a cloud-based data lakehouse using highly performant and cutting-edge Apache Spark capabilities using Azure Databricks, Azure Synapse Analytics, and Snowflake. You will learn to write efficient PySpark code for batch and streaming ELT jobs on Azure. And you will follow along with practical, scenario-based examples showing how to apply the capabilities of Delta Lake and Apache Spark to optimize performance, and secure, share, and manage a high volume, high velocity, and high variety of data in your lakehouse with ease.The patterns of success that you acquire from reading this book will help you hone your skills to build high-performing and scalable ACID-compliant lakehouses using flexible and cost-efficient decoupled storage and compute capabilities. Extensive coverage of Delta Lake ensures that you are aware of and can benefit from all that this new, open source storage layer can offer. In addition to the deep examples on Databricks in the book, there is coverage of alternative platforms such as Synapse Analytics and Snowflake so that you can make the right platform choice for your needs.After reading this book, you will be able to implement Delta Lake capabilities, including Schema Evolution, Change Feed, Live Tables, Sharing, and Clones to enable better business intelligence and advanced analytics on your data within the Azure Data Platform.What You Will LearnImplement the Data Lakehouse Paradigm on Microsoft's Azure cloud platformBenefit from the new Delta Lake open-source storage layer for data lakehouses Take advantage of schema evolution, change feeds, live tables, and moreWritefunctional PySpark code for data lakehouse ELT jobsOptimize Apache Spark performance through partitioning, indexing, and other tuning optionsChoose between alternatives such as Databricks, Synapse Analytics, and SnowflakeWho This Book Is ForData, analytics, and AI professionals at all levels, including data architect and data engineer practitioners. Also for data professionals seeking patterns of success by which to remain relevant as they learn to build scalable data lakehouses for their organizations and customers who are migrating into the modern Azure Data Platform.
Anbieter: AussieBookSeller, Truganina, VIC, Australien
Erstausgabe
EUR 82,53
Währung umrechnenAnzahl: 1 verfügbar
In den WarenkorbPaperback. Zustand: new. Paperback. Design and implement a modern data lakehouse on the Azure Data Platform using Delta Lake, Apache Spark, Azure Databricks, Azure Synapse Analytics, and Snowflake. This book teaches you the intricate details of the Data Lakehouse Paradigm and how to efficiently design a cloud-based data lakehouse using highly performant and cutting-edge Apache Spark capabilities using Azure Databricks, Azure Synapse Analytics, and Snowflake. You will learn to write efficient PySpark code for batch and streaming ELT jobs on Azure. And you will follow along with practical, scenario-based examples showing how to apply the capabilities of Delta Lake and Apache Spark to optimize performance, and secure, share, and manage a high volume, high velocity, and high variety of data in your lakehouse with ease.The patterns of success that you acquire from reading this book will help you hone your skills to build high-performing and scalable ACID-compliant lakehouses using flexible and cost-efficient decoupled storage and compute capabilities. Extensive coverage of Delta Lake ensures that you are aware of and can benefit from all that this new, open source storage layer can offer. In addition to the deep examples on Databricks in the book, there is coverage of alternative platforms such as Synapse Analytics and Snowflake so that you can make the right platform choice for your needs.After reading this book, you will be able to implement Delta Lake capabilities, including Schema Evolution, Change Feed, Live Tables, Sharing, and Clones to enable better business intelligence and advanced analytics on your data within the Azure Data Platform.What You Will LearnImplement the Data Lakehouse Paradigm on Microsofts Azure cloud platformBenefit from the new Delta Lake open-source storage layer for data lakehouses Take advantage of schema evolution, change feeds, live tables, and moreWritefunctional PySpark code for data lakehouse ELT jobsOptimize Apache Spark performance through partitioning, indexing, and other tuning optionsChoose between alternatives such as Databricks, Synapse Analytics, and SnowflakeWho This Book Is ForData, analytics, and AI professionals at all levels, including data architect and data engineer practitioners. Also for data professionals seeking patterns of success by which to remain relevant as they learn to build scalable data lakehouses for their organizations and customers who are migrating into the modern Azure Data Platform. Intermediate user level Shipping may be from our Sydney, NSW warehouse or from our UK or US warehouse, depending on stock availability.
EUR 64,19
Währung umrechnenAnzahl: 2 verfügbar
In den WarenkorbTaschenbuch. Zustand: Neu. Neuware -Design and implement a modern data lakehouse on the Azure Data Platform using Delta Lake, Apache Spark, Azure Databricks, Azure Synapse Analytics, and Snowflake. This book teaches you the intricate details of the Data Lakehouse Paradigm and how to efficiently design a cloud-based data lakehouse using highly performant and cutting-edge Apache Spark capabilities using Azure Databricks, Azure Synapse Analytics, and Snowflake. You will learn to write efficient PySpark code for batch and streaming ELT jobs on Azure. And you will follow along with practical, scenario-based examples showing how to apply the capabilities of Delta Lake and Apache Spark to optimize performance, and secure, share, and manage a high volume, high velocity, and high variety of data in your lakehouse with ease.The patterns of success that you acquire from reading this book will help you hone your skills to build high-performing and scalable ACID-compliant lakehouses using flexible and cost-efficient decoupled storage and compute capabilities. Extensive coverage of Delta Lake ensures that you are aware of and can benefit from all that this new, open source storage layer can offer. In addition to the deep examples on Databricks in the book, there is coverage of alternative platforms such as Synapse Analytics and Snowflake so that you can make the right platform choice for your needs.APress in Springer Science + Business Media, Heidelberger Platz 3, 14197 Berlin 488 pp. Englisch.
Anbieter: Rarewaves.com UK, London, Vereinigtes Königreich
Erstausgabe
EUR 65,59
Währung umrechnenAnzahl: 8 verfügbar
In den WarenkorbPaperback. Zustand: New. 1st ed. Design and implement a modern data lakehouse on the Azure Data Platform using Delta Lake, Apache Spark, Azure Databricks, Azure Synapse Analytics, and Snowflake. This book teaches you the intricate details of the Data Lakehouse Paradigm and how to efficiently design a cloud-based data lakehouse using highly performant and cutting-edge Apache Spark capabilities using Azure Databricks, Azure Synapse Analytics, and Snowflake. You will learn to write efficient PySpark code for batch and streaming ELT jobs on Azure. And you will follow along with practical, scenario-based examples showing how to apply the capabilities of Delta Lake and Apache Spark to optimize performance, and secure, share, and manage a high volume, high velocity, and high variety of data in your lakehouse with ease.The patterns of success that you acquire from reading this book will help you hone your skills to build high-performing and scalable ACID-compliant lakehouses using flexible and cost-efficient decoupled storage and compute capabilities. Extensive coverage of Delta Lake ensures that you are aware of and can benefit from all that this new, open source storage layer can offer. In addition to the deep examples on Databricks in the book, there is coverage of alternative platforms such as Synapse Analytics and Snowflake so that you can make the right platform choice for your needs.After reading this book, you will be able to implement Delta Lake capabilities, including Schema Evolution, Change Feed, Live Tables, Sharing, and Clones to enable better business intelligence and advanced analytics on your data within the Azure Data Platform.What You Will LearnImplement the Data Lakehouse Paradigm on Microsoft's Azure cloud platformBenefit from the new Delta Lake open-source storage layer for data lakehouses Take advantage of schema evolution, change feeds, live tables, and moreWritefunctional PySpark code for data lakehouse ELT jobsOptimize Apache Spark performance through partitioning, indexing, and other tuning optionsChoose between alternatives such as Databricks, Synapse Analytics, and SnowflakeWho This Book Is ForData, analytics, and AI professionals at all levels, including data architect and data engineer practitioners. Also for data professionals seeking patterns of success by which to remain relevant as they learn to build scalable data lakehouses for their organizations and customers who are migrating into the modern Azure Data Platform.
Anbieter: BuchWeltWeit Ludwig Meier e.K., Bergisch Gladbach, Deutschland
EUR 64,19
Währung umrechnenAnzahl: 2 verfügbar
In den WarenkorbTaschenbuch. Zustand: Neu. This item is printed on demand - it takes 3-4 days longer - Neuware -Design and implement a modern data lakehouse on the Azure Data Platform using Delta Lake, Apache Spark, Azure Databricks, Azure Synapse Analytics, and Snowflake. This book teaches you the intricate details of the Data Lakehouse Paradigm and how to efficiently design a cloud-based data lakehouse using highly performant and cutting-edge Apache Spark capabilities using Azure Databricks, Azure Synapse Analytics, and Snowflake. You will learn to write efficient PySpark code for batch and streaming ELT jobs on Azure. And you will follow along with practical, scenario-based examples showing how to apply the capabilities of Delta Lake and Apache Spark to optimize performance, and secure, share, and manage a high volume, high velocity, and high variety of data in your lakehouse with ease.The patterns of success that you acquire from reading this book will help you hone your skills to build high-performing and scalable ACID-compliant lakehouses using flexible and cost-efficient decoupled storage and compute capabilities. Extensive coverage of Delta Lake ensures that you are aware of and can benefit from all that this new, open source storage layer can offer. In addition to the deep examples on Databricks in the book, there is coverage of alternative platforms such as Synapse Analytics and Snowflake so that you can make the right platform choice for your needs.After reading this book, you will be able to implement Delta Lake capabilities, including Schema Evolution, Change Feed, Live Tables, Sharing, and Clones to enable better business intelligence and advanced analytics on your data within the Azure Data Platform.What You Will LearnImplement the Data Lakehouse Paradigm on Microsoft's Azure cloud platformBenefit from the new Delta Lake open-source storage layer for data lakehousesTake advantage of schema evolution, change feeds, live tables, and moreWritefunctional PySpark code for data lakehouse ELT jobsOptimize Apache Spark performance through partitioning, indexing, and other tuning optionsChoose between alternatives such as Databricks, Synapse Analytics, and SnowflakeWho This Book Is ForData, analytics, and AI professionals at all levels, including data architect and data engineer practitioners. Also for data professionals seeking patterns of success by which to remain relevant as they learn to build scalable data lakehouses for their organizations and customers who are migrating into the modern Azure Data Platform. 488 pp. Englisch.
Anbieter: Revaluation Books, Exeter, Vereinigtes Königreich
EUR 65,09
Währung umrechnenAnzahl: 1 verfügbar
In den WarenkorbPaperback. Zustand: Brand New. 487 pages. 10.00x7.01x0.98 inches. In Stock. This item is printed on demand.
Anbieter: moluna, Greven, Deutschland
EUR 52,37
Währung umrechnenAnzahl: Mehr als 20 verfügbar
In den WarenkorbZustand: New. Dieser Artikel ist ein Print on Demand Artikel und wird nach Ihrer Bestellung fuer Sie gedruckt. Design and implement a modern data lakehouse on the Azure Data Platform using Delta Lake, Apache Spark, Azure Databricks, Azure Synapse Analytics, and Snowflake. This book teaches you the intricate details of the Data Lakehouse Paradigm and how to effici.
Anbieter: AHA-BUCH GmbH, Einbeck, Deutschland
EUR 64,96
Währung umrechnenAnzahl: 1 verfügbar
In den WarenkorbTaschenbuch. Zustand: Neu. nach der Bestellung gedruckt Neuware - Printed after ordering - Design and implement a modern data lakehouse on the Azure Data Platform using Delta Lake, Apache Spark, Azure Databricks, Azure Synapse Analytics, and Snowflake. This book teaches you the intricate details of the Data Lakehouse Paradigm and how to efficiently design a cloud-based data lakehouse using highly performant and cutting-edge Apache Spark capabilities using Azure Databricks, Azure Synapse Analytics, and Snowflake. You will learn to write efficient PySpark code for batch and streaming ELT jobs on Azure. And you will follow along with practical, scenario-based examples showing how to apply the capabilities of Delta Lake and Apache Spark to optimize performance, and secure, share, and manage a high volume, high velocity, and high variety of data in your lakehouse with ease.The patterns of success that you acquire from reading this book will help you hone your skills to build high-performing and scalable ACID-compliant lakehouses using flexible and cost-efficient decoupled storage and compute capabilities. Extensive coverage of Delta Lake ensures that you are aware of and can benefit from all that this new, open source storage layer can offer. In addition to the deep examples on Databricks in the book, there is coverage of alternative platforms such as Synapse Analytics and Snowflake so that you can make the right platform choice for your needs.After reading this book, you will be able to implement Delta Lake capabilities, including Schema Evolution, Change Feed, Live Tables, Sharing, and Clones to enable better business intelligence and advanced analytics on your data within the Azure Data Platform.What You Will LearnImplement the Data Lakehouse Paradigm on Microsoft's Azure cloud platformBenefit from the new Delta Lake open-source storage layer for data lakehousesTake advantage of schema evolution, change feeds, live tables, and moreWritefunctional PySpark code for data lakehouse ELT jobsOptimize Apache Spark performance through partitioning, indexing, and other tuning optionsChoose between alternatives such as Databricks, Synapse Analytics, and SnowflakeWho This Book Is ForData, analytics, and AI professionals at all levels, including data architect and data engineer practitioners. Also for data professionals seeking patterns of success by which to remain relevant as they learn to build scalable data lakehouses for their organizations and customers who are migrating into the modern Azure Data Platform.