The concept of a lakehouse typically relies on the loading of raw data in native formats such as JSON, Parquet, Avro, etc., into the Delta format. A similar procedure is observed in Microsoft Fabric, where our initial task involves extracting files from source systems (utilizing tools such as Data Factory) and placing them in the […]
Author: Adrian Chodkowski
Six common mistakes made by Power BI developers
Power BI is a great platform that helps companies understand their data. However, it’s easy to make mistakes when using it. In this article, I will show six common mistakes made by Power BI developers, and most importantly, how to avoid them. Let’s get started! Skipping Documentation and Code Comments Alright, folks, here’s a mistake […]
Delta Lake 101 Part 3: Optimize ZOrdering and File Pruning
Today I would like to write few words about one of the most important techniques used to improve performance in your Lakehouse – two specific keywords: OPTIMIZE and ZORDER. Let’s start by introducing those two terms. The OPTIMIZE command in Delta Lake helps tidy up how data files are stored in a Delta table. It […]
Introduction to Lakehouses in Microsoft Fabric
In my latest post, I wrote a few words about the warehouses available within Fabric. Today, I would like to show you an end-to-end analytical scenario with the second most important item available on the platform, which is the Lakehouse. Microsoft’s Fabric Lakehouse serves as a cutting-edge data architecture platform that consolidates the storage, management, […]