Running a Power BI Performance load test using Visual Studio Code
This is the final part of my blog series for Power BI Performance testing, where I will finally run the Power BI Performance Load testing using Visual Studio Code. In this blog post I will show you how I set up the test, run the test and view the outputs from the performance testing. Previous Blog posts: Why should I…
Updating the PBIReport.JSON file with the Power BI Performance load testing Values
In the 3rd part of my blog series on Power BI Performance testing, we are now going to take all the values we complete in the previous blog post and update the PBIReport.JSON file. The reason I did this in a separate blog post is because working with JSON files can be tricky and adding to that putting in GUIDS…
How to get the details for Power BI Performance load testing
When completing the Power BI performance load testing, you will need to get details from your Power BI report and App Workspace, which will later be used in the PBIReport.JSON file. In this blog post I will show you how to find those details, so that when it comes time to add it to the PBIReport.JSON file, it will be…
Why should I complete Power BI Performance Load Testing & initial setup
It is becoming increasingly important to understand how the Power BI reports/Semantic Model that are being used in your organization are performing. When using Fabric Capacities this can potentially be of critical importance, because a single report that is not well designed could cripple or bring down your capacity. By completing Power BI Performance load testing before it goes into…
Using a Python Notebook Loop through a data frame and write once to a Lakehouse Table
In this blog post I am going to explain how to loop through a data frame to query data and write once to a Lakehouse table. The example I will use is to loop through a list of dates which I get from my date table, then query an API, append to an existing data frame and finally write once…
Using a Python Notebook to Query Multiple Lakehouse Tables into a Data frame
In this blog post I am going to explain how to query multiple Lakehouse tables into a data frame. The example I am going to use is when you want to load new data into your staging tables, but you need to know the max date from your previous data load. I am going to build on my previous blog…
Using a Python Notebook using Semantic Link Labs to write a DAX Query output to a Lakehouse Table
In this blog post I am going to explain how to use a Python Notebook using the Semantic Link module, to run a DAX query and write the output to a Lakehouse table. I will show you how to install a Python library and then use it within my python notebook. If you are interested in more details on why…
How to use SQL Code to write to Delta Tables in a Python Notebook
In this blog post I am going to explain how to use SQL Code to write to Delta Tables in a Microsoft Fabric Lakehouse. If you are interested in more details on why I am using a Python Notebook please read this blog post: What are the advantages of using Python Notebook vs a Spark Notebook in Microsoft Fabric –…
How to use SQL Code to read from Delta Tables in a Python Notebook
I come from a TSQL background, so using SQL makes it easy for me to work with data. There are multiple ways to use SQL in a PySpark notebook, and when I started using a Python notebook it was not so straightforward. In this blog post I will show you how I use SQL Code. As mentioned previously I am…
What are the advantages of using Python Notebook vs a Spark Notebook in Microsoft Fabric
I would like to start off because I am by no means a Python expert. I have been learning Python for about 2 years 😊 This is the first blog post in a series of blog posts where I dive into how to use Python notebooks instead of Spark notebooks. For example, I will show you how to run a…