How to view or track the progress of Notebook while it is running in Microsoft Fabric
How to view or track the progress of Notebook while it is running in Microsoft Fabric I was recently working with a Notebook in Microsoft Fabric that was started via a Data Pipeline. The challenge I had was that I had no idea how far the notebook had gone (as there were quite a lot of cells in this particular…
Using a Service Principal to get all Entra ID Group Members into JSON File using a Python Notebook
Using a Service Principal to get all Entra ID Group Members into JSON File using a Python Notebook Sometimes it is useful to get all Group Members into a JSON file so that this could be used for reporting purposes. Reference Notebook: Get Entra ID Group Members – Power BI In the steps below I will show you how I…
Downloading Scanner API data using a Microsoft Fabric Notebook
Downloading Scanner API data using a Microsoft Fabric Notebook I was recently working with a customer where they had more then 100 app workspaces and I was running into some challenges when using the Scanner API in Power Automate. I then discovered this blog post where they detailed how to download the Scanner API data (DataXbi – admin-scan.py), it was…
How to read a Lakehouse table in another App Workspace – Microsoft Fabric
I was doing some work recently for a customer and they had data stored in different Lakehouse’s which was in a different App Workspace. I was pleasantly surprised that this can be quite easy to do. In my example below I am going to show you how in my notebook I can read a table in a Lakehouse table when…
Semantic-Link-Labs – Automate updating your Incremental Refresh Policy for your Semantic Model
The scenario here is that quite often there is a requirement to only keep data from a specific start date, or where it should be keeping data for the last N number of years (which is the first day in January). Currently in Power BI using the default Incremental refresh settings this is not possible. Typically, you must keep more…
Loading Fabric Lakehouse Tables with partitions
When loading data, it is always important to load the data with performance and scalability in mind. For lakehouse tables to return queries quickly and to scale it is essential to load your lakehouse tables with partitions. What I am going to show you in my blog post today is how to load data into a Lakehouse table where the…
How to get the TopN rows using Python in Fabric Notebooks
How to get the TopN rows using Python in Fabric Notebooks When working with data there are sometimes weird and wonderful requirements which must be created in order to get to the desired solution. In today’s blog post I had a situation where I wanted to get a single row with the highest duration. This is how I did it…
Looping through data using PySpark notebook in Fabric
Fabric Notebooks – Looping through data using PySpark Continuing with my existing blog series on what I’m learning with notebooks and PySpark. Today, I’m going to explain to you how I found a way to loop through data in a notebook. In this example, I’m going to show you how I loop through a range of dates, which can then…
Using Sempy to Authenticate to Fabric/Power BI APIs using Service Principal and Azure Key Vault
I have been doing a fair amount of work lately with Fabric Notebooks. I am always conscious to ensure that when I am authenticating using a Service Principal, I can make sure it is as secure as possible. To do this I have found that I can use the Azure Key Vault and Azure identity to successfully authenticate. By using…
How to add current DateTime to existing PySpark data frame in a Fabric Notebook
How to add current DateTime to existing PySpark data frame in a Fabric Notebook In the blog post below, I am going to describe how to add the current Date Time to your existing Spark data frame. This is really useful when I am inserting data into a Fabric Lakehouse table, and I want to know when the data got…