I wrote a piece of SQL that had some new T-SQL syntax in it: IS DISTINCT FROM. This predicate checks if two expressions are equal taking NULL into account. I successfully ran the code on an Azure SQL Database. However, when I tried to deploy the code using an Azure Devops pipeline, it returned the […]
Speaking at dataMinds Connect 2024
I’m very excited to announce I’ve been selected as a speaker for dataMinds Connect 2024, Belgium’s (and maybe/probably/most likely mainland Europe’s) best and greatest Microsoft data platform conference. My session is about a topic I really love: writing window functions in T-SQL. Tackling the Gaps & Islands Problem with T-SQL Window Functions The gaps & […]
Dynamic Warehouse & Lakehouse Connections in Microsoft Fabric Data Pipelines
When you develop data pipelines in Microsoft Fabric (the Azure Data Factory equivalent in Fabric, not to be confused with deployment pipelines), you will most likely have some activities with a connection to a warehouse, a lakehouse or a KQL database (for the remainder of the blog post I’ll talk about a warehouse, but it […]
System Views in Microsoft Fabric – Query references an object that is not supported in distributed processing mode
I have a metadata-driven ELT framework that heavily relies on dynamic SQL to generate SQL statements that load data from views into a respective fact or dimension. Such a task is well suited for generation, since the pattern to load a type 1 SCD, type 2 SCD or a fact table is always the same. […]
PostgreSQL error in Fabric Pipelines – No pg_hba.conf Entry for Host
I came across a bit of a weird error when I was trying to connect to a PostGreSQL database (hosted as an Azure DB for PostGreSQL), where “xx.xx.xx.xxx” is some IP address: No pg_hba.conf entry for host “xx.xx.xx.xxx” Some context to get us started. I’m using Pipelines in Microsoft Fabric to copy data from an […]
Book Review – Fundamentals of Data Engineering
This book was recommended by some of my colleagues, so I decided to give it a try. After finishing the book, I was pleased that I did. As the title might give away, it’s a book about data engineering, but not from a pure technical perspective. It won’t teach you how to write Python or […]
Planview Portfolios REST API with ADF – Retrieving Actual Data
In the introduction of this blog post series, I explained the use case: extracting data from the Planview Portfolios REST API using Azure Data Factory. Any tool that can send HTTPS requests can work with the REST API, but our focus here is ADF (which can also be Synapse Pipelines or Fabric Pipelines). In the […]
Planview Portfolios REST API with ADF – Fetching the Access Token
In the introduction of this blog post series, I explained the use case: extracting data from the Planview Portfolios RESET API using Azure Data Factory. Any tool that can send HTTPS requests can work with the REST API, but our focus here is ADF (which can also be Synapse Pipelines or Fabric Pipelines). In the […]
Planview Portfolios REST API with ADF – The OData Feed
In the introduction, I explained the setup of the use case and why we need to use the OData feed: to get the list of Project IDs, because the REST API doesn’t have an endpoint for this. Weird, right? In this part of the series, we’ll dive into the Lookup activity, which uses the OData […]
Extracting Data from the Planview Portfolios REST API using ADF – Introduction
I have a client using Planview Portfolios to manage their PMO. As part of the data warehouse, we want to extract the project data from the Planview REST API. There are other methods to get data, such as using OData (which is used in the Power BI connector), SOAP Web Services (2005 called) and a […]