Skip to main content
cancel
Showing results for 
Search instead for 
Did you mean: 

Join us at FabCon Vienna from September 15-18, 2025, for the ultimate Fabric, Power BI, SQL, and AI community-led learning event. Save €200 with code FABCOMM. Get registered

Reply
jk-dbt
New Member

DBT Orchestration

Hi,

I have a question on dbt orchestration.

 

I'm not sure how I can trigger a production dbt run in a Fabric pipeline and hoping someone in the community had a suggestion.

The dbt-fabric adaptor is working perfectly.  The config is good and I can target dev and prod with a manual dbt run with no issues.

 

However, I would like to have run without me having to be at desk pressing go every morning.  Does Fabric have a way to pull the from the repo and trigger on a schedule?

 

Thanks a mil.

2 ACCEPTED SOLUTIONS

OK thanks.  At the moment you need to run dbt core (and therefore dbt run) on some non-fabric compute unfortunately, like a VM.  

View solution in original post

You can clone your code from git in the notebook and then run dbt in the notebook.
https://www.linkedin.com/pulse/dbt-microsoft-fabric-my-journ-nghia-tran/

 

Note: Change runtime version in Spark Setting from 1.2 to 1.1 (Runtime version 1.2 does not have git).

View solution in original post

9 REPLIES 9
sanetaboada
Regular Visitor

If you need an "all fabric" solution, I think the option there would be to use Fabic Data Factory's Data Workflows (Apache Airflow behind the hood). I.e. configure Data Workflow to point to a git repo where your dbt code is, and let it orchestrate your dbt jobs. This will involve learning Apache Airflow though.

 

I'm gearing towards that path since my team's skills are mostly dbt and I wanted to build a solution under the promise of Microsoft Fabric's "complete end-to-end data platform" without having to provision VM's, external integrations, etc.

 

But beware though, Data Factory Data Workflow is still in preview and is poorly documented. Until now I'm still scratching my head on how to make it work! 😅

There's a DBT core pipeline activity that is on the release plan for this quarter. Supposedly you can store your project in the files section of a lakehouse and it will run it for you. I have heard storing it in GitHub/DevOps will be supported later.

 

If you're on DBT cloud, you can trigger it by calling the job API from a pipeline using a web activity. 

Christophe93100
Frequent Visitor

Hi dears

 

Do you know ,how can we call a DBT cloud job ( for a project ) on a Microsoft

  Fabric datafactory pipeline tasks ? 

 

Via call to dbt cloud Api ?

 

WE would like Fabric datafactory pipelines to be full chain orchestrator 

AndyDDC
Super User
Super User

Hi @jk-dbt just a quick question: where are you running dbt core, is it on your desktop PC, or on a VM etc?

Hey @AndyDDC .  Appreciate you coming back on this.  I'm developing on my local machine.

The basic workflow I'd like to have if possible would be like: 

  • Analytics engineers develop on their local machine.
  • Push to github.
  • scheduled Fabric pipeline somehow clones the repo.
  • Fabric somehow does a dbt run
  • Alert messages get emailed ….

One idea I’ve seen floating around is to have a notebook do the lat 3 steps above.  I actually pip installed dbt in the notebook, got all excited but then got stuck trying to clone the repo.

You can clone your code from git in the notebook and then run dbt in the notebook.
https://www.linkedin.com/pulse/dbt-microsoft-fabric-my-journ-nghia-tran/

 

Note: Change runtime version in Spark Setting from 1.2 to 1.1 (Runtime version 1.2 does not have git).

That worked.....nice one Nghia!

OK thanks.  At the moment you need to run dbt core (and therefore dbt run) on some non-fabric compute unfortunately, like a VM.  

This is a great question and one I started to ponder myself about a year ago when I was working at Microsoft.

 

In my case I wanted to use native scheduling functionality in Fabric and avoid an external Python environment to host dbt. 

 

It led me to developing a new dbt-adapter that allows the dbt project to be built locally without connection to the fabric workspace and for it to generate notebooks instead of directly execute the sql statements against the Fabric endpoint.

 

 I wanted to use the lakehouse rather than the warehouse in fabric and I wanted to keep all or the transformation logic in spark sql rather than tsql.


It’s working very well so far and it gives the best of both worlds in my opinion. All of dbt’s benefits such as catalog, lineage, model dependency management, tests etc but with native Fabric artefacts generated. This also allows you to use native fabric and git functionality to migrate through dev, test, prod without having to setup a devops runner to host dbt. I am currently open sourcing the adapter so If you’d like more info hit me up on linked in

 

 https://www.linkedin.com/in/john-rampono-8a73729?utm_source=share&utm_campaign=share_via&utm_content...

Helpful resources

Announcements
Join our Fabric User Panel

Join our Fabric User Panel

This is your chance to engage directly with the engineering team behind Fabric and Power BI. Share your experiences and shape the future.

June FBC25 Carousel

Fabric Monthly Update - June 2025

Check out the June 2025 Fabric update to learn about new features.

June 2025 community update carousel

Fabric Community Update - June 2025

Find out what's new and trending in the Fabric community.