Power BI is turning 10, and we’re marking the occasion with a special community challenge. Use your creativity to tell a story, uncover trends, or highlight something unexpected.
Get startedJoin us at FabCon Vienna from September 15-18, 2025, for the ultimate Fabric, Power BI, SQL, and AI community-led learning event. Save €200 with code FABCOMM. Get registered
I am working on my feature branch using a copyjob to push data from a remote SQL into a lakehouse.
Once I merge my code to the main branch and open the production workspace the copyjob is broke, for obvious reasons, the lakehouse id changed.
Question is, how to mantain the code for jumping between branches and not breaking during merges.
Hi @yllsuarez76 ,
Did the above suggestions help with your scenario? if that is the case, you can consider Kudo or Accept the helpful suggestions to help others who faced similar requirements.
Regards,
Xiaoxin Sheng
HI @yllsuarez76,
AFAIK, current it seems not support these feature, perhaps you can submit an idea to improve the 'copy job' usage:
Regards
Xiaoxin Sheng
We maintain a table of workspace/lakehouse ids by name and environment and look these up at pipeline run time.
You can automate the creation of this table using sempy-labs, the API, and/or notebookutils.lakehouse.get.
Technically you could also dynamically look these up in a notebook that is ran before any other job that outputs in its exitValue the list of lakehouses and ids.
This is your chance to engage directly with the engineering team behind Fabric and Power BI. Share your experiences and shape the future.
Check out the June 2025 Fabric update to learn about new features.
User | Count |
---|---|
9 | |
5 | |
4 | |
3 | |
2 |