Join us at FabCon Atlanta from March 16 - 20, 2026, for the ultimate Fabric, Power BI, AI and SQL community-led event. Save $200 with code FABCOMM.
Register now!Calling all Data Engineers! Fabric Data Engineer (Exam DP-700) live sessions are back! Starting October 16th. Sign up.
I am working on my feature branch using a copyjob to push data from a remote SQL into a lakehouse.
Once I merge my code to the main branch and open the production workspace the copyjob is broke, for obvious reasons, the lakehouse id changed.
Question is, how to mantain the code for jumping between branches and not breaking during merges.
Hi @yllsuarez76 ,
Did the above suggestions help with your scenario? if that is the case, you can consider Kudo or Accept the helpful suggestions to help others who faced similar requirements.
Regards,
Xiaoxin Sheng
HI @yllsuarez76,
AFAIK, current it seems not support these feature, perhaps you can submit an idea to improve the 'copy job' usage:
Regards
Xiaoxin Sheng
We maintain a table of workspace/lakehouse ids by name and environment and look these up at pipeline run time.
You can automate the creation of this table using sempy-labs, the API, and/or notebookutils.lakehouse.get.
Technically you could also dynamically look these up in a notebook that is ran before any other job that outputs in its exitValue the list of lakehouses and ids.
Join the Fabric FabCon Global Hackathon—running virtually through Nov 3. Open to all skill levels. $10,000 in prizes!
Check out the September 2025 Fabric update to learn about new features.
User | Count |
---|---|
18 | |
5 | |
4 | |
3 | |
2 |