Don't miss your chance to take exam DP-600 or DP-700 on us!
Request nowFabric Data Days Monthly is back. Join us on March 26th for two expert-led sessions on 1) Getting Started with Fabric IQ and 2) Mapping & Spacial Analytics in Fabric. Register now
I am working on my feature branch using a copyjob to push data from a remote SQL into a lakehouse.
Once I merge my code to the main branch and open the production workspace the copyjob is broke, for obvious reasons, the lakehouse id changed.
Question is, how to mantain the code for jumping between branches and not breaking during merges.
Hi @yllsuarez76 ,
Did the above suggestions help with your scenario? if that is the case, you can consider Kudo or Accept the helpful suggestions to help others who faced similar requirements.
Regards,
Xiaoxin Sheng
HI @yllsuarez76,
AFAIK, current it seems not support these feature, perhaps you can submit an idea to improve the 'copy job' usage:
Regards
Xiaoxin Sheng
We maintain a table of workspace/lakehouse ids by name and environment and look these up at pipeline run time.
You can automate the creation of this table using sempy-labs, the API, and/or notebookutils.lakehouse.get.
Technically you could also dynamically look these up in a notebook that is ran before any other job that outputs in its exitValue the list of lakehouses and ids.
Share feedback directly with Fabric product managers, participate in targeted research studies and influence the Fabric roadmap.
Check out the February 2026 Fabric update to learn about new features.
| User | Count |
|---|---|
| 16 | |
| 7 | |
| 4 | |
| 3 | |
| 3 |