Power BI is turning 10! Tune in for a special live episode on July 24 with behind-the-scenes stories, product evolution highlights, and a sneak peek at what’s in store for the future.
Save the dateEnhance your career with this limited time 50% discount on Fabric and Power BI exams. Ends August 31st. Request your voucher.
Hi,
when Data copy will support updates and not just adding rows?
I have a lot of tables which can have updated records, so I want to update my destination not just adding new rows.
thanks.
Solved! Go to Solution.
ok... unfortunately it's not supporting deletes.
and the update appear to update all the rows matching the ID without option to select a hash or timestamp column.
well... I'll still use SQL statements for now. still faster with more control than relying on external tools like the copy job.
We often do a clear out of tables or certain rows in tables, but we have a few datasets that are so large that is slows things down.
This can de done with just pipeline activities (copy and script activity), notebooks are not always needed. I would only use a notebook when i need to do something with a lakehouse.
@smeetsh We always have Notebooks. No issues. Instead of Data Factory, Use Notebooks wherever we can. Databricks has implemented Delta Lake and then Microsoft follows Databricks. ( Delta Lake is based on Python and Scala ).
I could be wrong here, but I don't think a notebook can write to a warehouse, which for us at least would be the issue and it may be as well for the topic starter
Copy jobs in Data Factory means there is no need for updates. Overwrite where ever possible ( Data Lake and Data Lakehouse ) ( Apache Spark and Delta Lake ) or ( Merge in Microsoft Fabric or
Incremental Refresh with below code ( Append in Microsoft Fabric )
ok... unfortunately it's not supporting deletes.
and the update appear to update all the rows matching the ID without option to select a hash or timestamp column.
well... I'll still use SQL statements for now. still faster with more control than relying on external tools like the copy job.
What you could do is incorporate a where statement in the update statement, where you do something like
"WHERE hash in (select hash from TABLE) AND time_stamp_column > [your timestamp]
The above code is rough and probaly has some syntax errors, but I hope you understand the geste of the logic behind it 🙂
Cheers
Hans
Hi @Jerome22 ,
Hope you were able to get some clarity on the issue.
If any of the responses has addressed your query, kindly mark it as accepted solution so that the other members can find it easily.
Thank you
There is no option for warehouse for this yet, I did see an upsert (preview) button appear on a lakehouse though. The best way for now is to do it using SQL code, whwere you check for an existing record, which you than update (UPDATE TABLE, set column = etc etc ) and when a record does not exist you will do an insert
Hi @Jerome22 ,
Hope you were able to get some clarity on the issue. Let us know if need any help.
If the response has addressed your query, kindly mark it as accepted solution so that the other members can find it easily.
Thank you
Hi @Jerome22 ,
Unfortunately, in Microsoft Fabric, the Copy Data activity does support an "Upsert" behavior. You can however use workarounds like Dataflows or SQL logic.
You can refer-
upsert option missing in copy activity of microsoft fabric data pipeline
Hope this helps!
If the response has addressed your query, please accept it as a solution so other members can easily find it.
Thank You
This is your chance to engage directly with the engineering team behind Fabric and Power BI. Share your experiences and shape the future.
User | Count |
---|---|
5 | |
4 | |
2 | |
2 | |
2 |