Skip to main content
cancel
Showing results for 
Search instead for 
Did you mean: 

Find everything you need to get certified on Fabric—skills challenges, live sessions, exam prep, role guidance, and more. Get started

Reply
amien
Helper V
Helper V

reading huge table from Oracle (using On-Prem gateway)

What is the best approach in Fabric if i need to read 100 million of records from a Oracle using an on-prem gateway?

Do i have other options besides incremental reloads using pipelines/dataflowgen2

 

Thanks in advanced

 

 

5 REPLIES 5
amien
Helper V
Helper V

 

Thanks for your reply.

 

query is select * .. it takes hours and hours to load .. i stopped it. every day/week, more data is added to the table.

 

How many columns is * ?

Is the data slowly changing or can you use Incremental Refresh?

Is it any faster when you export to CSV (for example)?

has about 50 columns. 1 week of data already takes one hour. 

Incremental for each week would be possible. there are no smart fabric features that i can leverage right? I was looking for mirroring etc.

 

If i want to explore incremental and want to store the data in an ingestion layer (first layer). Then i have two questions:

 

* Shoud i use a pipleline for that? or better to use dataflow gen2?

* What is the best way to import the previous data from a CSV file into the lakehouse? Including adding the field mapping.

Here's another idea - create Parquet files from Oracle and then directly use them in the lakehouse.

lbendlin
Super User
Super User

Please provide more details. Is this a one time load or do you plan to refresh the data?  How fast is the Oracle source?  How complex is the query?

Helpful resources

Announcements
OCT PBI Update Carousel

Power BI Monthly Update - October 2024

Check out the October 2024 Power BI update to learn about new features.

October NL Carousel

Fabric Community Update - October 2024

Find out what's new and trending in the Fabric Community.

Top Solution Authors
Top Kudoed Authors