Skip to main content
cancel
Showing results for 
Search instead for 
Did you mean: 

Next up in the FabCon + SQLCon recap series: The roadmap for Microsoft SQL and Maximizing Developer experiences in Fabric. All sessions are available on-demand after the live show. Register now

Reply
Ramkishor
Helper II
Helper II

Capacity Recommendation for Migrating 80 AX2009 Reports to D365 F&O Using Microsoft Fabric

Hello all,

 

We need to migrate nearly 80 Project related AX2009 reports and embed them into D365 F&O. We are planning to use Microsoft Fabric to handle the data modelling anmd reporting layers and to Embed the PowerBI Reports into D365 F&O.

 

Here are the key points:

1. Number of reports - nearly 80(project related : budgets,invoices etc.)

2. Embedded report viewers 30

3. Estimated compressed data volume 50GB

4. Using Dataflows and Lakehouse for ETL from AX DB

* What is the recommended Microsoft Fabric capacity SKU to handle this?

* If anyone here is currently using Fabric to build Power BI reports and embed them into D365 F&O, your suggestions and real-world experiences would be **highly valuable. 

*  Also, which Fabric capacity SKU are you using in your project/reporting environments — and how is it performing?

2 ACCEPTED SOLUTIONS
v-saisrao-msft
Community Support
Community Support

Hi @Ramkishor,

Thank you for reaching out to the Microsoft Fabric Forum Community. 

Since you are working with multiple reports and users, handling larger volume of data, and using ETL through Dataflows and Lakehouse, I suggest starting with Microsoft Fabric F64 capacity. 

And if there are more frequent refreshes, complex models, or increased usage, you may want to scale up to F128 or higher capacity to maintain optimal performance for all workloads. 
Below is the link for your reference: 
Plan your capacity size - Microsoft Fabric | Microsoft Learn

 

Thank you. 

View solution in original post

burakkaragoz
Super User
Super User

Hi @Ramkishor ,

 

Based on what you shared, you're looking at a moderately heavy setup — 80 reports, 30 embedded viewers, and 50GB of compressed data is no small thing.

Here’s what I’d recommend:

  • F64 or F128 Fabric capacity would be a solid starting point. F64 gives you 64 CU (capacity units) which should handle your dataflows, Lakehouse refreshes, and embedded report usage pretty well.
  • If your reports are heavily interactive or refresh frequently, F128 might give you more breathing room, especially during peak hours.
  • Since you're embedding into D365 F&O, make sure your capacity is in the same region as your D365 tenant to avoid latency issues.

Also, keep in mind that Lakehouse and Dataflows can consume a good chunk of capacity during refreshes. You might want to schedule those during off-hours if possible.

Here’s a Microsoft doc that breaks down Fabric SKUs and use cases:
👉 Microsoft Learn 

Let me know if you want to run a quick sizing estimate based on refresh frequency or concurrency.

If my response resolved your query, kindly mark it as the Accepted Solution to assist others. Additionally, I would be grateful for a 'Kudos' if you found my response helpful.

Translation and formatting supported by AI

View solution in original post

3 REPLIES 3
v-saisrao-msft
Community Support
Community Support

Hi @Ramkishor,

checking in to see if your issue has been resolved. If the response provided was helpful, kindly mark it as the solution so that others with the same issue can benefit.
let us know if you still need assistance.

 

Thank you.

burakkaragoz
Super User
Super User

Hi @Ramkishor ,

 

Based on what you shared, you're looking at a moderately heavy setup — 80 reports, 30 embedded viewers, and 50GB of compressed data is no small thing.

Here’s what I’d recommend:

  • F64 or F128 Fabric capacity would be a solid starting point. F64 gives you 64 CU (capacity units) which should handle your dataflows, Lakehouse refreshes, and embedded report usage pretty well.
  • If your reports are heavily interactive or refresh frequently, F128 might give you more breathing room, especially during peak hours.
  • Since you're embedding into D365 F&O, make sure your capacity is in the same region as your D365 tenant to avoid latency issues.

Also, keep in mind that Lakehouse and Dataflows can consume a good chunk of capacity during refreshes. You might want to schedule those during off-hours if possible.

Here’s a Microsoft doc that breaks down Fabric SKUs and use cases:
👉 Microsoft Learn 

Let me know if you want to run a quick sizing estimate based on refresh frequency or concurrency.

If my response resolved your query, kindly mark it as the Accepted Solution to assist others. Additionally, I would be grateful for a 'Kudos' if you found my response helpful.

Translation and formatting supported by AI

v-saisrao-msft
Community Support
Community Support

Hi @Ramkishor,

Thank you for reaching out to the Microsoft Fabric Forum Community. 

Since you are working with multiple reports and users, handling larger volume of data, and using ETL through Dataflows and Lakehouse, I suggest starting with Microsoft Fabric F64 capacity. 

And if there are more frequent refreshes, complex models, or increased usage, you may want to scale up to F128 or higher capacity to maintain optimal performance for all workloads. 
Below is the link for your reference: 
Plan your capacity size - Microsoft Fabric | Microsoft Learn

 

Thank you. 

Helpful resources

Announcements
FabCon and SQLCon Highlights Carousel

FabCon &SQLCon Highlights

Experience the highlights from FabCon & SQLCon, available live and on-demand starting April 14th.

New to Fabric survey Carousel

New to Fabric Survey

If you have recently started exploring Fabric, we'd love to hear how it's going. Your feedback can help with product improvements.

Join our Fabric User Panel

Join our Fabric User Panel

Share feedback directly with Fabric product managers, participate in targeted research studies and influence the Fabric roadmap.

March Fabric Update Carousel

Fabric Monthly Update - March 2026

Check out the March 2026 Fabric update to learn about new features.