Power BI is turning 10! Tune in for a special live episode on July 24 with behind-the-scenes stories, product evolution highlights, and a sneak peek at what’s in store for the future.
Save the dateEnhance your career with this limited time 50% discount on Fabric and Power BI exams. Ends August 31st. Request your voucher.
Context: The below block of code works for the smaller Lakehouse Table = OrderData2021pt1, but when I try with a larger table (3M rows, maybe 20 columns) I get some errors on creating. Is this too large for a Pandas DataFrame?
My Follow Along Source For Testing/Reference 6:34 minute mark:
https://www.youtube.com/watch?v=8Xu1M-ORbK8&list=PLn1m_aBmgsbH5M_v7aZB_4GT9hecrDrEH&index=16
Error #1: Buffer overflow. Available: 0, required: 1506902. To avoid this, increase spark.kryoserializer.buffer.max value.
Error #2: Job was aborted due to user runtime error. This can be be for many reasons, a common cause is: 1. Ensure the files you are loading are of the format. If you're loading data via read.parquet, ensure the format of the data that is being read is indeed parquet. Consider gating wildcard loads with the file type suffix you intend to load to avoid. For example, instead of using a load string like /path/to/my/parquet/files/* Change this to: /path/to/my/parquet/files/*.parquet To avoid loading JSON files that might exist in the directory.
Hi @Anonymous
It would appear to me (I am not Fabric/notebook expert) that it is a limitation and too much data. Could you make the data into a smaller size or do it daily to then load it?
Check out the July 2025 Power BI update to learn about new features.
This is your chance to engage directly with the engineering team behind Fabric and Power BI. Share your experiences and shape the future.
User | Count |
---|---|
35 | |
22 | |
21 | |
14 | |
12 |
User | Count |
---|---|
28 | |
18 | |
16 | |
16 | |
14 |