<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Export entire delta table from lakehouse in Data Engineering</title>
    <link>https://community.fabric.microsoft.com/t5/Data-Engineering/Export-entire-delta-table-from-lakehouse/m-p/4131209#M3873</link>
    <description>&lt;P&gt;The Delta tables are already in Parquet format. it would be counterproductive to convert that to CSV.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;"the entire table"&amp;nbsp; - careful with that. Parquet supports versioning so you risk exporting all the prior versions too.&lt;/P&gt;</description>
    <pubDate>Tue, 03 Sep 2024 12:12:47 GMT</pubDate>
    <dc:creator>lbendlin</dc:creator>
    <dc:date>2024-09-03T12:12:47Z</dc:date>
    <item>
      <title>Export entire delta table from lakehouse</title>
      <link>https://community.fabric.microsoft.com/t5/Data-Engineering/Export-entire-delta-table-from-lakehouse/m-p/4131081#M3872</link>
      <description>&lt;P&gt;Is there any way to get a .csv onto my computer from a delta table in a lakehouse? I haven't found any viable approach so far despite trying.&lt;/P&gt;</description>
      <pubDate>Tue, 03 Sep 2024 11:10:57 GMT</pubDate>
      <guid>https://community.fabric.microsoft.com/t5/Data-Engineering/Export-entire-delta-table-from-lakehouse/m-p/4131081#M3872</guid>
      <dc:creator>Anonymous</dc:creator>
      <dc:date>2024-09-03T11:10:57Z</dc:date>
    </item>
    <item>
      <title>Re: Export entire delta table from lakehouse</title>
      <link>https://community.fabric.microsoft.com/t5/Data-Engineering/Export-entire-delta-table-from-lakehouse/m-p/4131209#M3873</link>
      <description>&lt;P&gt;The Delta tables are already in Parquet format. it would be counterproductive to convert that to CSV.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;"the entire table"&amp;nbsp; - careful with that. Parquet supports versioning so you risk exporting all the prior versions too.&lt;/P&gt;</description>
      <pubDate>Tue, 03 Sep 2024 12:12:47 GMT</pubDate>
      <guid>https://community.fabric.microsoft.com/t5/Data-Engineering/Export-entire-delta-table-from-lakehouse/m-p/4131209#M3873</guid>
      <dc:creator>lbendlin</dc:creator>
      <dc:date>2024-09-03T12:12:47Z</dc:date>
    </item>
    <item>
      <title>Re: Export entire delta table from lakehouse</title>
      <link>https://community.fabric.microsoft.com/t5/Data-Engineering/Export-entire-delta-table-from-lakehouse/m-p/4131215#M3874</link>
      <description>&lt;P&gt;One way to achieve this is to write a small Python script in a notebook that reads the delta table from your lakehouse and exports it to the Files section of your lakehouse. It may be convinient to make a pandas dataframe after you've read the delta table with spark and use pandas 'to_csv' functionality.&lt;BR /&gt;With the OneLake file explorer: &lt;A href="https://www.microsoft.com/en-us/download/details.aspx?id=105222" target="_blank"&gt;https://www.microsoft.com/en-us/download/details.aspx?id=105222&lt;/A&gt;, you could view and copy this csv to your local directory on your computer.&lt;/P&gt;</description>
      <pubDate>Tue, 03 Sep 2024 12:17:42 GMT</pubDate>
      <guid>https://community.fabric.microsoft.com/t5/Data-Engineering/Export-entire-delta-table-from-lakehouse/m-p/4131215#M3874</guid>
      <dc:creator>FabianSchut</dc:creator>
      <dc:date>2024-09-03T12:17:42Z</dc:date>
    </item>
    <item>
      <title>Re: Export entire delta table from lakehouse</title>
      <link>https://community.fabric.microsoft.com/t5/Data-Engineering/Export-entire-delta-table-from-lakehouse/m-p/4131219#M3875</link>
      <description>&lt;P&gt;There is only one version.&lt;/P&gt;</description>
      <pubDate>Tue, 03 Sep 2024 12:20:44 GMT</pubDate>
      <guid>https://community.fabric.microsoft.com/t5/Data-Engineering/Export-entire-delta-table-from-lakehouse/m-p/4131219#M3875</guid>
      <dc:creator>Anonymous</dc:creator>
      <dc:date>2024-09-03T12:20:44Z</dc:date>
    </item>
    <item>
      <title>Re: Export entire delta table from lakehouse</title>
      <link>https://community.fabric.microsoft.com/t5/Data-Engineering/Export-entire-delta-table-from-lakehouse/m-p/4131256#M3876</link>
      <description>&lt;P&gt;Brilliant, it works. FYI, if I use&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;df.write.csv("Files/my_csv.csv")&lt;/LI-CODE&gt;&lt;P&gt;to write the file, it keeps the delta table file structure, i.e. it makes a folder named "my_csv.csv". However, the file(s) are in csv, so for my purposes it just works.&lt;/P&gt;</description>
      <pubDate>Tue, 03 Sep 2024 12:36:40 GMT</pubDate>
      <guid>https://community.fabric.microsoft.com/t5/Data-Engineering/Export-entire-delta-table-from-lakehouse/m-p/4131256#M3876</guid>
      <dc:creator>Anonymous</dc:creator>
      <dc:date>2024-09-03T12:36:40Z</dc:date>
    </item>
    <item>
      <title>Re: Export entire delta table from lakehouse</title>
      <link>https://community.fabric.microsoft.com/t5/Data-Engineering/Export-entire-delta-table-from-lakehouse/m-p/4131282#M3877</link>
      <description>&lt;P&gt;But seriously, converting a Delta file back to CSV is a travesty. Can your application not work with Parquet?&lt;/P&gt;</description>
      <pubDate>Tue, 03 Sep 2024 12:47:57 GMT</pubDate>
      <guid>https://community.fabric.microsoft.com/t5/Data-Engineering/Export-entire-delta-table-from-lakehouse/m-p/4131282#M3877</guid>
      <dc:creator>lbendlin</dc:creator>
      <dc:date>2024-09-03T12:47:57Z</dc:date>
    </item>
    <item>
      <title>Re: Export entire delta table from lakehouse</title>
      <link>https://community.fabric.microsoft.com/t5/Data-Engineering/Export-entire-delta-table-from-lakehouse/m-p/4131303#M3879</link>
      <description>&lt;P&gt;My goals are beyond your understanding.&amp;nbsp;&lt;/P&gt;&lt;H1&gt;&lt;SPAN class=""&gt;¯\_(ツ)_/¯&lt;/SPAN&gt;&lt;/H1&gt;</description>
      <pubDate>Tue, 03 Sep 2024 12:58:15 GMT</pubDate>
      <guid>https://community.fabric.microsoft.com/t5/Data-Engineering/Export-entire-delta-table-from-lakehouse/m-p/4131303#M3879</guid>
      <dc:creator>Anonymous</dc:creator>
      <dc:date>2024-09-03T12:58:15Z</dc:date>
    </item>
    <item>
      <title>Re: Export entire delta table from lakehouse</title>
      <link>https://community.fabric.microsoft.com/t5/Data-Engineering/Export-entire-delta-table-from-lakehouse/m-p/4131637#M3884</link>
      <description>&lt;P&gt;If you use Pandas to_csv(), instead of PySpark write.csv(), you will get a single CSV file.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;A href="https://learn.microsoft.com/en-us/fabric/data-science/read-write-pandas" target="_blank" rel="noopener"&gt;https://learn.microsoft.com/en-us/fabric/data-science/read-write-pandas&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Pandas cannot work with such large data volumes as Spark.&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;A href="https://www.reddit.com/r/MicrosoftFabric/s/7EloDpd8wI" target="_blank" rel="noopener"&gt;https://www.reddit.com/r/MicrosoftFabric/s/7EloDpd8wI&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;If your amount of data is not too large volume, then I guess I would use Pandas for writing to a single csv.&lt;/P&gt;</description>
      <pubDate>Tue, 03 Sep 2024 15:57:57 GMT</pubDate>
      <guid>https://community.fabric.microsoft.com/t5/Data-Engineering/Export-entire-delta-table-from-lakehouse/m-p/4131637#M3884</guid>
      <dc:creator>frithjof_v</dc:creator>
      <dc:date>2024-09-03T15:57:57Z</dc:date>
    </item>
  </channel>
</rss>

