<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Best practice for loading data into snowflake in Data Engineering</title>
    <link>https://community.fabric.microsoft.com/t5/Data-Engineering/Best-practice-for-loading-data-into-snowflake/m-p/4345172#M5799</link>
    <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.fabric.microsoft.com/t5/user/viewprofilepage/user-id/782160"&gt;@Srisakthi&lt;/a&gt;&amp;nbsp;&lt;BR /&gt;&lt;SPAN&gt;We haven’t heard back since our last response and wanted to check if your query has been resolved. If not, please feel free to reach out for further assistance. If it has been resolved, kindly mark the helpful reply as the solution to make it easier for others to find. A kudos would also be greatly appreciated!&lt;/SPAN&gt;&lt;BR /&gt;&lt;SPAN&gt;Thank you.&lt;/SPAN&gt;&lt;/P&gt;</description>
    <pubDate>Sat, 28 Dec 2024 14:54:56 GMT</pubDate>
    <dc:creator>Anonymous</dc:creator>
    <dc:date>2024-12-28T14:54:56Z</dc:date>
    <item>
      <title>Best practice for loading data into snowflake</title>
      <link>https://community.fabric.microsoft.com/t5/Data-Engineering/Best-practice-for-loading-data-into-snowflake/m-p/4338008#M5707</link>
      <description>&lt;P&gt;Hello Everyone,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Can you please suggest best practice for loading data from SQL/Oracle to snowflake using Microsoft Fabric data pipeline.&lt;/P&gt;&lt;P&gt;Case 1 : Multiple tables at a time with huge volume of data&amp;nbsp;in each table&lt;/P&gt;&lt;P&gt;Case 2: Multiple tables at a time with less volume of data in each table&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;What is the recommended way for both performance and CU consumption&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Regards,&lt;/P&gt;&lt;P&gt;Srisakthi&lt;/P&gt;</description>
      <pubDate>Fri, 20 Dec 2024 05:26:33 GMT</pubDate>
      <guid>https://community.fabric.microsoft.com/t5/Data-Engineering/Best-practice-for-loading-data-into-snowflake/m-p/4338008#M5707</guid>
      <dc:creator>Srisakthi</dc:creator>
      <dc:date>2024-12-20T05:26:33Z</dc:date>
    </item>
    <item>
      <title>Re: Best practice for loading data into snowflake</title>
      <link>https://community.fabric.microsoft.com/t5/Data-Engineering/Best-practice-for-loading-data-into-snowflake/m-p/4338110#M5710</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.fabric.microsoft.com/t5/user/viewprofilepage/user-id/782160"&gt;@Srisakthi&lt;/a&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;BR /&gt;Thank you for reaching out Microsoft Fabric Community Forum.&lt;BR /&gt;&lt;BR /&gt;&lt;/P&gt;
&lt;UL&gt;
&lt;LI&gt;&lt;STRONG&gt;For high-volume data loads :&amp;nbsp;&lt;/STRONG&gt;We should&amp;nbsp;focus on &lt;STRONG&gt;parallel loading&lt;/STRONG&gt; using bulk loading techniques (COPY INTO), leveraging &lt;STRONG&gt;larger compute warehouses&lt;/STRONG&gt; for performance, and &lt;STRONG&gt;batch processing&lt;/STRONG&gt; to minimize compute and storage costs.&lt;/LI&gt;
&lt;LI&gt;&lt;STRONG&gt;For low-volume data loads :&lt;/STRONG&gt;&amp;nbsp; we aim to reduce compute overhead with &lt;STRONG&gt;smaller virtual warehouses&lt;/STRONG&gt;, &lt;STRONG&gt;sequential loads&lt;/STRONG&gt;, and &lt;STRONG&gt;incremental loading&lt;/STRONG&gt; (Streams and Tasks). Also, consider using &lt;STRONG&gt;Snowpipe&lt;/STRONG&gt; for real-time or near-real-time data loading to optimize cost and compute resource usage.&lt;BR /&gt;&lt;BR /&gt;&lt;SPAN&gt;If this solution helps, please consider giving us &lt;STRONG&gt;&lt;EM&gt;Kudos&lt;/EM&gt;&lt;/STRONG&gt; and &lt;STRONG&gt;&lt;EM&gt;accepting it as the solution&lt;/EM&gt;&lt;/STRONG&gt; so that it may assist other members in the community.&lt;BR /&gt;Thank you.&lt;/SPAN&gt;&lt;/LI&gt;
&lt;/UL&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Fri, 20 Dec 2024 06:55:05 GMT</pubDate>
      <guid>https://community.fabric.microsoft.com/t5/Data-Engineering/Best-practice-for-loading-data-into-snowflake/m-p/4338110#M5710</guid>
      <dc:creator>Anonymous</dc:creator>
      <dc:date>2024-12-20T06:55:05Z</dc:date>
    </item>
    <item>
      <title>Re: Best practice for loading data into snowflake</title>
      <link>https://community.fabric.microsoft.com/t5/Data-Engineering/Best-practice-for-loading-data-into-snowflake/m-p/4341493#M5752</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.fabric.microsoft.com/t5/user/viewprofilepage/user-id/782160"&gt;@Srisakthi&lt;/a&gt;&amp;nbsp;&lt;BR /&gt;&lt;SPAN&gt;Thank you for your patience. please let us know if anything was helpful to you, so that we can convert it into a formal answer. If so, we would appreciate it if you could&amp;nbsp;&lt;/SPAN&gt;&lt;STRONG&gt;Accept it as a solution&lt;/STRONG&gt;&lt;SPAN&gt;&amp;nbsp;and drop a '&lt;/SPAN&gt;&lt;STRONG&gt;Kudos&lt;/STRONG&gt;&lt;SPAN&gt;' so other members can find it more easily.&lt;/SPAN&gt;&lt;BR /&gt;&lt;SPAN&gt;Thank you.&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Tue, 24 Dec 2024 04:25:54 GMT</pubDate>
      <guid>https://community.fabric.microsoft.com/t5/Data-Engineering/Best-practice-for-loading-data-into-snowflake/m-p/4341493#M5752</guid>
      <dc:creator>Anonymous</dc:creator>
      <dc:date>2024-12-24T04:25:54Z</dc:date>
    </item>
    <item>
      <title>Re: Best practice for loading data into snowflake</title>
      <link>https://community.fabric.microsoft.com/t5/Data-Engineering/Best-practice-for-loading-data-into-snowflake/m-p/4345172#M5799</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.fabric.microsoft.com/t5/user/viewprofilepage/user-id/782160"&gt;@Srisakthi&lt;/a&gt;&amp;nbsp;&lt;BR /&gt;&lt;SPAN&gt;We haven’t heard back since our last response and wanted to check if your query has been resolved. If not, please feel free to reach out for further assistance. If it has been resolved, kindly mark the helpful reply as the solution to make it easier for others to find. A kudos would also be greatly appreciated!&lt;/SPAN&gt;&lt;BR /&gt;&lt;SPAN&gt;Thank you.&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Sat, 28 Dec 2024 14:54:56 GMT</pubDate>
      <guid>https://community.fabric.microsoft.com/t5/Data-Engineering/Best-practice-for-loading-data-into-snowflake/m-p/4345172#M5799</guid>
      <dc:creator>Anonymous</dc:creator>
      <dc:date>2024-12-28T14:54:56Z</dc:date>
    </item>
    <item>
      <title>Re: Best practice for loading data into snowflake</title>
      <link>https://community.fabric.microsoft.com/t5/Data-Engineering/Best-practice-for-loading-data-into-snowflake/m-p/4346145#M5816</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.fabric.microsoft.com/t5/user/viewprofilepage/user-id/782160"&gt;@Srisakthi&lt;/a&gt;&amp;nbsp;&lt;BR /&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;1. Firstly a). you will create a config table where we will mention the start and end range of date that choose either full and incremental load based on your datasets.&lt;/P&gt;&lt;P&gt;b). The data format should be &lt;STRONG&gt;parquet with snappy compression&lt;/STRONG&gt; that helps to optimized the data performance. data should be save into small chunk files.&lt;/P&gt;&lt;P&gt;c). If you are quite better in pyspark, so it would be better otherwise you will do it with the help of the copy activities in the pipleines&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;If this post helps, then please consider Accept it as the solution to help the other members find it more quickly.&lt;/P&gt;</description>
      <pubDate>Mon, 30 Dec 2024 10:00:50 GMT</pubDate>
      <guid>https://community.fabric.microsoft.com/t5/Data-Engineering/Best-practice-for-loading-data-into-snowflake/m-p/4346145#M5816</guid>
      <dc:creator>Ray_Minds</dc:creator>
      <dc:date>2024-12-30T10:00:50Z</dc:date>
    </item>
  </channel>
</rss>

