Skip to main content
cancel
Showing results for 
Search instead for 
Did you mean: 

Enhance your career with this limited time 50% discount on Fabric and Power BI exams. Ends September 15. Request your voucher.

Rufyda

Working with Data in Microsoft Fabric

Ingesting Data into Microsoft Fabric


Before any analysis or transformation can take place, data must first be ingested into the Fabric environment.

Fabric supports a wide range of data sources, allowing for seamless integration of both local and cloud-based data.

 

Supported Data Sources:

Local sources: such as CSV files or Excel sheets stored on your computer.

Cloudsources: including Azure Data Lake Storage Gen2 and other cloud-based repositories.

 

Once data is ingested, it is stored in a Lakehouse—a central, unified storage layer designed to handle structured, semi-structured, and unstructured data.

This architecture enables easy access for exploration and transformation, all while maintaining scalability and flexibility.

 

Exploring and Transforming Data with Notebooks:
Microsoft Fabric provides a familiar and powerful environment for data exploration through Notebooks, which are powered by Apache Spark—an open-source, distributed computing framework designed for large-scale data processing.

 

🔹 How Spark Works in Fabric:


Automatic Spark Compute: Each Notebook is automatically connected to a Spark runtime environment.

Session Management: When you run the first cell in a Notebook, a Spark session is initiated.

This session remains active during usage and is automatically paused during inactivity to optimize resource usage and control costs.

 

🔹 Supported Languages:
PySpark (Python)

SparkR (R)

 

Notebooks in Fabric allow you to explore your data using rich analysis libraries, custom scripts, and built-in visualization tools.

This interactive environment makes it easy to uncover insights, build models, and prepare data for downstream workflows.

 

Preparing Data with Data Wrangler
For users looking for a more guided, code-assisted transformation experience, Data Wrangler in Microsoft Fabric is an invaluable tool. It offers a user-friendly interface to quickly clean, shape, and transform datasets without writing complex code.

 

Key Features of Data Wrangler:

  • Data Overview: Get a statistical summary of your dataset to understand distributions, types, and patterns.
  • Issue Detection: Automatically identify missing, inconsistent, or invalid data.
  • Cleaning Operations:

Remove or impute missing values

Convert data types

Detect and handle outliers

 

  • Auto-generated Code: Every transformation you apply through the UI automatically generates reusable code—making it easy to reproduce or adjust later.
  • Data Output: Once transformed, your clean dataset can be saved directly back into the Lakehouse for analysis or sharing.


Microsoft Fabric streamlines the entire data workflow—from ingestion to transformation—into a single, cohesive platform. Whether you're a data scientist working with Spark notebooks or a business analyst using Data Wrangler, Fabric provides the tools and flexibility to work with data efficiently and at scale. By centralizing data storage and offering robust transformation tools, Microsoft Fabric enables faster, smarter decision-making for modern data teams.

 

Comments

Thanks for breaking this down so clearly. Love how you highlighted both the technical capabilities and user-friendly features of Microsoft Fabric. Very informative

Very well written, @Rufyda  —great draft! It’s clear and easy to follow. 

Adding on to this, OneLake is the real star—think “OneDrive for data”: the unified, native storage for all Fabric workloads, available the moment you sign in.

Shortcuts, Dataflow Gen2, and Pipelines are the key enablers to ingest and orchestrate data.
And once the data is in the Fabric environment, the story reveals itself quickly.

Thank you so much for this article.

Appreciate your kind feedback! Glad you found it useful @ArwaAldoud 

@shashiPaul1570_ Thanks a lot for pointing that out,, OneLake and the orchestration tools really do complete the Fabric story.