top of page
  • Writer's pictureHarini Mallawaarachchi

Ingest data with a pipeline in Microsoft Fabric


A data lakehouse is a common analytical data store for cloud-scale analytics solutions. One of the core tasks of a data engineer is to implement and manage the ingestion of data from multiple operational data sources into the lakehouse. In Microsoft Fabric, you can implement extract, transform, and load (ETL) or extract, load, and transform (ELT) solutions for data ingestion through the creation of pipelines.


Fabric also supports Apache Spark, enabling you to write and run code to process data at scale. By combining the pipeline and Spark capabilities in Fabric, you can implement complex data ingestion logic that copies data from external sources into the OneLake storage on which the lakehouse is based, and then uses Spark code to perform custom data transformations before loading it into tables for analysis.


This lab will take approximately 60 minutes to complete.



Note: You need a Microsoft school or work account to complete this exercise. If you don’t have one, you can sign up for a trial of Microsoft Office 365 E3 or higher.


Create a workspace

Create a Lakehouse

Create a pipeline

Create a notebook

Modify the pipeline

Clean up resources

* Using CoPilot




1 view0 comments

Recent Posts

See All

Comments


bottom of page