Storage Basics
Ingestion Choices
Pick the Tool
Load Patterns
100

ADLS Gen2 is implemented as a capability on top of this Azure service.

What is an Azure Storage account?

100

In Fabric, this option lets you reference data in OneLake or supported storage without copying it into your Lakehouse.

What is a Shortcut?

100

This Fabric tool is low code and great for repeatable data shaping using a Power Query style experience.

What are Dataflows Gen2?

100

This load pattern reprocesses the entire dataset every time it runs.

What is a full load?

200

In Azure, this feature enables folder like structure and fine grained access control, making Storage behave more like a data lake.

What is hierarchical namespace (ADLS Gen2)?

200

You would typically choose ingestion (copy) over Shortcut when you need heavy transformations and want data to live in your workspace for this reason.

What is performance and transformation control (local copy for optimized processing)?

200

This Fabric tool is best known for orchestration: scheduling, chaining activities, and moving data from A to B.

What are Pipelines?

200

This load pattern processes only new or changed records since the last run.

What is an incremental load?

300

This open format adds transactions and schema enforcement to a data lake, commonly used in Lakehouse designs.

What is Delta Lake?

300

Avoid making multiple copies of the same dataset across teams” is a reason to aim for this principle.

What is a single source of truth?

300

This option is best when you need code first control, custom logic, or Spark based transformations.

What are Notebooks?

300

A timestamp, ID, or version column used to track “what changed since last time” is often called a:

What is a watermark?

400

In Microsoft Fabric, this is the single, unified data lake for the entire tenant that workspaces use.

What is OneLake?

400

These two considerations often decide Shortcut vs ingest: who can access the data and how it is governed or secured.

What are access control and governance?

400

Scenario: “Run every night at 2 AM. Copy files from a landing zone into a Lakehouse table, then kick off a notebook.” Best primary tool to orchestrate.

What are Pipelines?

400

Capturing inserts, updates, and deletes from a source system is often described as:

What is CDC (change data capture)?

500

Bronze, Silver, Gold is a common way to describe these stages of data as it moves from raw to refined.

What is the medallion architecture (raw to curated layers)?

500

If the upstream system changes columns often, ingesting data can help protect downstream reports by reducing direct dependency. This problem is commonly called:

What is schema drift (or breaking schema changes)?

500

“Business analyst needs to connect to a SaaS source, clean columns, split fields, and load into a Lakehouse, no code.” Best tool.

What are Dataflows Gen2?

500

Incremental loads often reduce this compared to full loads, especially on large datasets.

What are compute cost and runtime?

M
e
n
u