00:20:00

Microsoft Fabric Data Engineer Associate Practice Test (DP-700)

Use the form below to configure your Microsoft Fabric Data Engineer Associate Practice Test (DP-700). The practice test can be configured to only include certain exam objectives and domains. You can choose between 5-100 questions and set a time limit.

Logo for Microsoft Fabric Data Engineer Associate DP-700
Questions
Number of questions in the practice test
Free users are limited to 20 questions, upgrade to unlimited
Seconds Per Question
Determines how long you have to finish the practice test
Exam Objectives
Which exam objectives should be included in the practice test

Microsoft Fabric Data Engineer Associate DP-700 Information

The Microsoft Fabric Data Engineer Associate (DP-700) exam shows that you know how to work with data in Microsoft Fabric. It tests your ability to collect, organize, and prepare data so it can be used for reports and dashboards. Passing the DP-700 means you can build and manage data pipelines, use tools like Power BI and Azure Synapse, and make sure data is clean and ready for analysis.

This exam is best for people who already have some experience working with data or databases and want to move into a data engineering role. If you enjoy working with numbers, building reports, or using SQL and Python to manage data, this certification can help you stand out to employers. It’s designed for anyone who wants to show their skills in data handling using Microsoft tools.

Before taking the real exam, it’s smart to use DP-700 practice exams, practice tests, and practice questions to prepare. These tools help you get used to the types of questions you’ll see on test day and show which topics you need to study more. By using practice tests often, you can build confidence, improve your score, and walk into the exam knowing what to expect.

Microsoft Fabric Data Engineer Associate DP-700 Logo
  • Free Microsoft Fabric Data Engineer Associate DP-700 Practice Test

  • 20 Questions
  • Unlimited
  • Implement and manage an analytics solution
    Ingest and transform data
    Monitor and optimize an analytics solution
Question 1 of 20

Your organization uses a Microsoft Fabric data warehouse named ContosoDW. Members of the EuropeanFinance security group must query data in the dbo.Sales table, but:

  • They should see only rows where Region = 'EU'.
  • They must not be able to read the UnitCost column, while all other columns remain visible.

You need to configure ContosoDW to meet both requirements by using built-in capabilities of Fabric.

Which approach should you implement?

  • Move the UnitCost column to a separate table protected by workspace item permissions, and use pipeline parameters to pass the required Region value at query time.

  • Apply dynamic data masking to the UnitCost column and assign the EuropeanFinance group the Viewer role at the workspace level.

  • Create a view that selects all columns except UnitCost, and give EuropeanFinance SELECT permissions only on the view; do not configure any RLS because workspace roles will enforce the row filter automatically.

  • Create a table-valued function that filters on Region = 'EU', add it to a row-level security (RLS) security policy for dbo.Sales, and grant SELECT on the allowed columns only, leaving UnitCost ungranted.

Question 2 of 20

A Microsoft Fabric workspace contains a dataflow Gen2 that ingests daily CSV files from an Azure Data Lake Storage Gen2 container into a lakehouse table. After a recent refresh, you must verify how many rows were written by the dataflow and whether any source files failed to load, without opening the lakehouse itself. Which Fabric feature should you use to view this information?

  • The Monitor hub's Dataflows page and its refresh history for the dataflow

  • The Fabric Capacity Metrics app in Power BI

  • The workspace's Lineage view for the lakehouse

  • OneLake Data Explorer for the lakehouse's Files section

Question 3 of 20

You are developing a PySpark notebook in Microsoft Fabric that joins a 2-TB fact table to three dimension tables, each about 100 MB. Execution metrics show most time is spent on shuffle reads during the joins. Without resizing the Spark pool, you want the dimension tables broadcast to executors to cut shuffle time. Which Spark configuration should you set before running the notebook?

  • Lower spark.sql.shuffle.partitions to 50 to reduce the number of shuffle partitions.

  • Enable adaptive query execution by setting spark.sql.adaptive.enabled to true.

  • Increase the value of spark.sql.autoBroadcastJoinThreshold to 134217728 (128 MB).

  • Set spark.sql.files.maxPartitionBytes to 134217728 bytes so that fewer input partitions are created.

Question 4 of 20

You are designing a lakehouse in Microsoft Fabric. To avoid duplicating data files that already reside in an Azure Data Lake Storage Gen2 account managed by another team, you decide to surface the folder /raw/finance in your lakehouse by using a shortcut. After the shortcut is created, analytics engineers will load the data with Spark, but the source team wants to guarantee that their files cannot be modified from your workspace.

Which statement about Fabric shortcuts satisfies the source team's requirement?

  • Write access is controlled by the lakehouse item role assignment, not by the shortcut type, so you must remove the Engineer role to prevent changes.

  • Shortcuts that reference Azure Data Lake Storage Gen2 are read-only, so Spark sessions in the lakehouse can read the files but cannot write or delete them.

  • Any shortcut becomes writable once the workspace owner is granted Contributor rights on the target storage account.

  • Fabric automatically creates a versioned copy of the target folder; engineers write to the copy while the original files stay untouched.

Question 5 of 20

Your organization uses Microsoft Purview sensitivity labels with Microsoft Fabric. You administer a workspace where engineers frequently create new lakehouses, dataflows, and reports. Compliance requires every new item in the workspace to be labeled "Contoso Confidential" automatically, but creators must still have the option to choose a different label if needed. Which workspace-level information-protection setting should you configure to meet this requirement?

  • Set a default sensitivity label of Contoso Confidential for the workspace and leave the "Users can change the applied label" option enabled.

  • Enable the "Inherit sensitivity label from data sources" option so new items copy labels from connected data.

  • Disable the workspace's default label and rely on tenant-level automatic labeling policies to tag items based on data content.

  • Enable the "Require users to apply a sensitivity label" option and make Contoso Confidential the only selectable label.

Question 6 of 20

You manage a Microsoft Fabric workspace that uses the built-in deployment pipeline with the default Development, Test, and Production stages. After several iterations, the Development stage contains updated lakehouse tables, a modified data warehouse schema, and new notebooks. You must deploy only the metadata changes to the Test stage without copying any of the underlying data files. Which action should you take before you select Deploy on the Development stage?

  • Select Deploy; no additional configuration is required because deployment pipelines move only metadata for lakehouses and warehouses by default.

  • Export the lakehouse and warehouse as .pbip project files and remove their data folders before deployment.

  • Disconnect the workspace from its Git repository to ensure only metadata is copied when deploying.

  • Enable data skipping for the lakehouse and warehouse in the Development stage so that data files are excluded during deployment.

Question 7 of 20

You manage a Microsoft Fabric warehouse that stores sales transactions in a table named Sales.Cards. The table has a varchar(16) column named CardNumber that contains customers' primary account numbers (PANs).

To limit data exposure, you execute the following T-SQL statement:

ALTER TABLE Sales.Cards
ALTER COLUMN CardNumber
ADD MASKED WITH (FUNCTION = 'partial(0,"XXXX-XXXX-XXXX-",4)');

Later, an analyst who has SELECT permission on the table-but does not hold the UNMASK permission-runs the query SELECT TOP 1 CardNumber FROM Sales.Cards;.

Which value will the analyst see if the stored PAN in that row is '1234567812345678'?

  • 5678-XXXX-XXXX-XXXX

  • XXXX-XXXX-XXXX-5678

  • 0000000000000000

  • XXXX-XXXX-5678-XXXX

Question 8 of 20

You administer a Microsoft Fabric workspace that contains a lakehouse with several Delta tables used by multiple analytics teams. You must ensure that data analysts can query the SalesOrders table but must never see rows where the column TerritoryID equals 60. Business analysts should have unrestricted access. You need to configure this requirement in the most granular and performant way without duplicating data or creating additional views. Which approach should you implement in the lakehouse?

  • Publish a Power BI semantic model on top of SalesOrders and use that model's row-level security instead of configuring security in the lakehouse.

  • Grant the data-analysts group the Viewer role on the lakehouse and deny them access to the TerritoryID column by defining column-level security.

  • Move SalesOrders to a separate folder secured with OneLake folder permissions that exclude the data-analysts group.

  • Create a row-level security policy on the SalesOrders Delta table that filters out rows with TerritoryID = 60 for the data-analysts security group.

Question 9 of 20

You manage a Microsoft Fabric workspace that hosts several lakehouses, pipelines, and Power BI reports. A group of data scientists needs to upload new datasets, create notebooks, and schedule pipeline runs in the workspace. However, they must not be able to change workspace permissions, delete the workspace, or publish the workspace as an app to other users. Which built-in workspace role should you assign to the data scientists to meet these requirements?

  • Viewer

  • Admin

  • Contributor

  • Member

Question 10 of 20

A semantic model named SalesModel refreshes hourly in Microsoft Fabric. The 02:00 refresh failed, and you must examine the step-by-step execution, including the processing time of each incremental-refresh partition, to locate the failure point. In the Fabric UI, which area should you open first to view this detailed refresh activity?

  • Open Data pipelines and select Semantic model refresh activity for SalesModel.

  • Use the Monitoring hub's Workspace Activity view for the workspace that hosts SalesModel.

  • Launch the Capacity Metrics app and filter by the 02:00 timeframe.

  • Open the SalesModel settings and view its Refresh history pane.

Question 11 of 20

You build a Microsoft Fabric pipeline that copies new rows from an Azure SQL Database table into a destination table in a Fabric lakehouse. The Azure SQL linked service authenticates by using the workspace's managed identity. When you trigger the pipeline, the Copy activity fails immediately and returns error code SqlErrorNumber 262 with the message "Permission denied in database 'Sales' to object 'Orders'." What should you do to resolve the failure so that the pipeline can complete successfully?

  • Enable staged copy with auto-create table in the Copy activity settings.

  • Grant the workspace's managed identity SELECT permission on dbo.Orders or add it to the db_datareader role in the Sales database.

  • Grant the workspace's managed identity INSERT permission on dbo.Orders in the Sales database.

  • Replace the lakehouse sink with a Data Lake Storage Gen2 linked service.

Question 12 of 20

You manage a Microsoft Fabric warehouse that contains a table named Customer with an nvarchar(256) column called EmailAddress. Data analysts query the warehouse by using the SQL Endpoint. Analysts must see e-mail addresses in the format j*****@contoso.com, but members of the FabricAdmins security group must see the full, unmasked values. You decide to use dynamic data masking. Which T-SQL actions should you perform to meet the requirements?

  • Create a view that returns LEFT(EmailAddress,1) + '*****' + SUBSTRING(EmailAddress, CHARINDEX('@', EmailAddress), LEN(EmailAddress)), and deny SELECT on dbo.Customer to the Analysts role.

  • Run: ALTER TABLE dbo.Customer ALTER COLUMN EmailAddress ADD MASKED WITH (FUNCTION = 'partial(1,"*****",12)'); GRANT UNMASK TO FabricAdmins;

  • Enable transparent data encryption on the warehouse and add FabricAdmins to the db_accessadmin database role.

  • Apply the Highly Confidential sensitivity label to EmailAddress and assign FabricAdmins as reviewers for the label.

Question 13 of 20

You maintain a database project (.sqlproj) that contains the Data Definition Language (DDL) scripts for a Microsoft Fabric Warehouse. The project is stored in Azure Repos Git. You need to configure an Azure DevOps CI/CD pipeline that will automatically:

  • Build the database project to produce a DACPAC file.
  • Compare the DACPAC with the target Fabric Warehouse through its T-SQL endpoint.
  • Apply only the required schema changes during deployment.

Which pipeline task or command should you add to perform the deployment to the Fabric Warehouse?

  • Add a Power BI deployment pipeline task that promotes the workspace to the test stage.

  • Add an Azure CLI task that calls the Microsoft Fabric REST API to import the DACPAC file as a dataflow.

  • Add a Copy Files task that copies all .sql script files to a blob container linked to the workspace.

  • Add an Azure SQL Database deployment task that runs the SqlPackage command with the Publish action against the warehouse's T-SQL endpoint.

Question 14 of 20

You must transform 200 GB of semi-structured logs in OneLake. Engineers need to use Python libraries like pandas and scikit-learn for feature engineering and visual analysis, then write the result to a Delta Lake table. The code should support iterative development and be runnable on a scheduled Fabric pipeline. Which Fabric transformation option best meets the requirements?

  • KQL database update policy

  • Spark notebook

  • Dataflow Gen2

  • T-SQL stored procedure in a Fabric warehouse

Question 15 of 20

You manage a Microsoft Fabric workspace that contains a data warehouse used by several analytics teams. The schema is expected to change frequently, and you must adopt a repeatable, source-controlled process that lets developers design tables offline, validate T-SQL syntax, and generate a deployment artifact that can be promoted by a pipeline to test and production workspaces. Which approach should you recommend to meet these requirements?

  • Author schema changes in Fabric notebooks and rely on automated retry logic in orchestration pipelines to apply them in downstream workspaces.

  • Create a SQL Server Database Project in Visual Studio with SQL Server Data Tools (SSDT) and publish directly from the IDE to each workspace.

  • Create a SQL database project in Visual Studio Code (or Azure Data Studio) by using the SQL Database Projects extension and commit the resulting .sqlproj and generated DACPAC to the Git repository.

  • Use the Export SQL scripts feature in the Fabric portal after every schema change and store the scripts in the Git repository for manual deployment.

Question 16 of 20

Your solution must ingest about one million IoT sensor events per second into Microsoft Fabric, auto-scale the compute, refresh operator dashboards in seconds, and let analysts run ad-hoc anomaly-detection queries by writing a familiar declarative language (not notebook code) against both real-time and historical data. Which Fabric streaming engine best meets these requirements?

  • T-SQL streaming ingestion into a lakehouse SQL endpoint

  • Eventstream with no downstream analytics engine

  • KQL streaming in Real-Time Analytics (Azure Data Explorer in Fabric)

  • Spark Structured Streaming in a Fabric notebook

Question 17 of 20

You manage a Microsoft Fabric workspace that contains a lakehouse used by finance analysts. The analysts frequently export tables to Excel and then email the files outside the organization. Management requests that every exported Excel file be automatically encrypted and marked "Confidential - Finance" without requiring the analysts to take any additional steps. Which action will meet this requirement with the current capabilities of Microsoft Fabric?

  • Set a default sensitivity label for the entire workspace so that all contained items inherit the label and its protection settings automatically.

  • Apply the "Confidential - Finance" Microsoft Purview sensitivity label directly to the lakehouse item in the Fabric workspace.

  • This requirement cannot be met because Fabric lakehouse exports do not yet support sensitivity-label-based encryption or content marking.

  • Enable Azure Information Protection on the OneLake storage account and configure an encryption policy that targets the finance folder.

Question 18 of 20

You manage a Microsoft Fabric workspace that contains a semantic model named Contoso Sales. The model is configured to refresh every hour. The data-quality team must receive an email whenever a scheduled refresh ends with a Failure status. You need a solution that can be configured directly in the Fabric (Power BI) service without building a custom monitoring pipeline. What should you do?

  • Pin the semantic model's refresh history to a dashboard and configure a data alert on the visual so that it emails the data-quality team when the value equals Failure.

  • Create an Azure Monitor metric alert that fires when the semantic model's refresh duration metric exceeds a threshold and add the distribution list to the action group.

  • Open the semantic model's Scheduled refresh settings, enable refresh failure notifications, and add the data-quality distribution list as email recipients.

  • Configure a Fabric capacity usage alert for the workspace and specify the distribution list as notification recipients.

Question 19 of 20

You are building a daily load from an on-premises SQL Server table into a Delta Lake table in a Microsoft Fabric lakehouse. The source table contains hundreds of millions of rows and has an incrementing bigint column named UpdateSequence that increases each time a row is inserted or modified. Loads must copy only new and changed rows, overwrite any existing versions in the destination, and be fully restartable without creating duplicate rows. Which loading pattern meets these requirements?

  • Implement a watermark-based incremental load that filters rows where UpdateSequence is greater than the last stored watermark, then upsert the data into the Delta Lake table by using MERGE INTO.

  • Use an eventstream to capture change events and append them directly to the Delta Lake table without performing updates.

  • Schedule a nightly pipeline that truncates the Delta Lake table and performs a full bulk INSERT of the entire source table.

  • Configure two alternating pipelines: the first copies rows whose UpdateSequence modulo 2 equals 0, and the second copies the remaining rows the next day.

Question 20 of 20

You have created a notebook in a Microsoft Fabric workspace that accepts two parameters and transforms raw JSON files into a curated Delta table. You must ensure that the notebook runs automatically whenever a new JSON file is written to the raw container of an Azure Data Lake Storage Gen2 account, and only after the file has been copied to OneLake. You want to design the solution by using Fabric features while writing as little custom code as possible. Which approach should you use?

  • Create a SQL job in the Lakehouse that listens for CREATE FILE events and, when triggered, uses dynamic SQL to call the notebook through the Fabric REST API.

  • Configure a scheduled run for the notebook that executes every five minutes and add code to the notebook to poll the raw container for new files and copy them before processing.

  • Develop a Dataflow Gen2 that copies data from the raw container to OneLake and adds a script step at the end to invoke the notebook.

  • Create a Data Factory pipeline that uses an Azure Storage event trigger, adds a Copy activity to move the file to OneLake, and then calls the parameterized notebook in a subsequent Notebook activity.